diff --git "a/train.jsonl" "b/train.jsonl" --- "a/train.jsonl" +++ "b/train.jsonl" @@ -42,8 +42,8 @@ {"id": "1230", "title": "Server safeguards tax service", "abstract": "Peterborough-based tax consultancy IE Taxguard wanted real-time failover protection for important Windows-based applications. Its solution was to implement a powerful failover server from UK supplier Neverfail in order to provide real-time backup for three core production servers", "keyphrases": ["tax consultancy", "IE Taxguard", "failover server", "Neverfail", "backup"], "prmu": ["P", "P", "P", "P", "P"]} {"id": "1275", "title": "Modeling dynamic objects in distributed systems with nested Petri nets", "abstract": "Nested Petri nets (NP-nets) is a Petri net extension, allowing tokens in a net marking to be represented by marked nets themselves. The paper discusses applicability of NP-nets for modeling task planning systems, multi-agent systems and recursive-parallel systems. A comparison of NP-nets with some other formalisms, such as OPNs of R. Valk (2000), recursive parallel programs of O. Kushnarenko and Ph. Schnoebelen (1997) and process algebras is given. Some aspects of decidability for object-oriented Petri net extensions are also discussed", "keyphrases": ["dynamic objects modelling", "distributed systems", "nested Petri nets", "multi-agent systems", "recursive-parallel systems", "process algebras", "decidability", "object-oriented Petri net"], "prmu": ["R", "P", "P", "P", "P", "P", "P", "P"]} {"id": "620", "title": "Adaptive image enhancement for retinal blood vessel segmentation", "abstract": "Retinal blood vessel images are enhanced by removing the nonstationary background, which is adaptively estimated based on local neighbourhood information. The result is a much better segmentation of the blood vessels with a simple algorithm and without the need to obtain a priori illumination knowledge of the imaging system", "keyphrases": ["adaptive image enhancement", "retinal blood vessel images", "local neighbourhood information", "nonstationary background removal", "image segmentation", "personal identification", "security applications"], "prmu": ["P", "P", "P", "R", "R", "U", "U"]} -{"id": "1094", "title": "Efficient allocation of knowledge in distributed business structures", "abstract": "Accelerated business processes demand new concepts and realizations of information systems and knowledge databases. This paper presents the concept of the collaborative information space (CIS), which supplies the necessary tools to transform individual knowledge into collective useful information. The creation of 'information objects' in the CIS allows an efficient allocation of information in all business process steps at any time. Furthermore, the specific availability of heterogeneous, distributed data is realized by a Web-based user interface, which enables effective search by a multidimensionally hierarchical composition", "keyphrases": ["distributed business structures", "efficient knowledge allocation", "accelerated business processes", "information systems", "knowledge databases", "collaborative information space", "information objects", "business process steps", "heterogeneous distributed data", "Web-based user interface", "multidimensionally hierarchical composition", "interactive system"], "prmu": ["P", "R", "P", "P", "P", "P", "R", "P", "R", "P", "P", "M"]} -{"id": "1445", "title": "Applying BGL to computational geometry", "abstract": "The author applies Boost Graph Library to the domain of computational geometry. First, he formulates a concrete problem in graph terms. Second, he develops a way to transform the output of an existing algorithm into an appropriate Boost Graph Library data structure. Finally, he implements two new algorithms for my Boost Graph Library graph. The first algorithm gets the job done, but could have been written in any programming language. The second algorithm, however, shows the power of Boost Graph Library's generic programming approach.Graphs, graphics, and generic programming combine in this novel use of the Boost Graph Library", "keyphrases": ["Boost libraries", "C++", "threads", "smart pointers", "Boost Graph Library", "graph-theoretic concepts", "directed graph", "file dependencies", "computational geometry", "BGL graph", "generic programming approach"], "prmu": ["R", "U", "U", "U", "P", "U", "M", "U", "P", "R", "M"]} +{"id": "1094", "title": "Efficient allocation of knowledge in distributed business structures", "abstract": "Accelerated business processes demand new concepts and realizations of information systems and knowledge databases. This paper presents the concept of the collaborative information space (CIS), which supplies the necessary tools to transform individual knowledge into collective useful information. The creation of 'information objects' in the CIS allows an efficient allocation of information in all business process steps at any time. Furthermore, the specific availability of heterogeneous, distributed data is realized by a Web-based user interface, which enables effective search by a multidimensionally hierarchical composition", "keyphrases": ["distributed business structures", "efficient knowledge allocation", "accelerated business processes", "information systems", "knowledge databases", "collaborative information space", "information objects", "business process steps", "heterogeneous distributed data", "Web-based user interface", "multidimensionally hierarchical composition", "interactive system"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "P", "R", "P", "P", "M"]} +{"id": "1445", "title": "Applying BGL to computational geometry", "abstract": "The author applies Boost Graph Library to the domain of computational geometry. First, he formulates a concrete problem in graph terms. Second, he develops a way to transform the output of an existing algorithm into an appropriate Boost Graph Library data structure. Finally, he implements two new algorithms for my Boost Graph Library graph. The first algorithm gets the job done, but could have been written in any programming language. The second algorithm, however, shows the power of Boost Graph Library's generic programming approach.Graphs, graphics, and generic programming combine in this novel use of the Boost Graph Library", "keyphrases": ["Boost libraries", "C++", "threads", "smart pointers", "Boost Graph Library", "graph-theoretic concepts", "directed graph", "file dependencies", "computational geometry", "BGL graph", "generic programming approach"], "prmu": ["R", "U", "U", "U", "P", "U", "M", "U", "P", "R", "P"]} {"id": "816", "title": "Accelerating filtering techniques for numeric CSPs", "abstract": "Search algorithms for solving Numeric CSPs (Constraint Satisfaction Problems) make an extensive use of filtering techniques. In this paper we show how those filtering techniques can be accelerated by discovering and exploiting some regularities during the filtering process. Two kinds of regularities are discussed, cyclic phenomena in the propagation queue and numeric regularities of the domains of the variables. We also present in this paper an attempt to unify numeric CSPs solving methods from two distinct communities, that of CSP in artificial intelligence, and that of interval analysis", "keyphrases": ["search algorithms", "Numeric CSPs", "Constraint Satisfaction Problems", "filtering techniques", "CSPs-solving", "artificial intelligence", "interval analysis", "extrapolation methods", "propagation", "pruning"], "prmu": ["P", "P", "P", "P", "U", "P", "P", "M", "P", "U"]} {"id": "853", "title": "Virtual Development Center", "abstract": "The Virtual Development Center of the Institute for Women and Technology seeks to significantly enhance the impact of women on technology. It addresses this goal by increasing the number of women who have input on created technology, enhancing the ways people teach and develop technology, and developing need-based technology that serves the community. Through activities of the Virtual Development Center, a pattern is emerging regarding how computing technologies do or do not satisfy the needs of community groups, particularly those communities serving women. This paper describes the Virtual Development Center program and offers observations on the impact of computing technology on non-technical communities", "keyphrases": ["Virtual Development Center", "women", "information technology", "teaching", "community groups", "gender issues", "computer science education"], "prmu": ["P", "P", "M", "P", "P", "U", "M"]} {"id": "778", "title": "Access matters", "abstract": "Discusses accessibility needs of people with disabilities, both from the perspective of getting the information from I&R programs (including accessible Web sites, TTY access, Braille, and other mechanisms) and from the perspective of being aware of accessibility needs when referring clients to resources. Includes information on ADA legislation requiring accessibility to public places and recommends several organizations and Web sites for additional information", "keyphrases": ["accessibility needs", "disabled people", "information and referral programs", "accessible Web sites", "TTY access", "Braille", "ADA legislation", "public places"], "prmu": ["P", "R", "M", "P", "P", "P", "P", "P"]} @@ -75,7 +75,7 @@ {"id": "1056", "title": "Eliminating recency with self-review: the case of auditors' 'going concern' judgments", "abstract": "This paper examines the use of self-review to debias recency. Recency is found in the 'going concern' judgments of staff auditors, but is successfully eliminated by the auditor's use of a simple self-review technique that would be extremely easy to implement in audit practice. Auditors who self-review are also less inclined to make audit report choices that are inconsistent with their going concern judgments. These results are important because the judgments of staff auditors often determine the type and extent of documentation in audit workpapers and serve as preliminary inputs for senior auditors' judgments and choices. If staff auditors' judgments are affected by recency, the impact of this bias may be impounded in the ultimate judgments and choices of senior auditors. Since biased judgments can expose auditors to significant costs involving extended audit procedures, legal liability and diminished reputation, simple debiasing techniques that reduce this exposure are valuable. The paper also explores some future research needs and other important issues concerning judgment debiasing in applied professional settings", "keyphrases": ["auditor going concern judgments", "recency debiasing", "self-review", "staff auditors", "audit report choices", "documentation", "audit workpapers", "senior auditors", "extended audit procedures", "legal liability", "diminished reputation", "judgment debiasing", "applied professional settings", "accountability", "probability judgments"], "prmu": ["R", "R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "U", "M"]} {"id": "1013", "title": "A scalable intelligent takeoff controller for a simulated running jointed leg", "abstract": "Running with jointed legs poses a difficult control problem in robotics. Neural controllers are attractive because they allow the robot to adapt to changing environmental conditions. However, scalability is an issue with many neural controllers. The paper describes the development of a scalable neurofuzzy controller for the takeoff phase of the running stride. Scalability is achieved by selecting a controller whose size does not grow with the dimensionality of the problem. Empirical results show that with proper design the takeoff controller scales from a leg with a single movable link to one with three movable links without a corresponding growth in size and without a loss of accuracy", "keyphrases": ["scalable intelligent takeoff controller", "simulated running jointed leg", "neural controllers", "changing environmental conditions", "scalability", "scalable neurofuzzy controller", "takeoff phase", "running stride", "intelligent robotic control"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "R"]} {"id": "60", "title": "Perceptual audio coding using adaptive pre- and post-filters and lossless compression", "abstract": "This paper proposes a versatile perceptual audio coding method that achieves high compression ratios and is capable of low encoding/decoding delay. It accommodates a variety of source signals (including both music and speech) with different sampling rates. It is based on separating irrelevance and redundancy reductions into independent functional units. This contrasts traditional audio coding where both are integrated within the same subband decomposition. The separation allows for the independent optimization of the irrelevance and redundancy reduction units. For both reductions, we rely on adaptive filtering and predictive coding as much as possible to minimize the delay. A psycho-acoustically controlled adaptive linear filter is used for the irrelevance reduction, and the redundancy reduction is carried out by a predictive lossless coding scheme, which is termed weighted cascaded least mean squared (WCLMS) method. Experiments are carried out on a database of moderate size which contains mono-signals of different sampling rates and varying nature (music, speech, or mixed). They show that the proposed WCLMS lossless coder outperforms other competing lossless coders in terms of compression ratios and delay, as applied to the pre-filtered signal. Moreover, a subjective listening test of the combined pre-filter/lossless coder and a state-of-the-art perceptual audio coder (PAC) shows that the new method achieves a comparable compression ratio and audio quality with a lower delay", "keyphrases": ["perceptual audio coding", "adaptive pre-filters", "adaptive post-filters", "lossless compression", "high compression ratio", "low encoding/decoding delay", "source signals", "music", "sampling rates", "redundancy reduction", "adaptive filtering", "predictive coding", "psycho-acoustically controlled adaptive linear filter", "irrelevance reduction", "predictive lossless coding", "weighted cascaded least mean squared", "WCLMS lossless coder", "subjective listening test", "pre-filter/lossless coder", "audio quality"], "prmu": ["P", "R", "R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} -{"id": "132", "title": "A unified view for vector rotational CORDIC algorithms and architectures based on angle quantization approach", "abstract": "Vector rotation is the key operation employed extensively in many digital signal processing applications. In this paper, we introduce a new design concept called Angle Quantization (AQ). It can be used as a design index for vector rotational operation, where the rotational angle is known in advance. Based on the AQ process, we establish a unified design framework for cost-effective low-latency rotational algorithms and architectures. Several existing works, such as conventional COordinate Rotational Digital Computer (CORDIC), AR-CORDIC, MVR-CORDIC, and EEAS-based CORDIC, can be fitted into the design framework, forming a Vector Rotational CORDIC Family. Moreover, we address four searching algorithms to solve the optimization problem encountered in the proposed vector rotational CORDIC family. The corresponding scaling operations of the CORDIC family are also discussed. Based on the new design framework, we can realize high-speed/low-complexity rotational VLSI circuits, whereas without degrading the precision performance in fixed-point implementations", "keyphrases": ["vector rotational CORDIC algorithms", "digital signal processing applications", "DSP applications", "angle quantization", "design index", "vector rotational operation", "unified design framework", "low-latency rotational algorithms", "greedy searching algorithm", "low-latency rotational architectures", "searching algorithms", "optimization problem", "scaling operations", "high-speed rotational VLSI circuits", "low-complexity rotational VLSI circuits", "fixed-point implementations", "trellis-based searching algorithm"], "prmu": ["P", "P", "M", "P", "P", "P", "P", "P", "M", "R", "P", "P", "P", "M", "M", "P", "M"]} +{"id": "132", "title": "A unified view for vector rotational CORDIC algorithms and architectures based on angle quantization approach", "abstract": "Vector rotation is the key operation employed extensively in many digital signal processing applications. In this paper, we introduce a new design concept called Angle Quantization (AQ). It can be used as a design index for vector rotational operation, where the rotational angle is known in advance. Based on the AQ process, we establish a unified design framework for cost-effective low-latency rotational algorithms and architectures. Several existing works, such as conventional COordinate Rotational Digital Computer (CORDIC), AR-CORDIC, MVR-CORDIC, and EEAS-based CORDIC, can be fitted into the design framework, forming a Vector Rotational CORDIC Family. Moreover, we address four searching algorithms to solve the optimization problem encountered in the proposed vector rotational CORDIC family. The corresponding scaling operations of the CORDIC family are also discussed. Based on the new design framework, we can realize high-speed/low-complexity rotational VLSI circuits, whereas without degrading the precision performance in fixed-point implementations", "keyphrases": ["vector rotational CORDIC algorithms", "digital signal processing applications", "DSP applications", "angle quantization", "design index", "vector rotational operation", "unified design framework", "low-latency rotational algorithms", "greedy searching algorithm", "low-latency rotational architectures", "searching algorithms", "optimization problem", "scaling operations", "high-speed rotational VLSI circuits", "low-complexity rotational VLSI circuits", "fixed-point implementations", "trellis-based searching algorithm"], "prmu": ["P", "P", "M", "P", "P", "P", "P", "P", "M", "R", "P", "P", "P", "R", "P", "P", "M"]} {"id": "971", "title": "Homogenization in L/sup infinity /", "abstract": "Homogenization of deterministic control problems with L/sup infinity / running cost is studied by viscosity solutions techniques. It is proved that the value function of an L/sup infinity / problem in a medium with a periodic micro-structure converges uniformly on the compact sets to the value function of the homogenized problem as the period shrinks to 0. Our main convergence result extends that of Ishii (Stochastic Analysis, control, optimization and applications, pp. 305-324, Birkhauser Boston, Boston, MA, 1999.) to the case of a discontinuous Hamiltonian. The cell problem is solved, but, as nonuniqueness occurs, the effective Hamiltonian must be selected in a careful way. The paper also provides a representation formula for the effective Hamiltonian and gives illustrations to calculus of variations, averaging and one-dimensional problems", "keyphrases": ["deterministic control", "L/sup infinity / running cost", "homogenization", "value function", "averaging", "calculus of variations", "convergence", "optimal control", "cell problem"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "R", "P"]} {"id": "934", "title": "Induced-shear piezoelectric actuators for rotor blade trailing edge flaps", "abstract": "Much of the current rotorcraft research is focused on improving performance by reducing unwanted helicopter noise and vibration. One of the most promising active rotorcraft vibration control systems is an active trailing edge flap. In this paper, an induced-shear piezoelectric tube actuator is used in conjunction with a simple lever-cusp hinge amplification device to generate a useful combination of trailing edge flap deflections and hinge moments. A finite-element model of the actuator tube and trailing edge flap (including aerodynamic and inertial loading) was used to guide the design of the actuator-flap system. A full-scale induced shear tube actuator flap system was fabricated and bench top testing was conducted to validate the analysis. Hinge moments corresponding to various rotor speeds were applied to the actuator using mechanical springs. The testing demonstrated that for an applied electric field of 3 kV cm/sup -1/ the tube actuator deflected a representative full-scale 12 inch flap +or-2.8 degrees at 0 rpm and +or-1.4 degrees for a hinge moment simulating a 400 rpm condition. The per cent error between the predicted and experimental full-scale flap deflections ranged from 4% (low rpm) to 12.5% (large rpm). Increasing the electric field to 4 kV cm/sup -1/ results in +or-2.5 degrees flap deflection at a rotation speed of 400 rpm, according to the design analysis. A trade study was conducted to compare the performance of the piezoelectric tube actuator to the state of the art in trailing edge flap actuators and indicated that the induced-shear tube actuator shows promise as a trailing edge flap actuator", "keyphrases": ["rotorcraft", "helicopter noise", "helicopter vibration", "vibration control", "active trailing edge flap", "lever-cusp hinge amplification device", "finite-element model", "aerodynamic loading", "12 inch flap", "inertial loading", "design", "shear tube actuator flap", "bench top testing", "piezoelectric tube actuator", "induced-shear tube actuator", "12 inch"], "prmu": ["P", "P", "R", "P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1232", "title": "Techniques for compiling and implementing all NAS parallel benchmarks in HPF", "abstract": "The NAS parallel benchmarks (NPB) are a well-known benchmark set for high-performance machines. Much effort has been made to implement them in High-Performance Fortran (HPF). In previous attempts, however, the HPF versions did not include the complete set of benchmarks, and the performance was not always good. In this study, we implement all eight benchmarks of the NPB in HPF, and parallelize them using an HPF compiler that we have developed. This report describes the implementation techniques and compiler features necessary to achieve good performance. We evaluate the HPF version on the Hitachi SR2201, a distributed-memory parallel machine. With 16 processors, the execution time of the HPF version is within a factor of 1.5 of the hand-parallelized version of the NPB 2.3 beta", "keyphrases": ["NAS parallel benchmarks", "high-performance machines", "compiler", "distributed-memory parallel supercomputers", "HPF compiler"], "prmu": ["P", "P", "P", "M", "P"]} @@ -139,7 +139,7 @@ {"id": "1297", "title": "Stochastic optimization of acoustic response - a numerical and experimental comparison", "abstract": "The objective of the work presented is to compare results from numerical optimization with experimental data and to highlight and discuss the differences between two fundamentally different optimization methods. The problem domain is minimization of acoustic emission and the structure used in the work is a closed cylinder with forced vibration of one end. The optimization method used in this paper is simulated annealing (SA), a stochastic method. The results are compared with those from a gradient-based method used on the same structure in an earlier paper (Tinnsten, 2000)", "keyphrases": ["numerical optimization", "acoustic emission minimization", "structure", "closed cylinder", "acoustic response", "forced vibration", "simulated annealing", "stochastic optimization", "gradient-based method"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "P", "P"]} {"id": "112", "title": "Revisiting Hardy's paradox: Counterfactual statements, real measurements, entanglement and weak values", "abstract": "Hardy's (1992) paradox is revisited. Usually the paradox is dismissed on grounds of counterfactuality, i.e., because the paradoxical effects appear only when one considers results of experiments which do not actually take place. We suggest a new set of measurements in connection with Hardy's scheme, and show that when they are actually performed, they yield strange and surprising outcomes. More generally, we claim that counterfactual paradoxes point to a deeper structure inherent to quantum mechanics", "keyphrases": ["Hardy paradox", "counterfactual statements", "real measurements", "entanglement", "gedanken-experiments", "weak values", "paradoxical effects", "quantum mechanics"], "prmu": ["R", "P", "P", "P", "U", "P", "P", "P"]} {"id": "951", "title": "How to drive strategic innovation [law firms]", "abstract": "Innovation. It has everything to do with organization and attitude. Marginal improvement isn't enough anymore. Convert your problem-solving skills into a new value for the entire firm. 10 initiatives", "keyphrases": ["law firms", "strategic innovation", "management", "change", "clients", "experiments"], "prmu": ["P", "P", "U", "U", "U", "U"]} -{"id": "914", "title": "A knowledge management framework for the support of decision making in humanitarian assistance/disaster relief", "abstract": "The major challenge in current humanitarian assistance/disaster relief (HA/DR) efforts is that diverse information and knowledge are widely distributed and owned by different organizations. These resources are not efficiently organized and utilized during HA/DR operations. We present a knowledge management framework that integrates multiple information technologies to collect, analyze, and manage information and knowledge for supporting decision making in HA/DR. The framework will help identify the information needs, be aware of a disaster situation, and provide decision-makers with useful relief recommendations based on past experience. A comprehensive, consistent and authoritative knowledge base within the framework will facilitate knowledge sharing and reuse. This framework can also be applied to other similar real-time decision-making environments, such as crisis management and emergency medical assistance", "keyphrases": ["knowledge management framework", "decision support system", "humanitarian assistance", "disaster relief", "organizations", "information technology", "information needs", "knowledge sharing", "knowledge reuse", "real-time decision-making environments", "crisis management", "emergency medical assistance", "case-based reasoning"], "prmu": ["P", "M", "R", "R", "P", "P", "P", "P", "R", "P", "P", "P", "U"]} +{"id": "914", "title": "A knowledge management framework for the support of decision making in humanitarian assistance/disaster relief", "abstract": "The major challenge in current humanitarian assistance/disaster relief (HA/DR) efforts is that diverse information and knowledge are widely distributed and owned by different organizations. These resources are not efficiently organized and utilized during HA/DR operations. We present a knowledge management framework that integrates multiple information technologies to collect, analyze, and manage information and knowledge for supporting decision making in HA/DR. The framework will help identify the information needs, be aware of a disaster situation, and provide decision-makers with useful relief recommendations based on past experience. A comprehensive, consistent and authoritative knowledge base within the framework will facilitate knowledge sharing and reuse. This framework can also be applied to other similar real-time decision-making environments, such as crisis management and emergency medical assistance", "keyphrases": ["knowledge management framework", "decision support system", "humanitarian assistance", "disaster relief", "organizations", "information technology", "information needs", "knowledge sharing", "knowledge reuse", "real-time decision-making environments", "crisis management", "emergency medical assistance", "case-based reasoning"], "prmu": ["P", "M", "P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "U"]} {"id": "586", "title": "A strategy for a payoff-switching differential game based on fuzzy reasoning", "abstract": "In this paper, a new concept of a payoff-switching differential game is introduced. In this new game, any one player at any time may have several choices of payoffs for the future. Moreover, the payoff-switching process, including the time of payoff switching and the outcome payoff, of any one player is unknown to the other. Indeed, the overall payoff, which is a sequence of several payoffs, is unknown until the game ends. An algorithm for determining a reasoning strategy based on fuzzy reasoning is proposed. In this algorithm, the fuzzy theory is used to estimate the behavior of one player during a past time interval. By deriving two fuzzy matrices GSM, game similarity matrix, and VGSM, variation of GSM, the behavior of the player can be quantified. Two weighting vectors are selected to weight the relative importance of the player's behavior at each past time instant. Finally a simple fuzzy inference rule is adopted to generate a linear reasoning strategy. The advantage of this algorithm is that it provides a flexible way for differential game specialists to convert their knowledge into a \"reasonable\" strategy. A practical example of guarding three territories is given to illustrate our main ideas", "keyphrases": ["payoff-switching differential game", "payoff switching", "outcome payoff", "reasoning strategy", "fuzzy reasoning", "fuzzy matrices", "game similarity matrix", "weighting vectors", "fuzzy inference", "differential game"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1196", "title": "Multiple shooting using a dichotomically stable integrator for solving differential-algebraic equations", "abstract": "In previous work by the first author, it has been established that a dichotomically stable discretization is needed when solving a stiff boundary-value problem in ordinary differential equations (ODEs), when sharp boundary layers may occur at each end of the interval. A dichotomically stable implicit Runge-Kutta method, using the 3-stage, fourth-order, Lobatto IIIA formulae, has been implemented in a variable step-size initial-value integrator, which could be used in a multiple-shooting approach. In the case of index-one differential-algebraic equations (DAEs) the use of the Lobatto IIIA formulae has an advantage, over a comparable Gaussian method, that the order is the same for both differential and algebraic variables, and there is no need to treat them separately. The ODE integrator has been adapted for the solution of index-one DAEs, and the resulting integrator (SYMDAE) has been inserted into the multiple-shooting code (MSHDAE) previously developed by R. Lamour for differential-algebraic boundary-value problems. The standard version of MSHDAE uses a BDF integrator, which is not dichotomically stable, and for some stiff test problems this fails to integrate across the interval of interest, while the dichotomically stable integrator SYMDAE encounters no difficulty. Indeed, for such problems, the modified version of MSHDAE produces an accurate solution, and within limits imposed by computer word length, the efficiency of the solution process improves with increasing stiffness. For some nonstiff problems, the solution is also entirely satisfactory", "keyphrases": ["multiple shooting", "stiff boundary-value problem", "ordinary differential equations", "implicit Runge-Kutta method", "Lobatto IIIA formulae", "initial-value integrator", "dichotomically stable integrator", "differential-algebraic equations"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "809", "title": "Edison's direct current influenced \"Broadway\" show lighting", "abstract": "During the early decades of the 20 th century, midtown Manhattan in New York City developed an extensive underground direct current (DC) power distribution system. This was a result of the original introduction of direct current by Thomas Edison's pioneering Pearl Street Station in 1882. The availability of DC power in the theater district, led to the perpetuation of an archaic form of stage lighting control through nearly three-quarters of the 20 th century. This control device was known as a \"resistance dimmer.\" It was essentially a series-connected rheostat, but it was wound with a special resistance \"taper\" so as to provide a uniform change in the apparent light output of typical incandescent lamps throughout the travel of its manually operated arm. The development and use of DC powered stage lighting is discussed in this article", "keyphrases": ["Broadway show lighting", "Manhattan", "New York City", "underground direct current power distribution system", "Thomas Edison's Pearl Street Station", "theater district", "stage lighting control", "resistance dimmer", "series-connected rheostat", "resistance taper", "apparent light output", "incandescent lamps", "DC powered stage lighting"], "prmu": ["R", "P", "P", "R", "R", "P", "P", "P", "P", "R", "P", "P", "P"]} @@ -164,7 +164,7 @@ {"id": "1316", "title": "Understanding Internet traffic streams: dragonflies and tortoises", "abstract": "We present the concept of network traffic streams and the ways they aggregate into flows through Internet links. We describe a method of measuring the size and lifetime of Internet streams, and use this method to characterize traffic distributions at two different sites. We find that although most streams (about 45 percent of them) are dragonflies, lasting less than 2 seconds, a significant number of streams have lifetimes of hours to days, and can carry a high proportion (50-60 percent) of the total bytes on a given link. We define tortoises as streams that last longer than 15 minutes. We point out that streams can be classified not only by lifetime (dragonflies and tortoises) but also by size (mice and elephants), and note that stream size and lifetime are independent dimensions. We submit that ISPs need to be aware of the distribution of Internet stream sizes, and the impact of the difference in behavior between short and long streams. In particular, any forwarding cache mechanisms in Internet routers must be able to cope with a high volume of short streams. In addition ISPs should realize that long-running streams can contribute a significant fraction of their packet and byte volumes-something they may not have allowed for when using traditional \"flat rate user bandwidth consumption\" approaches to provisioning and engineering", "keyphrases": ["Internet traffic streams", "dragonflies", "tortoises", "network traffic streams", "Internet stream size measurement", "Internet stream lifetime measurement", "traffic distributions", "mice", "elephants", "ISP", "forwarding cache mechanisms", "Internet routers", "long-running streams", "packet volume", "byte volume", "traffic provisioning", "traffic engineering"], "prmu": ["P", "P", "P", "P", "R", "R", "P", "P", "P", "P", "P", "P", "P", "R", "R", "R", "R"]} {"id": "868", "title": "Two quantum analogues of Fisher information from a large deviation viewpoint of quantum estimation", "abstract": "We discuss two quantum analogues of the Fisher information, the symmetric logarithmic derivative Fisher information and Kubo-Mori-Bogoljubov Fisher information from a large deviation viewpoint of quantum estimation and prove that the former gives the true bound and the latter gives the bound of consistent superefficient estimators. As another comparison, it is shown that the difference between them is characterized by the change of the order of limits", "keyphrases": ["quantum analogues", "quantum estimation", "Kubo-Mori-Bogoljubov Fisher information", "consistent superefficient estimators", "statistical inference", "large deviation viewpoint", "symmetric logarithmic derivative Fisher information"], "prmu": ["P", "P", "P", "P", "U", "P", "P"]} {"id": "1092", "title": "Ride quality evaluation of an actively-controlled stretcher for an ambulance", "abstract": "This study considers the subjective evaluation of ride quality during ambulance transportation using an actively-controlled stretcher (ACS). The ride quality of a conventional stretcher and an assistant driver's seat is also compared. Braking during ambulance transportation generates negative foot-to-head acceleration in patients and causes blood pressure to rise in the patient's head. The ACS absorbs the foot-to-head acceleration by changing the angle of the stretcher, thus reducing the blood pressure variation. However, the ride quality of the ACS should be investigated further because the movement of the ACS may cause motion sickness and nausea. Experiments of ambulance transportation, including rapid acceleration and deceleration, are performed to evaluate the effect of differences in posture of the transported subject on the ride quality; the semantic differential method and factor analysis are used in the investigations. Subjects are transported using a conventional stretcher with head forward, a conventional stretcher with head backward, the ACS, and an assistant driver's seat for comparison with transportation using a stretcher. Experimental results show that the ACS gives the most comfortable transportation when using a stretcher. Moreover, the reduction of the negative foot-to-head acceleration at frequencies below 0.2 Hz and the small variation of the foot-to-head acceleration result in more comfortable transportation. Conventional transportation with the head forward causes the worst transportation, although the characteristics of the vibration of the conventional stretcher seem to be superior to that of the ACS", "keyphrases": ["actively-controlled stretcher", "ambulance", "ride quality evaluation", "subjective evaluation", "ambulance transportation", "conventional stretcher", "assistant driver seat", "braking", "negative foot-to-head acceleration", "blood pressure variation", "patient head", "stretcher angle", "motion sickness", "nausea", "rapid acceleration", "rapid deceleration", "posture differences", "transported subject", "semantic differential method", "factor analysis", "head forward", "head backward", "comfortable transportation", "vibration"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "R", "R", "P", "P", "P", "R", "R", "P", "P", "P", "P", "P", "P", "P"]} -{"id": "1443", "title": "C and C++: a case for compatibility", "abstract": "Modern C and C++ are sibling languages descended from Classic C. In many people's minds, they are (wrongly, but understandably) fused into the mythical C/C++ programming language. There is no C/C++ language, but there is a C/C++ community. Previously the author described some of the incompatibilities that complicate the work of developers within that C/C++ community. In this article, he discusses some of the underlying myths that help perpetuate these incompatibilities. He also shows why more compatibility (ideally, full compatibility) is in the best interest of the C/C++ community. In the next paper, he presents some examples of how the incompatibilities in C and C++ might be resolved", "keyphrases": ["C++ language", "C language", "incompatibilities", "object-oriented programming", "class hierarchies", "low-level programming", "C++ libraries"], "prmu": ["R", "R", "P", "M", "U", "M", "M"]} +{"id": "1443", "title": "C and C++: a case for compatibility", "abstract": "Modern C and C++ are sibling languages descended from Classic C. In many people's minds, they are (wrongly, but understandably) fused into the mythical C/C++ programming language. There is no C/C++ language, but there is a C/C++ community. Previously the author described some of the incompatibilities that complicate the work of developers within that C/C++ community. In this article, he discusses some of the underlying myths that help perpetuate these incompatibilities. He also shows why more compatibility (ideally, full compatibility) is in the best interest of the C/C++ community. In the next paper, he presents some examples of how the incompatibilities in C and C++ might be resolved", "keyphrases": ["C++ language", "C language", "incompatibilities", "object-oriented programming", "class hierarchies", "low-level programming", "C++ libraries"], "prmu": ["P", "R", "P", "M", "U", "M", "M"]} {"id": "1406", "title": "Bluetooth bites back", "abstract": "It is now more than four years since we started to hear about Bluetooth, and from the user's point of view very little seems to have happened since then. Paul Haddlesey looks at the progress, and the role Bluetooth may eventually play in your firm's communications strategy", "keyphrases": ["Bluetooth", "communications strategy", "wireless connection", "mobile"], "prmu": ["P", "P", "U", "U"]} {"id": "810", "title": "Oracle's Suite grows up", "abstract": "Once a low-cost Web offering, Oracle's Small Business Suite now carries a price tag to justify VAR interest", "keyphrases": ["Oracle Small Business Suite", "NetLedger", "accounting", "resellers"], "prmu": ["R", "U", "U", "U"]} {"id": "855", "title": "Support communities for women in computing", "abstract": "This article highlights the many activities provided by the support communities available for women in computing. Thousands of women actively participate in these programs and they receive many benefits including networking and professional support. In addition, the organizations and associations help promote the accomplishments of women computer scientists and disseminate valuable information. This article surveys some of these organizations and concludes with a list of suggestions for how faculty members can incorporate the benefits of these organizations in their own institutions", "keyphrases": ["support communities", "women", "computing", "networking", "professional support", "information dissemination", "faculty members"], "prmu": ["P", "P", "P", "P", "P", "R", "P"]} @@ -227,7 +227,7 @@ {"id": "1031", "title": "Noise-constrained hyperspectral data compression", "abstract": "Storage and transmission requirements for hyperspectral data sets are significant. To reduce hardware costs, well-designed compression techniques are needed to preserve information content while maximizing compression ratios. Lossless compression techniques maintain data integrity, but yield small compression ratios. We present a slightly lossy compression algorithm that uses the noise statistics of the data to preserve information content while maximizing compression ratios. The adaptive principal components analysis (APCA) algorithm uses noise statistics to determine the number of significant principal components and selects only those that are required to represent each pixel to within the noise level. We demonstrate the effectiveness of these methods with airborne visible/infrared spectrometer (AVIRIS), hyperspectral digital imagery collection experiment (HYDICE), hyperspectral mapper (HYMAP), and Hyperion datasets", "keyphrases": ["noise-constrained hyperspectral data compression", "storage requirements", "transmission requirements", "hyperspectral data sets", "hardware costs", "information content", "compression ratios", "lossless compression techniques", "data integrity", "slightly lossy compression algorithm", "noise statistics", "adaptive principal components analysis algorithm", "noise level", "airborne visible/infrared spectrometer hyperspectral digital imagery collection experiment", "AVIRIS HYDICE", "hyperspectral mapper", "HYMAP", "Hyperion datasets", "Gaussian statistics"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "P", "R", "R", "P", "P", "P", "M"]} {"id": "798", "title": "ClioWeb, ClioRequest, and Clio database: enhancing patron and staff satisfaction", "abstract": "Faced with increased demand from students and faculty for a speedier and more user-friendly method of obtaining materials from other institutions, the interlibrary loan (ILL) department sought to implement a management system which would accomplish the task. Students wanted remote interconnectivity to the system and staff wanted increased workflow efficiency, reduced paper work, and better data management. This paper focuses on Washington College's experience in selecting and implementing an interlibrary loan system, which would enhance student satisfaction as well as that of the library staff", "keyphrases": ["Clio database", "ClioRequest", "ClioWeb", "staff satisfaction", "patron satisfaction", "faculty", "students", "interlibrary loan department", "user-friendly method", "management system", "remote interconnectivity", "workflow efficiency", "data management", "Washington College"], "prmu": ["P", "P", "P", "P", "R", "P", "P", "R", "P", "P", "P", "P", "P", "P"]} {"id": "765", "title": "Simulating fermions on a quantum computer", "abstract": "The real-time probabilistic simulation of quantum systems in classical computers is known to be limited by the so-called dynamical sign problem, a problem leading to exponential complexity. In 1981 Richard Feynman raised some provocative questions in connection to the \"exact imitation\" of such systems using a special device named a \"quantum computer\". Feynman hesitated about the possibility of imitating fermion systems using such a device. Here we address some of his concerns and, in particular, investigate the simulation of fermionic systems. We show how quantum computers avoid the sign problem in some cases by reducing the complexity from exponential to polynomial. Our demonstration is based upon the use of isomorphisms of algebras. We present specific quantum algorithms that illustrate the main points of our algebraic approach", "keyphrases": ["quantum computer", "fermions simulation", "real-time probabilistic simulation", "classical computers", "dynamical sign problem", "exponential complexity", "fermion systems", "sign problem", "isomorphisms", "algebras"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "P", "P", "P"]} -{"id": "720", "title": "19in monitors [CRT survey]", "abstract": "Upgrade your monitor from as little as Pounds 135. With displays on test and ranging up to Pounds 400, whether you're after the last word in quality or simply looking for again, this Labs holds the answer. Looks at ADI MicroScan M900, CTX PR960F, Eizo FlexScan T766, Hansol 920D, Hansol920P, Hitachi CM715ET, Hitachi CM721FET, liyama Vision Master Pro 454, LG Flatron 915FT Plus, Mitsubishi Diamond Pro 920, NEC MultiSync FE950+, Philips 109S40, Samsung SyncMaster 959NF, Sony Multiscan CPD-G420, and ViewSonic G90f", "keyphrases": ["19in monitors", "CRT survey", "ADI MicroScan M900", "CTX PR960F", "Eizo FlexScan T766", "Hansol 920D", "Hansol920P", "Hitachi CM715ET", "Hitachi CM721FET", "liyama Vision Master Pro 454", "LG Flatron 915FT Plus", "Mitsubishi Diamond Pro 920", "NEC MultiSync FE950", "Philips 109S40", "Samsung SyncMaster 959NF", "Sony Multiscan CPD-G420", "ViewSonic G90f", "19 in"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M", "P", "P", "P", "P", "M"]} +{"id": "720", "title": "19in monitors [CRT survey]", "abstract": "Upgrade your monitor from as little as Pounds 135. With displays on test and ranging up to Pounds 400, whether you're after the last word in quality or simply looking for again, this Labs holds the answer. Looks at ADI MicroScan M900, CTX PR960F, Eizo FlexScan T766, Hansol 920D, Hansol920P, Hitachi CM715ET, Hitachi CM721FET, liyama Vision Master Pro 454, LG Flatron 915FT Plus, Mitsubishi Diamond Pro 920, NEC MultiSync FE950+, Philips 109S40, Samsung SyncMaster 959NF, Sony Multiscan CPD-G420, and ViewSonic G90f", "keyphrases": ["19in monitors", "CRT survey", "ADI MicroScan M900", "CTX PR960F", "Eizo FlexScan T766", "Hansol 920D", "Hansol920P", "Hitachi CM715ET", "Hitachi CM721FET", "liyama Vision Master Pro 454", "LG Flatron 915FT Plus", "Mitsubishi Diamond Pro 920", "NEC MultiSync FE950", "Philips 109S40", "Samsung SyncMaster 959NF", "Sony Multiscan CPD-G420", "ViewSonic G90f", "19 in"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1375", "title": "Evaluation of the usability of digital maintenance manuals developed without either user input or a task analysis", "abstract": "The primary objective was to investigate the value that can be added to a low-cost digital maintenance manual by the addition of a navigational aid. Two versions of a digital maintenance manual were developed, the difference between them being the number of design heuristics observed when designing navigational aids. Neither version was based on an analysis of the tasks carried out by users, nor were users involved in the design process. Instead, the manuals were developed directly from the digital information used to produce the existing paper manual. Usability trials were carried out to test both versions according to the time taken and errors committed by users during typical information retrieval tasks. Users were questioned to determine their ease of use (EOU) perceptions for each manual. The main outcomes were that the navigation aid used in the second version reduced the time taken to use the manual but increased the number of errors made by users. The navigational aid also seemed to reduce the perceived EOU compared with the first version. In both cases, the perceived EOU was lower than for a previous digital manual that had been developed using a task analysis and user input. The paper concludes by recommending the development of a generic task model of user interaction with digital maintenance manuals", "keyphrases": ["digital maintenance manuals usability", "navigational aid", "usability trials", "information retrieval", "generic task model", "user interaction", "task analysis"], "prmu": ["R", "P", "P", "P", "P", "P", "P"]} {"id": "1330", "title": "Strobbe Graphics' next frontier: CTP for commercial printers", "abstract": "Strobbe is one of the more successful makers of newspaper platesetters, which are sold by Agfa under the Polaris name. But the company also has a growing presence in commercial printing markets, where it sells under its own name", "keyphrases": ["Strobbe Graphics", "Punch International", "Agfa", "commercial printing", "Polaris", "platesetters", "workflow"], "prmu": ["P", "U", "P", "P", "P", "P", "U"]} {"id": "1458", "title": "Direct gear tooth contact analysis for hypoid bevel gears", "abstract": "A new methodology for tooth contact analysis based on a very general mathematical model of the generating process is proposed. Considering the line of action as a first order singularity of a certain operator equation we develop first and second order conditions for a pair of generated gear tooth flanks to be in contact. The constructive approach allows the direct computation of the paths of contact as the solution of a nonlinear equation system including the exact determination of the bounds of the paths of contact. The transmission error as well as curvature properties in the contact points are obtained in a convenient way. The resulting contact ellipses approximate the bearing area. Through the use of automatic differentiation all the geometric quantities are calculable within the machine accuracy of the computer", "keyphrases": ["direct gear tooth contact analysis", "hypoid bevel gears", "mathematical model", "generating process", "first order singularity", "operator equation", "second order conditions", "first order conditions", "generated gear tooth flanks", "contact paths", "nonlinear equation system", "exact bound determination", "transmission error", "curvature properties", "contact ellipses", "bearing area", "automatic differentiation", "geometric quantities", "machine accuracy", "computer"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "R", "P", "R", "P", "R", "P", "P", "P", "P", "P", "P", "P", "P"]} @@ -239,7 +239,7 @@ {"id": "873", "title": "Programmatic efforts encouraging women to enter the information technology workforce", "abstract": "For over a decade the National Science Foundation (NSF) has been supporting projects designed to improve opportunities for women in computing. From an initial emphasis on increasing the number of women in graduate school studying computer science and engineering, NSF's current emphasis has broadened to include research studies examining the underlying reasons why women are underrepresented in the information technology (IT) workforce. This paper describes the recent history of NSF's activities in this area and the subsequent emergence of a research portfolio addressing the underrepresentation issue", "keyphrases": ["National Science Foundation", "women", "computing", "graduate school", "engineering", "IT workforce", "history"], "prmu": ["P", "P", "P", "P", "P", "R", "P"]} {"id": "1049", "title": "A typed representation for HTML and XML documents in Haskell", "abstract": "We define a family of embedded domain specific languages for generating HTML and XML documents. Each language is implemented as a combinator library in Haskell. The generated HTML/XML documents are guaranteed to be well-formed. In addition, each library can guarantee that the generated documents are valid XML documents to a certain extent (for HTML only a weaker guarantee is possible). On top of the libraries, Haskell serves as a meta language to define parameterized documents, to map structured documents to HTML/XML, to define conditional content, or to define entire Web sites. The combinator libraries support element-transforming style, a programming style that allows programs to have a visual appearance similar to HTML/XML documents, without modifying the syntax of Haskell", "keyphrases": ["typed representation", "HTML documents", "combinator library", "software libraries", "meta language", "parameterized documents", "conditional content", "Web sites", "element-transforming style", "functional programming", "syntax", "XML documents", "Haskell", "embedded domain specific languages"], "prmu": ["P", "R", "P", "M", "P", "P", "P", "P", "P", "M", "P", "P", "P", "P"]} {"id": "1111", "title": "The contiguity in R/M", "abstract": "An r.e. degree c is contiguous if deg/sub wtt/(A)=deg/sub wtt/(B) for any r.e. sets A,B in c. In this paper, we generalize the notation of contiguity to the structure R/M, the upper semilattice of the r.e. degree set R modulo the cappable r.e. degree set M. An element [c] in R/M is contiguous if [deg/sub wtt/(A)]=[deg/sub wtt/(B)] for any r.e. sets A, B such that deg/sub T/(A),deg/sub T/(B) in [c]. It is proved in this paper that every nonzero element in R/M is not contiguous, i.e., for every element [c] in R/M, if [c] not=[o] then there exist at least two r.e. sets A, B such that deg/sub T/(A), deg/sub T/(B) in [c] and [deg/sub wtt/(A)] not=[deg/sub wtt/(B)]", "keyphrases": ["contiguity", "Turing degree", "recursively enumerable set", "upper semilattice", "nonzero element", "recursion theory"], "prmu": ["P", "M", "M", "P", "P", "U"]} -{"id": "1154", "title": "The effect of voxel size on the accuracy of dose-volume histograms of prostate /sup 125/I seed implants", "abstract": "Cumulative dose-volume histograms (DVH) are crucial in evaluating the quality of radioactive seed prostate implants. When calculating DVHs, the choice of voxel size is a compromise between computational speed (larger voxels) and accuracy (smaller voxels). We quantified the effect of voxel size on the accuracy of DVHs using an in-house computer program. The program was validated by comparison with a hand-calculated DVH for a single 0.4-U iodine-125 model 6711 seed. We used the program to find the voxel size required to obtain accurate DVHs of five iodine-125 prostate implant patients at our institution. One-millimeter cubes were sufficient to obtain DVHs that are accurate within 5% up to 200% of the prescription dose. For the five patient plans, we obtained good agreement with the VariSeed (version 6.7, Varian, USA) treatment planning software's DVH algorithm by using voxels with a sup-inf dimension equal to the spacing between successive transverse seed implant planes (5 mm). The volume that receives at least 200% of the target dose, V/sub 200/, calculated by VariSeed was 30% to 43% larger than that calculated by our program with small voxels. The single-seed DVH calculated by VariSeed fell below the hand calculation by up to 50% at low doses (30 Gy), and above it by over 50% at high doses (>250 Gy)", "keyphrases": ["cumulative dose-volume histograms", "prostate /sup 125/I seed implants", "radioactive seed prostate implants", "voxel size", "computational speed", "in-house computer program", "hand-calculated dose-volume histograms", "single-seed dose-volume histograms", "/sup 125/I model", "/sup 125/I prostate implant patients", "VariSeed treatment planning software's dose-volume histogram algorithm", "I"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "R", "R", "R", "R", "U"]} +{"id": "1154", "title": "The effect of voxel size on the accuracy of dose-volume histograms of prostate /sup 125/I seed implants", "abstract": "Cumulative dose-volume histograms (DVH) are crucial in evaluating the quality of radioactive seed prostate implants. When calculating DVHs, the choice of voxel size is a compromise between computational speed (larger voxels) and accuracy (smaller voxels). We quantified the effect of voxel size on the accuracy of DVHs using an in-house computer program. The program was validated by comparison with a hand-calculated DVH for a single 0.4-U iodine-125 model 6711 seed. We used the program to find the voxel size required to obtain accurate DVHs of five iodine-125 prostate implant patients at our institution. One-millimeter cubes were sufficient to obtain DVHs that are accurate within 5% up to 200% of the prescription dose. For the five patient plans, we obtained good agreement with the VariSeed (version 6.7, Varian, USA) treatment planning software's DVH algorithm by using voxels with a sup-inf dimension equal to the spacing between successive transverse seed implant planes (5 mm). The volume that receives at least 200% of the target dose, V/sub 200/, calculated by VariSeed was 30% to 43% larger than that calculated by our program with small voxels. The single-seed DVH calculated by VariSeed fell below the hand calculation by up to 50% at low doses (30 Gy), and above it by over 50% at high doses (>250 Gy)", "keyphrases": ["cumulative dose-volume histograms", "prostate /sup 125/I seed implants", "radioactive seed prostate implants", "voxel size", "computational speed", "in-house computer program", "hand-calculated dose-volume histograms", "single-seed dose-volume histograms", "/sup 125/I model", "/sup 125/I prostate implant patients", "VariSeed treatment planning software's dose-volume histogram algorithm", "I"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "R", "R", "R", "R", "P"]} {"id": "993", "title": "A large deviations analysis of the transient of a queue with many Markov fluid inputs: approximations and fast simulation", "abstract": "This article analyzes the transient buffer content distribution of a queue fed by a large number of Markov fluid sources. We characterize the probability of overflow at time t, given the current buffer level and the number of sources in the on-state. After scaling buffer and bandwidth resources by the number of sources n, we can apply large deviations techniques. The transient overflow probability decays exponentially in n. In the case of exponential on/off sources, we derive an expression for the decay rate of the rare event probability under consideration. For general Markov fluid sources, we present a plausible conjecture. We also provide the \"most likely path\" from the initial state to overflow (at time t). Knowledge of the decay rate and the most likely path to overflow leads to (i) approximations of the transient overflow probability and (ii) efficient simulation methods of the rare event of buffer overflow. The simulation methods, based on importance sampling, give a huge speed-up compared to straightforward simulations. The approximations are of low computational complexity and are accurate, as verified by means of simulation experiments", "keyphrases": ["large deviations analysis", "Markov fluid inputs", "transient buffer content distribution", "buffer resources", "bandwidth resources", "approximations", "transient overflow probability", "simulation methods", "importance sampling", "computational complexity", "ATM multiplexers", "IP routers", "queuing theory"], "prmu": ["P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "U", "U", "U"]} {"id": "544", "title": "Virtual reality treatment of flying phobia", "abstract": "Flying phobia (FP) might become a very incapacitating and disturbing problem in a person's social, working, and private areas. Psychological interventions based on exposure therapy have proved to be effective, but given the particular nature of this disorder they bear important limitations. Exposure therapy for FP might be excessively costly in terms of time, money, and efforts. Virtual reality (VR) overcomes these difficulties as different significant environments might be created, where the patient can interact with what he or she fears while in a totally safe and protected environment, the therapist's consulting room. This paper intends, on one hand, to show the different scenarios designed by our team for the VR treatment of FP, and on the other, to present the first results supporting the effectiveness of this new tool for the treatment of FP in a multiple baseline study", "keyphrases": ["medical virtual reality", "psychology", "flying phobia", "patient treatment", "psychological interventions", "anxiety disorders", "virtual exposure", "exposure therapy"], "prmu": ["M", "P", "P", "R", "P", "M", "R", "P"]} {"id": "82", "title": "Bit-serial AB/sup 2/ multiplier using modified inner product", "abstract": "This paper presents a new multiplication algorithm and, based on this algorithm, proposes a hardware architecture, called modified inner-product multiplier (MIPM), which computes AB/sup 2/ multiplication based on a linear feedback shift register (LFSR). The algorithm is based on the property of the irreducible all one polynomial (AOP) over the finite field GF(2/sup m/). The proposed architecture reduces the time and space complexity for computing AB/sup 2/. The proposed architecture has a potential application to implementing exponentiation architecture for a public-key cryptosystem", "keyphrases": ["bit-serial AB/sup 2/ multiplier", "modified inner product", "multiplication algorithm", "hardware architecture", "modified inner-product multiplier", "linear feedback shift register", "irreducible all one polynomial", "space complexity", "time complexity", "public-key cryptosystem"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "R", "P"]} @@ -257,7 +257,7 @@ {"id": "1276", "title": "A comparative study of some generalized rough approximations", "abstract": "In this paper we focus upon a comparison of some generalized rough approximations of sets, where the classical indiscernibility relation is generalized to any binary reflexive relation. We aim at finding the best of several candidates for generalized rough approximation mappings, where both definability of sets by elementary granules of information as well as the issue of distinction among positive, negative, and border regions of a set are taken into account", "keyphrases": ["generalized rough approximations", "classical indiscernibility relation", "binary reflexive relation", "generalized rough approximation mappings", "elementary granules"], "prmu": ["P", "P", "P", "P", "P"]} {"id": "1233", "title": "Advanced optimization strategies in the Rice dHPF compiler", "abstract": "High-Performance Fortran (HPF) was envisioned as a vehicle for modernizing legacy Fortran codes to achieve scalable parallel performance. To a large extent, today's commercially available HPF compilers have failed to deliver scalable parallel performance for a broad spectrum of applications because of insufficiently powerful compiler analysis and optimization. Substantial restructuring and hand-optimization can be required to achieve acceptable performance with an HPF port of an existing Fortran application, even for regular data-parallel applications. A key goal of the Rice dHPF compiler project has been to develop optimization techniques that enable a wide range of existing scientific applications to be ported easily to efficient HPF with minimal restructuring. This paper describes the challenges to effective parallelization presented by complex (but regular) data-parallel applications, and then describes how the novel analysis and optimization technologies in the dHPF compiler address these challenges effectively, without major rewriting of the applications. We illustrate the techniques by describing their use for parallelizing the NAS SP and BT benchmarks. The dHPF compiler generates multipartitioned parallelizations of these codes that are approaching the scalability and efficiency of sophisticated hand-coded parallelizations", "keyphrases": ["Mgh-Performance Fortran", "legacy Fortran codes", "parallel performance", "HPF compilers", "compiler analysis", "compiler optimization", "Rice dHPF compiler", "multipartitioning", "automatic parallelization"], "prmu": ["M", "P", "P", "P", "P", "R", "P", "P", "M"]} {"id": "623", "title": "Stochastic recurrences of Jackpot Keno", "abstract": "We describe a mathematical model and simulation study for Jackpot Keno, as implemented by Jupiters Network Gaming (JNG) in the Australian state of Queensland, and as controlled by the Queensland Office of Gaming Regulation (QOGR) (http://www.qogr.qld.gov.au/keno.shtml). The recurrences for the house net hold are derived and it is seen that these are piecewise linear with a ternary domain split, and further, the split points are stochastic in nature. Since this structure is intractable (Brockett and Levine, Statistics & Probability & their Applications, CBS College Publishing, 1984), estimation of house net hold obtained through an appropriately designed simulator using a random number generator with desirable properties is described. Since the model and simulation naturally derives hold given payscale, but JNG and QOGR require payscale given hold, an inverse problem was required to be solved. This required development of a special algorithm, which may be described as a stochastic binary search. Experimental results are presented, in which the simulator is used to determine jackpot pay-scales so as to satisfy legal requirements of approximately 75% of net revenue returned to the players, i.e., 25% net hold for the house (JNG). Details of the algorithm used to solve this problem are presented, and notwithstanding the stochastic nature of the simulation, convergence to a specified hold for the inverse problem has been achieved to within 0.1% in all cases of interest to date", "keyphrases": ["stochastic recurrences", "Jackpot Keno", "mathematical model", "simulation", "Jupiters Network Gaming", "house net hold", "piecewise linear", "ternary domain split", "random number generator", "inverse problem", "stochastic binary search", "probability", "experimental results", "legal requirement", "Chinese lottery game"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M"]} -{"id": "908", "title": "Multivariable H/sub infinity // mu feedback control design for high-precision wafer stage motion", "abstract": "Conventional PID-like SISO controllers are still the most common in industry, but with performance requirements becoming tighter there is a growing need for advanced controllers. For the positioning devices in IC-manufacturing, plant interaction is a major performance-limiting factor. MIMO control can be invoked to tackle this problem. A practically feasible procedure is presented to design MIMO feedback controllers for electromechanical positioning devices, using H/sub infinity // mu techniques. Weighting filters are proposed to straightforwardly and effectively impose performance and uncertainty specifications. Experiments show that MIMO control can considerably improve upon the performance with multiloop SISO control. Some problems are highlighted that are important for industrial practice, but lacking a workable solution", "keyphrases": ["IC manufacture", "multivariable control systems", "weighting filters", "MIMO systems", "H/sub infinity / control", "feedback", "servo systems", "model uncertainty", "motion control", "mechatronics", "mu synthesis"], "prmu": ["U", "M", "P", "M", "M", "P", "U", "M", "R", "U", "M"]} +{"id": "908", "title": "Multivariable H/sub infinity // mu feedback control design for high-precision wafer stage motion", "abstract": "Conventional PID-like SISO controllers are still the most common in industry, but with performance requirements becoming tighter there is a growing need for advanced controllers. For the positioning devices in IC-manufacturing, plant interaction is a major performance-limiting factor. MIMO control can be invoked to tackle this problem. A practically feasible procedure is presented to design MIMO feedback controllers for electromechanical positioning devices, using H/sub infinity // mu techniques. Weighting filters are proposed to straightforwardly and effectively impose performance and uncertainty specifications. Experiments show that MIMO control can considerably improve upon the performance with multiloop SISO control. Some problems are highlighted that are important for industrial practice, but lacking a workable solution", "keyphrases": ["IC manufacture", "multivariable control systems", "weighting filters", "MIMO systems", "H/sub infinity / control", "feedback", "servo systems", "model uncertainty", "motion control", "mechatronics", "mu synthesis"], "prmu": ["U", "M", "P", "M", "R", "P", "U", "M", "R", "U", "M"]} {"id": "1177", "title": "Comparative statistical analysis of hole taper and circularity in laser percussion drilling", "abstract": "Investigates the relationships and parameter interactions between six controllable variables on the hole taper and circularity in laser percussion drilling. Experiments have been conducted on stainless steel workpieces and a comparison was made between stainless steel and mild steel. The central composite design was employed to plan the experiments in order to achieve required information with reduced number of experiments. The process performance was evaluated. The ratio of minimum to maximum Feret's diameter was considered as circularity characteristic of the hole. The models of these three process characteristics were developed by linear multiple regression technique. The significant coefficients were obtained by performing analysis of variance (ANOVA) at 1, 5 and 7% levels of significance. The final models were checked by complete residual analysis and finally were experimentally verified. It was found that the pulse frequency had a significant effect on the hole entrance diameter and hole circularity in drilling stainless steel unlike the drilling of mild steel where the pulse frequency had no significant effect on the hole characteristics", "keyphrases": ["comparative statistical analysis", "hole taper", "circularity", "laser percussion drilling", "stainless steel workpieces", "mild steel", "laser peak power", "laser pulse width", "pulse frequency", "assist gas pressure", "focal plane position", "central composite design", "process performance", "equivalent entrance diameter", "Ferets diameter", "linear multiple regression technique", "least squares procedure", "stepwise regression method", "analysis of variance", "ANOVA", "complete residual analysis"], "prmu": ["P", "P", "P", "P", "P", "P", "M", "M", "P", "U", "U", "P", "P", "M", "R", "P", "U", "M", "P", "P", "P"]} {"id": "1132", "title": "Semidefinite programming vs. LP relaxations for polynomial programming", "abstract": "We consider the global minimization of a multivariate polynomial on a semi-algebraic set Omega defined with polynomial inequalities. We then compare two hierarchies of relaxations, namely, LP relaxations based on products of the original constraints, in the spirit of the RLT procedure of Sherali and Adams (1990), and recent semidefinite programming (SDP) relaxations introduced by the author. The comparison is analyzed in light of recent results in real algebraic geometry on various representations of polynomials, positive on a compact semi-algebraic set", "keyphrases": ["polynomial programming", "semidefinite programming relaxations", "LP relaxations", "global minimization", "multivariate polynomial", "polynomial inequalities", "real algebraic geometry", "reformulation linearization technique", "semi-algebraic set", "constraint products", "RLT procedure"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "U", "P", "R", "P"]} {"id": "567", "title": "Hidden Markov model-based tool wear monitoring in turning", "abstract": "This paper presents a new modeling framework for tool wear monitoring in machining processes using hidden Markov models (HMMs). Feature vectors are extracted from vibration signals measured during turning. A codebook is designed and used for vector quantization to convert the feature vectors into a symbol sequence for the hidden Markov model. A series of experiments are conducted to evaluate the effectiveness of the approach for different lengths of training data and observation sequence. Experimental results show that successful tool state detection rates as high as 97% can be achieved by using this approach", "keyphrases": ["tool wear monitoring", "machining processes", "hidden Markov models", "vibration signals", "codebook", "vector quantization", "feature extraction", "tool state detection", "turning process", "HMM training", "discrete wavelet transform"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "P", "R", "R", "U"]} @@ -273,7 +273,7 @@ {"id": "890", "title": "Multiple criteria decision making without optimization", "abstract": "We present a development intended to make interactive decision making schemes accessible for a wider spectrum of decision makers. To this aim we propose to eliminate the need to solve optimization problems at successive iterations of interactive decision processes. We show that the need for optimization can be eliminated by the ability of establishing sufficiently tight bounds on criteria values for efficient decisions prior to explicit identification of such decisions. We present a technique, fully operational and numerically simple, for establishing such bounds. Bounds are dynamic, i.e., they become stronger with the growing number of decisions explicitly identified. They are also parametric with respect to weighting coefficients. We also point out how this technique can enhance the existing interactive decision making methods", "keyphrases": ["multiple criteria decision making", "interactive decision making schemes", "decision makers", "criteria values bounds", "efficient decisions"], "prmu": ["P", "P", "P", "R", "P"]} {"id": "1012", "title": "Evolving receptive-field controllers for mobile robots", "abstract": "The use of evolutionary methods to generate controllers for real-world autonomous agents has attracted attention. Most of the pertinent research has employed genetic algorithms or variations thereof. Research has applied an alternative evolutionary method, evolution strategies, to the generation of simple Braitenberg vehicles. This application accelerates the development of such controllers by more than an order of magnitude (a few hours compared to more than two days). Motivated by this useful speedup, the paper investigates the evolution of more complex architectures, receptive-field controllers, that can employ nonlinear interactions and, therefore, can yield more complex behavior. It is interesting to note that the evolution strategy yields the same efficacy in terms of function evaluations, even though the second class of controllers requires up to 10 times more parameters than the simple Braitenberg architecture. In addition to the speedup, there is an important theoretical reason for preferring an evolution strategy over a genetic algorithm for this problem, namely the presence of epistasis", "keyphrases": ["receptive-field controllers", "mobile robots", "evolutionary methods", "real-world autonomous agents", "evolution strategies", "simple Braitenberg vehicles", "nonlinear interactions", "complex behavior", "radial basis functions", "scalability"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "M", "U"]} {"id": "1057", "title": "Acceptance of a price discount: the role of the semantic relatedness between purchases and the comparative price format", "abstract": "Two studies are reported where people are asked to accept or not a price reduction on a target product. In the high (low) relative saving version, the regular price of the target product is low (high). In both versions, the absolute value of the price reduction is the same as well as the total of regular prices of planned purchases. As first reported by Tversky and Kahneman (1981), findings show that the majority of people accept the price discount in the high-relative saving version whereas the minority do it in the low one. In Study 1, findings show that the previous preference reversal disappears when planned purchases are strongly related. Also, a previously unreported preference reversal is found. The majority of people accept the price discount when the products are weakly related whereas the minority accept when the products are strongly related. In Study 2, findings show that the classic preference reversal disappears as a function of the comparative price format. Also, another previously unreported preference reversal is found. When the offered price reduction relates to a low-priced product, people are more inclined to accept it with a control than a minimal comparative price format. Findings reported in Studies 1 and 2 are interpreted in terms of mental accounting shifts", "keyphrases": ["price discount acceptance", "preference reversal", "semantic relatedness hypothesis", "mental accounting shifts", "high relative saving version", "low relative saving version", "planned purchases", "comparative price format", "low-priced product"], "prmu": ["R", "P", "M", "P", "R", "R", "P", "P", "P"]} -{"id": "911", "title": "Scheduling schemes for an integrated flight and propulsion control system", "abstract": "We describe two schemes for scheduling an integrated flight and propulsion control system for an experimental vertical/short take-off and landing (V/STOL) aircraft concept in the acceleration from hover (0-120 kn) flight phase. Multivariable integrated flight and propulsion controllers are designed at several points over the V/STOL envelope and implemented as exact plant observers with state feedback. In the first scheduling scheme, the values of the state feedback and observer gain matrices are interpolated between the fixed-point designs as a function of aircraft speed. In the second approach, the control signals produced by the different fixed-point controllers are blended, allowing a significant reduction in the order of the scheduled controllers. Both scheduling schemes are shown in nonlinear simulation to provide excellent handling qualities as the aircraft accelerates from the hover", "keyphrases": ["vertical short take-off landing aircraft", "VSTOL aircraft", "propulsion control", "flight control", "scheduling", "multivariable control systems", "observers", "state feedback", "fixed-point controllers"], "prmu": ["M", "M", "P", "R", "P", "R", "P", "P", "P"]} +{"id": "911", "title": "Scheduling schemes for an integrated flight and propulsion control system", "abstract": "We describe two schemes for scheduling an integrated flight and propulsion control system for an experimental vertical/short take-off and landing (V/STOL) aircraft concept in the acceleration from hover (0-120 kn) flight phase. Multivariable integrated flight and propulsion controllers are designed at several points over the V/STOL envelope and implemented as exact plant observers with state feedback. In the first scheduling scheme, the values of the state feedback and observer gain matrices are interpolated between the fixed-point designs as a function of aircraft speed. In the second approach, the control signals produced by the different fixed-point controllers are blended, allowing a significant reduction in the order of the scheduled controllers. Both scheduling schemes are shown in nonlinear simulation to provide excellent handling qualities as the aircraft accelerates from the hover", "keyphrases": ["vertical short take-off landing aircraft", "VSTOL aircraft", "propulsion control", "flight control", "scheduling", "multivariable control systems", "observers", "state feedback", "fixed-point controllers"], "prmu": ["R", "M", "P", "R", "P", "R", "P", "P", "P"]} {"id": "583", "title": "Neural networks in optimal filtration", "abstract": "The combined use and mutual influence of neural networks and optimal filtering is considered; the neural-network and filtering approaches are compared by solving two simple optimal-filtering problems: linear filtering and the filtering of a binary telegraph signal corresponding to observations in discrete white noise", "keyphrases": ["optimal filtering", "neural networks", "linear filtering", "binary telegraph signal", "observations", "discrete white noise"], "prmu": ["P", "P", "P", "P", "P", "P"]} {"id": "954", "title": "Caring for your new lawyers", "abstract": "In any given year, a striking number of lawyers are in a state of flux, from newly minted law school graduates looking for their first job, to senior litigators migrating to new challenges with new firms. The one certainty is that lawyers new to any firm need care and feeding in myriad ways. All of them need to know and understand three things: (1) the firm's culture; (2) the resources available to help them develop their practices; and (3) where to get help and guidance for research and practice purposes. Obtaining a thorough understanding of a new firm's workings may be the greatest research project lawyers face. How can a firm help its new lawyers learn what they need to know? To offer an example, here are programs in place at my firm", "keyphrases": ["new lawyers", "firm's culture", "resources"], "prmu": ["P", "P", "P"]} {"id": "117", "title": "Multiresolution Markov models for signal and image processing", "abstract": "Reviews a significant component of the rich field of statistical multiresolution (MR) modeling and processing. These MR methods have found application and permeated the literature of a widely scattered set of disciplines, and one of our principal objectives is to present a single, coherent picture of this framework. A second goal is to describe how this topic fits into the even larger field of MR methods and concepts-in particular, making ties to topics such as wavelets and multigrid methods. A third goal is to provide several alternate viewpoints for this body of work, as the methods and concepts we describe intersect with a number of other fields. The principle focus of our presentation is the class of MR Markov processes defined on pyramidally organized trees. The attractiveness of these models stems from both the very efficient algorithms they admit and their expressive power and broad applicability. We show how a variety of methods and models relate to this framework including models for self-similar and 1/f processes. We also illustrate how these methods have been used in practice", "keyphrases": ["multiresolution Markov models", "statistical multiresolution modeling", "wavelets", "multigrid methods", "pyramidally organized trees", "self-similar processes", "1/f processes"], "prmu": ["P", "R", "P", "P", "P", "R", "P"]} @@ -343,12 +343,12 @@ {"id": "1249", "title": "Aggregators versus disintermediators: battling it out in the information superhighstreet", "abstract": "Perhaps the future of large-scale content aggregators is now no longer in doubt but this was not the case 10 years ago, when many leading industry experts were much more pessimistic in their predictions. In the year that Dialog celebrates its thirtieth anniversary as the world's oldest and largest professional online information service, it is appropriate to look back at these changing perceptions, the reasons for these changes, and why the experts got it wrong. We also look at the present day; the value that large-scale content aggregators bring to the information supply chain; and we discuss why users would choose to use aggregators as opposed to going directly to the publishers", "keyphrases": ["large-scale content aggregators", "online information service", "information supply chain", "disintermediators"], "prmu": ["P", "P", "P", "P"]} {"id": "63", "title": "Geometric source separation: merging convolutive source separation with geometric beamforming", "abstract": "Convolutive blind source separation and adaptive beamforming have a similar goal-extracting a source of interest (or multiple sources) while reducing undesired interferences. A benefit of source separation is that it overcomes the conventional cross-talk or leakage problem of adaptive beamforming. Beamforming on the other hand exploits geometric information which is often readily available but not utilized in blind algorithms. We propose to join these benefits by combining cross-power minimization of second-order source separation with geometric linear constraints used in adaptive beamforming. We find that the geometric constraints resolve some of the ambiguities inherent in the independence criterion such as frequency permutations and degrees of freedom provided by additional sensors. We demonstrate the new method in performance comparisons for actual room recordings of two and three simultaneous acoustic sources", "keyphrases": ["geometric source separation", "geometric beamforming", "convolutive blind source separation", "adaptive beamforming", "cross-talk", "leakage problem", "blind algorithms", "cross-power minimization", "second-order source separation", "geometric linear constraints", "frequency permutations", "degrees of freedom", "sensors", "room recordings", "acoustic sources"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "659", "title": "Integration - no longer a barrier? [agile business]", "abstract": "Web services will be a critical technology for enabling the 'agile business'", "keyphrases": ["agile business", "Web services", "integration middleware", "Iona", "AMR Research"], "prmu": ["P", "P", "M", "U", "U"]} -{"id": "1148", "title": "Benchmarking of the Dose Planning Method (DPM) Monte Carlo code using electron beams from a racetrack microtron", "abstract": "A comprehensive set of measurements and calculations has been conducted to investigate the accuracy of the Dose Planning Method (DPM) Monte Carlo code for dose calculations from 10 and 50 MeV scanned electron beams produced from a racetrack microtron. Central axis depth dose measurements and a series of profile scans at various depths were acquired in a water phantom using a Scanditronix type RK ion chamber. Source spatial distributions for the Monte Carlo calculations were reconstructed from in-air ion chamber measurements carried out across the two-dimensional beam profile at 100 cm downstream from the source. The in-air spatial distributions were found to have full width at half maximum of 4.7 and 1.3 cm, at 100 cm from the source, for the 10 and 50 MeV beams, respectively. Energy spectra for the 10 and 50 MeV beams were determined by simulating the components of the microtron treatment head using the code MCNP4B. DPM calculations are on average within +or-2% agreement with measurement for all depth dose and profile comparisons conducted in this study. The accuracy of the DPM code illustrated in this work suggests that DPM may be used as a valuable tool for electron beam dose calculations", "keyphrases": ["dose planning method Monte Carlo code", "scanned electron beams", "racetrack microtron", "benchmarking", "central axis depth dose measurements", "profile scans", "water phantom", "ion chamber", "source spatial distributions", "two-dimensional beam profile", "in-air spatial distributions", "MCNP4B", "electron beam dose calculations", "radiotherapy treatment planning", "electron transport", "scoring parameters", "10 MeV", "50 MeV"], "prmu": ["R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M", "M", "U", "R", "P"]} +{"id": "1148", "title": "Benchmarking of the Dose Planning Method (DPM) Monte Carlo code using electron beams from a racetrack microtron", "abstract": "A comprehensive set of measurements and calculations has been conducted to investigate the accuracy of the Dose Planning Method (DPM) Monte Carlo code for dose calculations from 10 and 50 MeV scanned electron beams produced from a racetrack microtron. Central axis depth dose measurements and a series of profile scans at various depths were acquired in a water phantom using a Scanditronix type RK ion chamber. Source spatial distributions for the Monte Carlo calculations were reconstructed from in-air ion chamber measurements carried out across the two-dimensional beam profile at 100 cm downstream from the source. The in-air spatial distributions were found to have full width at half maximum of 4.7 and 1.3 cm, at 100 cm from the source, for the 10 and 50 MeV beams, respectively. Energy spectra for the 10 and 50 MeV beams were determined by simulating the components of the microtron treatment head using the code MCNP4B. DPM calculations are on average within +or-2% agreement with measurement for all depth dose and profile comparisons conducted in this study. The accuracy of the DPM code illustrated in this work suggests that DPM may be used as a valuable tool for electron beam dose calculations", "keyphrases": ["dose planning method Monte Carlo code", "scanned electron beams", "racetrack microtron", "benchmarking", "central axis depth dose measurements", "profile scans", "water phantom", "ion chamber", "source spatial distributions", "two-dimensional beam profile", "in-air spatial distributions", "MCNP4B", "electron beam dose calculations", "radiotherapy treatment planning", "electron transport", "scoring parameters", "10 MeV", "50 MeV"], "prmu": ["R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "U", "P", "M", "M", "U", "R", "P"]} {"id": "558", "title": "OS porting and application development for SoC", "abstract": "To deliver improved usability in high-end portable consumer products, the use of an appropriate consumer operating system (OS) is becoming far more widespread. Using a commercially supported OS also vastly increases the availability of supported applications. For the device developer, this trend adds major complexity to the problem of system implementation. Porting a complete operating system to a new hardware design adds significantly to the development burden, increasing both time-to-market and expense. Even for those familiar with the integration of a real-time OS, the porting, validation and support of a complex platform OS is a formidable task", "keyphrases": ["OS porting", "application development", "consumer operating system", "hardware design"], "prmu": ["P", "P", "P", "P"]} {"id": "892", "title": "Dementing disorders: volumetric measurement of cerebrospinal fluid to distinguish normal from pathologic finding - feasibility study", "abstract": "We have demonstrated that automated methods to describe the severity and distribution of cerebral atrophy are capable of providing diagnostic information in the classification of neurodegenerative diseases", "keyphrases": ["cerebrospinal fluid volumetric measurement", "magnetic resonance imaging technique", "medical diagnostic imaging", "healthy subjects", "dementing disorders", "normal-pathologic findings distinguishing", "diagnostic information", "neurodegenerative diseases classification", "automated methods", "cerebral atrophy distribution", "cerebral atrophy severity"], "prmu": ["R", "U", "M", "U", "P", "M", "P", "R", "P", "R", "R"]} {"id": "1010", "title": "Robust self-tuning PID controller for nonlinear systems", "abstract": "In this paper, we propose a robust self-tuning PID controller suitable for nonlinear systems. The control system employs a preload relay (P_Relay) in series with a PID controller. The P_Relay ensures a high gain to yield a robust performance. However, it also incurs a chattering phenomenon. In this paper, instead of viewing the chattering as an undesirable yet inevitable feature, we use it as a naturally occurring signal for tuning and re-tuning the PID controller as the operating regime digresses. No other explicit input signal is required. Once the PID controller is tuned for a particular operating point, the relay may be disabled and chattering ceases correspondingly. However, it is invoked when there is a change in setpoint to another operating regime. In this way, the approach is also applicable to time-varying systems as the PID tuning can be continuous, based on the latest set of chattering characteristics. Analysis is provided on the stability properties of the control scheme. Simulation results for the level control of fluid in a spherical tank using the scheme are also presented", "keyphrases": ["nonlinear systems", "robust self-tuning PID controller", "preload relay", "robust performance", "chattering phenomenon", "naturally occurring signal", "controller tuning", "controller re-tuning", "relay disabling", "operating regime", "time-varying systems", "continuous tuning", "stability properties", "simulation results", "fluid level control", "spherical tank"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "R", "R", "P", "P", "R", "P", "P", "R", "P"]} {"id": "1055", "title": "A re-examination of probability matching and rational choice", "abstract": "In a typical probability learning task participants are presented with a repeated choice between two response alternatives, one of which has a higher payoff probability than the other. Rational choice theory requires that participants should eventually allocate all their responses to the high-payoff alternative, but previous research has found that people fail to maximize their payoffs. Instead, it is commonly observed that people match their response probabilities to the payoff probabilities. We report three experiments on this choice anomaly using a simple probability learning task in which participants were provided with (i) large financial incentives, (ii) meaningful and regular feedback, and (iii) extensive training. In each experiment large proportions of participants adopted the optimal response strategy and all three of the factors mentioned above contributed to this. The results are supportive of rational choice theory", "keyphrases": ["probability learning task", "rational choice theory", "probability matching", "payoff probability", "response probabilities", "choice anomaly", "large financial incentives", "meaningful regular feedback", "extensive training", "optimal response strategy", "feedback", "rationality"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "P"]} -{"id": "701", "title": "High dynamic control of a three-level voltage-source-converter drive for a main strip mill", "abstract": "A high dynamic control system for the Alspa VDM 7000 medium-voltage drive was implemented, which provides fast torque response times of a few milliseconds despite the typically low switching frequency of gate-turn-off thyristors which is necessary to achieve high efficiency. The drive system consists of a three-level voltage-source converter with active front end and a synchronous motor. The drive has most recently been applied to a main strip mill. It provides a maximum of 8.3-MW mechanical power with a rated motor voltage of 3 kV. Besides motor torque as the main control objective, the control system has to comply with a number of additional objectives and constraints like DC-link voltage regulation and balancing, current and torque harmonics, motor flux, and excitation", "keyphrases": ["medium-voltage drive", "high dynamic control system", "gate-turn-off thyristors", "switching frequency", "efficiency", "three-level voltage-source converter", "synchronous motor", "strip mill", "mechanical power", "motor voltage", "control objective", "DC-link voltage regulation", "DC-link voltage balancing", "current harmonics", "torque harmonics", "motor flux", "excitation", "8.3 MW", "3 kV"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "R", "P", "P", "P", "U", "P"]} +{"id": "701", "title": "High dynamic control of a three-level voltage-source-converter drive for a main strip mill", "abstract": "A high dynamic control system for the Alspa VDM 7000 medium-voltage drive was implemented, which provides fast torque response times of a few milliseconds despite the typically low switching frequency of gate-turn-off thyristors which is necessary to achieve high efficiency. The drive system consists of a three-level voltage-source converter with active front end and a synchronous motor. The drive has most recently been applied to a main strip mill. It provides a maximum of 8.3-MW mechanical power with a rated motor voltage of 3 kV. Besides motor torque as the main control objective, the control system has to comply with a number of additional objectives and constraints like DC-link voltage regulation and balancing, current and torque harmonics, motor flux, and excitation", "keyphrases": ["medium-voltage drive", "high dynamic control system", "gate-turn-off thyristors", "switching frequency", "efficiency", "three-level voltage-source converter", "synchronous motor", "strip mill", "mechanical power", "motor voltage", "control objective", "DC-link voltage regulation", "DC-link voltage balancing", "current harmonics", "torque harmonics", "motor flux", "excitation", "8.3 MW", "3 kV"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "R", "P", "P", "P", "U", "M"]} {"id": "744", "title": "A virtual victory [virtual networks]", "abstract": "Newly fashionable virtual network operators look all set to clean up in the corporate sector", "keyphrases": ["virtual network operators", "corporate sector"], "prmu": ["P", "P"]} {"id": "1311", "title": "Blended implementation of block implicit methods for ODEs", "abstract": "In this paper we further develop a new approach for naturally defining the nonlinear splittings needed for the implementation of block implicit methods for ODEs, which has been considered by Brugnano [J. Comput. Appl. Math. 116 (2000) 41] and by Brugnano and Trigiante [in: Recent Trends in Numerical Analysis, Nova Science, New York, 2000, pp. 81-105]. The basic idea is that of defining the numerical method as the combination (blending) of two suitable component methods. By carefully choosing such methods, it is shown that very efficient implementations can be obtained. Moreover, some of them, characterized by a diagonal splitting, are well suited for parallel computers. Numerical tests comparing the performances of the proposed implementation with existing ones are also presented, in order to make evident the potential of the approach", "keyphrases": ["nonlinear splittings", "block implicit methods", "blended implementation", "ODEs", "numerical method", "diagonal splitting", "parallel computers", "numerical tests"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1354", "title": "Design and analysis of optimal material distribution policies in flexible manufacturing systems using a single AGV", "abstract": "Modern automated manufacturing processes employ automated guided vehicles (AGVs) for material handling, which serve several machine centres (MC) in a factory. Optimal scheduling of AGVs can significantly help to increase the efficiency of the manufacturing process by minimizing the idle time of MCs waiting for the raw materials. We analyse the requirements for an optimal schedule and then provide a mathematical framework for an efficient schedule of material delivery by an AGV. A mathematical model is developed and then a strategy for optimal material distribution of the available raw material to the MCs is derived. With this model, the optimal number of MCs to be utilized is also determined. Finally, the material delivery schedule employing multiple journeys to the MCs by the AGV is carried out. Through rigorous analysis and simulation experiments, we show that such a delivery strategy will optimize the overall performance", "keyphrases": ["optimal material distribution policies", "flexible manufacturing systems", "AGV", "automated guided vehicle", "material handling", "machine centres", "waiting time", "manufacturing lead time", "optimal scheduling", "idle time minimization", "material delivery"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "M", "P", "R", "P"]} @@ -373,7 +373,7 @@ {"id": "664", "title": "The agile revolution [business agility]", "abstract": "There is a new business revolution in the air. The theory is there, the technology is evolving, fast. It is all about agility", "keyphrases": ["business agility", "software design", "software deployment", "organisational structures", "supply chains"], "prmu": ["P", "U", "U", "U", "U"]} {"id": "621", "title": "MPEG-4 video object-based rate allocation with variable temporal rates", "abstract": "In object-based coding, bit allocation is performed at the object level and temporal rates of different objects may vary. The proposed algorithm deals with these two issues when coding multiple video objects (MVOs). The proposed algorithm is able to successfully achieve the target bit rate, effectively code arbitrarily shaped MVOs with different temporal rates, and maintain a stable buffer level", "keyphrases": ["MPEG-4 video coding", "bit allocation", "multiple video objects", "rate-distortion encoding", "object-based rate allocation", "variable temporal rates"], "prmu": ["R", "P", "P", "U", "P", "P"]} {"id": "1438", "title": "Three-dimensional particle image tracking for dilute particle-liquid flows in a pipe", "abstract": "A three-dimensional (3D) particle image tracking technique was used to study the coarse spherical particle-liquid flows in a pipe. The flow images from both the front view and the normal side view, which was reflected into the front view by a mirror, were recorded with a CCD camera and digitized by a PC with an image grabber card. An image processing program was developed to enhance and segment the flow image, and then to identify the particles. Over 90% of all the particles can be identified and located from the partially overlapped particle images using the circular Hough transform. Then the 3D position of each detected particle was determined by matching its front view image to its side view image. The particle velocity was then obtained by pairing its images in successive video fields. The measurements for the spherical expanded polystyrene particle-oil flows show that the particles, like the spherical bubbles in laminar bubbly flows, tend to conglomerate near the pipe wall and to line up to form the particle clusters. As liquid velocity decreases, the particle clusters disperse and more particles are distributed in the pipe centre region", "keyphrases": ["three-dimensional particle image tracking", "dilute particle-liquid flows", "two-phase flow", "pipe flow", "stereo-imaging technique", "phase distribution", "CCD camera", "3D position", "spherical expanded polystyrene particle", "particle clusters", "spherical bubble", "Wiener filter", "image segmentation", "region growing technique", "image recognition", "image matching", "Hough transform"], "prmu": ["P", "P", "M", "R", "M", "M", "P", "P", "R", "P", "P", "U", "R", "M", "M", "R", "P"]} -{"id": "705", "title": "Use of extra degrees of freedom in multilevel drives", "abstract": "Multilevel converters with series connection of semiconductors allow power electronics to reach medium voltages (1-10 kV) with relatively standard components. The increase of the number of semiconductors provides extra degrees of freedom, which can be used to improve different characteristics. This paper is focused on variable-speed drives and it is shown that with the proposed multilevel direct torque control strategy (DiCoIF) the tradeoff between the performances of the drive (harmonic distortions, torque dynamics, voltage step gradients, etc.) and the switching frequency of the semiconductors is improved. Then, a slightly modified strategy reducing common-mode voltage and bearing currents is presented", "keyphrases": ["degrees of freedom", "series connection", "semiconductors", "power electronics", "medium voltages", "variable-speed drives", "multilevel direct torque control strategy", "harmonic distortions", "torque dynamics", "voltage step gradients", "switching frequency", "common-mode voltage reduction", "bearing currents", "delay estimation", "industrial power systems", "insulated gate bipolar transistors", "state estimation", "multilevel drives", "fixed-frequency dynamic control", "1 to 10 kV"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M", "P", "U", "M", "U", "U", "P", "M", "M"]} +{"id": "705", "title": "Use of extra degrees of freedom in multilevel drives", "abstract": "Multilevel converters with series connection of semiconductors allow power electronics to reach medium voltages (1-10 kV) with relatively standard components. The increase of the number of semiconductors provides extra degrees of freedom, which can be used to improve different characteristics. This paper is focused on variable-speed drives and it is shown that with the proposed multilevel direct torque control strategy (DiCoIF) the tradeoff between the performances of the drive (harmonic distortions, torque dynamics, voltage step gradients, etc.) and the switching frequency of the semiconductors is improved. Then, a slightly modified strategy reducing common-mode voltage and bearing currents is presented", "keyphrases": ["degrees of freedom", "series connection", "semiconductors", "power electronics", "medium voltages", "variable-speed drives", "multilevel direct torque control strategy", "harmonic distortions", "torque dynamics", "voltage step gradients", "switching frequency", "common-mode voltage reduction", "bearing currents", "delay estimation", "industrial power systems", "insulated gate bipolar transistors", "state estimation", "multilevel drives", "fixed-frequency dynamic control", "1 to 10 kV"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M", "P", "U", "M", "U", "U", "P", "M", "R"]} {"id": "740", "title": "The Malaysian model", "abstract": "Japan's first third generation service, Foma, is unlikely to be truly attractive to consumers until 2005. That still falls well within the financial planning of its operator Docomo. But where does that leave European 3G operators looking for reassurance? Malaysia, says Simon Marshall", "keyphrases": ["3G operators", "Malaysia", "Maxis Communications", "Telekom Malaysia"], "prmu": ["P", "P", "U", "M"]} {"id": "1315", "title": "Traffic engineering with traditional IP routing protocols", "abstract": "Traffic engineering involves adapting the routing of traffic to network conditions, with the joint goals of good user performance and efficient use of network resources. We describe an approach to intradomain traffic engineering that works within the existing deployed base of interior gateway protocols, such as Open Shortest Path First and Intermediate System-Intermediate System. We explain how to adapt the configuration of link weights, based on a networkwide view of the traffic and topology within a domain. In addition, we summarize the results of several studies of techniques for optimizing OSPF/IS-IS weights to the prevailing traffic. The article argues that traditional shortest path routing protocols are surprisingly effective for engineering the flow of traffic in large IP networks", "keyphrases": ["IP routing protocols", "interior gateway protocols", "link weights configuration", "traffic routing", "network conditions", "user performance", "network resources", "intradomain traffic engineering", "network topology", "OSPF/IS-IS weights", "shortest path routing protocols", "IP networks", "TCP", "transmission control protocol", "Open Shortest Path First protocol", "Intermediate System-Intermediate System protocol"], "prmu": ["P", "P", "R", "R", "P", "P", "P", "P", "R", "P", "P", "P", "U", "M", "R", "R"]} {"id": "1350", "title": "Generalized mosaicing: wide field of view multispectral imaging", "abstract": "We present an approach to significantly enhance the spectral resolution of imaging systems by generalizing image mosaicing. A filter transmitting spatially varying spectral bands is rigidly attached to a camera. As the system moves, it senses each scene point multiple times, each time in a different spectral band. This is an additional dimension of the generalized mosaic paradigm, which has demonstrated yielding high radiometric dynamic range images in a wide field of view, using a spatially varying density filter. The resulting mosaic represents the spectrum at each scene point. The image acquisition is as easy as in traditional image mosaics. We derive an efficient scene sampling rate, and use a registration method that accommodates the spatially varying properties of the filter. Using the data acquired by this method, we demonstrate scene rendering under different simulated illumination spectra. We are also able to infer information about the scene illumination. The approach was tested using a standard 8-bit black/white video camera and a fixed spatially varying spectral (interference) filter", "keyphrases": ["generalized mosaicing", "wide field of view multispectral imaging", "spatially varying spectral bands", "spatially varying density filter", "image acquisition", "scene sampling rate", "registration method", "scene rendering", "simulated illumination spectra", "scene illumination", "hyperspectral imaging", "color balance", "image fusion", "physics-based vision", "image-based rendering"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M", "U", "M", "U", "M"]} @@ -449,14 +449,14 @@ {"id": "1077", "title": "Quantum learning and universal quantum matching machine", "abstract": "Suppose that three kinds of quantum systems are given in some unknown states |f>/sup (X)N/, |g/sub 1/>/sup (X)K/, and |g/sub 2/>/sup (X)K/, and we want to decide which template state |g/sub 1/> or |g/sub 2/>, each representing the feature of the pattern class C/sub 1/ or C/sub 2/, respectively, is closest to the input feature state |f>. This is an extension of the pattern matching problem into the quantum domain. Assuming that these states are known a priori to belong to a certain parametric family of pure qubit systems, we derive two kinds of matching strategies. The first one is a semiclassical strategy that is obtained by the natural extension of conventional matching strategies and consists of a two-stage procedure: identification (estimation) of the unknown template states to design the classifier (learning process to train the classifier) and classification of the input system into the appropriate pattern class based on the estimated results. The other is a fully quantum strategy without any intermediate measurement, which we might call as the universal quantum matching machine. We present the Bayes optimal solutions for both strategies in the case of K=1, showing that there certainly exists a fully quantum matching procedure that is strictly superior to the straightforward semiclassical extension of the conventional matching strategy based on the learning process", "keyphrases": ["quantum learning", "universal quantum matching machine", "pattern class", "pattern matching problem", "quantum domain", "qubit systems", "matching strategies", "semiclassical strategy", "two-stage procedure", "quantum strategy", "Bayes optimal solutions", "quantum matching procedure", "semiclassical extension", "matching strategy", "learning process"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "686", "title": "Technology CAD of SiGe-heterojunction field effect transistors", "abstract": "A 2D virtual wafer fabrication simulation suite has been employed for the technology CAD of SiGe channel heterojunction field effect transistors (HFETs). Complete fabrication process of SiGe p-HFETs has been simulated. The SiGe material parameters and mobility model were incorporated to simulate Si/SiGe p-HFETs with a uniform germanium channel having an L/sub eff/ of 0.5 mu m. A significant improvement in linear transconductance is observed when compared to control-silicon p-MOSFETs", "keyphrases": ["technology CAD", "heterojunction field effect transistors", "SiGe", "fabrication process", "material parameters", "mobility model", "uniform channel", "linear transconductance", "0.5 micron"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "P", "M"]} {"id": "1296", "title": "Development of visual design steering as an aid in large-scale multidisciplinary design optimization. I. Method development", "abstract": "A modified paradigm of computational steering (CS), termed visual design steering (VDS), is developed in this paper. The VDS paradigm is applied to optimal design problems to provide a means for capturing and enabling designer insights. VDS allows a designer to make decisions before, during or after an analysis or optimization via a visual environment, in order to effectively steer the solution process. The objective of VDS is to obtain a better solution in less time through the use of designer knowledge and expertise. Using visual representations of complex systems in this manner enables human experience and judgement to be incorporated into the optimal design process at appropriate steps, rather than having traditional black box solvers return solutions from a prescribed input set. Part I of this paper focuses on the research issues pertaining to the Graph Morphing visualization method created to represent an n-dimensional optimization problem using 2-dimensional and 3-dimensional visualizations. Part II investigates the implementation of the VDS paradigm, using the graph morphing approach, to improve an optimal design process. Specifically, the following issues are addressed: impact of design variable changes on the optimal design space; identification of possible constraint redundancies; impact of constraint tolerances on the optimal solution: and smoothness of the objective function contours. It is demonstrated that graph morphing can effectively reduce the complexity and computational time associated with some optimization problems", "keyphrases": ["visual design steering", "large-scale multidisciplinary design optimization", "computational steering", "optimal design problems", "designer decision making", "visual representations", "complex systems", "graph morphing visualization method", "3D visualizations", "2D visualizations", "n-dimensional optimization", "computational time", "complexity", "design variable changes", "constraint redundancies", "constraint tolerances", "objective function contour smoothness"], "prmu": ["P", "P", "P", "P", "R", "P", "P", "P", "M", "M", "P", "P", "P", "P", "P", "P", "R"]} -{"id": "915", "title": "A meteorological fuzzy expert system incorporating subjective user input", "abstract": "We present a fuzzy expert system, MEDEX, for forecasting gale-force winds in the Mediterranean basin. The most successful local wind forecasting in this region is achieved by an expert human forecaster with access to numerical weather prediction products. That forecaster's knowledge is expressed as a set of 'rules-of-thumb'. Fuzzy set methodologies have proved well suited for encoding the forecaster's knowledge, and for accommodating the uncertainty inherent in the specification of rules, as well as in subjective and objective input. MEDEX uses fuzzy set theory in two ways: as a fuzzy rule base in the expert system, and for fuzzy pattern matching to select dominant wind circulation patterns as one input to the expert system. The system was developed, tuned, and verified over a two-year period, during which the weather conditions from 539 days were individually analyzed. Evaluations of MEDEX performance for both the onset and cessation of winter and summer winds are presented, and demonstrate that MEDEX has forecasting skill competitive with the US Navy's regional forecasting center in Rota, Spain", "keyphrases": ["meteorological fuzzy expert system", "subjective user input", "MEDEX", "gale-force wind forecasting", "Mediterranean basin", "numerical weather prediction products", "rules-of-thumb", "fuzzy set theory", "subjective variables", "uncertainty", "rule specification", "fuzzy rule base", "fuzzy pattern matching", "wind circulation patterns"], "prmu": ["P", "P", "P", "R", "P", "P", "U", "P", "M", "P", "R", "P", "P", "P"]} +{"id": "915", "title": "A meteorological fuzzy expert system incorporating subjective user input", "abstract": "We present a fuzzy expert system, MEDEX, for forecasting gale-force winds in the Mediterranean basin. The most successful local wind forecasting in this region is achieved by an expert human forecaster with access to numerical weather prediction products. That forecaster's knowledge is expressed as a set of 'rules-of-thumb'. Fuzzy set methodologies have proved well suited for encoding the forecaster's knowledge, and for accommodating the uncertainty inherent in the specification of rules, as well as in subjective and objective input. MEDEX uses fuzzy set theory in two ways: as a fuzzy rule base in the expert system, and for fuzzy pattern matching to select dominant wind circulation patterns as one input to the expert system. The system was developed, tuned, and verified over a two-year period, during which the weather conditions from 539 days were individually analyzed. Evaluations of MEDEX performance for both the onset and cessation of winter and summer winds are presented, and demonstrate that MEDEX has forecasting skill competitive with the US Navy's regional forecasting center in Rota, Spain", "keyphrases": ["meteorological fuzzy expert system", "subjective user input", "MEDEX", "gale-force wind forecasting", "Mediterranean basin", "numerical weather prediction products", "rules-of-thumb", "fuzzy set theory", "subjective variables", "uncertainty", "rule specification", "fuzzy rule base", "fuzzy pattern matching", "wind circulation patterns"], "prmu": ["P", "P", "P", "R", "P", "P", "P", "P", "M", "P", "R", "P", "P", "P"]} {"id": "587", "title": "An improved self-organizing CPN-based fuzzy system with adaptive back-propagation algorithm", "abstract": "This paper describes an improved self-organizing CPN-based (Counter-Propagation Network) fuzzy system. Two self-organizing algorithms IUSOCPN and ISSOCPN, being unsupervised and supervised respectively, are introduced. The idea is to construct the neural-fuzzy system with a two-phase hybrid learning algorithm, which utilizes a CPN-based nearest-neighbor clustering scheme for both structure learning and initial parameters setting, and a gradient descent method with adaptive learning rate for fine tuning the parameters. The obtained network can be used in the same way as a CPN to model and control dynamic systems, while it has a faster learning speed than the original back-propagation algorithm. The comparative results on the examples suggest that the method is fairly efficient in terms of simple structure, fast learning speed, and relatively high modeling accuracy", "keyphrases": ["self-organizing fuzzy system", "Counter-Propagation Network", "neural-fuzzy system", "hybrid learning", "gradient descent", "structure learning", "initial parameters setting", "back-propagation learning scheme"], "prmu": ["R", "P", "P", "P", "P", "P", "P", "R"]} {"id": "950", "title": "Quantum sensitive dependence", "abstract": "Wave functions of bounded quantum systems with time-independent potentials, being almost periodic functions, cannot have time asymptotics as in classical chaos. However, bounded quantum systems with time-dependent interactions, as used in quantum control, may have continuous spectrum and the rate of growth of observables is an issue of both theoretical and practical concern. Rates of growth in quantum mechanics are discussed by constructing quantities with the same physical meaning as those involved in the classical Lyapunov exponent. A generalized notion of quantum sensitive dependence is introduced and the mathematical structure of the operator matrix elements that correspond to different types of growth is characterized", "keyphrases": ["quantum sensitive dependence", "wave functions", "bounded quantum systems", "time-independent potentials", "periodic functions", "time asymptotics", "classical chaos", "time-dependent interactions", "quantum control", "classical Lyapunov exponent", "operator matrix elements", "quantum complexity"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M"]} {"id": "113", "title": "Quantum limit on computational time and speed", "abstract": "We investigate if physical laws can impose limits on computational time and speed of a quantum computer built from elementary particles. We show that the product of the speed and the running time of a quantum computer is limited by the type of fundamental interactions present inside the system. This will help us to decide as to what type of interaction should be allowed in building quantum computers in achieving the desired speed", "keyphrases": ["quantum limit", "computational time", "computational speed", "quantum computer", "fundamental interactions"], "prmu": ["P", "P", "R", "P", "P"]} {"id": "1197", "title": "Numerical behaviour of stable and unstable solitary waves", "abstract": "In this paper we analyse the behaviour in time of the numerical approximations to solitary wave solutions of the generalized Benjamin-Bona-Mahony equation. This equation possesses an important property: the stability of these solutions depends on their velocity. We identify the error propagation mechanisms in both the stable and unstable case. In particular, we show that in the stable case, numerical methods that preserve some conserved quantities of the problem are more appropriate for the simulation of this kind of solutions", "keyphrases": ["numerical behaviour", "unstable solitary waves", "numerical approximations", "generalized Benjamin-Bona-Mahony equation", "error propagation mechanisms", "numerical methods", "stable solitary waves"], "prmu": ["P", "P", "P", "P", "P", "P", "R"]} {"id": "928", "title": "Weighted energy linear quadratic regulator vibration control of piezoelectric composite plates", "abstract": "In this paper on finite element linear quadratic regulator (LQR) vibration control of smart piezoelectric composite plates, we propose the use of the total weighted energy method to select the weighting matrices. By constructing the optimal performance function as a relative measure of the total kinetic energy, strain energy and input energy of the system, only three design variables need to be considered to achieve a balance between the desired higher damping effect and lower input cost. Modal control analysis is used to interpret the effects of three energy weight factors on the damping ratios and modal voltages and it is shown that the modal damping effect will increase with the kinetic energy weight factor, approaching square root (2/2) as the strain energy weight factor increases and decrease with the input energy weight factor. Numerical results agree well with those from the modal control analysis. Since the control problem is simplified to three design variables only, the computational cost will be greatly reduced and a more accurate structural control analysis becomes more attractive for large systems", "keyphrases": ["finite element linear quadratic regulator", "vibration control", "smart piezoelectric composite plates", "total weighted energy", "weighting matrices", "optimal performance function", "total kinetic energy", "strain energy", "damping effect", "modal control analysis", "damping ratios", "strain energy weight factor", "numerical results", "computational cost", "structural control analysis"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1256", "title": "High-speed consistency checking for hypothetical reasoning systems using inference path network", "abstract": "Hypothetical reasoning is popular in fault diagnostics and design systems, but slow reasoning speed is its drawback. The goal of the current study is developing hypothetical reasoning based on an inference path network, which would overcome this drawback. In hypothetical reasoning systems based on an inference path network, there is much room for improvement regarding the computing costs of connotation processing and consistency checking. The authors of this study demonstrate improvement ideas regarding one of these problems, namely, consistency checking. First, the authors obtained necessary and sufficient conditions under which inconsistencies occur during hypothesis composition. Based on the obtained results, the authors proposed an algorithm for speeding up the process of consistency checking. Processing with this algorithm in its core consists of transforming the inference path network in such a way that inconsistencies do not occur during the hypothesis composition, under the condition of unchanged solution hypotheses. The efficiency of this algorithm was confirmed by tests", "keyphrases": ["hypothetical reasoning", "fault diagnostics", "high-speed consistency checking", "inference path network", "reasoning speed", "inconsistencies", "hypothesis composition", "speed up"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P"]} -{"id": "1213", "title": "A knowledge intensive multi-agent framework for cooperative/collaborative design modeling and decision support of assemblies", "abstract": "Multi-agent modeling has emerged as a promising discipline for dealing with the decision making process in distributed information system applications. One of such applications is the modeling of distributed design or manufacturing processes which can link up various designs or manufacturing processes to form a virtual consortium on a global basis. This paper proposes a novel knowledge intensive multi-agent cooperative/collaborative framework for concurrent intelligent design and assembly planning, which integrates product design, design for assembly, assembly planning, assembly system design, and assembly simulation subjected to econo-technical evaluations. An AI protocol based method is proposed to facilitate the integration of intelligent agents for assembly design, planning, evaluation and simulation processes. A unified class of knowledge intensive Petri nets is defined using the OO knowledge-based Petri net approach and used as an AI protocol for handling both the integration and the negotiation problems among multi-agents. The detailed cooperative/collaborative mechanism and algorithms are given based on the knowledge object cooperation formalisms. As such, the assembly-oriented design system can easily be implemented under the multi-agent-based knowledge-intensive Petri net framework with concurrent integration of multiple cooperative knowledge sources and software. Thus, product design and assembly planning can be carried out simultaneously and intelligently in an entirely computer-aided concurrent design and assembly planning system", "keyphrases": ["knowledge intensive multi-agent framework", "collaborative design modeling", "decision support", "distributed information system applications", "distributed design", "virtual consortium", "cooperative framework", "concurrent intelligent design", "assembly planning", "product design", "design for assembly", "assembly simulation", "AI protocol", "knowledge intensive Petri nets", "agent negotiation", "knowledge object cooperation"], "prmu": ["P", "M", "P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "R", "P"]} +{"id": "1213", "title": "A knowledge intensive multi-agent framework for cooperative/collaborative design modeling and decision support of assemblies", "abstract": "Multi-agent modeling has emerged as a promising discipline for dealing with the decision making process in distributed information system applications. One of such applications is the modeling of distributed design or manufacturing processes which can link up various designs or manufacturing processes to form a virtual consortium on a global basis. This paper proposes a novel knowledge intensive multi-agent cooperative/collaborative framework for concurrent intelligent design and assembly planning, which integrates product design, design for assembly, assembly planning, assembly system design, and assembly simulation subjected to econo-technical evaluations. An AI protocol based method is proposed to facilitate the integration of intelligent agents for assembly design, planning, evaluation and simulation processes. A unified class of knowledge intensive Petri nets is defined using the OO knowledge-based Petri net approach and used as an AI protocol for handling both the integration and the negotiation problems among multi-agents. The detailed cooperative/collaborative mechanism and algorithms are given based on the knowledge object cooperation formalisms. As such, the assembly-oriented design system can easily be implemented under the multi-agent-based knowledge-intensive Petri net framework with concurrent integration of multiple cooperative knowledge sources and software. Thus, product design and assembly planning can be carried out simultaneously and intelligently in an entirely computer-aided concurrent design and assembly planning system", "keyphrases": ["knowledge intensive multi-agent framework", "collaborative design modeling", "decision support", "distributed information system applications", "distributed design", "virtual consortium", "cooperative framework", "concurrent intelligent design", "assembly planning", "product design", "design for assembly", "assembly simulation", "AI protocol", "knowledge intensive Petri nets", "agent negotiation", "knowledge object cooperation"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "R", "P"]} {"id": "646", "title": "Vibration control of structure by using tuned mass damper (development of system which suppress displacement of auxiliary mass)", "abstract": "In vibration control of a structure by using an active tuned mass damper (ATMD), stroke of the auxiliary mass is so limited that it is difficult to control the vibration in the case of large disturbance input. In this paper, two methods are proposed for the problem. One of the methods is a switching control system by two types of controllers. One of the controllers is a normal controller under small relative displacement of the auxiliary mass, and the other is not effective only for first mode of vibration under large relative displacement of the auxiliary mass. New variable gain control system is constructed by switching these two controllers. The other method is the brake system. In active vibration control, it is necessary to use actuator for active control. By using the actuator, the proposed system puts on the brake to suppress displacement increase of the auxiliary mass under large disturbance input. Finally, the systems are designed and the effectiveness of the systems is confirmed by the simulation", "keyphrases": ["tuned mass damper", "vibration control", "variable gain control system", "brake system", "actuator", "active control", "auxiliary mass displacement suppression", "controllers"], "prmu": ["P", "P", "P", "P", "P", "P", "R", "P"]} {"id": "603", "title": "PGE helps customers reduce energy costs", "abstract": "A new service from Portland General Electric (PGE, Portland, Oregon, US) is saving customers tens of thousands of dollars in energy costs. PGE created E-Manager to allow facility managers to analyze their energy consumption online at 15-minute intervals. Customers can go to the Web for complete data, powerful analysis tools and charts, helping them detect abnormal energy use and focus on costly problem areas", "keyphrases": ["energy costs reduction", "Portland General Electric", "Oregon", "E-Manager", "online energy consumption analysis", "abnormal energy use detection"], "prmu": ["M", "P", "P", "P", "R", "R"]} {"id": "81", "title": "A scalable and efficient systolic algorithm for the longest common subsequence problem", "abstract": "A longest common subsequence (LCS) of two strings is a common subsequence of two strings of maximal length. The LCS problem is that of finding an LCS of two given strings and the length of the LCS. This problem has been the subject of much research because its solution can be applied in many areas. In this paper, a scalable and efficient systolic algorithm is presented. For two given strings of length m and n, where m>or=n, the algorithm can solve the LCS problem in m+2r-1 (respectively n+2r-1) time steps with r100 times faster!). The DIPIE scheme requires separate mechanical and electrostatic field solvers. Therefore, it can be easily implemented in existing MOEMS CAD packages. Moreover, using the DIPIE scheme, the pull-in parameters extraction can be performed in a fully automated mode, and no user input for search bounds is required", "keyphrases": ["DIPIE algorithm", "MOEMS CAD packages", "electrostatically actuated MEMS devices", "pull-in parameters", "electrostatic actuators", "design geometries", "convergence rate", "displacement iteration pull-in extraction scheme", "mechanical field solver", "electrostatic field solver", "computer-aided design", "displacement iteration"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "R", "R", "P", "M", "P"]} {"id": "989", "title": "A dynamic checkpoint scheduling scheme for fault tolerant distributed computing systems", "abstract": "The selection of the optimal checkpointing interval has been a very critical issue in implementing checkpointing-recovery schemes for fault tolerant distributed systems. This paper presents a new scheme that allows a process to select the proper checkpointing interval dynamically. A process in the system evaluates the cost of checkpointing and possible rollback for each checkpointing interval and selects the proper time interval for the next checkpointing. Unlike the other schemes, the overhead incurred by both the checkpointing and rollback activities are considered for the cost evaluation, and the current communication pattern is reflected in the selection of the checkpointing interval. Moreover, the proposed scheme requires no extra message communication for the checkpointing interval selection and can easily be incorporated into the existing checkpointing coordination schemes", "keyphrases": ["dynamic checkpoint scheduling scheme", "fault tolerant computing", "distributed computing systems", "optimal checkpointing interval", "rollback recovery", "cost evaluation", "communication pattern"], "prmu": ["P", "R", "P", "P", "M", "P", "P"]} -{"id": "98", "title": "Automating the compliance and supervision process", "abstract": "New technology enables large broker/dealers to supervise and ensure compliance across multiple branches and managers", "keyphrases": ["compliance", "supervision", "brokers", "risk management"], "prmu": ["P", "P", "U", "M"]} +{"id": "98", "title": "Automating the compliance and supervision process", "abstract": "New technology enables large broker/dealers to supervise and ensure compliance across multiple branches and managers", "keyphrases": ["compliance", "supervision", "brokers", "risk management"], "prmu": ["P", "P", "P", "M"]} {"id": "898", "title": "Influence of advertising expenses on the characteristics of functioning of an insurance company", "abstract": "The basic characteristics of the functioning of an insurance company, including the average capital, ruin and survival probabilities, and the conditional time before ruin, are examined with allowance for advertising expenses", "keyphrases": ["advertising expenses influence", "insurance company functioning characteristics", "average capital", "ruin probabilities", "survival probabilities", "conditional time"], "prmu": ["R", "R", "P", "R", "P", "P"]} {"id": "820", "title": "Yet some more complexity results for default logic", "abstract": "We identify several new tractable subsets and several new intractable simple cases for reasoning in the propositional version of Reiter's default logic. The majority of our findings are related to brave reasoning. By making some intuitive observations, most classes that we identify can be derived quite easily from some subsets of default logic already known in the literature. Some of the subsets we discuss are subclasses of the so-called \"extended logic programs\". All the tractable subsets presented in this paper can be recognized in linear time", "keyphrases": ["reasoning", "default logic", "complexity results", "complexity classes", "nonmonotonic reasoning", "extended logic programs", "tractable subsets"], "prmu": ["P", "P", "P", "R", "M", "P", "P"]} {"id": "865", "title": "Setup cost and lead time reductions on stochastic inventory models with a service level constraint", "abstract": "The stochastic inventory models analyzed in this paper explore the problem of lead time associated with setup cost reductions for the continuous review and periodic review inventory models. For these two models with a mixture of backorders and lost sales, we respectively assume that their mean and variance of the lead time demand and protection interval (i.e., lead time plus review period) demand are known, but their probability distributions are unknown. We develop a minimax distribution free procedure to find the optimal solution-for each case", "keyphrases": ["setup cost reductions", "lead time reductions", "stochastic inventory models", "service level constraint", "continuous review inventory models", "periodic review inventory models", "backorders", "lost sales", "lead time demand", "protection interval", "probability distributions", "minimax distribution free procedure"], "prmu": ["P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P"]} @@ -652,7 +652,7 @@ {"id": "1200", "title": "From continuous recovery to discrete filtering in numerical approximations of conservation laws", "abstract": "Modern numerical approximations of conservation laws rely on numerical dissipation as a means of stabilization. The older, alternative approach is the use of central differencing with a dose of artificial dissipation. In this paper we review the successful class of weighted essentially non-oscillatory finite volume schemes which comprise sophisticated methods of the first kind. New developments in image processing have made new devices possible which can serve as highly nonlinear artificial dissipation terms. We view artificial dissipation as discrete filter operation and introduce several new algorithms inspired by image processing", "keyphrases": ["continuous recovery", "discrete filtering", "numerical approximations", "conservation laws", "numerical dissipation", "central differencing", "artificial dissipation", "finite volume schemes", "image processing", "highly nonlinear artificial dissipation terms", "discrete filter operation"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1245", "title": "A brief guide to competitive intelligence: how to gather and use information on competitors", "abstract": "The author outlines the processes involved in competitive intelligence, and discusses what it is, how to do it and gives examples of what happens when companies fail to monitor their competitive environment effectively. The author presents a case study, showing how the company that produced the pre-cursor to the Barbie doll failed to look at their business environment and how this led to the firm's failure. The author discusses what competitive intelligence is, and what it is not, and why it is important for businesses, and presents three models used to describe the competitive intelligence process, going through the various steps involved in defining intelligence requirements and collecting, analyzing, communicating and utilizing competitive intelligence", "keyphrases": ["competitive intelligence", "competitor information", "Barbie doll", "business environment", "intelligence collection", "intelligence analysis", "intelligence communication", "intelligence utilization"], "prmu": ["P", "R", "P", "P", "R", "M", "R", "R"]} {"id": "983", "title": "Limitations of delayed state feedback: a numerical study", "abstract": "Stabilization of a class of linear time-delay systems can be achieved by a numerical procedure, called the continuous pole placement method [Michiels et al., 2000]. This method can be seen as an extension of the classical pole placement algorithm for ordinary differential equations to a class of delay differential equations. In [Michiels et al., 2000] it was applied to the stabilization of a linear time-invariant system with an input delay using static state feedback. In this paper we study the limitations of such delayed state feedback laws. More precisely we completely characterize the class of stabilizable plants in the 2D-case. For that purpose we make use of numerical continuation techniques. The use of delayed state feedback in various control applications and the effect of its limitations are briefly discussed", "keyphrases": ["linear time-delay systems", "continuous pole placement method", "delay differential equations", "static state feedback", "delayed state feedback", "numerical continuation"], "prmu": ["P", "P", "P", "P", "P", "P"]} -{"id": "554", "title": "A scalable and lightweight QoS monitoring technique combining passive and active approaches: on the mathematical formulation of CoMPACT monitor", "abstract": "To make a scalable and lightweight QoS monitoring system, we (2002) have proposed a new QoS monitoring technique, called the change-of-measure based passive/active monitoring (CoMPACT Monitor), which is based on the change-of-measure framework and is an active measurement transformed by using passively monitored data. This technique enables us to measure detailed QoS information for individual users, applications and organizations, in a scalable and lightweight manner. In this paper, we present the mathematical foundation of CoMPACT Monitor. In addition, we show its characteristics through simulations in terms of typical implementation issues for inferring the delay distributions. The results show that CoMPACT Monitor gives accurate QoS estimations with only a small amount of extra traffic for active measurement", "keyphrases": ["quality of service", "change-of-measure", "passive monitoring", "active monitoring", "CoMPACT Monitor", "delay distributions", "Internet", "QoS monitoring", "network performance"], "prmu": ["M", "P", "P", "R", "P", "P", "U", "P", "U"]} +{"id": "554", "title": "A scalable and lightweight QoS monitoring technique combining passive and active approaches: on the mathematical formulation of CoMPACT monitor", "abstract": "To make a scalable and lightweight QoS monitoring system, we (2002) have proposed a new QoS monitoring technique, called the change-of-measure based passive/active monitoring (CoMPACT Monitor), which is based on the change-of-measure framework and is an active measurement transformed by using passively monitored data. This technique enables us to measure detailed QoS information for individual users, applications and organizations, in a scalable and lightweight manner. In this paper, we present the mathematical foundation of CoMPACT Monitor. In addition, we show its characteristics through simulations in terms of typical implementation issues for inferring the delay distributions. The results show that CoMPACT Monitor gives accurate QoS estimations with only a small amount of extra traffic for active measurement", "keyphrases": ["quality of service", "change-of-measure", "passive monitoring", "active monitoring", "CoMPACT Monitor", "delay distributions", "Internet", "QoS monitoring", "network performance"], "prmu": ["M", "P", "P", "P", "P", "P", "U", "P", "U"]} {"id": "1101", "title": "Evaluation of existing and new feature recognition algorithms. 1. Theory and implementation", "abstract": "This is the first of two papers evaluating the performance of general-purpose feature detection techniques for geometric models. In this paper, six different methods are described to identify sets of faces that bound depression and protrusion faces. Each algorithm has been implemented and tested on eight components from the National Design Repository. The algorithms studied include previously published general-purpose feature detection algorithms such as the single-face inner-loop and concavity techniques. Others are improvements to existing algorithms such as extensions of the two-dimensional convex hull method to handle curved faces as well as protrusions. Lastly, new algorithms based on the three-dimensional convex hull, minimum concave, visible and multiple-face inner-loop face sets are described", "keyphrases": ["feature recognition algorithms", "geometric models", "general-purpose feature detection techniques", "sets of faces", "depression faces", "protrusion faces", "National Design Repository", "single-face inner-loop technique", "concavity technique", "two-dimensional convex hull method", "curved faces", "three-dimensional convex hull", "CAD/CAM software", "geometric reasoning algorithms", "minimum concave", "visible inner-loop face sets", "multiple-face inner-loop face sets"], "prmu": ["P", "P", "P", "P", "R", "P", "P", "R", "P", "P", "P", "P", "U", "M", "P", "R", "P"]} {"id": "1144", "title": "Simultaneous iterative reconstruction of emission and attenuation images in positron emission tomography from emission data only", "abstract": "For quantitative image reconstruction in positron emission tomography attenuation correction is mandatory. In case that no data are available for the calculation of the attenuation correction factors one can try to determine them from the emission data alone. However, it is not clear if the information content is sufficient to yield an adequate attenuation correction together with a satisfactory activity distribution. Therefore, we determined the log likelihood distribution for a thorax phantom depending on the choice of attenuation and activity pixel values to measure the crosstalk between both. In addition an iterative image reconstruction (one-dimensional Newton-type algorithm with a maximum likelihood estimator), which simultaneously reconstructs the images of the activity distribution and the attenuation coefficients is used to demonstrate the problems and possibilities of such a reconstruction. As result we show that for a change of the log likelihood in the range of statistical noise, the associated change in the activity value of a structure is between 6% and 263%. In addition, we show that it is not possible to choose the best maximum on the basis of the log likelihood when a regularization is used, because the coupling between different structures mediated by the (smoothing) regularization prevents an adequate solution due to crosstalk. We conclude that taking into account the attenuation information in the emission data improves the performance of image reconstruction with respect to the bias of the activities, however, the reconstruction still is not quantitative", "keyphrases": ["image reconstruction", "positron emission tomography attenuation correction", "attenuation correction factors", "log likelihood distribution", "thorax phantom", "activity pixel values", "crosstalk", "iterative image reconstruction", "one-dimensional Newton-type algorithm", "maximum likelihood estimator", "activity distribution", "attenuation coefficients", "statistical noise", "smoothing", "attenuation information"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "92", "title": "Wireless-retail financial services: adoption can't justify the cost", "abstract": "Slow adoption by retail investors, costly services and bankrupt vendors has prompted banks and brokerage firms to turn off their wireless applications", "keyphrases": ["banks", "brokerage firms", "wireless applications"], "prmu": ["P", "P", "P"]} @@ -684,7 +684,7 @@ {"id": "887", "title": "Towards strong stability of concurrent repetitive processes sharing resources", "abstract": "The paper presents a method for design of stability conditions of concurrent, repetitive processes sharing common resources. Steady-state behaviour of the system with m cyclic processes utilising a resource with the mutual exclusion is considered. Based on a recurrent equations framework necessary and sufficient conditions for the existence of maximal performance steady-state are presented. It was shown that if the conditions hold then the m-process system is marginally stable, i.e., a steady-state of the system depends on the perturbations. The problem of finding the relative positions of the processes leading to waiting-free (maximal efficiency) steady-states of the system is formulated as a constraint logic programming problem. An example illustrating the solving of the problem for a 3-process system using object-oriented, constraint logic programming language Oz is presented. A condition sufficient for strong stability of the m-process system is given. When the condition holds then for any initial phases of the processes a waiting-free steady-state will be reached", "keyphrases": ["strong stability", "concurrent repetitive processes", "common resources", "steady-state behaviour", "cyclic processes", "mutual exclusion", "recurrent equations framework", "necessary and sufficient conditions", "maximal performance steady-state", "waiting-free steady-states", "constraint logic programming", "3-process system", "Oz language"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R"]} {"id": "1429", "title": "Online coverage of the Olympic Games", "abstract": "In 1956 a new medium was evolving which helped shape not only the presentation of the Games to a worldwide audience, but created entirely new avenues for marketing and sponsorship which changed the entire economic relevance of the Games. The medium in 1956 was television, and the medium now, of course, is the Internet. Not since 1956 has Olympic coverage been so impacted by the onset of new technology as the current Olympiad has been. But now the IOC finds itself in another set of circumstances not altogether different from 1956", "keyphrases": ["Olympic Games", "online coverage", "marketing", "sponsorship", "economic relevance", "Olympiad", "IOC", "online rights", "e-broadcast"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "M", "U"]} {"id": "1341", "title": "STEM: Secure Telephony Enabled Middlebox", "abstract": "Dynamic applications, including IP telephony, have not seen wide acceptance within enterprises because of problems caused by the existing network infrastructure. Static elements, including firewalls and network address translation devices, are not capable of allowing dynamic applications to operate properly. The Secure Telephony Enabled Middlebox (STEM) architecture is an enhancement of the existing network design to remove the issues surrounding static devices. The architecture incorporates an improved firewall that can interpret and utilize information in the application layer of packets to ensure proper functionality. In addition to allowing dynamic applications to function normally, the STEM architecture also incorporates several detection and response mechanisms for well-known network-based vulnerabilities. This article describes the key components of the architecture with respect to the SIP protocol", "keyphrases": ["Secure Telephony Enabled Middlebox", "STEM", "IP telephony", "network infrastructure", "firewalls", "network address translation devices", "dynamic applications", "STEM architecture", "network design", "static devices", "application layer", "detection mechanisms", "response mechanisms", "network-based vulnerabilities", "SIP protocol"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "P", "P", "P"]} -{"id": "1304", "title": "Center-crossing recurrent neural networks for the evolution of rhythmic behavior", "abstract": "A center-crossing recurrent neural network is one in which the null(hyper)surfaces of each neuron intersect at their exact centers of symmetry, ensuring that each neuron's activation function is centered over the range of net inputs that it receives. We demonstrate that relative to a random initial population, seeding the initial population of an evolutionary search with center-crossing networks significantly improves both the frequency and the speed with which high-fitness oscillatory circuits evolve on a simple walking task. The improvement is especially striking at low mutation variances. Our results suggest that seeding with center-crossing networks may often be beneficial, since a wider range of dynamics is more likely to be easily accessible from a population of center-crossing networks than from a population of random networks", "keyphrases": ["center-crossing recurrent neural networks", "rhythmic behavior evolution", "null surfaces", "symmetry", "activation function", "evolutionary algorithm", "random initial population", "evolutionary search", "high-fitness oscillatory circuits", "learning", "low mutation variance", "random networks"], "prmu": ["P", "R", "R", "P", "P", "M", "P", "P", "P", "U", "P", "P"]} +{"id": "1304", "title": "Center-crossing recurrent neural networks for the evolution of rhythmic behavior", "abstract": "A center-crossing recurrent neural network is one in which the null(hyper)surfaces of each neuron intersect at their exact centers of symmetry, ensuring that each neuron's activation function is centered over the range of net inputs that it receives. We demonstrate that relative to a random initial population, seeding the initial population of an evolutionary search with center-crossing networks significantly improves both the frequency and the speed with which high-fitness oscillatory circuits evolve on a simple walking task. The improvement is especially striking at low mutation variances. Our results suggest that seeding with center-crossing networks may often be beneficial, since a wider range of dynamics is more likely to be easily accessible from a population of center-crossing networks than from a population of random networks", "keyphrases": ["center-crossing recurrent neural networks", "rhythmic behavior evolution", "null surfaces", "symmetry", "activation function", "evolutionary algorithm", "random initial population", "evolutionary search", "high-fitness oscillatory circuits", "learning", "low mutation variance", "random networks"], "prmu": ["P", "R", "U", "P", "P", "M", "P", "P", "P", "U", "P", "P"]} {"id": "751", "title": "A new method of regression on latent variables. Application to spectral data", "abstract": "Several applications are based on the assessment of a linear model linking a set of variables Y to a set of predictors X. In the presence of strong colinearity among predictors, as in the case with spectral data, several alternative procedures to ordinary least squares (OLS) are proposed, We discuss a new alternative approach which we refer to as regression models through constrained principal components analysis (RM-CPCA). This method basically shares certain common characteristics with PLS regression as the dependent variables play a central role in determining the latent variables to be used as predictors. Unlike PLS, however, the approach discussed leads to straightforward models. This method also bears some similarity to latent root regression analysis (LRR) that was discussed by several authors. Moreover, a tuning parameter that ranges between 0 and 1 is introduced and the family of models thus formed includes several other methods as particular cases", "keyphrases": ["latent variables", "spectral data", "linear model", "near-IR spectroscopy", "predictors", "strong colinearity", "regression models through constrained principal components analysis", "dependent variables", "latent root regression analysis", "tuning parameter"], "prmu": ["P", "P", "P", "U", "P", "P", "P", "P", "P", "P"]} {"id": "714", "title": "Embeddings of planar graphs that minimize the number of long-face cycles", "abstract": "We consider the problem of finding embeddings of planar graphs that minimize the number of long-face cycles. We prove that for any k >or= 4, it is NP-complete to find an embedding that minimizes the number of face cycles of length at least k", "keyphrases": ["embeddings", "planar graphs", "long-face cycles", "NP-complete problem", "graph drawing"], "prmu": ["P", "P", "P", "R", "M"]} {"id": "124", "title": "High-speed CMOS circuits with parallel dynamic logic and speed-enhanced skewed static logic", "abstract": "In this paper, we describe parallel dynamic logic (PDL) which exhibits high speed without charge sharing problem. PDL uses only parallel-connected transistors for fast logic evaluation and is a good candidate for high-speed low-voltage operation. It has less back-bias effect compared to other logic styles, which use stacked transistors. Furthermore, PDL needs no signal ordering or tapering. PDL with speed-enhanced skewed static logic renders straightforward logic synthesis without the usual area penalty due to logic duplication. Our experimental results on two 32-bit carry lookahead adders using 0.25- mu m CMOS technology show that PDL with speed-enhanced skewed static (SSS) look reduces the delay over clock-delayed(CD)-domino by 15%-27% and the power-delay product by 20%-37%", "keyphrases": ["high-speed CMOS circuits", "parallel dynamic logic", "speed-enhanced skewed static logic", "parallel-connected transistors", "low-voltage operation", "logic synthesis", "carry lookahead adders", "delay", "power-delay product", "back-bias effect", "stacked transistors", "32 bit", "0.25 micron"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "U", "U"]} @@ -733,7 +733,7 @@ {"id": "657", "title": "The web services agenda", "abstract": "Even the most battle-scarred of CIOs have become excited at the prospect of what web services can do for their businesses. But there are still some shortcomings to be addressed", "keyphrases": ["web services", "transaction support", "security"], "prmu": ["P", "U", "U"]} {"id": "1202", "title": "More than the money [software project]", "abstract": "Experiences creating budgets for large software projects have taught manufacturers that it is not about the money - it is about what one really needs. Before a company can begin to build a budget for a software. project, it has to have a good understanding of what business issues need to be addressed and what the business objectives are. This step is critical because it defines the business goals, outlines the metrics for success, sets the scope for the project, and defines the criteria for selecting the right software", "keyphrases": ["software projects", "manufacturing industry", "budgeting", "management", "software requirements"], "prmu": ["P", "M", "P", "U", "M"]} {"id": "1247", "title": "The changing landscape for multi access portals", "abstract": "Discusses the factors that have made life difficult for consumer portal operators in recent years causing them, like others in the telecommunications, media and technology sector, to take a close look at their business models following the dot.com crash and the consequent reassessment of Internet-related project financing by the venture capital community. While the pressure is on to generate income from existing customers and users, portal operators must reach new markets and find realistic revenue streams. This search for real revenues has led to a move towards charging for content, a strategy being pursued by a large number of horizontal portal players, including MSN and Terra Lycos. This trend is particularly noticeable in China, where Chinadotcom operates a mainland portal and plans a range of fee-based services, including electronic mail. The nature of advertising itself is changing, with portals seeking blue-chip sponsorship and marketing deals that span a number of years. Players are struggling to redefine and reinvent themselves as a result of the changing environment and even the term \"portal\" is believed to be obsolete, partly due to its dot.com crash associations. Multi-access portals are expected to dominate the consumer sector, becoming bigger and better overall than their predecessors and playing a more powerful role in the consumer environment", "keyphrases": ["multi-access portals", "consumer portal operators", "revenue streams", "fee-based services", "advertising", "blue-chip sponsorship"], "prmu": ["P", "P", "P", "P", "P", "P"]} -{"id": "824", "title": "The Internet, knowledge and the academy", "abstract": "As knowledge is released from the bounds of libraries, as research becomes no longer confined to the academy, and education/certification is available, any time/any place, the university and the faculty must redefine themselves. Liberal studies, once the core, and currently eschewed in favor of science and technology, will be reborn in those institutions that can rise above the mundane and embrace an emerging \"third culture\"", "keyphrases": ["Internet", "knowledge", "academy", "education", "certification", "university", "faculty", "liberal studies"], "prmu": ["P", "P", "P", "U", "U", "P", "P", "P"]} +{"id": "824", "title": "The Internet, knowledge and the academy", "abstract": "As knowledge is released from the bounds of libraries, as research becomes no longer confined to the academy, and education/certification is available, any time/any place, the university and the faculty must redefine themselves. Liberal studies, once the core, and currently eschewed in favor of science and technology, will be reborn in those institutions that can rise above the mundane and embrace an emerging \"third culture\"", "keyphrases": ["Internet", "knowledge", "academy", "education", "certification", "university", "faculty", "liberal studies"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "861", "title": "The decision procedure for profitability of investment projects using the internal rate of return of single-period projects", "abstract": "The internal rate of return (IRR) criterion is often used to evaluate profitability of investment projects. In this paper, we focus on a single-period project which consists of two types of cash flows; an investment at one period and a return at a succeeding period, and a financing at one period and a repayment at a succeeding period. We decompose the given investment project into a series of the single-period projects. From the viewpoint of the single-period project, we point out the applicability issue of the IRR criterion, namely the IRR criterion cannot be applied in which a project is composed of both investment type and financing type. Investigating the properties of a series of the single-period projects, we resolve the applicability issue of the IRR criterion and propose the decision procedure for profitability judgment toward any type of investment project based on the comparison between the IRR and the capital cost. We develop a new algorithm to obtain the value of the project investment rate (PIR) for the given project, which is a function of the capital cost, only using the standard IRR computing routine. This outcome is a theoretical breakthrough to widen the utilization of IRR in practical applications", "keyphrases": ["decision procedure", "investment project profitability", "internal rate of return", "single-period projects", "profitability", "cash flows", "investment project decomposition", "IRR criterion", "project investment rate", "PIR"], "prmu": ["P", "R", "P", "P", "P", "P", "M", "P", "P", "P"]} {"id": "1432", "title": "To classify or not to classify, that is the question?", "abstract": "In addressing classification issues, the librarian needs to decide what best suits the purpose and requirements of the user group and the organisation they work in. The author has used the well-established Moys Classification Scheme. This gives the level of detail required for current stock and allows for the incorporation of new material as the firm's specialisations develop. The scheme is widely used in other firms as well as in the local law society library, so it will be familiar to many users", "keyphrases": ["Moys Classification Scheme", "law society library"], "prmu": ["P", "P"]} {"id": "1066", "title": "Application of artificial intelligence to search ground-state geometry of clusters", "abstract": "We introduce a global optimization procedure, the neural-assisted genetic algorithm (NAGA). It combines the power of an artificial neural network (ANN) with the versatility of the genetic algorithm. This method is suitable to solve optimization problems that depend on some kind of heuristics to limit the search space. If a reasonable amount of data is available, the ANN can \"understand\" the problem and provide the genetic algorithm with a selected population of elements that will speed up the search for the optimum solution. We tested the method in a search for the ground-state geometry of silicon clusters. We trained the ANN with information about the geometry and energetics of small silicon clusters. Next, the ANN learned how to restrict the configurational space for larger silicon clusters. For Si/sub 10/ and Si/sub 20/, we noticed that the NAGA is at least three times faster than the \"pure\" genetic algorithm. As the size of the cluster increases, it is expected that the gain in terms of time will increase as well", "keyphrases": ["artificial intelligence", "ground-state geometry", "atomic clusters", "global optimization procedure", "neural-assisted genetic algorithm", "artificial neural network", "population", "optimum solution", "silicon clusters", "Si/sub 10/", "Si/sub 20/", "cluster size"], "prmu": ["P", "P", "M", "P", "P", "P", "P", "P", "P", "P", "P", "R"]} @@ -766,7 +766,7 @@ {"id": "1259", "title": "A mechanism for inferring approximate solutions under incomplete knowledge based on rule similarity", "abstract": "This paper proposes an inference method which can obtain an approximate solution even if the knowledge stored in the problem-solving system is incomplete. When a rule needed for solving the problem does not exist, the problem can be solved by using rules similar to the existing rules. In an implementation using the SLD procedure, a resolution is executed between a subgoal and a rule if an atom of the subgoal is similar to the consequence atom of the rule. Similarities between atoms are calculated using a knowledge base of words with account of the reasoning situation, and the reliability of the derived solution is calculated based on these similarities. If many solutions are obtained, they are grouped into classes of similar solutions and a representative solution is then selected for each class. The proposed method was verified experimentally by solving simple problems", "keyphrases": ["inference method", "approximate solution", "incomplete knowledge", "rule similarity", "problem solving", "SLD procedure", "subgoal atom", "consequence atom", "word knowledge base", "reasoning", "reliability", "representative solution", "common sense knowledge"], "prmu": ["P", "P", "P", "P", "R", "P", "R", "P", "R", "P", "P", "P", "M"]} {"id": "1198", "title": "Post-projected Runge-Kutta methods for index-2 differential-algebraic equations", "abstract": "A new projection technique for Runge-Kutta methods applied to index-2 differential-algebraic equations is presented in which the numerical approximation is projected only as part of the output process. It is shown that for methods that are strictly stable at infinity, the order of convergence is unaffected compared to standard projected methods. Gauss methods, for which this technique is of special interest when some symmetry is to be preserved, are studied in more detail", "keyphrases": ["post-projected Runge-Kutta methods", "numerical approximation", "order of convergence", "projected methods", "index-2 differential-algebraic equations"], "prmu": ["P", "P", "P", "P", "P"]} {"id": "588", "title": "An accurate COG defuzzifier design using Lamarckian co-adaptation of learning and evolution", "abstract": "This paper proposes a design technique of optimal center of gravity (COG) defuzzifier using the Lamarckian co-adaptation of learning and evolution. The proposed COG defuzzifier is specified by various design parameters such as the centers, widths, and modifiers of MFs. The design parameters are adjusted with the Lamarckian co-adaptation of learning and evolution, where the learning performs a local search of design parameters in an individual COG defuzzifier, but the evolution performs a global search of design parameters among a population of various COG defuzzifiers. This co-adaptation scheme allows to evolve much faster than the non-learning case and gives a higher possibility of finding an optimal solution due to its wider searching capability. An application to the truck backer-upper control problem of the proposed co-adaptive design method of COG defuzzifier is presented. The approximation ability and control performance are compared with those of the conventionally simplified COG defuzzifier in terms of the fuzzy logic controller's approximation error and the average tracing distance, respectively", "keyphrases": ["optimal center of gravity defuzzifier", "learning", "evolution", "fuzzy logic controller", "local search"], "prmu": ["R", "P", "P", "P", "P"]} -{"id": "674", "title": "Portal payback", "abstract": "The benefits of deploying a corporate portal are well-documented: access to applications and content is centralised, so users do not spend hours searching for information; the management of disparate applications is also centralised, and by allowing users to access 'self-service' applications in areas such as human resources and procurement, organisations spend less time on manual processing tasks. But how far can prospective customers rely on the ROI figures presented to them by portal technology vendors? In particular, how reliable are the 'ROI calculators' these vendors supply on their web sites?", "keyphrases": ["corporate portal", "return on investment", "ROI calculator", "web sites", "metrics"], "prmu": ["P", "M", "R", "P", "U"]} +{"id": "674", "title": "Portal payback", "abstract": "The benefits of deploying a corporate portal are well-documented: access to applications and content is centralised, so users do not spend hours searching for information; the management of disparate applications is also centralised, and by allowing users to access 'self-service' applications in areas such as human resources and procurement, organisations spend less time on manual processing tasks. But how far can prospective customers rely on the ROI figures presented to them by portal technology vendors? In particular, how reliable are the 'ROI calculators' these vendors supply on their web sites?", "keyphrases": ["corporate portal", "return on investment", "ROI calculator", "web sites", "metrics"], "prmu": ["P", "M", "P", "P", "U"]} {"id": "631", "title": "A modified Fieller interval for the interval estimation of effective doses for a logistic dose-response curve", "abstract": "Interval estimation of the gamma % effective dose ( mu /sub gamma / say) is often based on the asymptotic variance of the maximum likelihood estimator (delta interval) or Fieller's theorem (Fieller interval). Sitter and Wu (1993) compared the delta and Fieller intervals for the median effective dose ( mu /sub 50/) assuming a logistic dose-response curve. Their results indicated that although Fieller intervals are generally superior to delta intervals, they appear to be conservative. Here an adjusted form of the Fieller interval for mu /sub gamma / termed an adjusted Fieller (AF) interval is introduced. A comparison of the AF interval with the delta and Fieller intervals is provided and the properties of these three interval estimation methods are investigated", "keyphrases": ["modified Fieller interval", "interval estimation", "effective doses", "logistic dose-response curve", "asymptotic variance", "maximum likelihood estimator", "delta interval", "Fieller's theorem", "median effective dose"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1264", "title": "Estimation of the vanishing point for automatic driving system using a cross ratio", "abstract": "This paper proposes a new method to estimate the vanishing point used as the vehicle heading, which is essential in automatic driving systems. The proposed method uses a cross ratio comprised of a ratio of lengths from four collinear points for extracting the edges that shape the vanishing point. Then, lines that intersect at one point are fitted to the edges in a Hough space. Consequently, the vanishing point is estimated robustly even when the lane markings are occluded by other vehicles. In the presence of lane markings, the road boundaries are also estimated at the same time. Experimental results from images of a real road scene show the effectiveness of the proposed method", "keyphrases": ["vanishing point estimation", "automatic driving system", "cross ratio", "automatic driving systems", "collinear points", "Hough space", "lane markings", "real road scene"], "prmu": ["R", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1221", "title": "An approach to developing computational supports for reciprocal tutoring", "abstract": "This study presents a novel approach to developing computational supports for reciprocal tutoring. Reciprocal tutoring is a collaborative learning activity, where two participants take turns to play the role of a tutor and a tutee. The computational supports include scaffolding tools for the tutor and a computer-simulated virtual participant. The approach, including system architecture, implementations of scaffolding tools for the tutor and of a virtual participant is presented herein. Furthermore, a system for reciprocal tutoring is implemented as an example of the approach", "keyphrases": ["reciprocal tutoring computational support", "collaborative learning", "scaffolding tools", "computer-simulated virtual participant", "system architecture", "intelligent tutoring system"], "prmu": ["R", "P", "P", "P", "P", "M"]} @@ -817,7 +817,7 @@ {"id": "591", "title": "Approximation theory of fuzzy systems based upon genuine many-valued implications - MIMO cases", "abstract": "It is constructively proved that the multi-input-multi-output fuzzy systems based upon genuine many-valued implications are universal approximators (they are called Boolean type fuzzy systems in this paper). The general approach to construct such fuzzy systems is given, that is, through the partition of the output region (by the given accuracy). Two examples are provided to demonstrate the way in which fuzzy systems are designed to approximate given functions with a given required approximation accuracy", "keyphrases": ["multi-input-multi-output fuzzy systems", "Boolean type fuzzy systems", "fuzzy systems", "many-valued implication", "universal approximator"], "prmu": ["P", "P", "P", "P", "P"]} {"id": "946", "title": "Entanglement measures with asymptotic weak-monotonicity as lower (upper) bound for the entanglement of cost (distillation)", "abstract": "We propose entanglement measures with asymptotic weak-monotonicity. We show that a normalized form of entanglement measures with the asymptotic weak-monotonicity are lower (upper) bound for the entanglement of cost (distillation)", "keyphrases": ["entanglement measures", "asymptotic weak-monotonicity", "entanglement of cost", "distillation"], "prmu": ["P", "P", "P", "P"]} {"id": "105", "title": "Greenberger-Horne-Zeilinger paradoxes for many qubits", "abstract": "We construct Greenberger-Horne-Zeilinger (GHZ) contradictions for three or more parties sharing an entangled state, the dimension of each subsystem being an even integer d. The simplest example that goes beyond the standard GHZ paradox (three qubits) involves five ququats (d = 4). We then examine the criteria that a GHZ paradox must satisfy in order to be genuinely M partite and d dimensional", "keyphrases": ["Greenberger-Horne-Zeilinger paradoxes", "many qubits", "GHZ contradictions", "entangled state", "GHZ paradox"], "prmu": ["P", "P", "R", "P", "P"]} -{"id": "1139", "title": "Development and evaluation of a case-based reasoning classifier for prediction of breast biopsy outcome with BI-RADS/sup TM/ lexicon", "abstract": "Approximately 70-85% of breast biopsies are performed on benign lesions. To reduce this high number of biopsies performed on benign lesions, a case-based reasoning (CBR) classifier was developed to predict biopsy results from BI-RADS/sup TM/ findings. We used 1433 (931 benign) biopsy-proven mammographic cases. CBR similarity was defined using either the Hamming or Euclidean distance measure over case features. Ten features represented each case: calcification distribution, calcification morphology, calcification number, mass margin, mass shape, mass density, mass size, associated findings, special cases, and age. Performance was evaluated using Round Robin sampling, Receiver Operating Characteristic (ROC) analysis, and bootstrap. To determine the most influential features for the CBR, an exhaustive feature search was performed over all possible feature combinations (1022) and similarity thresholds. Influential features were defined as the most frequently occurring features in the feature subsets with the highest partial ROC areas (/sub 0.90/AUC). For CBR with Hamming distance, the most influential features were found to be mass margin, calcification morphology, age, calcification distribution, calcification number, and mass shape, resulting in an /sub 0.90/AUC of 0.33. At 95% sensitivity, the Hamming CBR would spare from biopsy 34% of the benign lesions. At 98% sensitivity, the Hamming CBR would spare 27% benign lesions. For the CBR with Euclidean distance, the most influential feature subset consisted of mass margin, calcification morphology, age, mass density, and associated findings, resulting in /sub 0.90/AUC of 0.37. At 95% sensitivity, the Euclidean CBR would spare from biopsy 41% benign lesions. At 98% sensitivity, the Euclidean CBR would spare 27% benign lesions. The profile of cases spared by both distance measures at 98% sensitivity indicates that the CBR is a potentially useful diagnostic tool for the classification of mammographic lesions, by recommending short-term follow-up for likely benign lesions that is in agreement with final biopsy results and mammographer's intuition", "keyphrases": ["case-based reasoning classifier", "breast biopsy outcome", "BI-RADS lexicon", "benign lesions", "biopsy-proven mammographic cases", "CBR similarity", "Hamming distance measure", "Euclidean distance measure", "calcification distribution", "calcification morphology", "calcification number", "mass margin", "mass shape", "mass density", "mass size", "associated findings", "special cases", "age", "Round Robin sampling", "Receiver Operating Characteristic analysis", "bootstrap", "feature combinations", "similarity thresholds", "feature subsets", "highest partial ROC areas", "influential features", "diagnostic tool", "mammographic lesion classification", "short-term follow-up"], "prmu": ["P", "P", "M", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "R", "P"]} +{"id": "1139", "title": "Development and evaluation of a case-based reasoning classifier for prediction of breast biopsy outcome with BI-RADS/sup TM/ lexicon", "abstract": "Approximately 70-85% of breast biopsies are performed on benign lesions. To reduce this high number of biopsies performed on benign lesions, a case-based reasoning (CBR) classifier was developed to predict biopsy results from BI-RADS/sup TM/ findings. We used 1433 (931 benign) biopsy-proven mammographic cases. CBR similarity was defined using either the Hamming or Euclidean distance measure over case features. Ten features represented each case: calcification distribution, calcification morphology, calcification number, mass margin, mass shape, mass density, mass size, associated findings, special cases, and age. Performance was evaluated using Round Robin sampling, Receiver Operating Characteristic (ROC) analysis, and bootstrap. To determine the most influential features for the CBR, an exhaustive feature search was performed over all possible feature combinations (1022) and similarity thresholds. Influential features were defined as the most frequently occurring features in the feature subsets with the highest partial ROC areas (/sub 0.90/AUC). For CBR with Hamming distance, the most influential features were found to be mass margin, calcification morphology, age, calcification distribution, calcification number, and mass shape, resulting in an /sub 0.90/AUC of 0.33. At 95% sensitivity, the Hamming CBR would spare from biopsy 34% of the benign lesions. At 98% sensitivity, the Hamming CBR would spare 27% benign lesions. For the CBR with Euclidean distance, the most influential feature subset consisted of mass margin, calcification morphology, age, mass density, and associated findings, resulting in /sub 0.90/AUC of 0.37. At 95% sensitivity, the Euclidean CBR would spare from biopsy 41% benign lesions. At 98% sensitivity, the Euclidean CBR would spare 27% benign lesions. The profile of cases spared by both distance measures at 98% sensitivity indicates that the CBR is a potentially useful diagnostic tool for the classification of mammographic lesions, by recommending short-term follow-up for likely benign lesions that is in agreement with final biopsy results and mammographer's intuition", "keyphrases": ["case-based reasoning classifier", "breast biopsy outcome", "BI-RADS lexicon", "benign lesions", "biopsy-proven mammographic cases", "CBR similarity", "Hamming distance measure", "Euclidean distance measure", "calcification distribution", "calcification morphology", "calcification number", "mass margin", "mass shape", "mass density", "mass size", "associated findings", "special cases", "age", "Round Robin sampling", "Receiver Operating Characteristic analysis", "bootstrap", "feature combinations", "similarity thresholds", "feature subsets", "highest partial ROC areas", "influential features", "diagnostic tool", "mammographic lesion classification", "short-term follow-up"], "prmu": ["P", "P", "R", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "R", "P"]} {"id": "1280", "title": "Products and polymorphic subtypes", "abstract": "This paper is devoted to a comprehensive study of polymorphic subtypes with products. We first present a sound and complete Hilbert style axiomatization of the relation of being a subtype in presence of to , * type constructors and the For all quantifier, and we show that such axiornatization is not encodable in the system with to , For all only. In order to give a logical semantics to such a subtyping relation, we propose a new form of a sequent which plays a key role in a natural deduction and a Gentzen style calculi. Interestingly enough, the sequent must have the form E implies T, where E is a non-commutative, non-empty sequence of typing assumptions and T is a finite binary tree of typing judgements, each of them behaving like a pushdown store. We study basic metamathematical properties of the two logical systems, such as subject reduction and cut elimination. Some decidability/undecidability issues related to the presented subtyping relation are also explored: as expected, the subtyping over to , *, For all is undecidable, being already undecidable for the to , For all fragment (as proved in [15]), but for the *, For all fragment it turns out to be decidable", "keyphrases": ["polymorphic subtypes", "products subtypes", "Hilbert style axiomatization", "logical semantics", "Gentzen style calculi", "finite binary tree", "pushdown store", "metamathernatical properties", "decidability"], "prmu": ["P", "R", "P", "P", "P", "P", "P", "M", "P"]} {"id": "690", "title": "Robust Kalman filter design for discrete time-delay systems", "abstract": "The problem of finite- and infinite-horizon robust Kalman filtering for uncertain discrete-time systems with state delay is addressed. The system under consideration is subject to time-varying norm-bounded parameter uncertainty in both the state and output matrices. We develop a new methodology for designing a linear filter such that the error variance of the filter is guaranteed to be within a certain upper bound for any allowed uncertainty and time delay. The solution is given in terms of two Riccati equations. Multiple time-delay systems are also investigated", "keyphrases": ["robust Kalman filter", "discrete time-delay systems", "uncertain systems", "state delay", "time-varying parameter uncertainty", "norm-bounded parameter uncertainty", "state matrices", "output matrices", "linear filter", "Riccati equations", "robust state estimation"], "prmu": ["P", "P", "R", "P", "R", "P", "R", "P", "P", "P", "M"]} {"id": "859", "title": "Developing a hardware and programming curriculum for middle school girls", "abstract": "Techbridge provides experiences and resources that would teach girls technology skills as well as excite their curiosity and build their confidence. Funded by the National Science Foundation and sponsored by Chabot Space and Science Center in Oakland, California, Techbridge is a three-year program that serves approximately 200 girls annually. Techbridge is hosted at 8 middle and high schools in Oakland and at the California School for the Blind in Fremont, California generally as an after-school program meeting once a week. Techbridge comes at a critical time in girls' development when girls have many important decisions to make regarding classes and careers, but often lack the confidence and guidance to make the best choices. Techbridge helps girls plan for the next steps to high school and college with its role models and guidance. Techbridge also provides training and resources for teachers, counselors, and families", "keyphrases": ["middle school girls", "hardware and programming curriculum", "Techbridge", "technology skills teaching"], "prmu": ["P", "P", "P", "R"]} @@ -970,7 +970,7 @@ {"id": "966", "title": "Controlling in between the Lorenz and the Chen systems", "abstract": "This letter investigates a new chaotic system and its role as a joint function between two complex chaotic systems, the Lorenz and the Chen systems, using a simple variable constant controller. With the gradual tuning of the controller, the controlled system evolves from the canonical Lorenz attractor to the Chen attractor through the new transition chaotic attractor. This evolving procedure reveals the forming mechanisms of all similar and closely related chaotic systems, and demonstrates that a simple control technique can be very useful in generating and analyzing some complex chaotic dynamical phenomena", "keyphrases": ["Chen attractors", "Chen system", "Lorenz system", "tuning", "Lorenz attractor", "transition chaotic attractor"], "prmu": ["P", "P", "R", "P", "P", "P"]} {"id": "125", "title": "A fast implementation of correlation of long data sequences for coherent receivers", "abstract": "Coherent reception depends upon matching of phase between the transmitted and received signal. Fast convolution techniques based on fast Fourier transform (FFT) are widely used for extracting time delay information from such matching. The latency in processing a large data window of the received signal is a serious overhead for mission critical real time applications. The implementation of a parallel algorithm for correlation of long data sequences in multiprocessor environment is demonstrated here. The algorithm does processing while acquiring the received signal and reduces the computation overhead considerably because of inherent parallelism", "keyphrases": ["correlation", "long data sequences", "coherent receivers", "transmitted signal", "received signal", "fast Fourier transform", "time delay information", "latency", "mission critical real time applications", "parallel algorithm", "multiprocessor environment", "computation"], "prmu": ["P", "P", "P", "R", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "77", "title": "Modeling frequently accessed wireless data with weak consistency", "abstract": "To reduce the response times of wireless data access in a mobile network, caches are utilized in wireless handheld devices. If the original data entry has been updated, the cached data in the handheld device becomes stale. Thus, a mechanism is required to predict when the cached copy will expire. This paper studies a weakly consistent data access mechanism that computes the time-to-live (TTL) interval to predict the expiration time. We propose an analytic model to investigate this TTL-based algorithm for frequently accessed data. The analytic model is validated against simulation experiments. Our study quantitatively indicates how the TTL-based algorithm reduces the wireless communication cost by increasing the probability of stale accesses. Depending on the requirements of the application, appropriate parameter values can be selected based on the guidelines provided", "keyphrases": ["frequently accessed wireless data modeling", "weak consistency", "response time reduction", "wireless data access", "mobile network", "caches", "wireless handheld devices", "data entry", "time-to-live interval", "expiration time prediction", "analytic model", "simulation experiments", "wireless communication cost", "stale access probability"], "prmu": ["R", "P", "M", "P", "P", "P", "P", "P", "R", "R", "P", "P", "P", "R"]} -{"id": "608", "title": "How closely can a personal computer clock track the UTC timescale via the Internet?", "abstract": "Nowadays many software packages allow you to keep the clock of your personal computer synchronized to time servers spread over the internet. We present how a didactic laboratory can evaluate, in a statistical sense, the minimum synch error of this process (the other extreme, the maximum, is guaranteed by the code itself). The measurement set-up utilizes the global positioning system satellite constellation in 'common view' between two similar timing stations: one acts as a time server for the other, so the final timing difference at the second station represents the total synch error through the internet. Data recorded over batches of 10000 samples show a typical RMS value of 35 ms. This measurement configuration allows students to obtain a much better understanding of the synch task and pushes them, at all times, to look for an experimental verification of data results, even when they come from the most sophisticated 'black boxes' now readily available off the shelf", "keyphrases": ["personal computer clock", "UTC timescale", "internet", "software packages", "time servers", "didactic laboratory", "statistical sense", "global positioning system satellite constellation", "final timing difference", "synch error", "black boxes"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "M"]} +{"id": "608", "title": "How closely can a personal computer clock track the UTC timescale via the Internet?", "abstract": "Nowadays many software packages allow you to keep the clock of your personal computer synchronized to time servers spread over the internet. We present how a didactic laboratory can evaluate, in a statistical sense, the minimum synch error of this process (the other extreme, the maximum, is guaranteed by the code itself). The measurement set-up utilizes the global positioning system satellite constellation in 'common view' between two similar timing stations: one acts as a time server for the other, so the final timing difference at the second station represents the total synch error through the internet. Data recorded over batches of 10000 samples show a typical RMS value of 35 ms. This measurement configuration allows students to obtain a much better understanding of the synch task and pushes them, at all times, to look for an experimental verification of data results, even when they come from the most sophisticated 'black boxes' now readily available off the shelf", "keyphrases": ["personal computer clock", "UTC timescale", "internet", "software packages", "time servers", "didactic laboratory", "statistical sense", "global positioning system satellite constellation", "final timing difference", "synch error", "black boxes"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "P", "P"]} {"id": "1218", "title": "Knowledge acquisition for expert systems in accounting and financial problem domains", "abstract": "Since the mid-1980s, expert systems have been developed for a variety of problems in accounting and finance. The most commonly cited problems in developing these systems are the unavailability of the experts and knowledge engineers and difficulties with the rule extraction process. Within the field of artificial intelligence, this has been called the 'knowledge acquisition' (KA) problem and has been identified as a major bottleneck in the expert system development process. Recent empirical research reveals that certain KA techniques are significantly more efficient than others in helping to extract certain types of knowledge within specific problem domains. This paper presents a mapping between these empirical studies and a generic taxonomy of expert system problem domains. To accomplish this, we first examine the range of problem domains and suggest a mapping of accounting and finance tasks to a generic problem domain taxonomy. We then identify and describe the most prominent KA techniques employed in developing expert systems in accounting and finance. After examining and summarizing the existing empirical KA work, we conclude by showing how the empirical KA research in the various problem domains can be used to provide guidance to developers of expert systems in the fields of accounting and finance", "keyphrases": ["knowledge acquisition", "expert systems", "accounting", "finance", "rule extraction process", "artificial intelligence", "problem domain taxonomy"], "prmu": ["P", "P", "P", "P", "P", "P", "P"]} {"id": "1119", "title": "A component-based software configuration management model and its supporting system", "abstract": "Software configuration management (SCM) is an important key technology in software development. Component-based software development (CBSD) is an emerging paradigm in software development. However, to apply CBSD effectively in real world practice, supporting SCM in CBSD needs to be further investigated. In this paper, the objects that need to be managed in CBSD is analyzed and a component-based SCM model is presented. In this model, components, as the integral logical constituents in a system, are managed as the basic configuration items in SCM, and the relationships between/among components are defined and maintained. Based on this model, a configuration management system is implemented", "keyphrases": ["component-based software configuration management model", "software development", "integral logical constituents", "software reuse", "version control"], "prmu": ["P", "P", "P", "M", "U"]} {"id": "1004", "title": "Games machines play", "abstract": "Individual rationality, or doing what is best for oneself, is a standard model used to explain and predict human behavior, and von Neumann-Morgenstern game theory is the classical mathematical formalization of this theory in multiple-agent settings. Individual rationality, however, is an inadequate model for the synthesis of artificial social systems where cooperation is essential, since it does not permit the accommodation of group interests other than as aggregations of individual interests. Satisficing game theory is based upon a well-defined notion of being good enough, and does accommodate group as well as individual interests through the use of conditional preference relationships, whereby a decision maker is able to adjust its preferences as a function of the preferences, and not just the options, of others. This new theory is offered as an alternative paradigm to construct artificial societies that are capable of complex behavior that goes beyond exclusive self interest", "keyphrases": ["individual rationality", "human behavior", "game theory", "multiple-agent", "artificial social systems", "cooperation", "conditional preference relationships", "artificial societies", "self interest", "decision theory", "group rationality"], "prmu": ["P", "P", "P", "P", "P", "P", "P", "P", "P", "R", "R"]}