text
stringlengths 0
4.09k
|
|---|
Title: Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data
|
Abstract: When outcomes are missing for reasons beyond an investigator's control, there are two different ways to adjust a parameter estimate for covariates that may be related both to the outcome and to missingness. One approach is to model the relationships between the covariates and the outcome and use those relationships to predict the missing values. Another is to model the probabilities of missingness given the covariates and incorporate them into a weighted or stratified estimate. Doubly robust (DR) procedures apply both types of model simultaneously and produce a consistent estimate of the parameter if either of the two models has been correctly specified. In this article, we show that DR estimates can be constructed in many ways. We compare the performance of various DR and non-DR estimates of a population mean in a simulated example where both models are incorrect but neither is grossly misspecified. Methods that use inverse-probabilities as weights, whether they are DR or not, are sensitive to misspecification of the propensity model when some estimated propensities are small. Many DR methods perform better than simple inverse-probability weighting. None of the DR methods we tried, however, improved upon the performance of simple regression-based prediction of the missing values. This study does not represent every missing-data problem that will arise in practice. But it does demonstrate that, in at least some settings, two wrong models are not better than one.
|
Title: Comment: Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data
|
Abstract: Comment on ``Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data'' [arXiv:0804.2958]
|
Title: Comment: Performance of Double-Robust Estimators When ``Inverse Probability'' Weights Are Highly Variable
|
Abstract: Comment on ``Performance of Double-Robust Estimators When ``Inverse Probability'' Weights Are Highly Variable'' [arXiv:0804.2958]
|
Title: Comment: Understanding OR, PS and DR
|
Abstract: Comment on ``Understanding OR, PS and DR'' [arXiv:0804.2958]
|
Title: Comment: Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data
|
Abstract: Comment on ``Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data'' [arXiv:0804.2958]
|
Title: Rejoinder: Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data
|
Abstract: Rejoinder to ``Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data'' [arXiv:0804.2958]
|
Title: Measuring Traffic
|
Abstract: A traffic performance measurement system, PeMS, currently functions as a statewide repository for traffic data gathered by thousands of automatic sensors. It has integrated data collection, processing and communications infrastructure with data storage and analytical tools. In this paper, we discuss statistical issues that have emerged as we attempt to process a data stream of 2 GB per day of wildly varying quality. In particular, we focus on detecting sensor malfunction, imputation of missing or bad data, estimation of velocity and forecasting of travel times on freeway networks.
|
Title: The Epic Story of Maximum Likelihood
|
Abstract: At a superficial level, the idea of maximum likelihood must be prehistoric: early hunters and gatherers may not have used the words ``method of maximum likelihood'' to describe their choice of where and how to hunt and gather, but it is hard to believe they would have been surprised if their method had been described in those terms. It seems a simple, even unassailable idea: Who would rise to argue in favor of a method of minimum likelihood, or even mediocre likelihood? And yet the mathematical history of the topic shows this ``simple idea'' is really anything but simple. Joseph Louis Lagrange, Daniel Bernoulli, Leonard Euler, Pierre Simon Laplace and Carl Friedrich Gauss are only some of those who explored the topic, not always in ways we would sanction today. In this article, that history is reviewed from back well before Fisher to the time of Lucien Le Cam's dissertation. In the process Fisher's unpublished 1930 characterization of conditions for the consistency and efficiency of maximum likelihood estimates is presented, and the mathematical basis of his three proofs discussed. In particular, Fisher's derivation of the information inequality is seen to be derived from his work on the analysis of variance, and his later approach via estimating functions was derived from Euler's Relation for homogeneous functions. The reaction to Fisher's work is reviewed, and some lessons drawn.
|
Title: Generalized SURE for Exponential Families: Applications to Regularization
|
Abstract: Stein's unbiased risk estimate (SURE) was proposed by Stein for the independent, identically distributed (iid) Gaussian model in order to derive estimates that dominate least-squares (LS). In recent years, the SURE criterion has been employed in a variety of denoising problems for choosing regularization parameters that minimize an estimate of the mean-squared error (MSE). However, its use has been limited to the iid case which precludes many important applications. In this paper we begin by deriving a SURE counterpart for general, not necessarily iid distributions from the exponential family. This enables extending the SURE design technique to a much broader class of problems. Based on this generalization we suggest a new method for choosing regularization parameters in penalized LS estimators. We then demonstrate its superior performance over the conventional generalized cross validation approach and the discrepancy method in the context of image deblurring and deconvolution. The SURE technique can also be used to design estimates without predefining their structure. However, allowing for too many free parameters impairs the performance of the resulting estimates. To address this inherent tradeoff we propose a regularized SURE objective. Based on this design criterion, we derive a wavelet denoising strategy that is similar in sprit to the standard soft-threshold approach but can lead to improved MSE performance.
|
Title: Bayesian computation for statistical models with intractable normalizing constants
|
Abstract: This paper deals with some computational aspects in the Bayesian analysis of statistical models with intractable normalizing constants. In the presence of intractable normalizing constants in the likelihood function, traditional MCMC methods cannot be applied. We propose an approach to sample from such posterior distributions. The method can be thought as a Bayesian version of the MCMC-MLE approach of Geyer and Thompson (1992). To the best of our knowledge, this is the first general and asymptotically consistent Monte Carlo method for such problems. We illustrate the method with examples from image segmentation and social network modeling. We study as well the asymptotic behavior of the algorithm and obtain a strong law of large numbers for empirical averages.
|
Title: On the performance of approximate equilibria in congestion games
|
Abstract: We study the performance of approximate Nash equilibria for linear congestion games. We consider how much the price of anarchy worsens and how much the price of stability improves as a function of the approximation factor $\epsilon$. We give (almost) tight upper and lower bounds for both the price of anarchy and the price of stability for atomic and non-atomic congestion games. Our results not only encompass and generalize the existing results of exact equilibria to $\epsilon$-Nash equilibria, but they also provide a unified approach which reveals the common threads of the atomic and non-atomic price of anarchy results. By expanding the spectrum, we also cast the existing results in a new light. For example, the Pigou network, which gives tight results for exact Nash equilibria of selfish routing, remains tight for the price of stability of $\epsilon$-Nash equilibria but not for the price of anarchy.
|
Title: Technical Report - Automatic Contour Extraction from 2D Neuron Images
|
Abstract: This work describes a novel methodology for automatic contour extraction from 2D images of 3D neurons (e.g. camera lucida images and other types of 2D microscopy). Most contour-based shape analysis methods can not be used to characterize such cells because of overlaps between neuronal processes. The proposed framework is specifically aimed at the problem of contour following even in presence of multiple overlaps. First, the input image is preprocessed in order to obtain an 8-connected skeleton with one-pixel-wide branches, as well as a set of critical regions (i.e., bifurcations and crossings). Next, for each subtree, the tracking stage iteratively labels all valid pixel of branches, up to a critical region, where it determines the suitable direction to proceed. Finally, the labeled skeleton segments are followed in order to yield the parametric contour of the neuronal shape under analysis. The reported system was successfully tested with respect to several images and the results from a set of three neuron images are presented here, each pertaining to a different class, i.e. alpha, delta and epsilon ganglion cells, containing a total of 34 crossings. The algorithms successfully got across all these overlaps. The method has also been found to exhibit robustness even for images with close parallel segments. The proposed method is robust and may be implemented in an efficient manner. The introduction of this approach should pave the way for more systematic application of contour-based shape analysis methods in neuronal morphology.
|
Title: A Conversation with Seymour Geisser
|
Abstract: Seymour Geisser received his bachelor's degree in Mathematics from the City College of New York in 1950, and his M.A. and Ph.D. degrees in Mathematical Statistics at the University of North Carolina in 1952 and 1955, respectively. He then held positions at the National Bureau of Standards and the National Institute of Mental Health until 1961. From 1961 until 1965, he was Chief of the Biometry Section at the National Institute of Arthritis and Metabolic Diseases, and also held the position of Professorial Lecturer at the George Washington University from 1960 to 1965. From 1965 to 1970, he was the founding Chair of the Department of Statistics at the State University of New York, Buffalo, and in 1971, he became the founding Director of the School of Statistics at the University of Minnesota, remaining in that position until 2001. He held visiting professorships at Iowa State University, 1960; University of Wisconsin, 1964; University of Tel-Aviv (Israel), 1971; University of Waterloo (Canada), 1972; Stanford University, 1976, 1977, 1988; Carnegie Mellon University, 1976; University of the Orange Free State (South Africa), 1978, 1993; Harvard University, 1981; University of Chicago, 1985; University of Warwick (England), 1986; University of Modena (Italy), 1996; and National Chiao Tung University (Taiwan), 1998. He was the Lady Davis Visiting Professor, Hebrew University of Jerusalem, 1991, 1994, 1999, and the Schor Scholar, Merck Research Laboratories, 2002-2003. He was a Fellow of the Institute of Mathematical Statistics and the American Statistical Association.
|
Title: A Conversation with Monroe Sirken
|
Abstract: Born January 11, 1921 in New York City, Monroe Sirken grew up in a suburb of Pasadena, California. He earned B.A. and M.A. degrees in sociology at UCLA in 1946 and 1947, and a Ph.D. in 1950 in sociology with a minor in mathematics at the University of Washington in 1950 where Professor Z. W. Birnbaum was his mentor and thesis advisor. As a Post-Doctoral Fellow of the Social Science Research Council, Monroe spent 1950--1951 at the Statistics Laboratory, University of California at Berkeley and the Office of the Assistant Director for Research, U.S. Bureau of the Census in Suitland, Maryland. Monroe visited the Census Bureau at a time of great change in the use of sampling and survey methods, and decided to remain. He began his government career there in 1951 as a mathematical statistician, and moved to the National Office of Vital Statistics (NOVS) in 1953 where he was an actuarial mathematician and a mathematical statistician. He has held a variety of research and administrative positions at the National Center for Health Statistics (NCHS) and he was the Associate Director, Research and Methodology and the Director, Office of Research and Methodology until 1996 when he became a senior research scientist, the title he currently holds. Aside from administrative responsibilities, Monroe's major professional interests have been conducting and fostering survey and statistical research responsive to the needs of federal statistics. His interest in the design of rare and sensitive population surveys led to the development of network sampling which improves precision by linking multiple selection units to the same observation units. His interest in fostering research on the cognitive aspects of survey methods led to the establishment of permanent questionnaire design research laboratories, first at NCHS and later at other federal statistical agencies here and abroad.
|
Title: Phoneme recognition in TIMIT with BLSTM-CTC
|
Abstract: We compare the performance of a recurrent neural network with the best results published so far on phoneme recognition in the TIMIT database. These published results have been obtained with a combination of classifiers. However, in this paper we apply a single recurrent neural network to the same task. Our recurrent neural network attains an error rate of 24.6%. This result is not significantly different from that obtained by the other best methods, but they rely on a combination of classifiers for achieving comparable performance.
|
Title: A New Approach to Automated Epileptic Diagnosis Using EEG and Probabilistic Neural Network
|
Abstract: Epilepsy is one of the most common neurological disorders that greatly impair patient' daily lives. Traditional epileptic diagnosis relies on tedious visual screening by neurologists from lengthy EEG recording that requires the presence of seizure (ictal) activities. Nowadays, there are many systems helping the neurologists to quickly find interesting segments of the lengthy signal by automatic seizure detection. However, we notice that it is very difficult, if not impossible, to obtain long-term EEG data with seizure activities for epilepsy patients in areas lack of medical resources and trained neurologists. Therefore, we propose to study automated epileptic diagnosis using interictal EEG data that is much easier to collect than ictal data. The authors are not aware of any report on automated EEG diagnostic system that can accurately distinguish patients' interictal EEG from the EEG of normal people. The research presented in this paper, therefore, aims to develop an automated diagnostic system that can use interictal EEG data to diagnose whether the person is epileptic. Such a system should also detect seizure activities for further investigation by doctors and potential patient monitoring. To develop such a system, we extract four classes of features from the EEG data and build a Probabilistic Neural Network (PNN) fed with these features. Leave-one-out cross-validation (LOO-CV) on a widely used epileptic-normal data set reflects an impressive 99.5% accuracy of our system on distinguishing normal people's EEG from patient's interictal EEG. We also find our system can be used in patient monitoring (seizure detection) and seizure focus localization, with 96.7% and 77.5% accuracy respectively on the data set.
|
Title: A Method of Trend Extraction Using Singular Spectrum Analysis
|
Abstract: The paper presents a new method of trend extraction in the framework of the Singular Spectrum Analysis (SSA) approach. This method is easy to use, does not need specification of models of time series and trend, allows to extract trend in the presence of noise and oscillations and has only two parameters (besides basic SSA parameter called window length). One parameter manages scale of the extracted trend and another is a method specific threshold value. We propose procedures for the choice of the parameters. The presented method is evaluated on a simulated time series with a polynomial trend and an oscillating component with unknown period and on the seasonally adjusted monthly data of unemployment level in Alaska for the period 1976/01-2006/09.
|
Title: Natural pseudo-distance and optimal matching between reduced size functions
|
Abstract: This paper studies the properties of a new lower bound for the natural pseudo-distance. The natural pseudo-distance is a dissimilarity measure between shapes, where a shape is viewed as a topological space endowed with a real-valued continuous function. Measuring dissimilarity amounts to minimizing the change in the functions due to the application of homeomorphisms between topological spaces, with respect to the $L_\infty$-norm. In order to obtain the lower bound, a suitable metric between size functions, called matching distance, is introduced. It compares size functions by solving an optimal matching problem between countable point sets. The matching distance is shown to be resistant to perturbations, implying that it is always smaller than the natural pseudo-distance. We also prove that the lower bound so obtained is sharp and cannot be improved by any other distance between size functions.
|
Title: Isotropic PCA and Affine-Invariant Clustering
|
Abstract: We present a new algorithm for clustering points in R^n. The key property of the algorithm is that it is affine-invariant, i.e., it produces the same partition for any affine transformation of the input. It has strong guarantees when the input is drawn from a mixture model. For a mixture of two arbitrary Gaussians, the algorithm correctly classifies the sample assuming only that the two components are separable by a hyperplane, i.e., there exists a halfspace that contains most of one Gaussian and almost none of the other in probability mass. This is nearly the best possible, improving known results substantially. For k > 2 components, the algorithm requires only that there be some (k-1)-dimensional subspace in which the emoverlap in every direction is small. Here we define overlap to be the ratio of the following two quantities: 1) the average squared distance between a point and the mean of its component, and 2) the average squared distance between a point and the mean of the mixture. The main result may also be stated in the language of linear discriminant analysis: if the standard Fisher discriminant is small enough, labels are not needed to estimate the optimal subspace for projection. Our main tools are isotropic transformation, spectral projection and a simple reweighting technique. We call this combination isotropic PCA.
|
Title: Respect My Authority! HITS Without Hyperlinks, Utilizing Cluster-Based Language Models
|
Abstract: We present an approach to improving the precision of an initial document ranking wherein we utilize cluster information within a graph-based framework. The main idea is to perform re-ranking based on centrality within bipartite graphs of documents (on one side) and clusters (on the other side), on the premise that these are mutually reinforcing entities. Links between entities are created via consideration of language models induced from them. We find that our cluster-document graphs give rise to much better retrieval performance than previously proposed document-only graphs do. For example, authority-based re-ranking of documents via a HITS-style cluster-based approach outperforms a previously-proposed PageRank-inspired algorithm applied to solely-document graphs. Moreover, we also show that computing authority scores for clusters constitutes an effective method for identifying clusters containing a large percentage of relevant documents.
|
Title: Causal inference using the algorithmic Markov condition
|
Abstract: Inferring the causal structure that links n observables is usually based upon detecting statistical dependences and choosing simple graphs that make the joint measure Markovian. Here we argue why causal inference is also possible when only single observations are present. We develop a theory how to generate causal graphs explaining similarities between single objects. To this end, we replace the notion of conditional stochastic independence in the causal Markov condition with the vanishing of conditional algorithmic mutual information and describe the corresponding causal inference rules. We explain why a consistent reformulation of causal inference in terms of algorithmic complexity implies a new inference principle that takes into account also the complexity of conditional probability densities, making it possible to select among Markov equivalent causal graphs. This insight provides a theoretical foundation of a heuristic principle proposed in earlier work. We also discuss how to replace Kolmogorov complexity with decidable complexity criteria. This can be seen as an algorithmic analog of replacing the empirically undecidable question of statistical independence with practical independence tests that are based on implicit or explicit assumptions on the underlying distribution.
|
Title: Multiple Random Oracles Are Better Than One
|
Abstract: We study the problem of learning k-juntas given access to examples drawn from a number of different product distributions. Thus we wish to learn a function f : -1,1^n -> -1,1 that depends on k (unknown) coordinates. While the best known algorithms for the general problem of learning a k-junta require running time of n^k * poly(n,2^k), we show that given access to k different product distributions with biases separated by \gamma>0, the functions may be learned in time poly(n,2^k,\gamma^-k). More generally, given access to t <= k different product distributions, the functions may be learned in time n^k/t * poly(n,2^k,\gamma^-k). Our techniques involve novel results in Fourier analysis relating Fourier expansions with respect to different biases and a generalization of Russo's formula.
|
Title: A Quasi-Newton Approach to Nonsmooth Convex Optimization Problems in Machine Learning
|
Abstract: We extend the well-known BFGS quasi-Newton method and its memory-limited variant LBFGS to the optimization of nonsmooth convex objectives. This is done in a rigorous fashion by generalizing three components of BFGS to subdifferentials: the local quadratic model, the identification of a descent direction, and the Wolfe line search conditions. We prove that under some technical conditions, the resulting subBFGS algorithm is globally convergent in objective function value. We apply its memory-limited variant (subLBFGS) to L_2-regularized risk minimization with the binary hinge loss. To extend our algorithm to the multiclass and multilabel settings, we develop a new, efficient, exact line search algorithm. We prove its worst-case time complexity bounds, and show that our line search can also be used to extend a recently developed bundle method to the multiclass and multilabel settings. We also apply the direction-finding component of our algorithm to L_1-regularized risk minimization with logistic loss. In all these contexts our methods perform comparable to or better than specialized state-of-the-art solvers on a number of publicly available datasets. An open source implementation of our algorithms is freely available.
|
Title: Modelling coloured residual noise in gravitational-wave signal processing
|
Abstract: We introduce a signal processing model for signals in non-white noise, where the exact noise spectrum is a priori unknown. The model is based on a Student's t distribution and constitutes a natural generalization of the widely used normal (Gaussian) model. This way, it allows for uncertainty in the noise spectrum, or more generally is also able to accommodate outliers (heavy-tailed noise) in the data. Examples are given pertaining to data from gravitational wave detectors.
|
Title: Image Processing in Optical Guidance for Autonomous Landing of Lunar Probe
|
Abstract: Because of the communication delay between earth and moon, the GNC technology of lunar probe is becoming more important than ever. Current navigation technology is not able to provide precise motion estimation for probe landing control system Computer vision offers a new approach to solve this problem. In this paper, author introduces an image process algorithm of computer vision navigation for autonomous landing of lunar probe. The purpose of the algorithm is to detect and track feature points which are factors of navigation. Firstly, fixation areas are detected as sub-images and matched. Secondly, feature points are extracted from sub-images and tracked. Computer simulation demonstrates the result of algorithm takes less computation and fulfils requests of navigation algorithm.
|
Title: Hardware In The Loop Simulator in UAV Rapid Development Life Cycle
|
Abstract: Field trial is very critical and high risk in autonomous UAV development life cycle. Hardware in the loop (HIL) simulation is a computer simulation that has the ability to simulate UAV flight characteristic, sensor modeling and actuator modeling while communicating in real time with the UAV autopilot hardware. HIL simulation can be used to test the UAV autopilot hardware reliability, test the closed loop performance of the overall system and tuning the control parameter. By rigorous testing in the HIL simulator, the risk in the field trial can be minimized.
|
Title: Effects of Leaders Position and Shape on Aerodynamic Performances of V Flight Formation
|
Abstract: The influences of the leader in a group of V flight formation are dealt with. The investigation is focused on the effect of its position and shape on aerodynamics performances of a given V flight formation. Vortices generated the wing tip of the leader moves downstream forming a pair of opposite rotating line vortices. These vortices are generally undesirable because they create a downwash that increases the induced drag on leaders wing. However, this downwash is also accompanied by an upwash that can beneficial to the followers wing flying behind the leaders one, namely a favorable lift for the followers wing. How much contributions of the leaders wing to the followers wing in the V formation flight is determined by the strength of tip vortices generated by the leaders wing which is influenced by its position and shape including incidence angle, dihedral angle, aspect ratio and taper ratio. The prediction of aerodynamic performances of the V flight formation including lift, drag and moment coefficients is numerically performed by solving Navier Stokes equations with k e turbulence model. The computational domain is defined with multiblock topology to capture the complex geometry arrangement of the V flight formation.
|
Title: Automated Flight Test and System Identification for Rotary Wing Small Aerial Platform using Frequency Responses Analysis
|
Abstract: This paper proposes an autopilot system that can be used to control the small scale rotorcraft during the flight test for linear-frequency-domain system identification. The input frequency swept is generated automatically as part of the autopilot control command. Therefore the bandwidth coverage and consistency of the frequency swept is guaranteed to produce high quality data for system identification. Beside that we can set the safety parameter during the flight test (maximum roll or pitch value, minimum altitude, etc) so the safety of the whole flight test is guaranteed. This autopilot for automated flight test will be tested using hardware in the loop simulator for hover flight condition.
|
Title: Virtual Reality Simulation of Fire Fighting Robot Dynamic and Motion
|
Abstract: This paper presents one approach in designing a Fire Fighting Robot which has been contested annually in a robotic student competition in many countries following the rules initiated at the Trinity College. The approach makes use of computer simulation and animation in a virtual reality environment. In the simulation, the amount of time, starting from home until the flame is destroyed, can be confirmed. The efficacy of algorithms and parameter values employed can be easily evaluated. Rather than spending time building the real robot in a trial and error fashion, now students can explore more variation of algorithm, parameter and sensor-actuator configuration in the early stage of design. Besides providing additional excitement during learning process and enhancing students understanding to the engineering aspects of the design, this approach could become a useful tool to increase the chance of winning the contest.
|
Title: Heading Lock Maneuver Testing of Autonomous Underwater Vehicle
|
Abstract: In recent years, Autonomous Underwater Vehicle (UAV) research and development at Bandung Institute of Technology in Indonesia has achieved the testing stage in the field. This testing was still being classified as the early testing, since some of the preliminary tests were carried out in the scale of the laboratory. The paper would discuss the laboratory test and several tests that were done in the field. Discussions were stressed in the procedure and the aim that will be achieved, along with several early results. The testing was carried out in the lake with the area around 8300 Ha and the maximum depth of 50 meters. The location of the testing was chosen with consideration of minimizing the effect of the current and the wave, as well as the location that was not too far from the Laboratory. The type of testing that will be discussed in paper was Heading Lock Maneuver Testing. The vehicle was tested to move with a certain cruising speed, afterwards it was commanded by an arbitrarily selected heading direction. The response and the behavior of the vehicle were recorded as the data produced by the testing.
|
Title: Development of Architectures for Internet Telerobotics Systems
|
Abstract: This paper presents our experience in developing and implementing Internet telerobotics system. Internet telerobotics system refers to a robot system controlled and monitored remotely through the Internet. A robot manipulator with five degrees of freedom, called Mentor, is employed. Client-server architecture is chosen as a platform for our Internet telerobotics system. Three generations of telerobotics systems have evolved in this research. The first generation was based on CGI and two tiered architecture, where a client presents a Graphical User Interface to the user, and utilizes the user's data entry and actions to perform requests to robot server running on a different machine. The second generation was developed using Java. We also employ Java 3D for creating and manipulating 3D geometry of manipulator links and for constructing the structures used in rendering that geometry, resulting in 3D robot movement simulation presented to the users(clients) through their web browser. Recent development in our Internet telerobotics includes object recognition through image captured by a camera, which poses challenging problem, given the undeterministic latency of the Internet. The third generation is centered around the use of CORBA for development platform of distributed internet telerobotics system, aimed at distributing task of telerobotics system.
|
Title: Unmanned Aerial Vehicle Instrumentation for Rapid Aerial Photo System
|
Abstract: This research will proposed a new kind of relatively low cost autonomous UAV that will enable farmers to make just in time mosaics of aerial photo of their crop. These mosaics of aerial photo should be able to be produced with relatively low cost and within the 24 hours of acquisition constraint. The autonomous UAV will be equipped with payload management system specifically developed for rapid aerial mapping. As mentioned before turn around time is the key factor, so accuracy is not the main focus (not orthorectified aerial mapping). This system will also be equipped with special software to post process the aerial photos to produce the mosaic aerial photo map
|
Title: First Principle Approach to Modeling of Small Scale Helicopter
|
Abstract: The establishment of global helicopter linear model is very precious and useful for the design of the linear control laws, since it is never afforded in the published literatures. In the first principle approach, the mathematical model was developed using basic helicopter theory accounting for particular characteristic of the miniature helicopter. No formal system identification procedures are required for the proposed model structure. The relevant published literatures however did not present the linear models required for the design of linear control laws. The paper presents a step by step development of linear model for small scale helicopter based on first-principle approach. Beyond the previous work in literatures, the calculation of the stability derivatives is presented in detail. A computer program is used to solve the equilibrium conditions and then calculate the change in aerodynamics forces and moments due to the change in each degree of freedom and control input. The detail derivation allows the comprehensive analysis of relative dominance of vehicle states and input variables to force and moment components. Hence it facilitates the development of minimum complexity small scale helicopter dynamics model.
|
Title: Optimal Tracking Controller Design for a Small Scale Helicopter
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.