text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
It is important that one’s decision of their own life is respected.
[ 0.004605887457728386, 0.08484996110200882, -0.009248756803572178, -0.001936695072799921, 0.048571646213531494, -0.016934532672166824, 0.01783319190144539, -0.018416989594697952, -0.029644524678587914, -0.002143777906894684, 0.0022443737834692, 0.028797036036849022, 0.0010537616908550262, 0...
[ 0.44593942165374756, 0.20111925899982452, -0.21556605398654938, -0.23794446885585785, 0.13590413331985474, 0.08249932527542114, 0.4544777572154999, -0.19786319136619568, 0.015933671966195107, -0.582105278968811, -0.21693773567676544, 0.5569050908088684, -0.08633347600698471, -0.16184775531...
Some of the heterotic string models have an $E_6\otimes E_8$ symmetry. Examples include some orbifold models, some free fermionic models and Gepner models. We can break the gauge symmetry by including Wilson lines. Does anyone know if it is possible to break $E_6$ to $SO(10)$ while maintaining spacetime SUSY? I have seen examples in Gepner models where $E_6$ is broken to a smaller group than $SO(10)$ while keeping SUSY but none where $E_6$ is broken to $SO(10)$. Is this pure coincidence or is there a reason behind it? Any insights are appreciated!
[ -0.004212179221212864, 0.0084603950381279, -0.023897357285022736, 0.02037985250353813, 0.009908566251397133, -0.009320768527686596, 0.010335620492696762, 0.0031928885728120804, -0.009185781702399254, -0.015462785959243774, -0.00550902308896184, 0.011669768020510674, -0.028572097420692444, ...
[ 0.2092093974351883, -0.1730378419160843, 0.19458924233913422, 0.09149397909641266, -0.2426140457391739, -0.15440328419208527, 0.34892526268959045, -0.26770809292793274, -0.5208526253700256, -0.2531164288520813, -0.2853381335735321, 0.25010088086128235, -0.5264565944671631, 0.62934762239456...
The question goes like this : > A thin non conducting horizontal disc of mass $m$ having total charge $q$ > distributed uniformly over its surface, can rotate freely about its own > axis. Initially when the disk is stationary a magnetic field $B$ directed > perpendicular to the plane is switched on at $t=0$. Find the angular > velocity $\omega$ acquired by disc as a function of time, if $B=kt$, where > $k$ is an arbitrary non-zero constant and $t$ is time The given diagram : ![Disc in time varying magnetic field](http://i.stack.imgur.com/5bxM4.jpg) What I did was to find out the induced emf in the loop due to the changing magnetic flux, using Faraday's laws. These are my results: The emf:$$ \epsilon=k\pi R^2$$ The magnetic moment of the disc when it rotates with angular velocity $\omega$ : $$\mu=\frac{q\omega R^2}{2}$$ But now I am stuck. What should be my next step? Any hints??
[ -0.006184782832860947, 0.008701837621629238, -0.010118870995938778, 0.005541602149605751, -0.0003529407549649477, -0.0246315635740757, 0.007457131985574961, 0.0016772961243987083, -0.00982295535504818, -0.002434789901599288, -0.005319535732269287, 0.017683938145637512, 0.006502814590930939, ...
[ 0.1451752930879593, -0.2611885666847229, 0.9976155757904053, 0.328350692987442, 0.08454187959432602, 0.14253942668437958, -0.38255468010902405, -0.21941857039928436, -0.3385540246963501, -0.34173882007598877, -0.10718276351690292, 0.6139950752258301, -0.19772787392139435, 0.504362106323242...
Is there any syslog log analyzer script I can install on hosting to track user activity? I have giant user log and I need more specific log details.
[ -0.0032566883601248264, -0.002946414751932025, -0.011635493487119675, 0.028866181150078773, -0.020404646173119545, 0.016948478296399117, 0.014434624463319778, -0.0023214463144540787, -0.04776206612586975, -0.010477099567651749, -0.0008344577508978546, 0.003393421648070216, -0.004627582151442...
[ 0.39265644550323486, 0.13020265102386475, -0.026958301663398743, 0.3404944837093353, -0.05466710403561592, -0.4078192412853241, 0.24688971042633057, 0.38457930088043213, -0.2509784698486328, -0.797400951385498, 0.14698950946331024, 0.46110501885414124, -0.10260014235973358, 0.0260570179671...
I need help to understand the meaning of the following formula: $$\sum_{i=1}^n\sum_{j\neq i,j=1}^n(x_i-\mu_i)(x_j-\mu_j)c_{ij}$$ where C is positive semidefinite matrix, $\mu_i$ is the mean of $x_i$. I was thinking, since $c_{ij}$ is the covariance between $x_i$ and $x_j$, this formula kind of sums up the correlation between all the pairs of $x_i$ and $x_j$. Does that make sense? Do you have any other interpretation of the formula? Many thanks in advance, **Edited v1** Matrix $C$ is simply a positive semidefinite matrix, not necessarily covariance matrix. $\mu $ was given very general as well, not mentioned to be mean. That was my own interpretation of it. If you know of another formula using $\mu$ with different meaning, please do tell. Many thanks for Emre and Nesp for the response. So far, it really makes sense. **Edited v2** Thank you all very much for your responses. The Mahalanobis distance is a very good explanation of the formula given. However, I wonder if we take off the diagonal (all the i=j elements) from the formula, does it still make any sense? Without the condition $i\neq j$, the summation can even be negative, right? In that case, the interpretation of distance seems to be not appropriate anymore. (I made the modification to the formula above)
[ -0.006804558448493481, 0.0028584732208400965, -0.0016620877431705594, 0.01119222305715084, 0.006062987260520458, -0.00021986383944749832, 0.005336740054190159, -0.02413872256875038, -0.009209832176566124, -0.0013126530684530735, -0.007358673494309187, 0.008307093754410744, -0.008949653245508...
[ 0.08985602110624313, 0.1163126677274704, 0.5303542613983154, -0.1749504953622818, 0.11408548802137375, 0.00535102654248476, 0.04090064764022827, -0.3260931968688965, -0.07162681967020035, -0.3195745348930359, -0.01747174747288227, 0.43090519309043884, -0.19597148895263672, 0.45239192247390...
I am trying to create a timer that ever x minutes it will call to a server, get a JSON file and do something with it. The merits of that system are not in question right now, the question involves getting the JSON file. Currently I am using kimonify from KimonoLabs to create a JSON file from some other site. They suggest I use function kimonoCallback(data) { // do something with the data // please make sure the scope of this function is global } $.ajax({ "url":"http://www.kimonolabs.com/api/9ey3ljcs?apikey=d753b0c5546495826e7aaa5422f59e30&callback=kimonoCallback", "crossDomain":true, "dataType":"jsonp" }); At the same time I was pursuing other methods and was getting a parse error, and I found this solution. So far both used `&callback=` what is so special about `&callback=` that makes it more preferable to something like `.done()` or `success:` I am guessing it has something to deal with Access-Control-Allow-Origin.
[ 0.0047385613434016705, 0.014178010635077953, -0.004418360535055399, 0.0019733929075300694, 0.004747877828776836, -0.01527719758450985, 0.00830169953405857, 0.024376654997467995, -0.021532215178012848, -0.002480324823409319, -0.008434881456196308, 0.010040642693638802, 0.006653306540101767, ...
[ 0.5358798503875732, -0.1596335619688034, 0.15500852465629578, 0.10533999651670456, -0.12657131254673004, -0.33335569500923157, 0.06559804826974869, -0.10970142483711243, -0.2893056571483612, -0.3170207440853119, 0.021522758528590202, 0.42654329538345337, -0.06335817277431488, 0.14582920074...
How do I find the most probable value of position of a (non-Gaussian) wave function? Is it the same value as the expectation value of the position? And is it true that the most probable value of position is equal to the mean for a Gaussian?
[ -0.0010156198404729366, 0.01593218930065632, -0.00828466285020113, 0.011146127246320248, 0.010603046976029873, -0.0029065925627946854, 0.012985539622604847, -0.01457396149635315, -0.033163297921419144, -0.01945258304476738, -0.0031374034006148577, 0.04349004104733467, -0.02469341829419136, ...
[ 0.02284393273293972, -0.10014760494232178, 0.3569135069847107, 0.1625789999961853, 0.1502130925655365, 0.05664989352226257, -0.29333940148353577, -0.4235818684101105, 0.2356913834810257, -0.1432706117630005, 0.10935754328966141, 0.663385272026062, 0.08763725310564041, 0.15407294034957886, ...
I have read a lot of research papers about sentiment classification and related topics. Most of them use 10-fold cross validation to train and test classifiers. That means that no separate testing/validation is done. Why is that? What are the advantages/disadvantages of this approach, especially for those doing research?
[ 0.015485157258808613, 0.025948479771614075, -0.0055146533995866776, 0.02199164405465126, 0.01530012022703886, -0.00808528158813715, 0.011098728515207767, 0.01085295993834734, -0.015650875866413116, 0.008281623013317585, -0.004838881548494101, 0.012654483318328857, -0.02786778099834919, 0.0...
[ 0.4727207124233246, 0.09990495443344116, -0.11304052174091339, 0.3294283151626587, -0.18834611773490906, -0.17682363092899323, 0.1741395741701126, -0.062018536031246185, -0.02938239648938179, -0.31461814045906067, 0.3781856298446655, 0.05292550101876259, -0.19498209655284882, 0.07437393814...
we have several LVM-based VMs running on a Xen Opensource host. "LVM-based", in our case, means that there's a LVM volume for the root partition and another volume for the swap partition. The kernel and the initrd for the VM are outside of these volumes, as files in the Xen host. These VMs were created with xen-create-image, part of the xen-tools suite for Debian. We want to migrate these VMs to a XCP host, and we've found a script to do it (http://www.xen.org/files/xva/xva.py). But this script assumes that the disks listed in the configuration file are "virtual disks", ie., with their own partition table and stuff. This is not the case, obviously, and it doesn't work. What would you recommend to do this? Any homebrewn method? In the end, we could have to migrate the VMs by overwriting new ones with a big tar file from the old ones, but I'd like it to be our last resort. I thought that maybe we could create a disk like the one this script expects by using something that does the reverse of kpartx: instead of splitting a virtual disk into devices for each partition, bundle several partitions (or in our case, LVM volumes) in a virtual disk. But I don't know anything like that. Thanks in advance!
[ 0.024616364389657974, 0.011514321900904179, 0.0004820172907784581, 0.012653646990656853, -0.028595251962542534, -0.01319514773786068, 0.009771926328539848, 0.022355496883392334, -0.012327713891863823, 0.0041461121290922165, -0.014747073873877525, 0.015469569712877274, 0.005038904491811991, ...
[ 0.4259500205516815, -0.12934289872646332, 0.4083849787712097, 0.20221807062625885, 0.17680564522743225, -0.05987311899662018, -0.36748069524765015, 0.07601732760667801, -0.2971644699573517, -0.5458226799964905, -0.0570436455309391, 0.5724003314971924, -0.42064541578292847, 0.07418251037597...
I have some measurement data of estimates vs actuals (of some metric $m$). For every such data I also have the ratio of $\frac{actuals}{estimates} = r \space (say)$ - if this ratio is greater than 1 it implies underestimation (and overestimation if less than one). Given I have substantial data, about 100 or so data points, I calculate the mean $\mu$ and standard deviation $\sigma$, ($\forall \space r_i| i=1,...,100$ from above). Based on the data above I can say that historically derived confidence range of my estimate is $\mu \pm n\sigma$ - where 'n' is the number of standard deviations I wish to consider. Question - what can I say about my _confidence level_ for the 'next estimate' for $n=1,...,6$? Example: Let's say $\mu = 1.16$ and $\sigma = 0.38$ (data points are {$r_i|i = 1,...,100$}). For $n = 1$ the upper limit $u = 1.53$ and lower limit $v = 0.78$ (i.e., $\mu \pm n\sigma$) Now, say I make an estimate of 100 for $m$ and using the historical data with one-standard deviation as above I can say the following about a most likely estimate: * The estimate should be revised to $116$ (i.e., $100 \mu$) due to underestimation on 'average' * The value is quite likely to be within $78$ and $153$ ($100u$ and $100v$) So for $1\sigma$ (i.e., $n=1$) what is confidence level of the estimate? Is it simply just a ratio of $\frac{numbers Within Range}{Total No. of Elements}$? If I now make $n=2$ what can I say about the confidence level of the estimate for $m$? Basically if I have historically derived confidence ranges ($u, v$ and mean $\mu$) can I peg some sort of a percentage probability (or confidence level) of the likelihood value for the estimate of $m$? How would I do it? Should I restrict the max value of $n$ to 6? (why so?) PS: For those in the know, this is the cone of uncertainty data calculation for a particular metric. And I'd like to know the confidence value that I can associate with my estimate as per historic data.
[ 0.004083307925611734, 0.010123258456587791, -0.00890529714524746, 0.006991330534219742, 0.024458065629005432, -0.015503260307013988, 0.009490972384810448, -0.022470034658908844, -0.00691708829253912, -0.012750735506415367, 0.0007989155128598213, 0.010179256089031696, -0.014410298317670822, ...
[ 0.15371455252170563, 0.17950507998466492, 0.6368744373321533, -0.2524794340133667, 0.20765109360218048, 0.4136470556259155, 0.26987704634666443, -0.6412375569343567, 0.09040069580078125, -0.5430663228034973, -0.135587677359581, 0.7148659825325012, 0.25677961111068726, 0.21464167535305023, ...
Can I apply the Poisson distribution on the following type of data set? I have two types of processes, each generating events. The actual data that I posses are sets of these timestamps. The occurrences are independent and random. Unfortunately, events are clustered because the process generating them pauses for random periods (short AND long periods of pause). It is not the pause periods that interest me, but rather the periods during which the process is on (therefore the pause periods of time should not influence my analysis). Since the data is not uniformly distributed over time, but rather comes in clumps, I am not sure whether I can use the Poisson distribution here. My ultimate goal is to analyze the difference in frequency between two variants of such processes. I will compute the probability distribution for each process, for the interval of a day and see how the two differ. My fear is that the pause intervals would skew the distribution so badly that they will become meaningless to my goals. The following are histograms for the timestamps, each bin representing one day: https://dl.dropboxusercontent.com/u/17450714/i-1.jpg https://dl.dropboxusercontent.com/u/17450714/i-0.jpg If Poisson is not good for this data, how may I analyze it by excluding the huge pause periods? Cluster analysis? Thanks
[ -0.008570665493607521, 0.013837220147252083, -0.0057317353785037994, 0.02509447932243347, 0.006410543341189623, -0.015435603447258472, 0.010436088778078556, 0.010310003533959389, -0.01462789997458458, 0.03062526136636734, -0.008774016052484512, 0.011594148352742195, -0.0027520079165697098, ...
[ 0.23149242997169495, 0.0754525437951088, 0.29231756925582886, 0.1808343231678009, -0.1893569529056549, 0.04562784731388092, -0.014725933782756329, -0.3201298117637634, -0.4580274820327759, -0.3195163309574127, 0.14878633618354797, 0.38524869084358215, -0.3064838647842407, 0.352135837078094...
I'm coming from a winforms background and trying to get ahold of MVVM and XAML. Right now I'm wondering how I can take advantage from a data-shaping control (either native or provided by a third party) while maintaining separation of concerns intact. An example: I have a typical third party grid that comes with data grouping, sorting and filtering functionalities. I want the end user to be able to use these features so that he can get a custom set of data on which he can perform actions. These actions are exposed by the view model. Of course it knows about the record collection but it has absolutely no clue about how it is currently displayed in the UI. Given that, how can the view model know on which data actions should be performed, **without knowing about the control in which it is displayed**? Right now the only way I can think of is exposing an additional property in the ViewModel (something like `CurrentUIData`) and handling every single data- shaping events within the View so that the property is always representative of the UI state. This, however, feels kind of wrong to me. Is this how I should proceed, or am I missing something here?
[ -0.0039701201021671295, 0.011911772191524506, -0.009556290693581104, 0.012379268184304237, -0.012678075581789017, 0.005960637703537941, 0.005414350889623165, 0.010584909468889236, -0.013792812824249268, 0.0033759032376110554, -0.009603956714272499, 0.007305806502699852, -0.010040239430963993...
[ 0.6142231822013855, 0.1818436235189438, -0.1336016207933426, -0.04536202549934387, -0.0038475797045975924, -0.2017279416322708, -0.0703619122505188, -0.03932567313313484, -0.2535400092601776, -0.5517866015434265, -0.014163154177367687, 0.6080376505851746, -0.1892496794462204, 0.28934660553...
Is there a way to prevent deletion of system/shell aliased folders? I want these commands to fail. My shell is bash. Such as: $rm -rf . $rm -rf .. $rm -rf ~ $rm -rf *
[ 0.03516650199890137, 0.012279229238629341, -0.012281980365514755, 0.01895107887685299, 0.01475264597684145, 0.011753027327358723, 0.00909834261983633, 0.023173067718744278, -0.027317620813846588, 0.01229582354426384, -0.014906346797943115, -0.009201936423778534, -0.021229660138487816, 0.01...
[ 0.12412713468074799, 0.40723636746406555, 0.2681628167629242, -0.34914645552635193, -0.03536119684576988, -0.11920206248760223, 0.07342329621315002, -0.37305164337158203, -0.1939328908920288, -0.3433349132537842, 0.13057394325733185, 1.0871049165725708, -0.12624599039554596, -0.16033478081...
Is there an easy way to convert this table (with complete lines) ![completelines](http://i.stack.imgur.com/v2x2o.png) into that one (with incomplete lines; the image is "Photoshopped" for illustration)? ![incompletelines](http://i.stack.imgur.com/QY3Kb.png) Note that I want to get rid of the initial part of the lines (vertical and horizontal). All solutions are welcomed but I would prefer a solution that doesn't require additional packages, if possible. This is the code for the first image: \documentclass{article} \begin{document} \begin{tabular}{l|c|c|} & A & B \\ \hline a & 0 & 1 \\ \hline b & 10 & 11\\ \hline c & 20 & 21\\ \hline d & 30 & 31\\ \hline \end{tabular} \end{document}
[ 0.0072274040430784225, 0.0014346718089655042, -0.0053306566551327705, 0.029861651360988617, -0.02118051052093506, -0.0091893645003438, 0.004934776108711958, 0.012001787312328815, -0.01751522719860077, 0.0029906800482422113, -0.009410228580236435, 0.0033736147452145815, -0.008545669727027416,...
[ 0.08629380911588669, -0.0063517275266349316, 0.4156423807144165, 0.10922057926654816, -0.07126278430223465, 0.07979819178581238, 0.04950624704360962, -0.019971199333667755, -0.4169340133666992, -0.5282875299453735, -0.13405358791351318, 0.17442117631435394, -0.431081622838974, -0.401082605...
It can be another layer of security..
[ 0.03313169628381729, 0.07751987874507904, -0.055932771414518356, 0.02027490921318531, 0.05840348079800606, 0.0007461548084393144, 0.021408047527074814, -0.04491991922259331, -0.08033167570829391, 0.04867257922887802, -0.09332507103681564, 0.05938531085848808, -0.06651630997657776, 0.053147...
[ 0.5555796027183533, 0.08136295527219772, 0.3957556188106537, 0.35342124104499817, 0.3549964129924774, -0.44436782598495483, 0.11104380339384079, -0.025294339284300804, -0.14245586097240448, -0.385469913482666, -0.24898256361484528, 0.3952024579048157, -0.06955229490995407, -0.0782208815217...
I'm using scientific workplace 5.5 and if write `\begin{thebibliography}{LongestLabel}` the bibliography still seems to use 99 instead of `LongestLabel`. Is there a way of getting aroung this issue?
[ 0.014283616095781326, 0.017141779884696007, -0.02906353585422039, 0.021073147654533386, -0.0331234447658062, 0.0009981427574530244, 0.011607719585299492, 0.030249183997511864, -0.009582090191543102, -0.039179619401693344, -0.00636646244674921, 0.005753831472247839, -0.01908710040152073, 0....
[ 0.26919662952423096, 0.3021509647369385, 0.37177717685699463, 0.19673487544059753, -0.1167105957865715, -0.5166690349578857, 0.09468211978673935, 0.3617938756942749, -0.3250032365322113, -0.39127615094184875, 0.18891607224941254, 0.1867782473564148, -0.06563302129507065, 0.1977119594812393...
I am trying to convert latitude and longitude into meters in polar regions. I am looking for higher accuracy upto 0.00 meters preferably. I tried the haversine formula but it doesnt seem to be accurate enough for small increments in latitude. does anyone know a better formula that can be used near poles?
[ 0.001060668146237731, 0.014099955558776855, -0.02385515347123146, 0.010790282860398293, -0.013987057842314243, 0.010142004117369652, 0.012392484582960606, -0.018825046718120575, -0.027830878272652626, -0.012727562338113785, 0.01588272675871849, 0.010924752801656723, -0.01919107511639595, 0...
[ 0.3379765748977661, 0.3656773269176483, 0.4674398899078369, 0.0009495233534835279, -0.22441498935222626, -0.08777017146348953, 0.4509017765522003, -0.12916892766952515, -0.007914029993116856, -0.6256504654884338, 0.4244506061077118, 0.10852928459644318, 0.2741641104221344, 0.06637104600667...
I'm writing my thesis and have used the geometry package to control the size of the text body so that it follows 'good' typesetting practices as described in the KOMA-script documentation. Many of my figures are full-page floats that are taller than the text height. Consequently I have been using the newgeometry/restoregeometry and afterpage commands to adjust the page layout and ensure that the figure is appropriately placed in the document. This works fine except that on the page immediately following the figure, the text 'runs' into the bottom margin. I've have provided a MWE below. Note how the fourth page is formatted incorrectly. (I am using pdfLaTeX). Does anyone know how to correct this behaviour? (or have an alternative method of accomplishing the same layout?) It seems to me as though the restoregeometry command is not being executed correctly. \documentclass{scrbook} \usepackage{calc} \usepackage[showframe]{geometry} % Calculate type area for page \geometry{ bindingoffset = 8mm, heightrounded = true, body = {125mm, 125mm * 16 / 9}, marginratio = {9:16, 9:16}, marginparsep = 5mm, marginparwidth = 30mm, headsep = 10mm, headheight = 10mm, footskip = 15mm } \KOMAoptions{fontsize = 11pt} % Chapters open on right. \KOMAoptions{open = right} % \cleardoublepage command uses the empty page style \KOMAoptions{cleardoublepage = empty} % Don't typeset the Chapter and Appendix prefixes \KOMAoptions{chapterprefix = false} \KOMAoptions{appendixprefix = false} % We also avoid having running heads and page numbers on float pages by using the floatpage % package to make float pages use the empty page style. \usepackage{floatpag} \floatpagestyle{empty} \rotfloatpagestyle{empty} % For dummy text \usepackage{lipsum} % \usepackage{afterpage} % For good control of floats \usepackage[demo]{graphicx} \usepackage{floatrow} \floatsetup{facing=yes, capbesideposition={outside,center}, footnoterule=none} \floatsetup[table]{style=plaintop} \floatsetup[subfigure]{style=plain} % Allow floats to hang into the outer margin \DeclareMarginSet{minhangoutside}{% \setfloatmargins*{\relax} {\hfill\hskip-\marginparwidth\hskip-\marginparsep}% } \floatsetup[figure]{margins=minhangoutside} % For control of caption formatting \usepackage{caption} \usepackage{subcaption} \captionsetup{font=normal, format=hang, labelfont=normal, labelsep=colon} % End Preamble % %------------------------------------------------------------------------ \begin{document} \chapter{First Chapter} % \section{First Section}% \lipsum% % Add a full page figure. I want the preceeding page to be full of text so % use the afterpage command. Also as the figure is taller than the page height % I use the newgeometry/savegeometry/restoregeometry commands. % \afterpage{% \newgeometry{bindingoffset = 8mm, body = {125mm, 125mm * 16 / 9 + 10mm + 15mm}, headsep = 0mm, footskip = 0mm}% \savegeometry{extraheight} % \begin{figure}[pc]% \centering% \ffigbox% {\includegraphics*[width=160mm,height=20mm]{}% \vspace*{1ex}% \begin{subfloatrow}% \ffigbox[\FBwidth]% {\includegraphics*[width=160mm,height=96.0mm]{}}% {\caption{First sub-figure}}% \end{subfloatrow}% \vspace*{1ex}% \begin{subfloatrow}% \ffigbox[\FBwidth]% {\includegraphics*[width=160mm,height=96.0mm]{}}% {\caption{Second sub-figure }}% \end{subfloatrow}}% {\caption{Main caption of the figure. In reality this runs to two lines long so I'm just filling in space here with meaningless text.}}% \end{figure} % \restoregeometry% } % % Next section \section{Second Section}% \lipsum% % \section{Third Section}% \lipsum% \end{document}
[ -0.010699338279664516, 0.009637907147407532, -0.008819084614515305, 0.016474317759275436, 0.011391576379537582, 0.00654249731451273, 0.009445523843169212, -0.010203046724200249, -0.015417708083987236, -0.01592543162405491, -0.015492529608309269, 0.004046161659061909, -0.007570450659841299, ...
[ 0.34926387667655945, 0.30766117572784424, 0.7499428987503052, 0.016851412132382393, -0.13569697737693787, 0.07299041748046875, 0.02454940602183342, -0.342637300491333, -0.1603967547416687, -0.4343331754207611, 0.15670102834701538, 0.35881534218788147, 0.5312251448631287, -0.174062758684158...
We can write: MorphologicalComponents[image] // Colorize To assign a random set of colors to each component in some image. The example used for reference by Mathematica is a map of the United States: http://reference.wolfram.com/mathematica/ref/Colorize.html?q=Colorize&lang=en My question is - is it possible to ask Mathematica to specify that any two adjacent colors should have some minimum threshold distance in in their hue or RGB values (e.g. http://en.wikipedia.org/wiki/File:Map_of_United_States_vivid_colors_shown.png)? I believe asking for a four-coloring of a map is computationally non-trivial, but are there any built-in routines, perhaps with the graph analytics packages in Mathematica, to allow for this? Let's define two morphological components as adjacent if the line segments spanning the shortest distance between the two components fails to intersect any other morphological components. We can also include a distance $D$ cutoff for defining two components as adjacent. I'm certainly open to stricter definitions.
[ 0.01534110028296709, 0.011522875167429447, -0.005748513620346785, 0.022434286773204803, -0.01038927212357521, -0.007822582498192787, 0.009332846850156784, -0.00898231752216816, -0.014146341010928154, 0.0265263132750988, -0.012858850881457329, 0.006899633444845676, -0.009565836749970913, 0....
[ 0.07001745700836182, -0.08164861053228378, 0.2884037494659424, 0.20858003199100494, -0.06545989960432053, 0.10827749967575073, -0.18743647634983063, -0.35873493552207947, -0.13641752302646637, -0.8678305149078369, 0.034211885184049606, 0.1460534930229187, -0.2045615166425705, -0.0775171890...
You know that for a potential function (conservative force/fields) that $\nabla U = \pm \vec{F}$ In math, we don't have that minus sign, we have only the plus one. What does it mean if you get rid of the plus sign? I remember the minus signs tells us that the force is always in the opposite direction of the increasing potential function. Does math tell me there exists a potential function such that it's force also increases as potential goes up?
[ -0.012834941036999226, 0.015576990321278572, -0.004100144375115633, 0.02093084529042244, -0.010724091902375221, -0.02078160271048546, 0.007450270466506481, -0.013676994480192661, -0.015403320081532001, 0.014006290584802628, -0.008382382802665234, 0.011652966029942036, 0.014051692560315132, ...
[ 0.29273492097854614, -0.27771106362342834, 0.3220108449459076, -0.16047579050064087, 0.11087557673454285, 0.002183973789215088, -0.10139799863100052, -0.10954747349023819, -0.4303276538848877, -0.2913709282875061, -0.06692281365394592, 0.757591724395752, -0.4224222004413605, 0.233331322669...
I have a number of features loaded on a mxd I would like to create a script or model builder to assign each row containing a value in the coloum "ENTE" with a RGB color which must be represented in the map and the corresponding color name (unique). The solution could be to have a dBase file attached that contains the RGB values ​​in a column and in the other, the name of the color. In dBase file will contains about 3000 lines in which the colors are repeated in groups of 20 (ie, have a base of 20 colors) I can not understand how I could assign color and assign its name to the single feature not randomly. thanks
[ 0.012638324871659279, 0.01737024262547493, -0.019748438149690628, -0.0016399831511080265, -0.007490966934710741, 0.004655898083001375, 0.007989279925823212, 0.006564726121723652, -0.013141313567757607, 0.0022711821366101503, -0.011075567454099655, 0.006345841102302074, 0.009429145604372025, ...
[ 0.34663283824920654, 0.2733568251132965, 0.509831964969635, 0.17982949316501617, -0.28082865476608276, 0.23217903077602386, 0.04177110642194748, -0.333326518535614, -0.18660332262516022, -0.6636353731155396, 0.1659879833459854, 0.4156642556190491, -0.47122886776924133, -0.03766370192170143...
I have a doubt about permanent magnets. If a magnet is permanent it can attract some materials permanently. Attracting something involves energy. If a permanent magnet can do this forever, from where does this energy come from? How can it not run out of energy? Doesn't it contradict with the laws of energy?
[ 0.021250445395708084, 0.01341031864285469, -0.0008730702102184296, 0.0236391369253397, 0.017122328281402588, -0.02857603318989277, 0.012694189324975014, -0.01110096275806427, -0.02776801958680153, 0.005404761992394924, -0.014523093588650227, 0.027097495272755623, 0.026231518015265465, -0.0...
[ 0.6840176582336426, 0.15274043381214142, 0.5046124458312988, 0.3248387575149536, 0.14106842875480652, -0.403997004032135, -0.03113844059407711, 0.1815970093011856, -0.3408277630805969, -0.12094006687402725, 0.0526091493666172, 0.420718252658844, -0.36094531416893005, 0.6055148243904114, ...
I want to typeset the combination m̐llo in Devanagari with XeLaTeX. When I use the correct Unicode ल्ँलो, there is no ligature of the ls. When I place the candrabindu on the ligature the o is not connected with theligature: ल्लँो. How can I get XeLaTeX to place the candrabindu on the ligature and get still the o in the correct position? \documentclass{article} \usepackage{fontspec} \usepackage{polyglossia} \setmainlanguage{sanskrit} \setmainfont{Siddhanta} \begin{document} ल्ँलो ल्लँो \end{document} Here is an example from a textbook, how it should look like: ![enter image description here](http://i.stack.imgur.com/7fSOL.png)
[ 0.018175868317484856, 0.01515051256865263, -0.010796927846968174, 0.01700753904879093, 0.01515413448214531, 0.014816210605204105, 0.012338164262473583, 0.011827715672552586, -0.011720011942088604, -0.018588602542877197, -0.02256142348051071, 0.0067882416769862175, -0.024739284068346024, 0....
[ 0.12800724804401398, 0.0641176849603653, 0.48822593688964844, -0.1866113543510437, 0.06595348566770554, 0.7523524165153503, -0.08911062031984329, -0.25262439250946045, -0.02412838488817215, -0.522911548614502, 0.19003763794898987, 0.3674638271331787, -0.4188678562641144, -0.174591109156608...
I have recently heard the term "Acme developer of IPhone." I was just wondering what it stands for? I think I know the meaning: Is it when a developer goes to a client and presents his/her ideas for the app?
[ -0.027736905962228775, 0.01405983604490757, 0.01728283055126667, 0.02809740975499153, -0.0655178353190422, -0.0008124676533043385, 0.012521715834736824, 0.005126672331243753, -0.03220297396183014, 0.00842453446239233, -0.01030806079506874, 0.024912508204579353, 0.08627229183912277, 0.01475...
[ 0.4345501661300659, 0.23118145763874054, -0.19766908884048462, -0.12864546477794647, -0.025838688015937805, -0.1489167958498001, -0.11459103971719742, 0.40983110666275024, -0.16216349601745605, 0.11473483592271805, -0.15623100101947784, 0.3545241355895996, -0.10232428461313248, 0.094736762...
In this video, a pool of liquid nitrogen in a metal bowl can cool it enough so that oxygen from the air condenses at the bottom: http://www.youtube.com/watch?v=Ojl10zMCYkY&list=PLmyvBUHJ4pVC3FrSXsT_5S7dFMKPLG_pS&index=1 Thermodynamically, it seems that the bowl can't get cold enough. The bowl can't cool off below the temperature at which the nitrogen pressure in the air (80kPa, I got all the vapor pressure information form Air Liquide) equals the vapor pressure of the liquid nitrogen; at that point the rate of evaporation equals the rate of condensation from the pool. This happens at ~76K. This is also the threshold where oxygen can condense (vapor pressure of 20kPa). However, 76K is a lower bound. The actual temperature will be higher when you include condensation of water into/onto the bowl, convective/radiative heat transfer, and the fact that there is additional nitrogen/oxygen above the surface of the pool beyond the 80kPa used in the calculations. In the face of these calculations, why does it seem to work? My reasoning is that it still gets cold enough for a _mixture_ of nitrogen and oxygen to condense which has enough oxygen to burn the tissue paper, is this correct?
[ 0.012862047180533409, -0.004338162951171398, -0.00855998694896698, 0.006790261249989271, -0.008114529773592949, -0.021620647981762886, 0.005220241378992796, -0.01468595676124096, -0.014527015388011932, 0.01030349638313055, -0.0018478992860764265, 0.021507449448108673, -0.007502877153456211, ...
[ 0.41158193349838257, 0.01836339198052883, 0.35246002674102783, 0.5401448011398315, 0.4488455057144165, 0.03248808905482292, 0.2663661539554596, -0.2730207145214081, -0.2923201322555542, -0.31390032172203064, -0.11085532605648041, -0.3106080889701843, -0.027546770870685577, 0.19351898133754...
I apologize if this questions has already been answered, I have a feeling I'm just not using the right terminology for my search. I'm in the process of building a base theme for my employer. One of the things we want to do with this new theme is have the option to switch from a standard template to one that's been Zurbified. The way I'd prefer to handle this is by having an admin panel checkbox. Within the theme itself we want to have 2 separate folders that contain the actual theme template files (Ex: `theme_folder/normal/page.php || theme_folder/zurb/page.php`). How would I go about properly processing this? Is there a hook/filter that allows you to specify which sub folder to use in the specific situation? From what I read its possible to have the sub directories house the template page files, http://nacin.com/2012/03/29/page-templates-in-subdirectories-new-in- wordpress-3-4/ , but I'm not finding the call I need to tell Wordpress to use the specific subs.
[ 0.01205914095044136, 0.015064530074596405, 0.006753726862370968, 0.018152307718992233, 0.008093208074569702, -0.00023649167269468307, 0.005695807281881571, -0.0062028272077441216, -0.009642606601119041, 0.004712574649602175, -0.012464095838367939, 0.007232392206788063, 0.00517623033374548, ...
[ 0.6096423864364624, -0.018023841083049774, 0.6773898005485535, 0.1257648468017578, -0.044100020080804825, -0.3544464409351349, 0.08228658884763718, 0.05061497539281845, -0.4200533330440521, -0.43275296688079834, -0.1563294380903244, 0.21889446675777435, -0.15643279254436493, 0.394736409187...
I am trying to set up my system and to get some practice with scripting and the like. I am running Linux Mint 16 on a Samsung RC520 laptop (`Linux 3.11.0-12-generic x86_64`). I want to implement a script that will lower my screen brightness to a pre-set value when I unplug the power (and raise it when I plug it back in). I can do some very basic scripting but not near this level. I have "xbacklight" which will set the brightness from the command line, so when I unplug the power the line I want to implement would be `xbacklight -set 20`. I also know that unplugging the power is registered by acpi (this may sound completely stupid but hopefully it is clear what I mean); when I plug out and then back in the power I get the following with `acpi_listen`: $ acpi_listen ac_adapter ADP1 00000080 00000000 processor CPU0 00000081 00000000 processor CPU1 00000081 00000000 processor CPU2 00000081 00000000 processor CPU3 00000081 00000000 ac_adapter ADP1 00000080 00000001 processor CPU0 00000081 00000000 processor CPU1 00000081 00000000 processor CPU2 00000081 00000000 processor CPU3 00000081 00000000 battery BAT1 00000080 00000001 So I know I have to use this but I have two main problems: I wouldn't know where to put the script so that it is automatically implemented, and I don't know how I would recognise the plug in/out in acpi because all I know how to do is `acpi_listen`.
[ 0.032522622495889664, -0.0014016602654010057, -0.004327655769884586, 0.0006086551584303379, -0.03263409063220024, -0.01853329688310623, 0.006550440564751625, -0.006657418794929981, -0.015128396451473236, 0.0010662837885320187, -0.01735016331076622, 0.006846909876912832, -0.003677457105368375...
[ 0.400799423456192, 0.4100905656814575, 0.45832017064094543, -0.31475692987442017, 0.02038809284567833, 0.011528869159519672, 0.3936792314052582, 0.30682095885276794, 0.09151215851306915, -0.7769691348075867, 0.10575970262289047, 0.7289682626724243, -0.20096419751644135, 0.02182289958000183...
Circuit in series: $10\,\text{V}$ power supply, 2 resistors connected. Each resistor is $10\,\Omega$, total of $20\,\Omega$ resistance. If I put an ammeter at the end of the circuit, will the current be calculated using 20 Ohms or 10? Electrons flow through the first 10 Ohm resistor and are slowed down, yes? Then they flow through the second 10 Ohm resistor. They have passed through a total of 20 Ohms BUT at 10 Ohms per resistor. Will the current be the same as if it had passed through 1 20 Ohm resistor? That violates everything I understand.
[ -0.0008540453854948282, 0.009636754170060158, -0.011162050068378448, 0.013215708546340466, -0.007818490266799927, -0.006460584234446287, 0.009691765531897545, -0.0083999652415514, -0.0195515938103199, 0.018134979531168938, 0.0010120172519236803, 0.011595244519412518, -0.016329968348145485, ...
[ 0.5028307437896729, -0.5162975192070007, 0.8744061589241028, 0.05030803754925728, -0.10448700189590454, 0.39156731963157654, 0.07517696917057037, -0.6586458086967468, -0.28544676303863525, -0.422556072473526, 0.3437187969684601, 0.347247838973999, -0.293194442987442, 0.4052671790122986, ...
In Gears of Way Judgement, you can re-up - similar to the 'prestiging' in Call of Duty. The problem is though, I would like to know how many times my opponents have re-uped. The current symbol for re-up #1 looks like a bronze star. I think the #2 re-up looks like a bronze star with a circle around it. Can anyone help me out with some sort of key or reference ? I heard you can re-up 11 times.
[ -0.0004152710025664419, 0.019556082785129547, -0.015501189976930618, 0.00627412972971797, -0.008367384783923626, -0.013505618087947369, 0.010951969772577286, -0.023128533735871315, -0.014980585314333439, 0.014700494706630707, 0.0019047037931159139, 0.014042076654732227, -0.009813924320042133...
[ -0.07761786133050919, 0.01314701046794653, 0.4782901406288147, 0.2784724831581116, -0.2778250277042389, 0.4482005536556244, 0.1716308295726776, -0.18537938594818115, -0.5273072123527527, -0.534514307975769, 0.19338557124137878, 0.8695390224456787, 0.13083864748477936, 0.11893102526664734, ...
I'm trying to do something with finding solutions to a quantum mechanics problem with n wells. If there are 40 wells, I need to find the solution to an equation in the form: {a,b}.MatrixPower[A,40].{c,d} So, I use FindRoot and everything works out great, in a decent amount of time. The actual code used is something along these lines: lowPropagator[ dist_, α_] := {{Cos[α dist],Sin[α dist]/α}, {-α Sin[α dist], Cos[α dist]}}; highPropagator[ dist_, β_] := {{Cosh[β dist], Sinh[β dist]/β}, {β Sinh[β dist],Cosh[β dist]}}; root[d_, n_, start_] := FindRoot[{α, 1}. MatrixPower[lowPropagator[d, α].highPropagator[d,α], n - 1]. lowPropagator[d, \[Alpha]].{1,α}, {α,start}, MaxIterations -> 50 ]; Manipulate[ root[d, n,α0], {{d, 1}, 0, 5}, {{n, 2}, 1, 10, 1}, {{α0, 1}, 0, 5}] It's responsive even for larger n. Then, I try to plot the equation. The plotting function is a large/complicated piecewise function, alternating between the form $f=A \cos(\alpha x)+B\sin(\alpha x)$, and $f=A \cosh(\beta x)+B\sinh(\beta x)$ (where beta is a function of alpha). The matrices "lowPropagator" and "highPropagator" allow me to find the function and its derivative at each point where the form of the piecewise function changes, so I can use a matrix inverse to solve for A and B, and then plug all of this into a giant Piecewise[]. So there's a lot of computation going on in the function. The function is in the form: getWaveFunction[α_, β_, \[Rho]_, a_, b_, n_] := Module[{...}, ... totalFunction[x_] := Piecewise[Table[{eqlist[[i]][x], x < wavePositions[[i]]}, {i, 1, 2 n}], eqlist[[2 n + 1]][x]]; totalFunction[#] & ]; So the function isn't trivial, but it still runs fast, faster than the root finding even, especially when n is high. (it's doing at least 2n matrix multiplications) However, when I do something like this: plotIt[d_, n_, start_] := Module[{αnear, f}, αnear = root[d, n, start]; f = getWaveFunction[αnear, ..., d, n]; Plot[f[x], {x, -5, 5}]] For some values, the function takes forever to draw the graph. The black evaluation bar doesn't pop up, so alt+. doesn't do anything. Basically, if I choose the input values incorrectly, most of the time Mathematica becomes completely unresponsive and I have to restart. I think Mathematica IS still running and calculating, because sometimes after 30 seconds or so it will return a result. Out of curiosity I left it running overnight with a different value and it was still frozen when I came back, so, something is not right. So the biggest issue is that these two functions are fast on their own, at ALL input values, but at certain input values the combination of the two takes forever. My question is, how can I fix this to make the function run at a decent speed, and if I can't, how can I make it fail gracefully so that the notebook doesn't crash? Relevant things: * It fails at large n and large d or "dist". * It IS taking a matrix with hyperbolic functions in it, to the nth power. So, the values can get pretty huge (understatement), and this is exactly what happens with the values at the crash. However, the root finding function on its own doesn't seem to crash at these values. Neither does the plotting function. It's only both together that causes problems. * I figure evaluation might be the problem, maybe the root and/or the graph are trying to be found many times inside the plot function, but adding "//Evaluate" to the end of whatever f and αnear equal doesn't solve the problem. Let me know if I should post the full code. The relevant bits add up to ~180 lines, with a good amount of spacing/comments. The full code produces graphics like this: ![Solution of the time independent schrodinger equation](http://i.stack.imgur.com/Vqz1o.png) (The red line corresponds to the α result of FindRoot. The purpose of FindRoot is to find all solutions that are exponentially decaying on both sides of the graph, so the red line is exponentially decaying on both sides, but the dashed line [which corresponds to the "start" value in FindRoot] isn't a solution, since it diverges on the right side. I'm trying to find solutions to the time independent schrodinger equation, for a single particle, with n wells, and this is for the end of chapter 2 of "A First Course in Computational Physics" by DeVries [not homework, & the book doesn't ask for solutions this general].) * * * ## Update with full code So, I haven't been able to simplify the code to much significant extent. If I simplify it too much, to the point where it seems to work but isn't calculating what I want it to calculate, the problem seems to disappear! I do think the problem could be with evaluation though. If you comment out the "Show" at the end and just write "f2" in its place, (f2 is the function that is a result of getWaveFunction, with FindRoot passed in as an argument), it prints out, "totalFunction$30552[#1] &" (or some similar number on the end). Since totalFunction doesn't exist in the scope it's at... I'm guessing that this is an evaluation problem, where Plot re-evaluates something it doesn't really have to a bunch of times. Code: (And, keep in mind while running it, it can crash) (* \[Alpha] is the frequency of open regions \[Beta] is the coefficient of high energy regions (should be \ \[Beta]=Sqrt[vm2h-\[Alpha]^2]) \[Rho] is the height at the leftmost wall a is the width of the open region b is the width of the high energy regions n is the number of wells. This function returns a pure function, that is supposed to be the \ wavefunction \[Psi][x]. (does not depend on time) *) getWaveFunction[\[Alpha]_, \[Beta]_, \[Rho]_, a_, b_, n_] := Module[{left, eqLeft, eqRight, eqlist, lowPropagator, highPropagator, lowInverseCoefficients, highInverseCoefficients, waveRecursiveFunction, waveValues, wavePositions, waveConstants, totalFunction}, lowPropagator[ dist_] := {{Cos[\[Alpha] dist], Sin[\[Alpha] dist]/\[Alpha]}, {-\[Alpha] Sin[\[Alpha] dist], Cos[\[Alpha] dist]}}; highPropagator[ dist_] := {{Cosh[\[Beta] dist], Sinh[\[Beta] dist]/\[Beta]}, {\[Beta] Sinh[\[Beta] dist], Cosh[\[Beta] dist]}}; (* The wavefunction and its derivative can be represented as: A.{aconst,bconst}={\[Psi],\[Psi]'}, so to solve for aconst and bconst you just take the inverse of the \ matrix A, where A either corresponds to the low solution or the high solution. *) lowInverseCoefficients[ x_] := {{Cos[\[Alpha] x], -Sin[\[Alpha] x]/\[Alpha]}, {Sin[\ \[Alpha] x], Cos[\[Alpha] x]/\[Alpha]}}; highInverseCoefficients[ x_] := {{Cosh[\[Beta] x], -Sinh[\[Beta] x]/\[Beta]}, {-Sinh[\ \[Beta] x], Cosh[\[Beta] x]/\[Beta]}}; (* Define the leftmost point so that everything is centered around \ 0 *) left = -(n*(a + b) - b)/2; (* waveValues is to be an array giving the values of the \ wavefunction an its derivative. ie: waveValues[[1]]=={\[Psi][left],\[Psi]'[left]}; waveValues[[2]]=={\[Psi][left+\[CapitalDelta]x],\[Psi]'[ left+\[CapitalDelta]x]}; etc. wavePositions is just the position of the nth change in height, with wavePositions[[1]] being the start of the leftmost dip. waveConstants holds the values of coefficients to equations. So, you'll have each waveConstants[[ n]] be equal to some vector {aconst, bconst}. Whether the equation is in the zero energy region or high energy \ region depends on whether n is odd or even respectively. So if n is odd the constants correspond to the equation (aconst* Cos[\[Alpha] x]+bconst*Sin[\[Alpha] x]), and if n is even the constants correspond to the equation (aconst* Cosh[\[Beta] x]+bconst*Sinh[\[Beta] x]). *) waveValues = wavePositions = waveConstants = Table[0, {q, 1, 2 n}]; waveRecursiveFunction[1] := ( waveValues[[1]] = {\[Rho], \[Rho] \[Beta]}; wavePositions[[1]] = left; waveConstants[[1]] = lowInverseCoefficients[left].{\[Rho], \[Rho] \[Beta]}; ); waveRecursiveFunction[j_] := ( (* fill out previous values *) waveRecursiveFunction[j - 1]; If[Mod[j, 2] == 0, (* j=2m, so to the left of it is a zero region and to the right is a \ high energy region. *) waveValues[[j]] = lowPropagator[a].waveValues[[j - 1]]; wavePositions[[j]] = wavePositions[[j - 1]] + a; (* width of the high energy region is a. *) waveConstants[[j]] = highInverseCoefficients[0].waveValues[[j]]; , (* j=2m+1, so to the left of it is a high energy region and to the right \ is a zero energy region. *) waveValues[[j]] = highPropagator[b].waveValues[[j - 1]]; wavePositions[[j]] = wavePositions[[j - 1]] + b; waveConstants[[j]] = lowInverseCoefficients[wavePositions[[j]]].waveValues[[j]]; ]; ); waveRecursiveFunction[2 n]; eqlist = Join[{Evaluate[(highInverseCoefficients[ 0].{\[Rho], \[Rho] \[Beta]}).{Cosh[\[Beta] (# - left)], Sinh[\[Beta] (# - left)]}] & }, Table[If[Mod[m, 2] == 0, (Evaluate[ waveConstants[[m]].{Cosh[\[Beta] (# - wavePositions[[m]])], Sinh[\[Beta] (# - wavePositions[[m]])]}]) &, (Evaluate[ waveConstants[[m]].{Cos[\[Alpha] #], Sin[\[Alpha] #]}]) & ], {m, 1, 2 n} ] ]; totalFunction[x_] := Piecewise[ Table[{eqlist[[i]][x], x < wavePositions[[i]]}, {i, 1, 2 n}], eqlist[[2 n + 1]][x]]; totalFunction[#] & (*Show[Plot[totalFunction[x],{x,left-1,-left+1},PlotRange->{-1,1}], drawPotential[a,b,1,n]]*) (*{Plot[totalFunction[x],{x,-4,4}],waveValues,wavePositions, waveConstants,eqlist,totalFunction[x]}//Column*) ]; Manipulate[ Module[{f1, f2, \[Alpha]near, left, lowPropagator, highPropagator}, c = Rationalize[c]; d = Rationalize[d]; a = Rationalize[a]; \[Rho] = Rationalize[\[Rho]]; left = -(n*(a + d) - d)/2; f1 = getWaveFunction[\[Alpha], Sqrt[c - \[Alpha]^2], \[Rho], a, d, n]; lowPropagator[ dist_, \[Alpha]_] := {{Cos[\[Alpha] dist], Sin[\[Alpha] dist]/\[Alpha]}, {-\[Alpha] Sin[\[Alpha] dist], Cos[\[Alpha] dist]}}; highPropagator[ dist_, \[Beta]_] := {{Cosh[\[Beta] dist], Sinh[\[Beta] dist]/\[Beta]}, {\[Beta] Sinh[\[Beta] dist], Cosh[\[Beta] dist]}}; \[Alpha]near = FindRoot[{Sqrt[c - \[Alpha]v^2], 1}. MatrixPower[ lowPropagator[a, \[Alpha]v].highPropagator[d, Sqrt[c - \[Alpha]v^2]], n - 1] .lowPropagator[a, \[Alpha]v].{1, Sqrt[c - \[Alpha]v^2]}, {\[Alpha]v, \[Alpha]}, MaxIterations -> 50] // Evaluate; \[Alpha]near = (\[Alpha]v /. \[Alpha]near); If[Abs[Im[\[Alpha]near]] < 0.00001, \[Alpha]near = Re[\[Alpha]near], \[Alpha]near = 0]; If[\[Alpha]near > 0, f2 = Evaluate[ getWaveFunction[\[Alpha]near, Sqrt[c - \[Alpha]near^2], \[Rho], a, d, n]], f2 = 0 &]; Show[Plot[f1[x],{x,left-1,-left+1},PlotStyle->Dashed], Plot[f2[x],{x,left-1,-left+1},PlotStyle->Directive[Thick,Red], Exclusions->None], PlotRange->{-1.5,1.5} ] ], {{\[Alpha], 2}, 0.001, Sqrt[c]}, {{c, 10}, 0.001, 20}, {{a, 1}, 0.001, 10}, {{d, 0.2}, 0.001, 10}, {{\[Rho], 0.2}, 0, 1}, {n, 2, 10, 1}]
[ 0.015951475128531456, 0.008494902402162552, -0.015087089501321316, 0.0035040592774748802, -0.025153858587145805, 0.007952727377414703, 0.005726504139602184, 0.007679983973503113, -0.012614469975233078, -0.005233274772763252, 0.005611109547317028, 0.01142280362546444, -0.01462344266474247, ...
[ -0.04794160649180412, 0.048110056668519974, 0.222477525472641, -0.04541018605232239, 0.31723567843437195, 0.13073575496673584, 0.16725961863994598, -0.419851690530777, 0.30520161986351013, -0.45091572403907776, 0.20090949535369873, 0.7419123649597168, -0.19104453921318054, -0.0786081627011...
The node directly off the bell tower is giving me trouble. It's a tree on a pyramid surrounded by water with a hole on one side that makes i noise but i dont know what to do. If anyone knows it would be appreciated.
[ -0.053252361714839935, -0.014566589146852493, -0.003471540752798319, 0.0038044725079089403, -0.06341961026191711, -0.015047984197735786, 0.013276924379169941, 0.013832069002091885, -0.03486643731594086, -0.013379741460084915, -0.027010716497898102, -0.0125120272859931, -0.0313471294939518, ...
[ 0.3255472481250763, -0.031992167234420776, 0.07770974189043045, 0.42018598318099976, 0.1870303899049759, -0.21847206354141235, 0.44244086742401123, 0.09935132414102554, -0.013124783523380756, -0.5650612711906433, 0.15559443831443787, 0.11783704161643982, 0.0054459343664348125, -0.068036817...
I am trying to color a simple 2D plot (using `ListLinePlot`) based on data in the form of 3-vectors. ListLinePlot[data[[All, {1, 2}]]] I would like to use one of the `ColorData` color schemes (like "Temperature") for the line color, depending on this third value in my table 3-vectors. That is,`Data[[All, 3]]` will be the source of the color-coding values.
[ 0.006148429121822119, 0.011736101470887661, -0.020236419513821602, 0.0020370641723275185, 0.007406344637274742, 0.0014693689299747348, 0.008156977593898773, -0.0008847179124131799, -0.012991965748369694, 0.009371726773679256, 0.0007157553336583078, 0.005630438216030598, 0.0022451526019722223...
[ 0.39506807923316956, 0.024719838052988052, 0.3755102753639221, -0.13510525226593018, -0.010918022133409977, 0.3431957960128784, 0.10563955456018448, -0.23995596170425415, 0.22919563949108124, -0.7660772204399109, -0.1767793446779251, 0.06366395205259323, -0.4199132025241852, 0.094152070581...
\documentclass{article} \usepackage[fleqn]{amsmath} \usepackage{mathrsfs} \usepackage{amsfonts} \pagestyle{empty} \setlength\mathindent{0pt} \begin{document} \[3x^3\] \[3x^2\] \[3x\] \[3\] \end{document} When I render this document, there is a lot more space between the first two options than between the second and third options. Is there a way to make sure it's consistent? Thanks!
[ 0.01522502489387989, -0.002055806340649724, 0.0036265719681978226, 0.01336399931460619, -0.005033997353166342, -0.005761736538261175, 0.008561339229345322, 0.0013896524906158447, -0.009906662628054619, -0.027755029499530792, -0.009013361297547817, 0.0032813968136906624, 0.0004710555076599121...
[ -0.3218407928943634, 0.19436392188072205, 0.5956851243972778, -0.06610625982284546, 0.30451643466949463, 0.10402984172105789, 0.28037765622138977, -0.3514058589935303, -0.020285626873373985, -0.9115144610404968, -0.40168556571006775, 0.6325391530990601, 0.02254735492169857, -0.307969212532...
I was trying to figure out how to add a new SSH key to my amazon server - I found this question and answer How do you copy the public key to a ssh-server? However I am so new to Linux/Unix and I overthink things so I thought the following 1. Since my clients access this server if I mess this up I am going to have serious problems recovering. 2. How could this possibly work because I have no connection to that other server. I read the first response and it suggests that I could send my newly generated public key to any server in the land. Does my second point make sense. The command from the shell (which I just installed) reads to me as 1. ssh - use an ssh session 2. copy the key to (in my case it would be ubuntu@ourspecial.server.com (hosted by Amazon though I am not sure it matters) In other words - for a totally new user the command suggests that if I use the following command $ ssh-copy-id tim@just.some.other.server if there is actually a user named tim and a server named just.some.other.server I would then have some privileges at that server And so I finally tried the command and clearly there is something missing as when I keyed in $ ssh-copy-id ubuntu@my.amazonserver.com the shell returned /usr/bin/ssh-copy-id: ERROR: No identities found
[ 0.006029397714883089, 0.0030942526645958424, 0.006674183066934347, 0.025826290249824524, -0.01461340207606554, 0.010054315440356731, 0.0063375867903232574, -0.0017219344154000282, -0.02938876301050186, 0.0024105797056108713, 0.0007556690834462643, 0.02818455547094345, -0.012225101701915264, ...
[ 0.5374826192855835, -0.08035777509212494, 0.15397998690605164, 0.48668745160102844, -0.11834198981523514, -0.3038276731967926, 0.1389707624912262, -0.03374878317117691, -0.03876030072569847, -0.4903104305267334, 0.23553340137004852, 0.2663416564464569, -0.2561453878879547, 0.41627267003059...
We've went ahead and purchased Microsoft Dynamics (see edit history for why) in order to write 2 functions to interact with one tiny part of it to fulfill one big need. What I want to know is, from other programmers who have integrated their software with MS Dynamics GP. Is there any way to connect with it (fully) without having to write an interface in C#? We want to use GP SOAP services so we can communicate to it from Delphi and eventually Java, but it seems like the only way we're going to be able to maximize it's usefulness is to write some sort of middleware program in C#.
[ -0.008978249505162239, 0.003891129046678543, -0.005353956948965788, 0.009265348315238953, -0.00907028280198574, -0.004681074060499668, 0.006637980230152607, 0.011163242161273956, -0.015274918638169765, -0.027408553287386894, -0.003978406079113483, 0.020114436745643616, -0.009956419467926025,...
[ 0.3205006420612335, 0.2842644453048706, 0.34346702694892883, 0.08405362069606781, -0.3382163643836975, 0.03639752417802811, 0.0883302241563797, 0.07111228257417679, -0.1310306042432785, -0.5392959713935852, 0.2788214385509491, 0.29804307222366333, -0.1826142966747284, 0.09617637097835541, ...
Our site has recored significantly drop in visitors from China from Google Analytics recently. Do you have the same observation?
[ -0.024093428626656532, 0.034280404448509216, 0.015342196449637413, 0.07036066800355911, -0.0024811620824038982, -0.01493392325937748, 0.014641945250332355, -0.013396053574979305, -0.054874520748853683, -0.037090566009283066, -0.01481504738330841, 0.044695522636175156, -0.016053928062319756, ...
[ 0.6546104550361633, 0.21705855429172516, 0.1752568483352661, 0.24568957090377808, 0.024281151592731476, -0.1751202493906021, 0.14863280951976776, 0.7020303606987, -0.3966594338417053, -0.3872206211090088, 0.22725677490234375, -0.16213726997375488, 0.09964019805192947, 0.20972302556037903, ...
I am using on my Android Gtalk where i can only do chat text send and receive. But how can i do all the features of WhatsAPP from Gtalk? Features like: * file transfer (image pictures, music, video clips) * Audio replay from the chat by history * Video reply from the chat by history Is there any way to make Gtalk by default to do file transfer features?? So that pictures in mobile can be sent, musics in mobile can be sent etc can be sent from mobile directly ??
[ -0.011385580524802208, 0.0029377415776252747, 0.007617665454745293, 0.011253172531723976, -0.005489980801939964, -0.01730485074222088, 0.010181382298469543, 0.017120111733675003, -0.024276334792375565, -0.0036975000984966755, -0.01763956807553768, 0.009908841922879219, 0.032741546630859375, ...
[ -0.10001532733440399, 0.042034659534692764, 0.47313064336776733, 0.3050368130207062, -0.30708929896354675, -0.11767839640378952, 0.1394520252943039, -0.05418579652905464, -0.35573333501815796, -0.4748755991458893, 0.04306470975279808, 0.7931755185127258, -0.34064939618110657, -0.0370979085...
I use the `glossaries` package to create a list of acronyms as well as a list of symbols. \usepackage[acronym,style=long]{glossaries} For the list of symbols the standard-glossary is used. Now I want to sort the list of acronyms alphabetically (`sort=standard`), but the list of symbols in order of appearance (`sort=def`). How do I do that? I use MikTeX 2.9 with TeXnicCenter 2 (beta) on Windows 7. My project includes the `hyperref` package.
[ 0.0257373359054327, 0.006778073031455278, -0.021327074617147446, 0.010386616922914982, -0.023238901048898697, 0.00517995934933424, 0.007111010141670704, -0.0057319654151797295, -0.013755204156041145, -0.0002776003093458712, -0.009136222302913666, -0.007588764652609825, -0.000635149423032999,...
[ 0.2521818280220032, 0.14949658513069153, 0.6900712251663208, -0.018696226179599762, -0.013599051162600517, -0.12748418748378754, 0.437369704246521, -0.08556730300188065, 0.21678419411182404, -0.8487824201583862, 0.009094550274312496, 0.9453462362289429, -0.47444841265678406, -0.20282688736...
I'm a bit OCD about efficiency and I have noticed that at least one of my two Android devices tells me to unplug it after charging is complete to save energy. My questions are: 1. Is it true that a chargeable USB device may or will continue to consume similar amounts of power if it remains plugged in after fully charging? 2. Does this vary based on hardware and/or software (Is it possible my LG Optimus T phone continues to consume similar power after it is charged, but my Nexus 7 doesn't)? 3. If a device continues to "consume" power after it is charged, where does it go (laws of thermodynamics and all)? In searching for my answer, I have found the following questions relevant but not entirely applicable to my question: * Does constant charging harm my Android cellphone? * AC-USB Charging Device Makes Sounds when Device Fully Charged * What component of my phone charger produces chirps and how does it do that?
[ 0.011742061004042625, 0.0130698811262846, 0.008420338854193687, 0.005360274109989405, -0.019122503697872162, -0.015287610702216625, 0.006413870491087437, -0.008275400847196579, -0.011861561797559261, -0.00521403644233942, -0.008697249926626682, 0.017494134604930878, 0.006234201602637768, 0...
[ 0.59258633852005, -0.02960139326751232, 0.3316439688205719, 0.14523258805274963, 0.0365665964782238, -0.037465762346982956, 0.09110645949840546, -0.3193278908729553, -0.2743394672870636, -0.4002962112426758, 0.07704737782478333, 0.5974159240722656, -0.2609306275844574, -0.11831304430961609...
In an answer to another question, I suggested creating a randomized value for input to a specific method. In my experience this has been useful for making tests more readable and it lets you skip the "trivial phase" where you hardcode specific results in a method. Here's the unit test code pasted from the other answer, for quick reference: [Test] public void FullHouseReturnsTrue() { var roll = AnyFullHouse(); Assert.That(sut.IsFullHouse(roll)); } [Test] public void StraightReturnsFalse() { var roll = AnyStraight(); Assert.That(sut.IsFullHouse(roll), Is.False); } A couple of comments in response suggested that this strategy would not work well, because the helper method would need to be tested also. In my experience, I've never had a need to test methods like this, since creating the corresponding production code also tests my test code. Do `AnyStraight()` and `AnyFullHouse()` need to have their own unit tests? If so, how do you solve the chicken-and-egg problem that presents? **EDIT** Would I still want to create dedicated unit tests for the `AnyFullHouse` algorithm if I inlined the method? [Test] public void FullHouseReturnsTrue() { var roll = listOfHardCodedFullHouseRolls[_random.Next(0, listOfHardCodedFullHouseRolls.Length)]; Assert.That(sut.IsFullHouse(roll)); }
[ 0.02265065349638462, 0.01640789583325386, -0.009982453659176826, 0.0006006893236190081, -0.017083261162042618, 0.011596825905144215, 0.0073227956891059875, 0.011580303311347961, -0.015786826610565186, 0.00958833284676075, -0.0033218017779290676, 0.007610333152115345, 0.017242584377527237, ...
[ 0.603861391544342, -0.13582827150821686, 0.25362420082092285, -0.039096493273973465, 0.21569381654262543, 0.23134228587150574, 0.3805583715438843, -0.6404026746749878, -0.009153070859611034, -0.11702621728181839, 0.028134342283010483, 0.3747863471508026, -0.10378455370664597, 0.09199360013...
I am trying to concatenate two fields into one new field in GRASS, but I am unable to get the syntax to work. Ideally, I would like "FIELD1" and "FIELD2" concatenated together in "FIELD3". All fields are VARCHARs. I have tried: v.db.update layername col=FIELD2 value=FIELD1+FIELD2 using many combinations of +, ||, &, and wrapping in single and double quotes all to no avail. Can someone provide some guidance on the required syntax?
[ -0.00805182196199894, 0.021539218723773956, -0.008520172908902168, 0.013043791055679321, -0.014483822509646416, -0.0035853427834808826, 0.008027737960219383, 0.006017948966473341, -0.015371643006801605, 0.006512241903692484, -0.0064224377274513245, 0.011599390767514706, -0.02004324458539486,...
[ 0.4383867681026459, 0.018544912338256836, 0.2404252588748932, -0.4141639471054077, -0.28735649585723877, -0.08524169027805328, 0.2670363485813141, -0.2958969175815582, 0.26032009720802307, -0.6092150807380676, -0.14561913907527924, 0.4832904040813446, -0.2741207182407379, -0.02086650952696...
I am using python in Arcgis 10.2, but it seems Python cannot meet all my developing demands, from the forum, others recommend to use python to access arc objects,but what they mentioned is the old version, so I want to confirm whether I can make it in 10.2, any extension needed? If can, I can use vb.net and c# right? Everybody has experience on it?
[ -0.013635542243719101, -0.004354681354016066, -0.026223033666610718, 0.007300478871911764, -0.002747179241850972, -0.002747005084529519, 0.01024186983704567, 0.01927841454744339, -0.025403635576367378, -0.06528127193450928, 0.011481440626084805, 0.008718764409422874, -0.010224816389381886, ...
[ 0.7640007138252258, 0.30290544033050537, 0.14418314397335052, -0.15592944622039795, -0.20646804571151733, -0.4092216491699219, 0.35764819383621216, 0.03769376873970032, -0.2178458869457245, -0.4669301211833954, 0.16088628768920898, 0.7138077616691589, 0.01796591468155384, -0.29875406622886...
I have a problem getting the `\framezoom` command to work when compiling with XeLaTeX, using fully updated MikTeX 2.9. Here is an example: \documentclass{beamer} \usepackage{mwe} \begin{document} \begin{frame}<1>[label=zoom] \frametitle<1>{Test zoom} \framezoom<1><2>[border](1cm,1cm)(3cm,2cm) \includegraphics[width=0.5\textwidth]{example-image-a.png} \end{frame} \againframe<2->[plain]{zoom} \end{document} This works fine with pdfLaTeX and LuaLaTeX, but not with XeLaTeX: there, it creates the two slides as expected, but both the hyperlinks and the border on frame 1 are missing. I have checked the log-file and the hyperref uses the hxetex driver, i.e. the same as if I pass the option `xetex` to hyperref manually. Any idea how to fix this? (For some reasons, I need to use XeLaTeX.) Thanks PS: I hade to use a PNG version of the picture, as with PDF-version I run into a pdf-version problem and the picture is not shown at all.
[ -0.003125096671283245, 0.003781059989705682, 0.0029776040464639664, 0.011186396703124046, -0.008504549041390419, 0.012689312919974327, 0.0069189053028821945, -0.0011344669619575143, -0.010561199858784676, 0.0012142648920416832, -0.010538325645029545, 0.003229066263884306, 0.00014909100718796...
[ 0.21695712208747864, 0.09381897002458572, 0.5725207328796387, -0.1390201449394226, 0.02693510241806507, -0.10182948410511017, 0.21469435095787048, -0.490642786026001, -0.06426962465047836, -0.6745139956474304, 0.14967651665210724, 0.4935310482978821, -0.3100135624408722, 0.0923539474606514...
Okay, so I have just gotten back into Pokemon, just managed to evolve my Charmeleon into a Charizard. It wants to learn Wing Attack and I was just wondering if you can delete that move but still keep the skill? Is it possible to delete a HM move like Cut, but still be able to use the ability it grants? (cut down trees)
[ 0.010672683827579021, 0.018415043130517006, -0.0003711195313371718, 0.01652720756828785, -0.02324984036386013, -0.009995226748287678, 0.009199672378599644, 0.010059577412903309, -0.02153150364756584, -0.004240741021931171, -0.011423243209719658, 0.016386663541197777, -0.005040650721639395, ...
[ 0.318987101316452, -0.33316245675086975, -0.06319817900657654, 0.42264506220817566, -0.4177919328212738, -0.13072191178798676, 0.10331107676029205, 0.2914193570613861, -0.247384175658226, -0.22757336497306824, 0.3404633104801178, 0.17209599912166595, 0.25571301579475403, 0.198557049036026,...
A few weeks ago I found out about console commands. So I started playing around with them. I changed speed, size, things like that just to mess around. They usually didn't last through saves. Then I played around with my level, just to see what it would be like to be "all powerful" using a forceav command. I changed carry weight, too. A few days ago I decided I wanted to go back to my original, non-cheated game, but the levels and carry weight are stuck. Now when I try to fix it, nothing happens. Does anyone know how to undo forceav commands?
[ 0.0002386195119470358, 0.017132962122559547, 0.00006858690176159143, -0.0016513426089659333, 0.008354688063263893, -0.011676890775561333, 0.004965732805430889, -0.0047795590944588184, -0.014607439748942852, 0.007679441943764687, -0.0008805716643109918, 0.024227574467658997, 0.002854943508282...
[ 0.06281360238790512, -0.22885407507419586, 0.33463364839553833, 0.3290460407733917, -0.10242685675621033, 0.17679601907730103, 0.030489392578601837, -0.5498044490814209, -0.3910556137561798, -0.2644173800945282, 0.40407174825668335, 0.43169108033180237, -0.029005272313952446, -0.0996471717...
On my DV from MediaTemple running Centos 5.8 I installed APC. I then added `extension = "apc.so"` to `/etc/php.ini` so APC will be enabled for all existing and future sites on the server. When I checked `php.info` still no APC showing up. I removed the line and added it to `/var/www/vhosts/example.com/etc/php.ini` and then checked phpinfo which was running in example.com and APC showed up. However, I don't of course want to have to add this each time to every domain. Is there a reason why /etc/php.ini isn't picking it up? I restart httpd, do I need to restart it more? Or if for some reason with FastCGI it has to be in each domain php.ini is there a way to get it added to that template? Also the server uses the default plesk 11.
[ -0.02052130363881588, 0.009967667981982231, 0.003946973942220211, 0.014979483559727669, -0.02365666627883911, -0.0030781920067965984, 0.0077915857546031475, -0.02704729698598385, -0.013078592717647552, -0.01955856941640377, -0.009846743196249008, 0.013178110122680664, -0.020570944994688034, ...
[ 0.280594140291214, 0.11307641863822937, 0.8118048906326294, -0.22351554036140442, -0.24618875980377197, 0.28217586874961853, 0.36443111300468445, 0.09729024767875671, -0.3454801142215729, -0.6892927885055542, -0.009522886015474796, 0.39378097653388977, -0.49708229303359985, 0.0097999302670...
Will I lose my statue? Will I lose my Hall of Monuments point progression? Or is it safe? What happens if I destroy a minipet dedicated at the Hall of Monuments?
[ 0.03303147107362747, 0.019434846937656403, 0.024393750354647636, 0.0037659413646906614, 0.023451626300811768, 0.019815290346741676, 0.017180079594254494, 0.03223381191492081, -0.029437623918056488, -0.013567447662353516, -0.007683930452913046, 0.05450030788779259, -0.0020570093765854836, 0...
[ 0.5730176568031311, 0.4265992343425751, 0.5054230690002441, 0.3328825533390045, 0.10580676048994064, 0.03850393369793892, 0.6673001646995544, 0.09330049157142639, -0.04138736426830292, -0.43556955456733704, 0.19658347964286804, 0.5045337080955505, 0.02718069776892662, 0.2927585542201996, ...
How do I make stairs in Terraria? It's getting really annoying that NPC's can't go up platforms.
[ 0.021221209317445755, 0.018781542778015137, -0.0020442502573132515, 0.006331838201731443, -0.041043054312467575, -0.017574774101376534, 0.01474696397781372, -0.017689721658825874, -0.05384054034948349, -0.022109802812337875, -0.046948499977588654, 0.03229855000972748, -0.008929111994802952, ...
[ 0.28568387031555176, 0.0639319196343422, 0.31306615471839905, 0.5190388560295105, 0.17692141234874725, -0.15196073055267334, 0.4224606156349182, -0.2725706696510315, -0.7005471587181091, -0.6018344163894653, 0.24115000665187836, -0.06328246742486954, 0.17036205530166626, 0.0438669510185718...
In the Embassies area of the Citadel, I talked to an officer named Delk. He's requesting I find some Cerberus ciphers for him. The specific quest text is as follows: > A C-Sec officer needs updated ciphers to break Cerberus codes. Find the > ciphers and deliver them to him, at the Citadel embassy. **Where are the ciphers, and what is the reward for returning them to Delk?**
[ 0.01564621552824974, 0.00684752594679594, -0.016441600397229195, 0.00003357750392751768, -0.017505621537566185, 0.0176126379519701, 0.01119824219495058, -0.013607374392449856, -0.023467442020773888, -0.017200347036123276, -0.0025811914820224047, 0.019402796402573586, 0.004615537356585264, ...
[ 0.37797674536705017, 0.4749465584754944, 0.3571913540363312, 0.5681947469711304, 0.24577979743480682, 0.06252364069223404, 0.5181414484977722, -0.23856836557388306, -0.2296711802482605, -0.687846839427948, -0.5740777850151062, -0.0070618740282952785, -0.288566529750824, 0.18191049993038177...
I'm trying to create a UDF partition on a flash drive to install grub2 (to boot ISOs from) ~$ export DSK=/dev/sdd # wipe the disk ~$ sudo dd if=/dev/zero of=${DSK} bs=512 count=2048 ~$ sudo sgdisk --zap-all ${DSK} # create partition for udf ~$ #sudo parted -s ${DSK} mklable msdos ~$ #sudo parted -s ${DSK} -a optimal unit MB -- mkpart primary 1 -1 ~$ (echo o; echo n; echo p; echo 1; echo ; echo ; echo t ; echo 6 ; echo w) | sudo fdisk -b 512 ${DSK} ~$ sudo mkudffs --media-type=hd --blocksize=512 ${DSK}1 trying to change type of multiple extents
[ -0.023930413648486137, 0.0023815990425646305, -0.01115044392645359, 0.016862493008375168, -0.021352123469114304, 0.013902124017477036, 0.007891726680099964, 0.001905153039842844, -0.01334613561630249, 0.005528877954930067, -0.0049164872616529465, -0.0036524306051433086, -0.006713752634823322...
[ -0.1154155507683754, -0.07308895885944366, 0.21853144466876984, -0.018552981317043304, 0.06644093245267868, 0.13933376967906952, 0.26104363799095154, -0.11284393817186356, 0.03329882398247719, -0.9924895763397217, -0.02391556091606617, 0.12600809335708618, -0.3949221074581146, 0.3910012841...
I was looking at this page http://www.illustratingshadows.com/www-formulae-collection.pdf and found the equation used for estimating the solar declination (sun declination). One problem, though, I wish to cite this equation in my work and was hoping there would be a textbook that shows this same equation? Can anyone suggest a reference for this other than a website?
[ -0.015223096124827862, -0.0005994853563606739, -0.008285300806164742, 0.019582398235797882, -0.058065108954906464, -0.011774437502026558, 0.008890654891729355, -0.017385469749569893, -0.02937932126224041, -0.04300031438469887, 0.01988154463469982, 0.019036486744880676, -0.046421073377132416,...
[ 0.4062398374080658, -0.060161691159009933, 0.22712385654449463, 0.1346496045589447, -0.0854172483086586, -0.06714087724685669, -0.28358393907546997, 0.33471599221229553, -0.3392661213874817, -0.6539754271507263, 0.3791249990463257, 0.20999261736869812, -0.032420966774225235, 0.298112601041...
I'm writing a paper in LaTeX, and I'm getting a strange "`Missing $ inserted`" error when I try to compile the paper using `pdflatex`, but only if it contains the following code. \begin{algorithm*}[tbh] \begin{haskell} translate :: \textsc{Featherweight Java} \Rightarrow \textsc{Feather-Sub terms}\\ translate (class C_1 extends C_2 \{T_i f_i ^ \{i \in 1...n\}\; K M_j ^ \{j \in 1...m\}\}) = \hsbody{\langle C_1 = \{(translate C_2), f_i \: (translate T_i) ^ \{i \in 1...n\}\} \rangle where (translate K) and (translate M_j ^ \{j \in 1...m\}) are emitted as separate translations into the same unit}\\ translate (C(C_i f_i) \{ super(f_i); this.f_i = f_i; \} ^ \{i \in 1...n\}) = \hsbody{\hslet{C = \lambda f_i.\langle C = \{l_i = f_i\} ^ \{i \in 1...n\}}{rest of translation unit}}\\ translate (C' m(C_i x_i ^ \{i \in 1...n\}) \{ return e; \} in class C) = \hslet{ m = \lambda t.case t of \langle C=this \rangle \Rightarrow \lambda x_i.(translate ((C')e)) for each subclass C' of C with overriding method m' \langle C' = this \rangle \Rightarrow m' t x_i}{rest of translation unit}\\ translate x = x translate t.f \hswhere{t is known to have type C} = case t of \langle C = x \rangle x.f for each subclass C' of C \langle C' = x \rangle x.f\\ translate t_1.m_i ^ \{i \in 1...n\} = m (translate t_1) (translate t_i) ^ \{i \in 1...n \} \\ translate (new C(t_i ^ \{i \in 1...n\})) = C (translate t_i) ^ \{i \in 1...n \} \\ translate ((C)t) = \hslet{t' = (translate t)}{\langle C = \{for each f_i : \tau_i \in (translate C) f_i = t'.f_i\} \rangle} \end{haskell} \caption{Encoding of Featherweight Java into Feather Sub\label{fig:Encoding-of-Featherweight}} \end{algorithm*} I already have `\usepackage` commands in my prelude for the `algorithm*` and `haskell` environments, and I use them earlier in the paper. This algorithm is listed in an appendix section, could that change things somehow? What can I do to fix it? I would really appreciate any help I can get with this, as the submission deadline is this Friday! Everything else about the paper is ready for submission except this, so, once again, would really appreciate any help.
[ 0.00024412153288722038, 0.019458265975117683, -0.0026541207917034626, 0.019330915063619614, 0.00907975621521473, 0.011342509649693966, 0.009126776829361916, -0.00810270756483078, -0.015432214364409447, -0.013129159808158875, -0.006674647331237793, -0.0016916438471525908, -0.00642328895628452...
[ 0.046401720494031906, 0.2224130779504776, 0.4187890589237213, -0.12659087777137756, 0.20518526434898376, -0.2588031589984894, 0.10381713509559631, -0.1196388378739357, 0.14341233670711517, -0.4147709906101227, -0.10366139560937881, 0.0969708040356636, -0.23297064006328583, -0.1864193379878...
When Thane got stabbed my map showed that he is in the hospital.When I arrived on the Citadel he was still in the hospital.But when I entered the hospital he was gone.Was I too late?Or is this a glitch?
[ -0.0005463421111926436, 0.03950846567749977, 0.010870087891817093, 0.0172542966902256, -0.0308939628303051, -0.03132212907075882, 0.010991762392222881, 0.008607624098658562, -0.023237189278006554, -0.0012884271563962102, -0.006903586909174919, 0.03187032416462898, 0.0326337032020092, 0.018...
[ 0.3515488803386688, -0.2739425301551819, 0.29080769419670105, 0.29283976554870605, -0.10939175635576248, 0.1198103204369545, 0.22307375073432922, -0.19920606911182404, 0.08872058987617493, -0.6416600942611694, 0.11153474450111389, -0.1494724005460739, -0.5270722508430481, 0.350740343332290...
In his famous The Free Lunch Is Over article from 2005, Herb Sutter predicted a Concurrent Programming Revolution as big as Object-Oriented Revolution. Has this revolution really happend in years 2005 - 2013? * * * Key points in the article: * Processor manufacturers have run out of room with most of their traditional approaches to boosting CPU performance. Instead of driving clock speeds ever higher, they are instead turning to hyperthreading and multicore architectures. * Applications will increasingly need to be concurrent if they want to fully exploit CPU throughput gains. * “Oh, performance doesn’t matter so much, computers just keep getting faster” statement will be wrong. * Efficiency and performance optimization will get more, not less, important. Those languages that already lend themselves to heavy optimization will find new life; those that don’t will need to find ways to compete and become more efficient and optimizable. Expect long-term increased demand for performance-oriented languages and systems. * Programming languages and systems will increasingly be forced to deal well with concurrency. We desperately need a higher-level programming model for concurrency than languages offer today.
[ -0.007600401993840933, 0.006386307533830404, -0.01362280547618866, 0.02121003158390522, -0.005915320944041014, -0.024051757529377937, 0.008124805986881256, -0.0072351545095443726, -0.012160640209913254, -0.016710704192519188, -0.0007837224984541535, 0.02726890705525875, -0.000900140847079455...
[ 0.13949468731880188, 0.03183585777878761, 0.031028805300593376, 0.3796829581260681, -0.2523290812969208, 0.02502402849495411, -0.060936689376831055, 0.5841434597969055, -0.6249014735221863, -0.43280211091041565, 0.022877728566527367, 0.1507672816514969, -0.2685273289680481, 0.3386562764644...
As $E=hf=\frac{hc}{\lambda}$, blue light - with a smaller wavelength - should have a higher energy. However, it is the case that blue light scatters the most. Why is it that higher energy rays scatter more?
[ 0.008222323842346668, -0.003980140667408705, -0.020713986828923225, -0.0069329096004366875, -0.004196311812847853, -0.037263933569192886, 0.010016473941504955, -0.003985254094004631, -0.017637217417359352, -0.0815504938364029, -0.010892662219703197, 0.012710785493254662, -0.02000260911881923...
[ 0.6201006174087524, 0.15808530151844025, 0.4218474328517914, 0.12581411004066467, 0.12739793956279755, -0.1775566041469574, -0.008895009756088257, -0.35542404651641846, -0.4515136480331421, -0.6022536754608154, -0.29356539249420166, 0.5302978754043579, -0.48703303933143616, 0.2494467645883...
I have the following MWE \documentclass[pstricks,border=12pt]{standalone} \usepackage{pst-node, pst-plot} \psset{unit=0.1cm} \begin{document} \begin{pspicture}[showgrid=false](-3.0, -25)(110, 24.0) \psaxes[Ox=0,Dx=10,yAxis=false]{->}(-2.0,-20)(105,23) \end{pspicture} \end{document} The axis runs from `0..100`, as desired - but I would like for it to start at e.g. -2 or -3, such that there is a "buffer" before the axis starts. The ticks should be as they are now, i.e. at 0, 10, 20, .... All I want is a "buffer- axis" that starts at -2. If I use `Ox=-2`, then the ticks run from -2, 8, ...
[ -0.0037663918919861317, -0.0010882765054702759, -0.0030144713819026947, 0.01942799985408783, 0.0029956025537103415, -0.0030503706075251102, 0.007227764464914799, -0.000886192312464118, -0.009928696788847446, 0.006083489395678043, -0.006548166740685701, -0.003184097120538354, -0.0015708014834...
[ 0.10889130085706711, 0.006207896396517754, 0.8398816585540771, -0.0089946324005723, 0.03668104112148285, 0.2625501751899719, 0.3886323869228363, -0.4266088604927063, 0.3747074007987976, -0.7522907853126526, 0.1418459415435791, 0.2288079857826233, -0.05928317457437515, 0.2516563832759857, ...
We all know the effects of speed on time, and we have a formula for that, but I have a hunch, that time also is relative to pressure, and that by increasing the pressure on matter, in fact time will speed up for that matter. I am not a physics expert ( I have a Masters Degree in Computer Engineering, and have hence taken several physics, math and chemistry classes ) so don't slaughter me, I would appreciate a talkative tone. So I will say some "ideas" in my head that might be stupid, or assumptions from my understanding of how things are that are in fact wrong from an expert view, but please try to understand my **intention** or follow my idea of what I am thinking instead of focusing on details. So, my thinking is that once something starts travelling near the speed of light, or the speed of light, it will start to dissolve, before turning into energy and becoming the lightest of all while dissolving = light, and hence is why it can travel at the speed of light. So if something is to travel at the speed of light, it actually have to be dissolved into energy. The idea I have with pressure, is that if matter, in a pressure chamber, where pressure is turned up, the matter will eventually turn into pure energy as well = light. I also have this thought, that once pressure is applied to matter, the atoms starts to vibrate, move faster, and a reaction that would normally take hundreds of years, can be achieved with quicker results in a pressure chamber because time for all that materia increases, not because of the pressure, but because time is happening faster, which supports my idea of time actually speeding up for those particles. What happens with used reactor fuel in a pressure chamber today? How much can we speed up the halftime? It appears to me that if this was true, it will be the contradiction to speed. A twin X travelling near the speed of light, will come back to visit an older twin Y on earth. However, with pressure, it might just be the opposite, that twin X in the pressure chamber will in fact be older than twin Y outside. Surely turning up pressure on a twin might not be so healthy, but neither is acceleration from 0 to the speed of light in 1s. So pressure is probably not a healthy way to speed up time for matter, it will probably destroy certain things in it. Am I on something here? No?
[ -0.011457931250333786, 0.012916108593344688, -0.008451297879219055, 0.008901848457753658, -0.0020926117431372404, -0.010117116384208202, 0.0036711059510707855, -0.010160685516893864, -0.009737798944115639, -0.007985246367752552, 0.004472790285944939, 0.015782106667757034, 0.00214409735053777...
[ 0.5269378423690796, 0.13612408936023712, 0.03095298632979393, 0.11829718202352524, 0.06389135867357254, -0.09316664934158325, 0.25662150979042053, 0.33101293444633484, -0.12537521123886108, -0.37635883688926697, 0.41617918014526367, 0.633428156375885, 0.17064660787582397, 0.081347331404685...
This question may well be too open ended to get a definitive answer, but hopefully not. Machine learning algorithms, such as SVM, GBM, Random Forest etc, generally have some free parameters that, beyond some rule of thumb guidance, need to be tuned to each data set. This is generally done some kind of re-sampling technique (bootstrap, CV etc) in order to fit the set of parameters that give the best generalisation error. My question is, can you go too far here? People talk about doing grid searches as so forth, but why not simply treat this as an optimisation problem and drill down to the best possible set of parameters? I asked about some mechanics of this in this question, but it hasn't received much attention. Maybe the question was badly asked, but perhaps the question itself represents a bad approach that people generally do not do? What bothers me is the lack of regularisation. I might find by re-sampling that that best number of trees to grow in a GBM for this data set is 647 with an interaction depth of 4, but how sure can I be that this will be true of new data (assuming the new population is identical to the training set)? With no reasonable value to 'shrink' to (or if you will, no informative prior information) re-sampling seems like the best we can do. I just don't hear any talk about this, so it makes me wonder if there is something I'm missing. Obviously there is a large computational cost associated with doing many many iterations to squeeze every last but of predictive power out of a model, so clearly this is something you would do if you've got the time/grunt to do the optimisation and every bit of performance improvement is valuable.
[ 0.005566637963056564, 0.010346041060984135, 0.008574429899454117, 0.008668802678585052, -0.009022366255521774, 0.005527514964342117, 0.007806027308106422, 0.006141234654933214, -0.01151964906603098, -0.0032870490103960037, -0.008524568751454353, 0.0136971864849329, -0.002402530051767826, 0...
[ 0.08768032491207123, -0.201145738363266, -0.052235063165426254, 0.5421154499053955, -0.04381738230586052, -0.36073869466781616, 0.2005818635225296, -0.0393391028046608, -0.28379538655281067, -0.47181445360183716, 0.1048976480960846, 0.31658148765563965, -0.2965763509273529, 0.2886616587638...
I want to try the new oxygen to gtk port for kde 4.6. But I'm not sure how to enable it. I currently use qtcurve for gtk.
[ -0.001258234609849751, -0.006225822493433952, -0.03481394425034523, -0.006485546473413706, -0.048199351876974106, -0.015689732506871223, 0.014419682323932648, -0.018168890848755836, -0.01969427429139614, -0.032846737653017044, -0.0030348608270287514, 0.006417948752641678, -0.0306881591677665...
[ 0.2943061888217926, 0.1900745928287506, 0.09571400284767151, -0.03449080511927605, -0.17086590826511383, -0.16867193579673767, 0.31933271884918213, 0.38554683327674866, 0.39475035667419434, -0.8076854944229126, 0.024326497688889503, 0.3586606979370117, -0.30051755905151367, 0.2683719694614...
I have an equation of state as a function of $p$ (pressure), $T$ (temperature) and $V$ (volume) for a real gas and need to find an expression for the internal energy $U$. I know this is given by the first law $$U=Q+W=TS+pV$$ I am not sure how to start. Do I use the differential for $$dU=TdS+pdV$$ and the Maxwell relations to get an expression for $dU$ in terms of $p$,$V$ and $T$? Or do I integrate $pdV$? But then how do I integrate $TdS$?
[ -0.009274609386920929, 0.00015470257494598627, -0.0034695956856012344, 0.007961606606841087, -0.030529983341693878, -0.0026029443833976984, 0.0077338507398962975, -0.007032450288534164, -0.009879897348582745, -0.017681725323200226, 0.001550585962831974, 0.008635247126221657, -0.0209356062114...
[ 0.2587568461894989, 0.23467819392681122, 0.38271400332450867, -0.1895737498998642, 0.0718182921409607, 0.3558691143989563, 0.0029252218082547188, -0.6711997985839844, 0.49843505024909973, -0.5452647805213928, 0.08312127739191055, 0.623608410358429, -0.4036242365837097, 0.600376546382904, ...
I have a sentence: > However, until then it is obvious that there lie intermediate “targets” — > foo, bar and baz. Is the use of _lie_ in this way correct? The targets still exist, so I don’t want the past tense. I don’t think that _lies_ sounds correct, because there ~~’s~~ are multiple targets and I'd say _the dogs lie_ rather than _the dogs lies_.
[ 0.01043908204883337, 0.006811023224145174, -0.01692330650985241, 0.011913765221834183, -0.0073500438593328, -0.009199796244502068, 0.007239954546093941, 0.003985993564128876, -0.011039016768336296, -0.020384980365633965, -0.004729752894490957, 0.0019308588234707713, -0.0013860291801393032, ...
[ 0.056617334485054016, 0.25561660528182983, 0.27656319737434387, -0.3139655888080597, -0.2559363543987274, 0.24116250872612, 0.46256986260414124, -0.2289539873600006, -0.1549903005361557, -0.5643736720085144, -0.1991218626499176, 0.5874223709106445, -0.12389620393514633, -0.0278862211853265...
Is there a possibility to count to how many categories a post belongs to?
[ -0.04265444353222847, 0.05424221605062485, 0.003819801611825824, 0.08013147115707397, 0.07422331720590591, 0.014515991322696209, 0.014741952531039715, -0.006147283595055342, -0.010943722911179066, 0.06840912252664566, -0.049928322434425354, 0.008058634586632252, 0.0026833496522158384, 0.02...
[ 0.3252200782299042, 0.00583474338054657, -0.12527336180210114, 0.25008150935173035, 0.08587946742773056, 0.2537551522254944, -0.0662447139620781, 0.11453352123498917, -0.44995206594467163, -0.28261810541152954, 0.0667332336306572, -0.15579532086849213, -0.12896384298801422, 0.0033786771818...
The POSIX description of the `-b` flag for the `sort` command is > Ignore leading characters when determining the starting and ending positions > of a restricted sort key. I can understand the use for the starting position of the key, but what about the ending positions? Can anybody give an example? For instance, with the locale set to POSIX, the file with content x z x a would be sorted differently under `sort -k 2` and `sort -k 2b`, but I haven't been able to come up with a case where, say, `sort -k 2,3b` and `sort -k 2,3` would make a difference.
[ 0.01045167725533247, 0.009122630581259727, -0.01584039255976677, 0.01939784735441208, -0.012827439233660698, 0.017098288983106613, 0.009277674369513988, 0.015263160690665245, -0.02048613131046295, 0.05086131393909454, -0.018015054985880852, 0.002041897037997842, -0.010382783599197865, 0.02...
[ -0.005028335843235254, 0.03882977366447449, 0.1438288390636444, -0.26659443974494934, -0.03569793328642845, -0.2544296383857727, 0.041980937123298645, -0.10833052545785904, -0.1873299926519394, -0.267505407333374, -0.3980132043361664, 0.2961992025375366, -0.5448391437530518, -0.10230398178...
**Question 1** What's the computational complexity of the Groovy `unique()` method? **Question 2** How could I have figured it out by myself? The `unique()` method is defined in the class `DefaultGroovyMethods`. The source code can be found here: org.codehaus.groovy.runtime.DefaultGroovyMethods. Can you point me to the piece of code which illustrates the answer to question 1?
[ 0.004252092447131872, 0.012618759647011757, -0.0019424697384238243, 0.005594807211309671, -0.009413567371666431, -0.0233388002961874, 0.007267645560204983, -0.02026309072971344, -0.013187945820391178, 0.012003278359770775, -0.0056825545616447926, 0.011390084400773048, 0.018881823867559433, ...
[ 0.11798218637704849, 0.33176189661026, -0.0416354201734066, 0.34782424569129944, 0.03891284391283989, 0.0252851415425539, -0.0166157353669405, -0.20678269863128662, -0.15244358777999878, -0.3224763870239258, 0.024067910388112068, 0.2175576090812683, -0.19320596754550934, 0.0579164698719978...
I'm trying to use the ManPageView addon to view info documentation within `vim`. Although I find ManPageView very useful for, well, viewing manpages, its info- reading functionality doesn't seem to work at all. But then I almost never read info documentation, mostly because it doesn't seem worth taking the time to learn the `info` interface. So I may be doing it wrong. Anyhow: running `:Man info.i` as suggested by the ManPageView documentation gives me info: No menu item `(info)Top' in node `(dir)Top'. info: No menu item `(info)Top' in node `(dir)Top'. info: No menu item `(info)Top' in node `(dir)Top'. ***warning*** sorry, no manpage exists for <(info)Top> I see some success if I delete the following lines, which start at line 345 in the file `autoload/manpageview.vim`: " special exception for info {{{3 if a:viamap == 0 && ext == "i" let s:manpageview_pfx_i = "(".manpagetopic.")" let manpagetopic = "Top" " call Decho("top-level info: manpagetopic<".manpagetopic.">") endif Deleting those lines will allow `:Man info.i` to work, but I still can't follow links in the displayed page. It works by calling `info` on the command line and capturing the output, so perhaps it's just that `info`'s CLI has changed? Specifically, `$ info info Advanced` will bring up the topic "Advanced" within the "info" node. I dug around in the addon's source code for a while but didn't find any obvious way to direct it to pass the node as an extra argument instead of doing what it seems to be supposed to do, which is to wrap the node name in parentheses and prefix it to the topic, passing the combination as a single command-line argument to `info`. I've not tried to use this addon to view info documentation before. I'm running an updated ubuntu 10.10, using the vim-gnome package. I've filed some semblance of a bug report at what seems to be the recommended location Have others had success using this addon to view info documentation? I feel like I'm lost in the mists that eternally shroud the outer reaches of Obscurity. **note** : Whilst composing this message, I've discovered the info addon, which seems to work acceptably, at least at first glance.
[ -0.00887242704629898, 0.017870761454105377, 0.0026449463330209255, 0.010580725967884064, -0.015783485025167465, 0.001971364254131913, 0.008338632062077522, 0.005645121447741985, -0.018620500341057777, 0.02764708921313286, -0.02162245102226734, 0.007375660352408886, -0.000931537477299571, 0...
[ 0.26895639300346375, 0.15606245398521423, 0.3886738419532776, 0.09629574418067932, 0.13341431319713593, -0.37041395902633667, 0.3919726312160492, 0.08104486763477325, -0.09763262420892715, -0.6540477871894836, 0.08001989126205444, 0.8794618248939514, -0.38380172848701477, 0.422555297613143...
My code is given below. In the for loop I am getting unexpected output i.e. before completing the execution of for loop by 1st thread the 2nd thread comes in for loop and shows incremented value. public class ThreadSafe { public static void main(String[] args) throws InterruptedException { System.out.println("main()"); B b=new B("1st Thread"); B b1=new B("2nd Thread"); B b2=new B("3rd Thread"); b.start(); b1.start(); b2.start(); } } class MyCounter { private static int count; public static int getCount(){ for(int m=0;m<2;m++){ System.out.println(count+" "+Thread.currentThread().getName()); } return count++; } } class B extends Thread{ public B(String tname) { super(tname); } public void run() { MyCounter.getCount(); } } The question is when 1st thread executes for loop then other threads executing return statement is this the reason of showing increment value or something else? How other thread can show the incremented value unless the 1st thread reaches the return statement?
[ -0.013405181467533112, 0.026876911520957947, 0.0004803610499948263, 0.01122301071882248, -0.00003060465678572655, -0.004040890838950872, 0.0049142539501190186, 0.013265365734696388, -0.010098563507199287, 0.00878944806754589, -0.011525120586156845, 0.018206045031547546, -0.011949408799409866...
[ 0.2156306803226471, -0.10890565067529678, 0.2976049482822418, -0.18101392686367035, 0.41971060633659363, 0.18343134224414825, 0.47606775164604187, 0.11929158866405487, -0.3665159344673157, -0.4945879578590393, -0.10706929117441177, 0.46361106634140015, -0.4143831431865692, 0.44737362861633...
As far as I can see `PrincipalComponents[data]` just gives data in the principal component basis. The problem is that after this, I would like to transform new data into the same basis, which requires that I know the actual principal components. There is the option of calculating `Eigenvectors[Covariance[data]]` (given the data is centered on the origin) but for the eigenvectors I get this way the signs are usually different from the ones that the `PrincipalComponents[]` function uses. Is there an easy way of getting the actual transformation used or transforming new data into the PC basis of the previous dataset?
[ 0.005339752417057753, 0.009563842788338661, -0.008152758702635765, 0.02286561205983162, -0.004628789611160755, -0.005424029193818569, 0.006512245163321495, -0.017392940819263458, -0.012910265475511551, 0.01967025175690651, -0.004614444449543953, 0.02044537663459778, -0.003280818462371826, ...
[ 0.22855636477470398, -0.1390543431043625, 0.37171727418899536, 0.055392637848854065, -0.22118601202964783, -0.051785148680210114, -0.08735360950231552, 0.05236561968922615, -0.23517322540283203, -0.388734370470047, 0.07831548899412155, 0.42302757501602173, 0.01754922606050968, 0.4402067661...
I have a wavey looking curve and would like to show the corresponding normal coordinate system, which is defined by arclength along the curve and (signed) distance from the curve. So there should be lines normal to the curve (for arclength) and curves parallel to the original curve (for distance from the curve). I am using the solution from Easy curves in TikZ to draw my wavey curve and How to draw tangent line of an arbitrary point on a path in TikZ to get the normal lines. I can draw a single parallel line using `double distance=1cm` from How to raise a generic curve. Problem with \pgfdeclaredecoration, but I'm not sure how to get a whole set of parallel lines. Eventually I would also like nodes that label the lines/curves `s=1`, `s=2`, ..., `r=1`, `r=2`,...,`r=0`,`r=-1`,... Here is my current code: \documentclass[10pt]{amsart} \usepackage{tikz} \usetikzlibrary{decorations.markings} \tikzstyle{tangent}=[decoration={markings,% switch on markings mark= at position #1 with { \coordinate (tangent point-\pgfkeysvalueof{/pgf/decoration/mark info/sequence number}) at (0pt,0pt); \coordinate (tangent unit vector-\pgfkeysvalueof{/pgf/decoration/mark info/sequence number}) at (1,0pt); \coordinate (tangent orthogonal unit vector-\pgfkeysvalueof{/pgf/decoration/mark info/sequence number}) at (0pt,1); }}, postaction=decorate] \tikzstyle{use tangent}=[ shift=(tangent point-#1), x=(tangent unit vector-#1), y=(tangent orthogonal unit vector-#1)] \tikzset{forcedist/.style={decorate, decoration={markings, mark=between positions 0 and 1 step 0.0999 with {\draw[<-] (0,-#1) -- (0,#1); }}}}%Here \tikzset{use tangent/.default=1} \begin{document} \begin{figure} \begin{tikzpicture}[scale=.75] \draw [cyan, double distance = 1cm, tangent=0.001,tangent=0.1,tangent=0.2,tangent=0.3,tangent=0.4,tangent=0.5,tangent=0.6,tangent=0.7,tangent=0.8,tangent=0.9,tangent=0.999] plot [smooth, tension=1] coordinates { (0,.5) (3,2) (7,1) (10,3)}; \foreach \i in {1,...,11} { \draw [very thin,gray,use tangent=\i] (0,-4) -- (0,4); } \end{tikzpicture} \end{figure} \end{document} ![Valid XHTML](http://i1294.photobucket.com/albums/b611/Hanmyo/sample_image.png)
[ -0.01093280129134655, 0.010065194219350815, -0.02082357555627823, 0.017438659444451332, -0.03187624365091324, -0.026686377823352814, 0.010952360928058624, 0.004444882273674011, -0.02031697705388069, -0.02211049199104309, -0.004779982380568981, 0.0038186938036233187, -0.008952700532972813, ...
[ 0.12704886496067047, 0.02438282035291195, 0.9058170318603516, -0.04332523047924042, -0.37621167302131653, 0.08211617916822433, -0.29296958446502686, -0.1942662000656128, -0.2734820246696472, -0.4549315869808197, 0.28903135657310486, 0.49940919876098633, -0.10452932119369507, 0.112133398652...
When building a CART model (specifically classification tree) using rpart (in R), it is often interesting to know what is the importance of the various variables introduced to the model. Thus, my question is: **What common measures exists for ranking/measuring variable importance of participating variables in a CART model? And how can this be computed using R (for example, when using the rpart package)** For example, here is some dummy code, created so you might show your solutions on it. This example is structured so that it is clear that variable x1 and x2 are "important" while (in some sense) x1 is more important then x2 (since x1 should apply to more cases, thus make more influence on the structure of the data, then x2). set.seed(31431) n <- 400 x1 <- rnorm(n) x2 <- rnorm(n) x3 <- rnorm(n) x4 <- rnorm(n) x5 <- rnorm(n) X <- data.frame(x1,x2,x3,x4,x5) y <- sample(letters[1:4], n, T) y <- ifelse(X[,2] < -1 , "b", y) y <- ifelse(X[,1] < 0 , "a", y) require(rpart) fit <- rpart(y~., X) plot(fit); text(fit) info.gain.rpart(fit) # your function - telling us on each variable how important it is (references are always welcomed)
[ 0.008131962269544601, 0.016781756654381752, -0.01172487623989582, 0.009326189756393433, 0.0009954813867807388, 0.010514415800571442, 0.008715031668543816, -0.010412808507680893, -0.015103420242667198, -0.01795484498143196, -0.008286042138934135, 0.01112445630133152, -0.009252206422388554, ...
[ -0.09665236622095108, 0.035867538303136826, 0.3464411795139313, 0.49093061685562134, -0.2944033443927765, 0.21724027395248413, -0.2384023666381836, -0.26883161067962646, -0.3083913028240204, -0.42079126834869385, 0.012981836684048176, 0.5612630248069763, -0.25974422693252563, -0.0465436950...
I'm sure you guys are wondering, "why the hell?" - basically it is because it does not work with a plugin by NextSCRIPTS that pulls comments from facebook. When I'm in the admin panel I see the comments properly, but on the blog posts it just keeps showing the admin instead of the users. This happened after updating to 3.6, NextSCRIPTS told me they are aware of this issue but it has been one month and we have over 8,000 comments and I'd like to find a work around. So really, you don't need any code from me, just please if anyone knows, can you share how to grab the comments without using wp_list_comments? there has to be another way. Thanks, Matt
[ -0.014108831062912941, -0.008797879330813885, -0.0008310999255627394, 0.025596899911761284, -0.014283050782978535, 0.00504846777766943, 0.005362268537282944, 0.01131644006818533, -0.011837615631520748, 0.0036381524987518787, -0.004975688178092241, 0.007849283516407013, 0.006994239985942841, ...
[ 1.1660219430923462, 0.18837210536003113, -0.03109906241297722, 0.10425682365894318, -0.49993354082107544, -0.1959134340286255, 0.47199949622154236, 0.2904721200466156, -0.41315069794654846, -0.5212199091911316, 0.24405507743358612, 0.052425213158130646, -0.1889336258172989, 0.1757488101720...
When reading the topic modeling tutorial written by Blei, KDD 2011 tutorial I was confused about a set of diagrams which aim to show the effect of $\alpha$ in Dirichlet distribution. For example, for the plot with $\alpha=1$, what am I suppose to discover? What does item mean here? Do those 15 items mean a 15-dimensional probability vector? The $\alpha=1$ is assumed to lead to a uniform distribution. But I feel confused about how to connect this plot with a uniform distribution. I also list other plots with $\alpha=100$ and $\alpha=0.01$. ![enter image description here](http://i.stack.imgur.com/jO3Bo.jpg) ![enter image description here](http://i.stack.imgur.com/QikYy.jpg) ![enter image description here](http://i.stack.imgur.com/BmSoz.jpg)
[ -0.01671449840068817, 0.026827290654182434, -0.006558444816619158, 0.02146337553858757, -0.006205726880580187, -0.01482870988547802, 0.009091028943657875, -0.00965538527816534, -0.018561353906989098, -0.016987772658467293, -0.019917190074920654, 0.008227149024605751, -0.008327112533152103, ...
[ -0.14701680839061737, -0.4284687042236328, 0.25292015075683594, 0.18821552395820618, -0.46964821219444275, 0.02938811667263508, -0.00526985852047801, -0.3675025403499603, -0.32491379976272583, -0.5153903961181641, 0.015762075781822205, 0.4233834445476532, -0.2603648006916046, 0.28592747449...
I am constantly reading comments and documents from people who use the verb _redact_ to refer to the act of _deleting_ or otherwise _censoring_ content. This never _seemed_ correct to me, but until today it never actually occurred to me to look up the dictionary definition. I just did, and this is what it says: > _**–verb (used with object)_** > **1.** to put into suitable literary form; revise; edit. > **2.** to draw up or frame (a statement, proclamation, etc.). Nowhere does this seem to indicate anything about _deletion_. I suppose that, in a very limited set of circumstances, redaction in the context of "editing for publication" might indeed involve the deletion of certain content, but that aspect seems incidental rather than fundamental. That said, I accept that plenty of "common English" hasn't made it to the dictionaries or style guides yet, so I'm wondering if maybe I missed a memo, because the "delete" definition seems to come up almost everywhere I look. For example: 1. > ...before sharing board minutes (or shareholder or committee minutes) with auditors or other third parties, carefully review the minutes and **redact (delete)** sections containing privileged information to avoid waiving the attorney-client privilege. [AllBusiness] 2. > The term **redaction** may not be a household term, but is often used in the legal community. **It's the practice of removing confidential or sensitive data before giving the document to others.** [ProductivityPortfolio] 3. On synonyms.net, it is actually defined as "edit" but has several "delete" synonyms listed: > cast, **edit out** , edit, blue-pencil, **cut** , redact, put, frame, couch, > **delete** 4. Adobe Acrobat even has a feature named Search and Redact that _deletes_ results. So... what exactly is going on here? Is this usage: * Inane corp-speak parroted by people who have no idea what it means? * Jargon that escaped from the legal profession and made it into common use? * The result of most types of redaction being deletion, leading to natural confusion due to frequent use? * A more mundane form of confusion - possibly with the similar-sounding _retract_? * Just another case of the dictionary being at odds with the informal/spoken language? Basically, what I'm trying to understand is: **Is this really an acceptable usage of the word, and if so, then what is its real origin and/or justification?** Note: I'm having a little trouble making sense of the tagging conventions here, so please feel free to retag this if necessary.
[ 0.0009735152125358582, 0.0015738662332296371, -0.010023208335042, 0.010766694322228432, -0.0009187553077936172, -0.026211939752101898, 0.0065903570502996445, 0.008622126653790474, -0.01590845361351967, 0.03082745335996151, 0.004891235381364822, 0.0004580065724439919, 0.014126473106443882, ...
[ 0.4499742090702057, 0.34421342611312866, 0.38502034544944763, -0.12552331387996674, -0.282088965177536, -0.15053929388523102, 0.2330314964056015, 0.04746131971478462, -0.22019577026367188, -0.6557784676551819, -0.24825914204120636, 0.09685738384723663, -0.13528957962989807, 0.2787629663944...
Is it possible for information (like 1 and 0s) to be transmitted faster than light? For instance, take a rigid pole of several AU in length. Now say you have a person on each end, and one of them starts pulling and pushing on his/her end. The person on the opposite end should receive the pushes and pulls instantaneously as no particle is making the full journey. Would this actually work?
[ 0.004119472112506628, 0.01771780289709568, -0.01777128130197525, 0.007426896132528782, -0.014067213982343674, -0.015455368906259537, 0.007526294328272343, -0.004624172113835812, -0.017098428681492805, -0.023078659549355507, -0.00732428440824151, 0.01227568555623293, 0.00029163077124394476, ...
[ 0.5359463095664978, -0.3367939591407776, 0.2630569636821747, 0.455689400434494, -0.15122807025909424, 0.3010009229183197, -0.10511504113674164, -0.06523873656988144, -0.2893720865249634, -0.3420369625091553, 0.2537219524383545, 0.03555747866630554, -0.18803755939006805, 0.4154968857765198,...
In order to import/extract data from tons of DWG files, I'd like to know experiences with ArcGIS Data Interoperability, and its strong points versus FME or GeoKettle. Thank you very much.
[ -0.030121056362986565, 0.005354409571737051, -0.03216998651623726, 0.02914033643901348, 0.018049899488687515, 0.035496678203344345, 0.015175958164036274, 0.02235301025211811, -0.03226852789521217, -0.025438811630010605, 0.009835640899837017, 0.01615525409579277, -0.0009212935692630708, 0.0...
[ 0.4733256697654724, 0.0792699083685875, 0.08001846075057983, 0.3871581554412842, -0.2493753284215927, -0.3662121891975403, -0.04230395331978798, 0.4386875629425049, -0.08273269981145859, -0.48757821321487427, -0.11088789254426956, 0.4699350595474243, 0.4037768244743347, 0.16928580403327942...
I was at a data visualization workshop a few days ago and someone there asked if there are any symbols/colours that might offend people when making charts, the presenter didn't really understand the answer so I helped to explain how in cartography using certain colours to represent certain groups could offend. For example using yellow to represent East Asian's, Red to represent Native Americans, or Red to represent basically any minority (since red has a connotation of being negative/bad/dangerous) The only symbols I can think of that have negative connotations are yellow six pointed stars to identify Jewish people, and pink triangles to identify gay people as these were used in Nazi Germany. I realized I only ever learned this from working with other researchers and was wondering if there is some comprehensive list out there for these types of colours/symbols.
[ 0.002289075404405594, 0.011754763312637806, -0.02058357559144497, 0.016549143940210342, -0.03794272989034653, 0.003940825816243887, 0.008521776646375656, 0.013150379993021488, -0.011295579373836517, 0.0032822364009916782, -0.007623634301126003, 0.006982225459069014, -0.008026767522096634, ...
[ 0.5552387833595276, 0.237299844622612, -0.289787620306015, 0.23471198976039886, -0.4181385338306427, 0.14191268384456635, 0.3137296736240387, 0.4651362895965576, -0.16709573566913605, -0.6412767767906189, 0.025110313668847084, -0.40250247716903687, -0.422916054725647, 0.4767163395881653, ...
It's pretty easy to track when we fix security vulnerabilities in existing code. But to make sure the whole team is staying on their toes about writing secure code, I'd like to also track how well we are preventing and avoiding writing new security vulnerabilities. What is the right way to measure that?
[ 0.031064528971910477, 0.03784717619419098, -0.021971797570586205, -0.01866455376148224, 0.017073465511202812, 0.023174509406089783, 0.01231269259005785, -0.00529205659404397, -0.024086570367217064, -0.024370914325118065, 0.008246684446930885, 0.026490138843655586, -0.02290382981300354, 0.0...
[ 0.497906357049942, 0.26974037289619446, -0.1150422990322113, 0.3785206973552704, 0.0918664038181305, -0.25322672724723816, 0.5749951601028442, 0.08887695521116257, -0.19303806126117706, -0.21664774417877197, 0.24704550206661224, 0.5856330990791321, 0.30690962076187134, -0.26444336771965027...
A year ago I bought a Galaxy ACE GT-5830i which came with Gingerbread 2.3.6, recently I wanted to root it, I tried following the instructions in Root,Unroot,CWM,and how to Flash Galaxy Ace S5830i, without success. I wiped cache, Dalvik cache, rooted again, again and again but did not work. After the root process I found the _Superuser_ icon on the launcher, I checked for root and even tried to use _link2sd_ and nothing. I tried to update the superuser app from Google play and ... nothing. How to root this device?
[ 0.004645911976695061, -0.0023854179307818413, -0.009838523343205452, 0.006265374831855297, -0.02050909399986267, 0.001240831334143877, 0.00816982239484787, -0.0005956823006272316, -0.011137280613183975, 0.0004174418281763792, -0.00806502252817154, 0.007428441196680069, -0.019700024276971817,...
[ -0.13151228427886963, 0.19260667264461517, 0.6284133791923523, 0.03648998588323593, -0.19200412929058075, 0.057515937834978104, 0.44285646080970764, 0.211430624127388, -0.48959648609161377, -0.46843987703323364, 0.05748022720217705, 0.26589059829711914, -0.3930414021015167, 0.3063420057296...
> **Possible Duplicate:** > Why doesn't PatternTest work with Composition? My function is something like this, a[x_Real?(# > 0) &] := Plot[x*Sin[y], {y, 0, 10}]; But, `a[5]` returns nothing but `a[5]`. This function doesn't work either, a[x_Real/;(# > 0) &] := Plot[x*Sin[y], {y, 0, 10}]; I'm very much confused. Can anybody tell me why these won't work?
[ -0.0035035330802202225, -0.0003812005161307752, -0.006067296024411917, 0.008833996020257473, 0.009258242323994637, 0.002350933849811554, 0.004772908985614777, -0.0026682917959988117, -0.013079321011900902, -0.00832537654787302, -0.008166145533323288, 0.0033130445517599583, -0.001233614282682...
[ 0.296058714389801, -0.06623124331235886, -0.029612962156534195, 0.06550822407007217, 0.022797489538788795, 0.08301810175180435, 0.34848934412002563, 0.008199630305171013, -0.330574095249176, -0.8993529081344604, 0.4303562343120575, 0.3125174343585968, -0.5161499977111816, 0.139954850077629...
I already understand how overtime dmg/sec works for elemental weapons (Shock/Corrosive/Fire), but if I decide to switch between two weapons of the _same_ element, such as the Heartbreaker (81.6 dmg/sec) and the Gospel (180.4 dmg/sec), both fire, and shot the _same_ enemy, would the damage stack? Replace the other?
[ -0.018327787518501282, 0.024542994797229767, -0.019264034926891327, -0.0002457794325891882, 0.025205474346876144, 0.005900425370782614, 0.010152916423976421, 0.002186050172895193, -0.019676288589835167, -0.021060384809970856, -0.000504647265188396, 0.027363717555999756, -0.010951196774840355...
[ 0.5646748542785645, 0.14088775217533112, -0.0902358740568161, 0.13795247673988342, -0.2002752125263214, 0.18951337039470673, 0.6617398858070374, -0.6912988424301147, -0.041995227336883545, -0.04774579778313637, 0.12192831188440323, 0.7783657312393188, -0.14288996160030365, -0.3488422930240...
I often need align environments where the first line is centered, e.g., when first introducing the types of the functions to be defined, then defining them. I essentially want to obtain the effect that `\multicolumn{2}{c}` has in a table, but in an align environment. The below solution is as far as I got, but the first line is then not centered. I could put them into separate environments, but this would mess with spacing in a way that I would like to avoid. \documentclass{article} \usepackage{amsmath} \begin{document} \begin{align*} \omit \rlap{$f: A \to B, g: B \to C$}\\ f(a) &= \ldots \\ g(b) &= \ldots \end{align*} \end{document} Any insight is appreciated. I don't really need the full power of a multicolumn command, but I want to be able to insert centered lines in an align environment.
[ 0.01044226624071598, 0.013057807460427284, -0.009123872965574265, 0.02118985913693905, -0.00453173229470849, -0.011574344709515572, 0.005643260665237904, 0.01649177074432373, -0.012443017214536667, 0.008324050344526768, -0.010461818426847458, 0.00994315929710865, -0.004738989751785994, 0.0...
[ -0.034873247146606445, 0.19173581898212433, 0.43182358145713806, -0.03656056523323059, 0.1674327701330185, -0.39454931020736694, -0.11403107643127441, -0.3575771450996399, -0.0399472750723362, -0.7376612424850464, -0.1354905068874359, 0.7836276888847351, -0.31557509303092957, -0.3296500146...
My android (Sony Xperia 10a) spontaneously reboots on a 1m to 15m cycle. I had a Nokia before and I think it rebooted more often than expected as well. My Google contacts, a SIM and AT&T are the only things I brought forward from one to the other so I suspect all three. Does anyone know if a bad SIM can cause rebooting?
[ 0.0003142119385302067, 0.009885990060865879, -0.014171544462442398, 0.023488378152251244, -0.028846167027950287, -0.023054536432027817, 0.007618683855980635, 0.032648950815200806, -0.019038759171962738, -0.010881848633289337, -0.01607384905219078, 0.0029533125925809145, 0.025465022772550583,...
[ 0.22568169236183167, 0.18220025300979614, 0.41475436091423035, 0.20358479022979736, -0.03930467739701271, 0.27179062366485596, 0.3092593848705292, 0.3073025345802307, -0.10968901216983795, -0.5182111859321594, 0.06302458792924881, 0.5109801888465881, -0.17463037371635437, -0.04095187038183...
I saw a great documentary last night about 'nothing'. It's about vacuums, and how if you have a total vacuum atoms will pop out of nowhere! Pretty crazy stuff. Atoms literally coming out of nowhere, then disappearing quickly. It got me thinking, before the big bang if there was nothing, then presumably atoms would be popping out of nowhere still. If enough atoms suddenly popped simultaneously into the same space could it create the big bang? The chance of this is 1 in trillions on trillions but if time is infinite then surely the chance of this happening is 1? Also if time is infinite and atoms really do pop out of nowhere, if you go back in time far enough was there ever a moment when a donkey appeared out of nowhere in a giant vacuum simply as a result of a complex pattern of atoms appearing out of nowhere? I'm a Physics noob but love thinking about all this stuff, is this something people have thought about before? And what other theories of how the big bang happened in the first place are there?
[ -0.01024394016712904, 0.017310835421085358, 0.011459734290838242, 0.003057994646951556, -0.011608615517616272, -0.02102736569941044, 0.006820966023951769, 0.005249625071883202, -0.01614432781934738, -0.006600270513445139, -0.0021273349411785603, 0.015004037879407406, 0.030839981511235237, ...
[ 0.6314646005630493, -0.2371007651090622, -0.01875125803053379, 0.273105651140213, -0.11879657953977585, -0.23695409297943115, 0.13247625529766083, 0.5529194474220276, -0.6685022711753845, -0.28176891803741455, 0.08167677372694016, 0.32947635650634766, -0.13724888861179352, 0.80044269561767...
I'm getting the Bibliography in the following format: > 1. Firstname Lastname. Article.. > What should I change to get the Bibliography in the following format: > 1. Lastname, Firstname. Article.. > This is the bibtex file I'm referring to: @ARTICLE{Clark2010, author = {Clark, Michael James and Homer, Nils and O'Connor, Brian D. and Chen, Zugen and Eskin, Ascia and Lee, Hane and Merriman, Barry and Nelson, Stanley F}, title = {U87MG decoded: the genomic sequence of a cytogenetically aberrant human cancer cell line}, journal = {PLoS genetics}, year = {2010}, volume = {6}, } And a `.tex` MWE \documentclass[11pt]{book} \usepackage{cite} \bibliographystyle{unsrt} \begin{document} Getting the Lastname, Firstname format \cite{Clark2010}\\ \bibliography{bibfile} \end{document} Thank you for your replies!
[ 0.021952293813228607, 0.00915367342531681, -0.008157800883054733, 0.019532769918441772, -0.004292056430131197, -0.0016543434467166662, 0.007891669869422913, 0.018337450921535492, -0.016838015988469124, -0.020634248852729797, -0.0012968251248821616, 0.004772603511810303, -0.010307621210813522...
[ -0.008656145073473454, 0.42854663729667664, 0.5486534833908081, 0.26623860001564026, 0.14015452563762665, 0.10239235311746597, -0.12338686734437943, 0.21352100372314453, -0.46064719557762146, -0.3623317778110504, -0.0961880087852478, 0.04929722473025322, -0.3134639263153076, 0.331028938293...
I had a user recently who reported an issue with my plugin's options page not loading. They would click on the "settings" link but only got a blank page. There was also an issue with the rich text editor showing up on the category edit screen. I could not figure out why, on almost every other site but their's, the plugins loaded fine. We disabled all plugins except mine and still could not get the plugins to load properly. Some time passed and I got an email from the user that he had resolved the issue (WP 3.1 site) by editing the file wp-admin/includes/misc.php changing... $got_rewrite = apache_mod_loaded('mod_rewrite', true); To... $got_rewrite = true; Anyone have any insights into why the unedited file might cause issues with plugins?
[ 0.015244640409946442, 0.00498502841219306, 0.003769906237721443, 0.02017507702112198, -0.026122713461518288, -0.016913391649723053, 0.005724828224629164, 0.005027255043387413, -0.014473261311650276, 0.004931514151394367, -0.015431712381541729, 0.01304652076214552, -0.0035080478992313147, 0...
[ 0.42515403032302856, 0.02125244215130806, 0.31388959288597107, -0.1596393883228302, -0.3838118314743042, -0.1570568084716797, 0.44936344027519226, -0.12418470531702042, -0.26895931363105774, -0.46679359674453735, 0.043058428913354874, 0.8114578127861023, -0.30905792117118835, 0.29315242171...
I have created a class from within which I would like to call a private method (from within the same class) from the `__construct` method as an action callback. When I would like to use a public method I can access it by: add_action( 'init', array( $this, 'action_callback' ) ); However, this causes an error when the method is private. I have also tried unsuccessfully: add_action('init', $this->action_callback() ); How do I access a private method? The class looks something like: class My_class { public function __construct() { add_action( 'init', array( $this, 'action_callback' ) ); } private function action_callback() { // do something } }
[ -0.0009003239683806896, 0.005416255444288254, -0.004218068905174732, 0.014197577722370625, 0.005894349422305822, -0.0002873600460588932, 0.008897192776203156, -0.010330390185117722, -0.01193478237837553, 0.03591492772102356, -0.009255846962332726, 0.014277268201112747, -0.008400040678679943,...
[ 0.00517435185611248, -0.0916501060128212, 0.11967746168375015, -0.06396403163671494, -0.07286684960126877, 0.20862792432308197, 0.2613210678100586, -0.2327258437871933, -0.141397163271904, -0.6069040298461914, 0.0169804859906435, 0.49400487542152405, -0.3922516405582428, 0.2150013446807861...
I have a train track loop on a map with two trains on it. Orders for both trains are identical, and they move clockwise around the loop: station A -> waypoint A -> station B -> waypoint B -> ... However one of the two trains is always parked at the depot. When the currently active train enters the depot for maintenance the other takes off. What changes do I need to make such that I can have more than one train operating on the loop simultaneously?
[ -0.024714920669794083, 0.04138705134391785, -0.008248421363532543, 0.015129227191209793, 0.0003188113041687757, 0.00019438829622231424, 0.010079044848680496, 0.012068822048604488, -0.01870548166334629, -0.006339962128549814, -0.01045498251914978, 0.02421957068145275, -0.02893776074051857, ...
[ 0.6273942589759827, 0.06331822276115417, 0.11093685775995255, 0.039018433541059494, 0.15048183500766754, 0.12410970032215118, 0.4691256880760193, -0.03045874647796154, -0.3869185745716095, -0.5870345234870911, 0.22595013678073883, 0.13255320489406586, -0.2739603519439697, 0.492869079113006...
I did correlation tests between a number of variables consisting of both interval and categorical scales ($n = 1274$, one sample). I used $\chi ^2$ test for categorical vs ordinal and Spearman's $\rho$ for interval vs ordinal data. The correlation between climate zone (zone $1, 2, 3, 4$) and daily exposure to air-conditioning (low, medium, high) is $0.35$ (Cramer's V). The correlation between exposure to AC and pro-environmental attitude score ($1-5$, interval) is $0.21$ (Spearman's $\rho$). Can I compare the strength of the relationship if I used different correlation tests, say the relationship between climate and exposure to AC is stronger than that between the pro-environmental attiudes and exposure to AC?
[ 0.006678162142634392, 0.016470788046717644, -0.03001335635781288, 0.014763940125703812, 0.029583685100078583, -0.014254805631935596, 0.011237050406634808, -0.03233237564563751, -0.0103914150968194, -0.015283973887562752, -0.011296230368316174, 0.010214678943157196, -0.017877016216516495, 0...
[ 0.691325843334198, 0.05763065069913864, 0.2020866423845291, 0.08805302530527115, 0.1780424565076828, 0.15558166801929474, 0.45164015889167786, -0.4072149097919464, -0.006335534155368805, -0.1545826941728592, 0.012727375142276287, 0.5204965472221375, -0.16822339594364166, 0.1414437890052795...
> **Possible Duplicate:** > “Can/may/will you help me with this?” Here is the sentence: If you like you {can|may} translate this text to English. What word should I use here and why.
[ -0.0025737301912158728, 0.02732507325708866, 0.000497152388561517, 0.03463270515203476, -0.022878626361489296, 0.001747358008287847, 0.012402149848639965, 0.0021402521524578333, -0.02484317310154438, 0.03658038750290871, -0.02130923606455326, 0.019212622195482254, 0.011791178956627846, 0.0...
[ 0.07692320644855499, 0.15132074058055878, 0.27337244153022766, 0.18335632979869843, 0.1827118992805481, 0.1473434865474701, 0.429934024810791, 0.3201163709163666, 0.20596377551555634, -0.8193030953407288, -0.05934157222509384, 0.39887964725494385, 0.05383579060435295, -0.21492739021778107,...
When I first launched my site, it I had MediaWiki in the root and a wordpress blog in a subfolder, /blog/. After a while spammers took over the MediaWiki site and I had to remove it. The problem is I lost a lot of good content and inbound links from 3rd party sites. My goal is to redirect the root folder to /blog/ while keeping the rest of the URL string intact. I have a Wordpress plug-in going that will automatically track the 404 errors. As these come in I'll be able to pull the old content from archive.org and give it a new home on a wordpress page.
[ 0.0053405859507620335, 0.012933621183037758, -0.0006519719609059393, 0.015440557152032852, -0.017419399693608284, 0.008851509541273117, 0.008300096727907658, 0.0001490800641477108, -0.01412127073854208, -0.021254470571875572, -0.008643945679068565, 0.010670943185687065, -0.017953194677829742...
[ 0.3332795798778534, 0.1144765168428421, 0.5543238520622253, 0.11501701176166534, -0.20942558348178864, -0.140328511595726, 0.39478787779808044, 0.3701101541519165, -0.14404511451721191, -0.5083416700363159, 0.020274050533771515, 0.33464255928993225, -0.31886377930641174, 0.5943093299865723...
I recently turned my Raspberry Pi into a web server, and now I installed Wordpress on it. Everything is going excellently, but when I went to change the theme from default, I found I need my hostname, FTP username, and FTP password. I didn't see this as a major problem, so since I host my webpage on my 'Pi I put in "localhost" and my MySQL username/password for the server I set up. It popped up an error saying it failed to connect to the FTP server. Since I am still fairly new to Wordpress and webhosting, could someone please tell me what I _should_ have used?
[ 0.012048608623445034, -0.008461802266538143, 0.011527027934789658, 0.016156034544110298, -0.04739130288362503, 0.0051819379441440105, 0.010642483830451965, 0.02934299409389496, -0.01481305155903101, -0.03420284390449524, -0.011921495199203491, 0.013322724029421806, 0.0028552557341754436, 0...
[ -0.13179811835289001, 0.2928341329097748, 0.7780545949935913, -0.16494889557361603, -0.011816689744591713, 0.021107368171215057, 0.35362979769706726, 0.6017036437988281, -0.20309002697467804, -0.8335371017456055, 0.12566952407360077, 0.42915651202201843, -0.25584372878074646, 0.36575129628...
Anyone could give me the whole Lasso Algorithm? The step by step procedure in determining Lasso solutions? I barely need this for my thesis. Thank You.
[ 0.05005738139152527, 0.038022302091121674, -0.015895728021860123, 0.034989405423402786, -0.010297060012817383, 0.021900895982980728, 0.014891917817294598, 0.015065700747072697, -0.06613641232252121, -0.018485965207219124, 0.010049528442323208, 0.01114105898886919, -0.03328874334692955, 0.0...
[ 0.24026618897914886, 0.3193155527114868, 0.09975975751876831, 0.4764842689037323, 0.07264722883701324, -0.383514940738678, 0.20434990525245667, 0.18877938389778137, -0.07221106439828873, -0.3996604084968567, 0.3678569197654724, 0.08025093376636505, 0.3786845803260803, 0.15961682796478271, ...
I am stacking one header by using this macro: `\newcommand*{\textstackrel}[2]{\ensuremath{\stackrel{\hbox{#1}}{\hbox{#2}}}}`. This makes the other single line headers bottom aligned, which looks bad. So I tried a few multirow tricks I know to no avail, e.g.: `assign column name/.code={\multicolumn{1}{c}{\multirow{2}{*}{#1}}}` Most of the time I ran into this error message: ! siunitx error: "duplicate-exponent-token" ! Duplicate exponent marker token '\token_to_str:N d' in input. How do I do this correctly? PS: MWE will follow when I have (more) time and if there is no answer before that.
[ 0.013538971543312073, 0.012405894696712494, -0.007803607266396284, 0.019429730251431465, 0.004217206034809351, 0.00002332986332476139, 0.007316478528082371, -0.0013817029539495707, -0.010859908536076546, 0.007557007949799299, -0.0004744025645777583, -0.0021231211721897125, 0.0042351172305643...
[ 0.2619820237159729, 0.12496326118707657, 0.4672892987728119, -0.15847070515155792, 0.029747046530246735, -0.18273906409740448, 0.02326023392379284, -0.23781780898571014, -0.34190815687179565, -0.5368495583534241, 0.10633765906095505, 0.3506101071834564, -0.5743494033813477, 0.0786313340067...
In the first function below, inside the foreach, I'm calling out to the 2nd function in order to test for the occurence of matching category ids that I want to remove from the category listing. However, the way I'm doing it in the 2nd function seems to me to be pretty crude, for lack of a better term. How could I improve this lookup? function admin_trim_category_description( $terms, $taxonomies ) { if( 'category' != $taxonomies[0] )return $terms; foreach( $terms as $key=>$term ) { $terms[$key]->description = strip_tags(substr( $term->description, 0, 75 ))."..."; if(ce4_get_utility_cats($terms[$key]->term_id)) { unset($terms[$key]); } } return $terms; } function ce4_get_utility_cats($cat_id) { if($cat_id == get_cat_ID('category1') OR $cat_id == get_cat_ID('hidden') OR $cat_id == get_cat_ID('category2') OR $cat_id == get_cat_ID('category3')) { return true; } else { return false; } } EDIT: The above functions are called in the following manner... add_action( 'admin_head-edit-tags.php', 'admin_edit_tags' ); function admin_edit_tags(){ add_filter( 'get_terms', 'admin_trim_category_description', 10, 2 ); }
[ 0.00784820131957531, 0.01747247204184532, 0.00018491479568183422, 0.02274361625313759, -0.0029641627334058285, 0.0047484394162893295, 0.0061089228838682175, 0.0174739807844162, -0.018325544893741608, 0.022239284589886665, -0.018680516630411148, 0.003915225155651569, -0.01423089113086462, 0...
[ -0.1878768652677536, -0.18717609345912933, 0.29576918482780457, -0.15445822477340698, -0.026852348819375038, -0.1685706228017807, 0.16450382769107819, -0.31550008058547974, -0.025870468467473984, -0.6958239078521729, -0.1669873744249344, 0.9491442441940308, -0.4793540835380554, 0.344032973...
I'm using OGR to perform a union using UnionCascaded. All my polygons have been created by GDAL polygonize which ensures that all polygons follow pixel edges and should therefore align nicely. But I get these strange flat polygons looking like lines: ![Example of union result with strange flat polygons](http://i.stack.imgur.com/8G4eu.jpg) They are positioned at locations where two polygons met before they where unioned. **Can you recognise this pattern and what lead to it?**
[ -0.00894198752939701, 0.01915811002254486, 0.00032925716368481517, 0.02820564992725849, -0.014984317123889923, -0.01282979641109705, 0.007906407117843628, -0.004602177068591118, -0.012623581103980541, 0.018194008618593216, -0.008883623406291008, 0.005827505607157946, -0.0023581390269100666, ...
[ 0.24837584793567657, 0.08672858029603958, 0.3307424485683441, 0.01405305229127407, -0.23030677437782288, 0.2979174554347992, -0.15586711466312408, -0.024931423366069794, -0.33221060037612915, -0.7954756021499634, 0.30261942744255066, 0.1861262023448944, -0.36224693059921265, 0.374134689569...
I'm confused by the new customization system in COD: Ghosts. Do I have to create a new soldier for every customization i want? How does it compare to the "Pick 10" system from Black ops 2?
[ -0.0038689905777573586, 0.021617278456687927, -0.002364781219512224, 0.009468194097280502, 0.006342080421745777, -0.03125810623168945, 0.013358400203287601, -0.027384132146835327, -0.03410767391324043, -0.016545608639717102, -0.027849534526467323, 0.04487978294491768, 0.02583944797515869, ...
[ 0.2195386290550232, 0.20773909986019135, 0.32082754373550415, -0.03273481875658035, -0.16604328155517578, 0.11856039613485336, 0.09305693209171295, -0.03762279078364372, -0.6987003684043884, -0.6924721598625183, 0.23346973955631256, 0.34000980854034424, 0.1671692281961441, 0.24711114168167...
> **Possible Duplicate:** > Android USB reverse tethering: How to fool the apps I use USB tethering to connect my HTC Hero to a computer. And the cellphone can surf the Internet through the computer. But some programs cannot find the network connection. I can ping any IP address on the Internet in ConnectBot, which means the connection doesn't have a problem. I think it's due to some options not being toggled that makes these stupid programs not being able to find the connection. More specifically, the notice is: > Connection failed > > This application requires network access. Enable mobile network or Wi-Fi to > download data. How can I make these programs use the USB connection rather than mobile network or Wi-Fi?
[ -0.004091163165867329, 0.0046036457642912865, -0.0029577347449958324, -0.0017111576162278652, -0.02994752675294876, -0.014800770208239555, 0.008543277159333229, -0.00925956666469574, -0.012113882228732109, -0.027318209409713745, -0.01055135764181614, 0.004655902739614248, -0.0021575952414423...
[ 0.23359006643295288, 0.03615890070796013, 0.7117004990577698, 0.4255896210670471, 0.04622123762965202, -0.027489958330988884, 0.48980242013931274, -0.11843102425336838, -0.38547155261039734, -0.3970370888710022, 0.16380850970745087, 0.38526278734207153, -0.2692333161830902, -0.005443115718...
Are there any custom maps for BFBC2? If so, how do you play them?
[ 0.011356912553310394, 0.03993793949484825, -0.0032566236332058907, -0.003511584596708417, 0.10161715745925903, -0.04815569892525673, 0.01529684104025364, 0.009885953739285469, -0.030060240998864174, 0.033604059368371964, 0.0006464010220952332, 0.05578942596912384, -0.0269466582685709, -0.0...
[ 0.23652219772338867, 0.08341507613658905, 0.19331534206867218, -0.05169984698295593, 0.42034393548965454, 0.03315655514597893, -0.2727779746055603, 0.15067800879478455, -0.5506608486175537, -0.3854498863220215, 0.23136313259601593, 0.5428450703620911, -0.07654083520174026, -0.3635458648204...
When I started with a fresh SD card, there were only very few folders in it. As time passed by when I tried many apps, I could see the folder list in my SD card seems to grow. It is understandable that while an app is being uninstalled, it leaves behind its folder where it would have kept few files, which may be the settings file or game scores etc. Is there any way I can remove all those orphaned folders in a batch or short operations? I have seen the question "How do one know which folder is created by which android apps" and can manually do it, but I am interested in kind of automatic solution. My phone info: LG Optimus One (P500) running stock 2.3.3 V20D (rooted).
[ -0.001194787910208106, 0.013256467878818512, 0.01305217482149601, 0.01869431883096695, 0.011960112489759922, -0.005124474409967661, 0.007065111771225929, 0.022340189665555954, -0.013897843658924103, -0.0028876238502562046, -0.013412034139037132, 0.017385873943567276, 0.004098539240658283, ...
[ 0.470535010099411, 0.17265856266021729, -0.05449344590306282, 0.3147936165332794, 0.07843177765607834, 0.0774383693933487, 0.05068320035934448, -0.13361909985542297, -0.3892889618873596, -0.6143953204154968, 0.2006969153881073, 0.6536911129951477, -0.3012084364891052, 0.06687425822019577, ...
All, I have a feature class in ArcCatalog. I need to save the feature class to a Map and then publish the map to the Arcserver. All programatically and hopefully in C#. I can't seem to find any examples at ESRI or by googling. Any help and examples of this would be absolutely amazing. But just leading me down the right path is cool too. Thanks guys!
[ -0.012307778000831604, -0.00025847743381746113, -0.010614725761115551, 0.02370188571512699, 0.01859717257320881, 0.0040966360829770565, 0.008985198102891445, 0.02959216944873333, -0.029425717890262604, -0.030460063368082047, 0.01272693183273077, 0.014370488002896309, -0.004656573757529259, ...
[ 0.3907097280025482, 0.23907364904880524, 0.2959985136985779, 0.4484516978263855, -0.36804550886154175, -0.44399294257164, 0.5166215300559998, 0.43471062183380127, 0.003130848053842783, -0.6630266308784485, -0.13739906251430511, 0.8120002150535583, 0.4285920262336731, -0.1875731498003006, ...
If the words _computation_ and _calculation_ are not perfect synonyms what is the difference between them? Which one describes more accurately what is done by a person computing or calculating something on a piece of paper?
[ -0.011692211031913757, 0.02823370322585106, 0.0027102946769446135, 0.02002217248082161, -0.08111629635095596, -0.012607353739440441, 0.018154747784137726, 0.016348207369446754, -0.01840738207101822, -0.04296636953949928, -0.00946300383657217, 0.007072495296597481, 0.05257076397538185, -0.0...
[ 0.4944610595703125, 0.36573895812034607, -0.18961307406425476, 0.21128183603286743, -0.16533039510250092, 0.02665601298213005, -0.07065137475728989, -0.07789136469364166, -0.3168284595012665, -0.40845349431037903, 0.29716116189956665, 0.2618832290172577, 0.14435362815856934, 0.227669402956...