text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
Very frequently, I am stuck when choosing the best design decision. Even for small details, such as function definitions, control flow, and variable names, I spend unusually long periods perusing the benefits and trade-offs of my choices. I feel like I am losing a lot of efficiency by spending my hours on insignificant details like these. Even though, I know in the back of my mind that I can change these things if my current design doesn't work out, I have trouble deciding firmly on one choice. What should I do to combat this problem?
[ 0.030126960948109627, 0.018021322786808014, -0.02727978117763996, 0.009345809929072857, -0.0252668596804142, -0.017315082252025604, 0.0057542091235518456, -0.00172223593108356, -0.01407280657440424, 0.028500549495220184, -0.009152515791356564, 0.007019855547696352, -0.026057301089167595, 0...
[ 0.43768447637557983, 0.4427877366542816, -0.05404438450932503, 0.16857224702835083, -0.021447965875267982, 0.20683011412620544, 0.3111109137535095, 0.2572126090526581, -0.26435795426368713, -0.8068851232528687, -0.008516756817698479, 0.5355433225631714, 0.11902999132871628, 0.6105183362960...
I have seen several people use the word _backslash_ , it sounds like something similar to disaster, bad result, etc. But I am not quite sure what it means and when I should use it. Examples: > * But if big layoffs result at home, corporations and the US government > will face a backslash. > * Celeb Boutique saw a major backslash as result after the inappropriate > usage of the word _Aurora_. >
[ -0.002710264176130295, 0.011237232945859432, -0.008445380255579948, 0.03653787076473236, -0.001159613486379385, -0.017991259694099426, 0.008333676494657993, -0.0033328686840832233, -0.014362521469593048, -0.0211782269179821, -0.005822519306093454, 0.0057258219458162785, 0.020170805975794792,...
[ 0.5999576449394226, -0.11553658545017242, 0.1363242119550705, 0.015240561217069626, -0.17422938346862793, -0.017482098191976547, 0.21228185296058655, 0.6702534556388855, -0.6069796085357666, -0.12031560391187668, 0.012519482523202896, 0.03420073911547661, -0.016082925722002983, 0.503312170...
I would be thankful if anyone can guide me in writing commands for a fibre product in LaTeX. I would also like to know how to change the angle of the diagram as I want it to be.
[ 0.017429469153285027, 0.019165270030498505, -0.044889189302921295, 0.03523740544915199, -0.035219889134168625, 0.021909039467573166, 0.012803391553461552, 0.017407895997166634, -0.03836730867624283, -0.000707933446392417, -0.0035374746657907963, -0.003160045947879553, -0.004274039063602686, ...
[ 0.6975117325782776, -0.026842745020985603, 0.39268258213996887, -0.17863279581069946, 0.04844704270362854, 0.2314249873161316, 0.016069071367383003, 0.38479501008987427, -0.603762686252594, -0.870502233505249, 0.3512810468673706, 0.38410380482673645, 0.3014795184135437, 0.05520208925008774...
**First of all:** From what I understood, bootstrapping residuals works as follows: 1. Fit model to data 2. Calculate the residuals 3. Resample the residuals and add them to 1. 4. Fit model to new dataset from 3. 5. Repeat `n` times, but always add the resampled residuals to the fit from 1. Is that correct so far? * * * **What I want to do** is something slightly different: I want to estimate parameter and prediction uncertainty for an algorithm that estimates some environmental variable. What I have is a error-free time-series (from a simulation) of that variable, `x_true`, to which I add some noise, `x_noise`, in order to generate a synthetic dataset `x`. I then try to find optimal parameters by fitting my algorithm with the sum of squares `sum((x_estimate - x_true)^2)` (! not `x_estimate - x` !) as an objective function. In order to see how my algorithm performs and to create samples of my parameters' distributions, I want to resample `x_noise`, add it to `x_true`, fit my model again, rinse and repeat. Is that a valid approach to assess parameter uncertainty? Can I interpret the fits to the bootstrapped datasets as prediction uncertainty, or do I have to follow the procedure I posted above? Sorry if I'm not very clear with terminology, English isn't my first language and I'm pretty new to this. /edit: I think I haven't really made clear what my model does. Think of it as essentially something like a de-noising method. It's not a predictive model, it's an algorithm that tries to extract the underlying signal of a noisy time- series of environmental data. /edit^2: **For the MATLAB-Users** out there, I wrote down some quick & dirty linear regression example of what I mean. This is what I believe "ordinary" bootstrapping of residuals is (please correct me if I'm wrong): http://pastebin.com/C0CJp3d1 This is what I want to do: http://pastebin.com/mbapsz4c
[ 0.020639672875404358, 0.01973903737962246, -0.00728572066873312, 0.009147631004452705, 0.005879602394998074, -0.00852404534816742, 0.006697231903672218, -0.000740406452678144, -0.013234935700893402, -0.0005805781111121178, 0.001300922129303217, 0.009203784167766571, -0.01782320626080036, -...
[ 0.588092565536499, -0.03722730278968811, 0.1916777640581131, 0.12086635082960129, -0.06603974103927612, -0.08464697003364563, 0.27647244930267334, -0.5504105687141418, -0.269065797328949, -0.6452296376228333, 0.3994199335575104, 0.8079609870910645, -0.2739940583705902, 0.0711294636130333, ...
How can I save the output of the function `the_content_rss` into a string variable?
[ -0.003378205467015505, 0.02416876144707203, -0.011319637298583984, 0.02094579115509987, -0.05919799953699112, 0.02960919588804245, 0.019870657473802567, 0.014688984490931034, -0.023480882868170738, -0.045862071216106415, -0.031093068420886993, 0.020938048139214516, -0.029705027118325233, 0...
[ 0.27652114629745483, 0.03446558117866516, 0.03557779639959335, 0.03867039084434509, 0.1995316594839096, -0.18856661021709442, 0.08051788061857224, -0.06717292964458466, 0.01383240520954132, -0.22361576557159424, 0.09638771414756775, 0.3177538514137268, -0.4676186442375183, 0.17720922827720...
Imagine a sphere of uniform density with similar volume and average density of our Earth. There is a bore leading to the center of the sphere from the surface with a scale at regular intervals. At what part of your journey towards the center would the scale read your greatest weight? There are some caveats to consider. The sphere is in a very isolated region of space and has no angular motion. * * * I wrote this question down last night while I was having trouble sleeping. My intuition tells me that the surface of the sphere will have the greatest force of gravity. Any distance beyond the surface and the mass of the sphere above you will be reducing your weight. Is this correct? * * * If however the core is much more massive similar to Earth, one would feel a greater strain to stand upright as they descend.
[ -0.017294690012931824, 0.002279178239405155, -0.0028777350671589375, 0.012602286413311958, -0.018395371735095978, -0.006240102928131819, 0.005088423378765583, -0.01248311996459961, -0.013919605873525143, -0.020332802087068558, 0.001573042361997068, 0.019229844212532043, -0.012661203742027283...
[ 0.30195876955986023, 0.13168680667877197, 0.5368332266807556, 0.0532342903316021, 0.04329534247517586, 0.26539739966392517, 0.08995462954044342, -0.1488838791847229, -0.3525146245956421, -0.7654332518577576, 0.21057787537574768, 0.11444366723299026, 0.25665196776390076, 0.36599427461624146...
I know there are ways I can copy these files on my own and avoid this question completely, but I'm curious for the less tech saavy... If I were to uninstall Mass Effect 2 would I lose my saves? Is there a dialog box on uninstallation that gives me an option? Does it not mention it at all and just leaves it in the directory where they were originally while deleting other game content? Does Steam save any of this info in case of a total fail on my SSD's part. I'm not taking any chances on this considering the hours I put in on 2... note: my question pertains specifically to the PC title.
[ 0.003208132926374674, 0.012345915660262108, 0.01578143984079361, 0.015974514186382294, 0.016315294429659843, 0.0012118405429646373, 0.005836073309183121, -0.008213726803660393, -0.014267412945628166, 0.006153757218271494, -0.0042317211627960205, 0.015237619169056416, 0.004197381436824799, ...
[ 0.6081374287605286, -0.04582489654421806, 0.14727811515331268, 0.3421007990837097, -0.1338822841644287, -0.2271701991558075, 0.1343139111995697, 0.14189618825912476, -0.24700585007667542, -0.558722734451294, 0.19640174508094788, 0.32719942927360535, -0.10615407675504684, 0.301292359828949,...
I'm building out a custom taxonomy archive page that needs to list terms in order by their associated post's `post_date`. I wrote this, it's working, but I hate resorting to custom SQL: global $wpdb; $post_types = array( 'series' ); $taxonomy = array( 'taxonomy' ); $order = 'DESC'; $sql = "SELECT DISTINCT t.* FROM $wpdb->terms AS t INNER JOIN $wpdb->term_taxonomy AS tt ON t.term_id = tt.term_id JOIN $wpdb->term_relationships AS tr ON tr.term_taxonomy_id = tt.term_taxonomy_id JOIN $wpdb->posts AS p ON p.ID = tr.object_id WHERE tt.taxonomy IN ( '" . implode( "','", $taxonomy ) . "' ) AND p.post_type IN ( '" . implode( "','", $post_types ) . "' ) ORDER BY p.post_date {$order};"; $terms_ordered_by_post_date = $wpdb->get_results( $sql ); Curious if anyone has a better solution?
[ 0.009758898988366127, -0.0015677884221076965, -0.0100106131285429, 0.024196555837988853, 0.009557897225022316, 0.016482895240187645, 0.008675677701830864, 0.0359831303358078, -0.015509871765971184, 0.005596164148300886, -0.013009465299546719, 0.004097197204828262, -0.009457908570766449, 0....
[ 0.10450877994298935, -0.10806067287921906, 0.6938996315002441, -0.005585071165114641, 0.035591237246990204, -0.06213325262069702, 0.13924111425876617, -0.09943830966949463, -0.1400405764579773, -0.606924295425415, 0.01457949634641409, 0.43739065527915955, -0.010123376734554768, 0.677096545...
I have a simple equation, much like the following (I use the `braket` package) : \begin{equation} \mathcal{Y} = \Set{ x \in \mathcal{G} | \Vert x \Vert > d } \end{equation} My gripe with this equation's rendering is that the | and || are close to eachother and of the exact same height - I'd like the | bar to be bigger to make the symbols more distinct. One way to do this would be to make the whole `Set`, or left-right environment, a bit taller. How can I do this? Can I insert an invisible character that will make the whole thing bigger? Should I insert a higher | bar manually?
[ 0.008993430994451046, 0.008409890346229076, -0.007995945401489735, 0.004090355709195137, 0.0004711211659014225, 0.007863868959248066, 0.006072808988392353, -0.00430644303560257, -0.010780688375234604, -0.030217397958040237, -0.002725201426073909, 0.004058041609823704, -0.014117314480245113, ...
[ 0.11743103712797165, 0.24395087361335754, 0.5578171014785767, -0.338492751121521, -0.0839841365814209, 0.22244328260421753, 0.18215957283973694, -0.32613977789878845, -0.08018699288368225, -0.669614851474762, 0.20150360465049744, 0.5618453025817871, 0.05178495869040489, -0.1752915531396865...
I don't mean the set of smileys we use today which are surely a result of the keys on a keyboard. What I do mean is: In casual conversation before the creation of the personal computer, did people who wrote letters or notes have the ability to attach an emotion to things that they wrote in a simple way? For example, today I might write: > Don't worry about it :) which establishes very many things in very concise way. Did the writers of the past have some mechanism for doing this, or did they write out everything?
[ -0.011602744460105896, 0.004227885976433754, -0.002825220348313451, 0.009539618156850338, 0.0015462345909327269, -0.014786173589527607, 0.006039791740477085, -0.009026345796883106, -0.016433430835604668, -0.005243879742920399, -0.007781419903039932, 0.010267958976328373, 0.026926644146442413...
[ 0.2477491945028305, 0.10379835218191147, -0.08776126056909561, 0.1389903724193573, -0.2802128195762634, 0.06311354041099548, 0.3499389588832855, 0.3986705541610718, -0.2844509482383728, -0.5736181735992432, 0.21162691712379456, 0.3258915841579437, -0.24722681939601898, -0.03545315191149711...
I'm thinking of using multiple web servers with a load balancer in a big WP install. The database will be on a separate server and that's the easy part. Trickier is files in blogs.dir, and to a lesser extent, the themes and plugins folders. So I plan to use a separate NFS server, either for all of wp_content or just for blogs.dir. But I'm worried about the performance impact of having these files available over the local network. If you have done this I would love to hear your thoughts on... * Is it wise to share all of wp-content in this way? If not, how about just blogs.dir? * How much of a performance trade-off was it? * Do you have any suggestions based on your experience? Thanks.
[ 0.009317316114902496, 0.01654129847884178, 0.0024782787077128887, 0.02706276625394821, 0.006193429231643677, -0.00279123499058187, 0.007692362647503614, 0.00567603949457407, -0.014605456031858921, -0.020662788301706314, -0.011500249616801739, 0.01625748723745346, -0.0051058693788945675, 0....
[ 0.7013998627662659, -0.013726787641644478, 0.2882918417453766, 0.10927114635705948, -0.09619962424039841, -0.20022563636302948, 0.07557892799377441, 0.04752092808485031, -0.5134074091911316, -0.8402098417282104, 0.2942523658275604, 0.4309576749801636, -0.16684234142303467, 0.22549875080585...
When did these design patterns originate? Balking, Builder, Delegation, Facade, Memento. I have looked for days across the net, so if someone points me to a simple google search I may shoot myself. The real question I have is only which is the oldest and which is the youngest, so the specific dates for each are not important. I have already found that the Balking pattern originated in 2002.
[ 0.012323617935180664, 0.02062414027750492, -0.018720529973506927, 0.0061723836697638035, 0.01520464662462473, 0.0233469195663929, 0.009232649579644203, 0.016530167311429977, -0.01899670995771885, 0.019116492941975594, -0.018267899751663208, 0.022693946957588196, 0.016225457191467285, -0.00...
[ 0.5524015426635742, 0.10851293802261353, 0.16577425599098206, -0.11620771884918213, 0.0015192891005426645, 0.23669832944869995, 0.043677106499671936, 0.23507849872112274, -0.5653405785560608, -0.6125515103340149, 0.3755456507205963, -0.5145771503448486, -0.3785831332206726, 0.7838583588600...
If I register a really good .com domain name, should I also register: * alternate spellings * mis-spellings * hyphenated versions * other variations: iExample.com, eExample.com, myExample.com, etc. * other TLDs: .org, .net, .biz, .info * international TLDs Which definitely, which maybe and which no? How do you know where to draw the line?
[ 0.013294353149831295, 0.0060143922455608845, -0.0023810029961168766, 0.015403343364596367, 0.0021507905330508947, 0.037915777415037155, 0.011810027062892914, 0.02364814467728138, -0.01677691750228405, -0.013841607607901096, -0.0031283951830118895, -0.006096949800848961, 0.004717643838375807,...
[ 0.7809687852859497, -0.23088930547237396, 0.4880037307739258, 0.13776050508022308, 0.012352915480732918, -0.4143497347831726, 0.30231189727783203, 0.6256311535835266, 0.01706617884337902, -0.5080753564834595, -0.05752754583954811, 0.018803460523486137, -0.16194725036621094, 0.3028282225131...
In Grand Theft Auto V, you can buy and sell shares on the Stock Market via the in-game web browser. I have some questions about this. * Is the rise and fall of any of the stocks tied to the game's storyline? (if so, which/when?) * Is there anything that the characters can do in the game that influences the price of stock? (what?) (I realise that asking multiple questions in one post is discouraged, but I think these are related enough to be a single question. Please let me know if you disagree and I'll post them separately.)
[ 0.0017523791175335646, 0.007670809049159288, 0.008930528536438942, 0.0004369435482658446, 0.014799261465668678, 0.0038730017840862274, 0.0069139013066887856, 0.013179842382669449, -0.015401458367705345, -0.019252195954322815, -0.019214075058698654, 0.019335174933075905, -0.001768759102560579...
[ 0.4719025492668152, -0.27071166038513184, 0.3741711974143982, 0.257638156414032, -0.10684439539909363, -0.02715982496738434, -0.12902115285396576, 0.3252467215061188, -0.16733597218990326, -0.09747287631034851, -0.0153583362698555, 0.8248486518859863, -0.16421297192573547, -0.0270919837057...
I originally bought the domain (www.iconicphotos.co.uk) a while ago, I then decided to try and sell it on Sedo. The domain was up for sale for a few months before I noticed that the ads being pulled through on the parked domain were porn based (Not happy about this!). Anyhow recently created a gallery site to sell my artwork from but I am finding now that the site isn't being indexed at all! I had a couple of pages appear a few days ago but for only a couple of hours before they were removed. If you try the non www version of the domain using (site) there are a couple of pages but they are they show the old parked domain info. I have created a WMT account and updated it with a new sitemap. I have also submitted the site again using the fetch as google bot option. Not sure what to do now?
[ -0.004502701107412577, 0.007981601171195507, -0.0016130651347339153, 0.020768098533153534, -0.0045541031286120415, 0.001651096623390913, 0.006007447373121977, 0.01999645121395588, -0.013310045003890991, 0.02038143388926983, -0.0199725404381752, 0.009835179895162582, -0.0013926965184509754, ...
[ 0.7758994698524475, -0.2654697299003601, 0.5709643959999084, -0.18638014793395996, -0.23725394904613495, 0.18171480298042297, -0.10079842060804367, 0.6127135753631592, -0.5023414492607117, -0.04763328284025192, 0.6422755122184753, -0.0018750907620415092, -0.04445059224963188, 0.40468657016...
The following quotes are from the Wikipedia article. It seems to me that they all use "they" for generic person. For example, in the Chesterfield's example: "If a person is born of a . . . gloomy temper . . . they cannot help it.", "a person" appears to be singular but it represents any person. It is essentially plural. 'Tis meet that some more audience than a mother, since nature makes them partial, should o'erhear the speech."— Shakespeare, Hamlet (1599); "If a person is born of a . . . gloomy temper . . . they cannot help it."— Chesterfield, Letter to his son (1759); "Now nobody does anything well that they cannot help doing"— Ruskin, The Crown of Wild Olive (1866); "Nobody in their senses would give sixpence on the strength of a promissory note of the kind."— Bagehot, The Liberal Magazine (1910); "I would have every body marry if they can do it properly."— Austen, Mansfield Park (1814); Caesar: "No, Cleopatra. No man goes to battle to be killed." Cleopatra: "But they do get killed" —Shaw, Caesar and Cleopatra (1901); "A person can't help their birth."— W. M. Thackeray, Vanity Fair (1848); "When in the Course of human events, it becomes necessary for one people to dissolve the political bands which have connected them with another . . ." —United States Declaration of Independence; **My question** Is the use of singular they in the following passage grammatically correct? Someone was approaching my room. I could see that they were alone judging from their footsteps. They knocked on my door. I didn't answer. They knocked again. I still didn't answer so they left.
[ 0.0026783409994095564, 0.0007630152394995093, -0.014768211171030998, 0.014430535957217216, -0.008618942461907864, -0.006301420740783215, 0.005390774458646774, -0.00014893594197928905, -0.007927833124995232, -0.008610119111835957, 0.006544961128383875, 0.006640721578150988, 0.0175316166132688...
[ -0.22085747122764587, 0.12130947411060333, -0.16382558643817902, -0.46723473072052, -0.26283013820648193, 0.2961418032646179, 0.8368065357208252, 0.25506770610809326, 0.0022609608713537455, -0.24422402679920197, -0.1638568788766861, 0.32305458188056946, -0.22989022731781006, 0.525229752063...
I'm trying to add a 'postinst' script to my debian package to clean up something a previous version may have left behind. I've simply created a 'postinst' file in the debian/ directory, but the file is not present in the .deb file. The permissions: -rwxrwxr-x 1 kp kp 354 Sep 16 20:00 postinst* The content: #!/bin/sh set -e set -x ...
[ -0.0017717017326503992, 0.004543909803032875, -0.008633892051875591, 0.016740333288908005, 0.026493707671761513, 0.008318983018398285, 0.008438043296337128, 0.016644654795527458, -0.016153980046510696, -0.024813182651996613, -0.008842683397233486, 0.0017548721516504884, -0.018604811280965805...
[ 0.6124051809310913, 0.2501458525657654, 0.4933844208717346, -0.14116545021533966, 0.009952686727046967, -0.4171323776245117, 0.5877354741096497, -0.10096859931945801, 0.009400409646332264, -0.6655819416046143, -0.2651398181915283, 0.2020973563194275, -0.5237729549407959, 0.2552591264247894...
I have several shp files which are the components (or tiles) of a map, I have joint all these tiles together to have a large shp file. Then I realized some of the polygon is now splitted into 2 halves, as the upper part belngs to a tile, while the lower part belongs to another tile. This output disturb me a lot as I need to fill some polygon, but not part of the "splitted polygon", so I need to find a way to merge these splitted polygon back to one (preferrably mechanical way, as I have 10000+ polygons, not knowing which are splitted, which are not). I need to know if the two or more polygons orginate from the larger polygon. And I think I can some how use union and spRbind to create an algorithm for this task. But I doubt if I am really the first one who encounter this problem, so I wonder if there is any existing tool in R that can solve this problem. Thanks.
[ -0.011738887056708336, 0.01445366907864809, -0.001647253055125475, 0.022261343896389008, 0.0023286747746169567, 0.013445548713207245, 0.008246384561061859, 0.014481221325695515, -0.01850138232111931, 0.020918672904372215, -0.010294348001480103, 0.012201606296002865, -0.015576634556055069, ...
[ 0.29985150694847107, -0.03170099854469299, 0.3990683853626251, 0.12080352753400803, -0.2635550796985626, 0.1827000230550766, 0.04896009340882301, -0.2956452965736389, -0.4574962258338928, -1.0996583700180054, 0.19134998321533203, 0.1509876400232315, -0.041127387434244156, 0.432896256446838...
I would like to dump some minds map in latex. I have them following this structure: ![http://www.wikilearning.com/imagescc/sinop2.jpg](http://i.stack.imgur.com/gpcMV.jpg) Although it does not need need to be exactly like in that chart, just some nice way of drawing hierarchical outline summaries. Thanks.
[ -0.004952423740178347, 0.014914472587406635, -0.014170296490192413, 0.01882990077137947, 0.011288358829915524, 0.007913068868219852, 0.006398119498044252, -0.002913658507168293, -0.02317008003592491, -0.008282065391540527, -0.0036589791998267174, 0.00021426595048978925, -0.007512437179684639...
[ 0.30884334444999695, 0.15426881611347198, 0.14752084016799927, 0.14832505583763123, -0.28513476252555847, -0.12004721164703369, 0.058010563254356384, 0.09943537414073944, -0.16204161942005157, -0.7851790189743042, 0.3484104573726654, 0.28227394819259644, 0.13859978318214417, 0.077460110187...
Now that I've gotten Will Robertson's excellent unicode-math package working to change the colour and style of letters in my mathematics (trust me, there _is_ a reason) thanks to his answer to my previous question, I want to figure out how to implement this properly. The key step seems to be that the command to change how a letter is displayed needs the unicode numerical value of the letter. The example that Will gave was: \setmathfont[range="66,math-style=upright,Colour=FF0000]{xits-math.otf} (font name changed to work with xelatex as per Will's comment on that question) which changes the letter `e`. But I want a command that I can invoke as `\type{constR}{e}` (which should declare `e` to be a constant real number). So I need to convert `e` to its unicode value. Now it seems possible that there isn't a general "charater to unicode" command, so I'd be content if I could convert a standard letter to some number since the characters that I'm most going to use this on are the alphanumerics. Thus: > What's the best way to convert an alphanumerical character to a position in > the alphabet?
[ 0.00980368535965681, 0.010307151824235916, -0.01099594496190548, 0.005458222236484289, -0.015072006732225418, -0.002427159808576107, 0.006930254399776459, -0.015156719833612442, -0.01662478595972061, -0.024039655923843384, -0.01672735996544361, 0.0007392882835119963, 0.005827857181429863, ...
[ 0.2873937785625458, 0.2592339813709259, 0.5498946905136108, -0.37717753648757935, 0.03848212957382202, 0.43304261565208435, 0.04256826266646385, -0.38532930612564087, -0.29282116889953613, -0.6234727501869202, -0.24782581627368927, 0.44342759251594543, -0.29190486669540405, 0.0052059637382...
I saw this answer, which uses the `etoolbox` package to create a macro with an optional argument. Unfortunately, I can't get it to work. Here is a minimal working example: \documentclass[12pt]{article} \usepackage{etoolbox} \newcommand\mymacrotwo[2][]{ \ifstrempty{#2}{ #1: none }{ #1: #2 } } \begin{document} \mymacrotwo{hello} \mymacrotwo{hello}{world} \end{document} This outputs: : hello : hello world Not only are the colons in the wrong place, but it doesn't display the `none` if the second argument isn't defined. What's the problem? I also tried this variant: \documentclass[12pt]{article} \usepackage{etoolbox} \newcommand{\mymacro}[2]{#1 \ifboolexpr{ test {\ifdef{#2}} } {#2} {none}} \begin{document} \mymacro{hello} \mymacro{hello}{world} \end{document} but this outputs: hello hello orldworld which clearly isn't right either.
[ 0.013660293072462082, 0.018188122659921646, 0.0018434112425893545, 0.007760742679238319, 0.0039376066997647285, 0.002336585894227028, 0.0058830794878304005, -0.012823278084397316, -0.009506347589194775, -0.014932937920093536, -0.02256717160344124, -0.00025882176123559475, -0.0151158012449741...
[ -0.021043920889496803, 0.16946709156036377, 0.19860924780368805, -0.06811971217393875, 0.2992781102657318, -0.08224467933177948, 0.31284549832344055, -0.04747672379016876, 0.09400879591703415, -0.49850860238075256, 0.095811627805233, 0.5568803548812866, -0.36487919092178345, 0.011051771230...
I've finished my first playthrough and I'm pretty sure i went through all areas in the game, but i can't seem to get the location based achievements like: Urban Explorer Arctic Explorer Blight Explorer I open my map and all areas are discovered.... Any help?
[ 0.005014956463128328, 0.009192593395709991, 0.003180001862347126, 0.010595738887786865, 0.021229948848485947, 0.00040812522638589144, 0.009209280833601952, 0.03385256603360176, -0.03347576782107353, 0.040469296276569366, -0.010238188318908215, 0.029644755646586418, -0.025444146245718002, -...
[ 0.21660779416561127, 0.2706732451915741, 0.4057782292366028, 0.14216183125972748, 0.2570582926273346, -0.4372237026691437, 0.2954120934009552, 0.42573457956314087, -0.4190358817577362, -0.7943207621574402, -0.12426665425300598, 0.2593197822570801, 0.1205168068408966, 0.2658074200153351, ...
I want to be able to have a macro access a specific element of a list. The following code works, however requires me to have two separate macros: one that accepts a list, and the other that accepts the name of a macro that contains the list. \documentclass{article} \usepackage{tikz} \usetikzlibrary{calc} \usepackage{xstring} \begin{document} \newcommand*{\GetListMemberA}[2]{% \foreach \a [count=\i] in {#1} {% \IfEq{\i}{#2}{\a\breakforeach}{}% }% \par% }% % This is same as above except does not have {} around #1 \newcommand*{\GetListMemberB}[2]{% \foreach \a [count=\i] in #1 {% \IfEq{\i}{#2}{\a\breakforeach}{}% }% \par% }% \newcommand*{\MyList}{a1,b2,c3,d4,e5}% % These work if the foreach uses {#1} (ie, with the curly braces) This should print "b": \GetListMemberA{a,b,c,d,e}{2}% This should be blank: \GetListMemberA{a,b,c,d,e}{6}% % These work if the foreach uses #1 (ie, without the curly braces} This should be "a1": \GetListMemberB{\MyList}{1}% This should be "c3": \GetListMemberB{\MyList}{3}% This should be blank: \GetListMemberB{\MyList}{6}% \end{document} This question on TikZ \foreach loop with macro-defined list suggest to remove the braces around the macro when using a foreach, and that is how I was able to come up with this solution. So, how can I change this so I don't have to have two macros? I don't think this issue is related to Using Macro Defined Lists in TikZ/PGFplots as that was specific to lists used to label tick marks on axis.
[ 0.01674487069249153, 0.019136909395456314, -0.0032061291858553886, 0.0015697246417403221, 0.02717704325914383, 0.011037486605346203, 0.006065413821488619, 0.007642535492777824, -0.01874368265271187, 0.0039208452217280865, -0.01521183829754591, -0.003103291615843773, -0.009935775771737099, ...
[ 0.2070450484752655, 0.018782952800393105, 0.6095207929611206, -0.21513107419013977, 0.3034149408340454, 0.07168831676244736, 0.03935621306300163, -0.3681986629962921, -0.13274767994880676, -0.7072713375091553, 0.11709427833557129, 0.5269367098808289, -0.31679025292396545, -0.09212997555732...
Are **Small Basic** ( **SB** ) and **Microsoft Small Basic** ( **MSB** ) related or is this a confusion based upon similar names? I have read that **SB** is written in C and is GPL. The oldest date on SourceForge is 2004 but the facebook group states its over 10years old. **MSB** is written in .NET 3.5 by Microsoft DevLabs. Wikipedia states it originated in 2008. They seem completely unrelated. Is this another Microsoft "Embrace, Extend, and Extinguish" tactic? I have read that **MSB** does not allow distribution: > The license is presented when you install the software. Small Basic includes > a command line compiler that can be used as part of the product. You may > install and use any number of copies of the software on your premises to > test how it runs with your programs, distribution is not allowed. Please > review the license terms for additional use rights/restrictions. Thank you. -MSB Forums
[ -0.0063263606280088425, -0.0005821216618642211, -0.011427639052271843, 0.01706765592098236, 0.03534599393606186, 0.020549636334180832, 0.010736614465713501, 0.01755286008119583, -0.018052052706480026, -0.03465666621923447, -0.012691725045442581, 0.017453549429774284, 0.01421294454485178, -...
[ -0.03623124212026596, 0.1899213045835495, 0.43410974740982056, -0.343215674161911, -0.2832329571247101, -0.012760068289935589, 0.37706515192985535, 0.15574157238006592, -0.027925213798880577, -0.5662961006164551, 0.0009756241925060749, 0.250016450881958, -0.21599198877811432, 0.02931210957...
I have a 2D road segment and convert it into 3D using a raster in ARCGIS 9.3. So, what i see is that the length of 3D road segment is still the same as the 2D road. How can this be? As due to elevation aspect, the 3D length should be longer than the 2D counterpart.
[ -0.046614281833171844, 0.018744444474577904, -0.028436951339244843, 0.03258269280195236, -0.014651772566139698, 0.003142950590699911, 0.013216616585850716, -0.009499878622591496, -0.025025980547070503, -0.04411447048187256, -0.004219159483909607, 0.01969386450946331, -0.009689822793006897, ...
[ 0.28656256198883057, 0.30759918689727783, 0.4332510828971863, -0.13197892904281616, -0.17835633456707, 0.06524444371461868, 0.1443069726228714, 0.09634365886449814, -0.13172480463981628, -0.8722659349441528, 0.16480258107185364, 0.6914112567901611, 0.087155781686306, 0.15004050731658936, ...
I need a book as a reference for the "principle of indifference" or "principle of insufficient reason". Any suggestion?
[ 0.0064084334298968315, 0.013431230559945107, -0.002691693603992462, 0.02467874437570572, -0.007863471284508705, 0.021704016253352165, 0.014064441435039043, -0.011021899059414864, -0.012820000760257244, 0.004508793354034424, -0.003171563148498535, 0.010127020068466663, -0.035469260066747665, ...
[ 0.49038901925086975, 0.12921379506587982, 0.0421886220574379, 0.34568050503730774, 0.026642881333827972, -0.01815822720527649, 0.13506059348583221, 0.2731820344924927, 0.054914690554142, -0.3580112159252167, 0.3981677293777466, 0.3219117820262909, -0.15614308416843414, 0.09832525253295898,...
I'm developing a website that allows users to share their location via JavaScript. Based on that location, either the content of the current page will change, or it will redirect a user to a page with content relevant to their location. For users that don't share, I'll present relevant content that I would expect to also get indexed by search bots. Since it is fairly well known that Google's crawlers do process JavaScript, is there any evidence to suggest that Googlebot shares geo-location information? I ask mainly because I'm wondering what the impact would be on my site's ability to be indexed correctly.
[ -0.024420369416475296, 0.0033124075271189213, 0.007455827668309212, 0.01446752343326807, 0.016814060509204865, -0.0047721900045871735, 0.008062159642577171, 0.019552081823349, -0.01636357791721821, -0.00034059397876262665, 0.0034558854531496763, 0.031117774546146393, 0.009445461444556713, ...
[ 0.6392786502838135, 0.4120684564113617, 0.1267787218093872, -0.0021617042366415262, 0.34365546703338623, 0.006896168924868107, -0.0032110409811139107, 0.15958373248577118, -0.136934295296669, -0.5186473727226257, -0.10239614546298981, 0.08778016269207001, -0.33184048533439636, 0.4792662858...
I don't want to ask about potential resolutions to the arrow of time. On the contrary, I want to take it as a given that it is the case for the universe, but then ask a rather odd question about its applicability (or lack thereof) to the surface of the earth. The surface of the earth is certainly not a closed system, in that it receives tremendous energy from the sun. For that reason alone, there is no reason to believe its entropy should increase over time. Shouldn't I be able to conclude that, for processes _that are only sensitive to conditions on the surface of the earth_ , there is no arrow of time ? Shouldn't be this be the case for life as a whole ? If I were studying large scale structure formation in the universe, it would make sense to keep in mind that there will be an asymmetry in the change in entropy between one direction and another given the low entropy of the initial conditions. The evolution of life clearly also follows this asymmetry in that if I were to ask in what direction of time I would have to go to find the common ancestor of 2 animals, it would be the same as the low entropy direction of the universe, but there's no clear reason why that should be the case (in other words, if I were to reconstruct the tree of life, there would be a direction from which branches can flow out of points, but never into them). The bacteria floating around a puddle of water are hardly sensitive to the entropy conditions of the whole universe, and only care about their immediate surroundings. More generally, as I've pointed out, there is no reason to think that the surface of the earth will have greater entropy in a billion years. So why can't life appear in the future, and evolve "backwards" relative to us, leading to a separate time reversed tree of life from us ? Why isn't there a second tree of life that sees time "backwards" relative to us, with living beings that seem to multiply "backwards" in time ? More generally, if this cannot be the case for the surface of the earth if the energy from the sun simply isn't enough to accomplish it, can there exist local systems whose arrow of time will be the opposite of the global arrow of time ?
[ -0.00036376016214489937, 0.012670244090259075, -0.010699974372982979, 0.009267237037420273, -0.005566901993006468, -0.01428079605102539, 0.005324167665094137, 0.005018559284508228, -0.007802305743098259, -0.02525128424167633, -0.008191270753741264, 0.01725231483578682, -0.012173698283731937,...
[ 0.4468846917152405, 0.25440219044685364, 0.044016703963279724, 0.27914297580718994, 0.09816531836986542, 0.0817565768957138, -0.0010558604262769222, 0.005945790559053421, -0.5040761232376099, -0.6496278643608093, 0.08591750264167786, 0.3385762870311737, 0.00866825133562088, 0.7693246603012...
I have this custom query that Im working on and I need to be able to query output values as well as actual stored values. Example: $string = "This is #*&^% a test"; $string = preg_replace('/[^(\x20-\x7F)]*/','', $string); $string_clean = $string So if I echo $string_clean, I'll get this: This is a test Which is fine, but the issue is that the $string_clean value isn't a stored value (just an output), so I dont know how to use its value in a query, being that the query results are based on stored database values. Im using a multi array capability of the wordpress query ( **here**). Basically, I use two keys in the array with AND as the meta_compare. Ideally, I would like to be able to query both value (stored and output). Im interested in this as a search functionality for my query. Can someone help me with this? How do I do it?. Or if someone knows of a better solution please let me know. **UPDATE:** This is the code that Im working with. I couldn't find any reliable solutions to implement REGEXP. Basically, what this should do, in theory, is use the php `preg_replace` function to remove/replace any special characters in the string and query the new clean string, that will then be used in the query's argument. Here it is: *The initial value of the variable are submitted through a simple html form (not included). if(isset($_POST['providername'])) { $providername = $_POST['providername']; if (preg_match('/[\'^£$%&*()}{@#~?><>,|=_+¬-]/', $providername)) { $clean = $providername; // Replace other special chars $specialCharacters = array( ',' => '', '&amp' => 'and', '@' => 'at', '.' => '', '+' => '', '=' => '', '/' => '', '*' => '', '!' => '', '(' => '', ')' => '', '$' => '', '#' => '', '%' => '', '\'' => '', '_' => '', '^' => '', '-' => '' ); while (list($character, $replacement) = each($specialCharacters)) { $clean = str_replace($character, '-' . $replacement . '-', $clean); } // Remove all remaining other unknown characters $clean = preg_replace('/[^a-zA-Z0-9\-]/', ' ', $clean); $clean = preg_replace('/^[\-]+/', '', $clean); $clean = preg_replace('/[\-]+$/', '', $clean); $clean = preg_replace('/[\-]{2,}/', '', $clean); if (preg_match('/[\'^£$%&*()}{@#~?><>,|=_+¬-]/', $providername)) { $providername = $clean; } else { $providername = $_POST['providername']; } } } if(isset($_POST['providerloc'])) { $provider_location = $_POST['providerloc']; } This is the multi array argument: $args = array( 'count_total' => true, 'role' => 'subscriber', 'offset' => 8, 'number' => 8, 'meta_query' => array( 'relation' => 'AND', array( 'key' => 'jabber', 'value' => $provider_location, 'compare' => 'LIKE' ) , array( 'key' => 'keywordcontent', 'value' => $providername, 'compare' => 'LIKE' ) ) ); After testing this solution, I get no results when trying to search for the "clean" outputted value. For example: If... $value = "G-P's Workin' Barn"; After applying `preg_replace`, the output will be... GPs Workin Barn Now...if I search for "GPs Workin Barn" (the output value), I get no results. But when I search for "G-P's Workin' Barn" (the original value), it shows the result for that term. So it seems that the query isn't able to process the output value. As I thought, this makes sense because the query is searching against stored values in the database...There should be a way for the query to interpret/translate the output value and compare it against the stored value. After searching online, the closest solution that relates to what I need is the post_where filter for posts [credit @Rarst]. There doesn't seem to be a similar filter for users. Even so, I need a way to search using REGEXP against actual variable values. Any thoughts or solutions? Thanks
[ 0.00666152173653245, 0.018641401082277298, -0.0018280865624547005, -0.004805768374353647, -0.007765146903693676, 0.00983024574816227, 0.006908487528562546, 0.006819744594395161, -0.01638149470090866, -0.01969706453382969, -0.0028213607147336006, -0.0006374333752319217, -0.021551664918661118,...
[ 0.5013702511787415, 0.05012281984090805, 0.11689155548810959, -0.08466579765081406, -0.21845419704914093, -0.26666101813316345, 0.5803722739219666, -0.41837576031684875, 0.029492255300283432, -0.6073430776596069, 0.23355381190776825, 0.6257861852645874, -0.31823137402534485, 0.281727969646...
Imagine there are two observers $A$ and $B$ and a particle $P$. $A$ and $B$ are at about the same point, $P$ is some distance away. From $A$'s point of view, $P$ has velocity $V$ and $B$ has velocity $-V$. P----> A <----B Suppose $A$ and $B$ have the same rest mass $m$ and $P$ has rest mass $m_P$. It seems to me that $A$ will observe a gravitational force between $P$ and $A$ of $F_a$ and between $P$ and $B$ of $F_b$ with $F_a\neq F_b$ because the relativistic masses of $A$ and $B$ are different. However, B will observe forces $F_a'$ and $F_b'$ and it seems to me that $F_a' \neq F_a$ and $F_b' \neq F_b$ because $A$ and $B$ observe different relativistic masses for everything involved. I've tried demonstrating $F_a=F_a'$ on paper and gotten into a tangle. Intuitively, I think this equality is false and expect the difference to be resolved elsewhere. I have a qualitative understanding of general relativity, but can't handle tensors etc. Can this be explained with high-school maths?
[ -0.02488415688276291, 0.022105617448687553, 0.007605298887938261, 0.018015922978520393, 0.005281845107674599, -0.016110606491565704, 0.009135948494076729, -0.019318414852023125, -0.009110813029110432, 0.002811633050441742, -0.016983434557914734, 0.024414584040641785, -0.01966547593474388, ...
[ 0.215730682015419, 0.004635656718164682, 0.337433785200119, 0.05512957274913788, -0.004298940300941467, 0.4247550964355469, -0.14193852245807648, -0.14048779010772705, -0.35394713282585144, -0.6661463379859924, 0.03408593684434891, 0.15594330430030823, -0.5057137608528137, 0.55538535118103...
I added a shape file of Iran to Qgis. 'Tehran' has 35.6833° N, 51.4167° E coordiantes, however Qgis shows Tehran with 4772860,31296767. How i can view coordinates of a city or location in latitude and longitude form in qgis? Thanks.
[ -0.009444156661629677, 0.011955717578530312, -0.017790207639336586, 0.014486528001725674, -0.012815551832318306, 0.002896682359278202, 0.012088884599506855, 0.006333679426461458, -0.024262042716145515, -0.03388671949505806, 0.011020490899682045, 0.017699571326375008, -0.029326628893613815, ...
[ 0.2221796065568924, 0.12187068909406662, 0.5197423696517944, 0.39604583382606506, -0.35810717940330505, -0.008679581806063652, 0.053891073912382126, 0.5093584060668945, 0.04796764254570007, -0.5517749190330505, -0.22768662869930267, -0.24789714813232422, -0.014404895715415478, 0.4315097928...
I want to use rgdal to plot a shapefile graduating the colour of features according to the values of one field (`density`). I couldn't find online detailed indications on how to do it and the package documentation doesn't explain the options for `plot()`. This is what I composed so far library(rgdal) shapefile = "shapefile.shp" map <- readOGR(dsn = shapefile, layer = "shapefile", verbose=FALSE) plot(map, xlim = c(6.70, 18.32), ylim = c(35.2, 47.6), border=NA, add=TRUE) This is the style in QGIS I am trying to replicate: ![enter image description here](http://i.stack.imgur.com/hihky.png)
[ -0.0021077031269669533, 0.010186711326241493, -0.013698037713766098, 0.014362046495079994, -0.005770749878138304, -0.006875531282275915, 0.009120258502662182, -0.00037975108716636896, -0.011882748454809189, -0.006321084685623646, 0.008345011621713638, 0.0034220507368445396, -0.00494795106351...
[ 0.19975093007087708, -0.3046957552433014, 0.7380343675613403, -0.16638857126235962, 0.1525157392024994, 0.11177201569080353, -0.21457140147686005, -0.29186713695526123, 0.017090721055865288, -0.8045839667320251, -0.06574586033821106, 0.31692853569984436, 0.07113935053348541, -0.00593259697...
Why does the Zerg AI player sometimes get stuck at the very beginning of the game and doesn't ever evolve beyond producing drones? This happened to me in StarCraft: Brood War, in a melee game against computer player(s). I'm not sure it is map-specific, but I'm pretty sure it happens on the **Big game hunters** map. This is a mystery for me since I was a child, and now that I can understand some AI, I would really like to understand this.
[ -0.020590728148818016, -0.0011970626655966043, 0.006150522734969854, 0.008177492767572403, 0.001005274709314108, -0.015109970234334469, 0.007307304535061121, -0.013238825835287571, -0.020024677738547325, 0.00994864758104086, -0.004802231211215258, 0.027254078537225723, -0.014683415181934834,...
[ -0.11692728847265244, 0.0643962100148201, -0.12816721200942993, -0.05509311705827713, -0.3066648244857788, -0.116806261241436, -0.1840643286705017, 0.2810581624507904, -0.3141196370124817, -0.6293215155601501, 0.3292536437511444, 0.19680912792682648, -0.3579501211643219, 0.5006832480430603...
Im trying unhook the comment form found in the functions.php of the deafault buddypress theme so I can change it a bit. I placed this piece of code into my child theme functions.php but it didn't change anything. What did I do wrong? remove_filter( 'comment_form_defaults', 'bp_dtheme_comment_form', 10 ); if ( !function_exists( 'my_dtheme_comment_form' ) ) : function my_dtheme_comment_form( $default_labels ) { global $user_identity; $commenter = wp_get_current_commenter(); $req = get_option( 'require_name_email' ); $aria_req = ( $req ? " aria-required='true'" : '' ); $fields = array( 'author' => '<p class="comment-form-author">' . '<label for="author">' . __( 'Name', 'buddypress' ) . ( $req ? '<span class="required"> *</span>' : '' ) . '</label> ' . '<input id="author" name="author" type="text" value="' . esc_attr( $commenter['comment_author'] ) . '" size="30"' . $aria_req . ' /></p>', 'email' => '<p class="comment-form-email"><label for="email">' . __( 'Email', 'buddypress' ) . ( $req ? '<span class="required"> *</span>' : '' ) . '</label> ' . '<input id="email" name="email" type="text" value="' . esc_attr( $commenter['comment_author_email'] ) . '" size="30"' . $aria_req . ' /></p>', 'url' => '<p class="comment-form-url"><label for="url">' . __( 'Website', 'buddypress' ) . '</label>' . '<input id="url" name="url" type="text" value="' . esc_attr( $commenter['comment_author_url'] ) . '" size="30" /></p>', ); $new_labels = array( 'comment_field' => '<p class="form-textarea"><textarea name="comment" id="comment" cols="60" rows="10" aria-required="true"></textarea></p>', 'fields' => apply_filters( 'comment_form_default_fields', $fields ), 'logged_in_as' => '', 'must_log_in' => '<p class="alert">' . sprintf( __( 'You must be <a href="%1$s">logged in</a> to post a comment.', 'buddypress' ), wp_login_url( get_permalink() ) ) . '</p>', 'title_reply' => __( '', 'buddypress' ) ); return apply_filters( 'my_dtheme_comment_form', array_merge( $default_labels, $new_labels ) ); } add_filter( 'comment_form_defaults', 'my_dtheme_comment_form', 10 ); endif; To test that nothing changed I removed the label text for 'title_reply'
[ 0.004407104104757309, 0.0070936609990894794, 0.002392185851931572, 0.019422587007284164, -0.0024979980662465096, -0.008382471278309822, 0.007005964405834675, 0.0132429338991642, -0.014301301911473274, -0.0034328047186136246, -0.01204661838710308, 0.007115188054740429, 0.0002012341283261776, ...
[ -0.11879982054233551, -0.04334680363535881, 0.5154819488525391, -0.028776533901691437, 0.13594408333301544, 0.14787647128105164, 0.48644837737083435, -0.07982129603624344, -0.04536417871713638, -0.8901804089546204, -0.03537549450993538, 0.646841824054718, -0.5718452334403992, 0.23482219874...
If you were tasked with performing due diligence for a software product/IP you/your company was considering buying. What would be your steps to make sure the purchase was acceptable from a technical risk level? The scenario is one company has a product you or your company is interested in buying in it's entirety (IP and code) so that your company can now sell that product or it's services. The upfront assumption is the product has some value, but you want to assess the technical foundation of the product to ensure it's not a complete hack and nightmare to support. Examples of this might include * Compile/Build and install the software on a clean computer to verify the build and deploy cycle. * Run a code profiling tool against the code base to get an idea of overall code quality. What else? Are there any resources/checklists on the web which might help?
[ 0.010319344699382782, 0.011988312937319279, -0.001770353177562356, 0.01265889871865511, 0.003213347401469946, 0.012377063743770123, 0.008382931351661682, -0.002179152797907591, -0.01126162987202406, -0.002948448061943054, -0.009496050886809826, 0.018104568123817444, -0.011449184268712997, ...
[ 0.5749620795249939, 0.044809725135564804, -0.22163896262645721, 0.2891494631767273, 0.15419085323810577, -0.3163565397262573, 0.04079441726207733, 0.0589279904961586, -0.29926979541778564, -0.3170950710773468, -0.12529447674751282, 0.5677199959754944, 0.08933393657207489, 0.181978449225425...
We have a complicated database query and we would like to exclude some posts. We've tried to do it with SQL, but it's really complicated. The simplest solution seems to loop through the results after the query's made and then exclude a few posts. The problem is, i don't know where to do it! I've tried to unset posts in $wp_query->posts, but they still show up. Is there a proper way to do this? Is there another array WordPress uses that I should unset from?
[ 0.036211010068655014, 0.013798006810247898, 0.0010623808484524488, 0.01713515818119049, 0.0026055993512272835, 0.008872274309396744, 0.0049818712286651134, 0.02375722862780094, -0.0212322436273098, -0.005033756140619516, -0.006968334782868624, 0.01340140774846077, -0.011556747369468212, -0...
[ 0.4870383143424988, 0.2224263846874237, 0.11981797963380814, 0.3240630030632019, 0.22790735960006714, -0.2922012209892273, 0.48416146636009216, -0.04545077309012413, -0.11227595061063766, -0.5339867472648621, 0.35456329584121704, 0.12956774234771729, -0.20451728999614716, 0.413042753934860...
I am about to start a new ASP.NET MVC5 project and I am planning the authentication / authorization requirements at present. The client wants Windows authentication, to prevent their users having to remember another new password. The site is web facing, so the downside would be an ugly pop up box asking for their credentials when they are accessing it offsite. Worse yet, on mobile, this box would be a problem. The Active Directory Authentication options from out of the box are new to me, but after some reading appear to be more about controlling roles and authorization through your AD groups. I intend to keep all authorization concerns internal to the application. Ideally, users will have Window authentication but with a nice login page where they can select their domain from a dropdown box and enter their domain login credentials. From some reading I thought possibly ActiveDirectoryMembershipProvider is the answer. However with the new available options I want to be sure there is not other options before blindly taking this route.
[ -0.002775247674435377, 0.0063729798421263695, -0.005286687985062599, 0.017229177057743073, -0.009937521070241928, -0.0012771994806826115, 0.006511378102004528, 0.008796609938144684, -0.01565508544445038, 0.00383806717582047, -0.011643904261291027, 0.01701914332807064, 0.013403619639575481, ...
[ 0.29542168974876404, 0.25636744499206543, 0.6049399375915527, -0.03505140170454979, -0.02977932058274746, -0.21230946481227875, 0.14315485954284668, -0.16656643152236938, -0.22902774810791016, -0.6339390277862549, 0.034589365124702454, 0.5231370329856873, 0.05216803401708603, 0.07941672205...
I'm facing a problem that's about to drive me crazy. I'm running a Debian Linux distribution and currently trying to establish a connection to my wireless network (WPA2 security, wpa_supplicant already installed). The wireless dongle connects to the network, but it only responds to pings and lets me connect through SSH from another computer when the ethernet cable is plugged in. It is still reachable via wireless connection after unplugging the ethernet connection, but it won't work until a connection to a wired network had been created. I'm not sure if anything's wrong with my configuration here... Output of 'ifconfig wlan0': wlan0 IEEE 802.11bgn ESSID:"*censored*" Nickname:"<WIFI@REALTEK>" Mode:Managed Frequency:2.457 GHz Access Point: *censored* Bit Rate:72.2 Mb/s Sensitivity:0/0 Retry:off RTS thr:off Fragment thr:off Encryption key:****-****-****-****-****-****-****-**** Security mode:open Power Management:off Link Quality=89/100 Signal level=58/100 Noise level=0/100 Rx invalid nwid:0 Rx invalid crypt:0 Rx invalid frag:0 Tx excessive retries:0 Invalid misc:0 Missed beacon:0 Contents of /etc/network/interfaces auto lo iface lo inet loopback allow-hotplug eth0 iface eth0 inet static address 192.168.178.130 netmask 255.255.255.0 allow-hotplug wlan0 iface wlan0 inet static wpa-ssid "*censored*" wpa-key-mgmt WPA-PSK wpa-group TKIP CCMP wpa-psk *censored* address 192.168.178.131 netmask 255.255.255.0 gateway 192.168.178.1
[ -0.023491907864809036, 0.0042366222478449345, -0.00784394796937704, 0.008961455896496773, -0.017446614801883698, -0.02496947906911373, 0.007905520498752594, -0.018165670335292816, -0.010789492167532444, -0.022511720657348633, -0.01777045428752899, 0.0010262608993798494, -0.022798141464591026...
[ 0.17243871092796326, 0.22840705513954163, 0.2526519000530243, 0.10735112428665161, 0.28178635239601135, -0.2102390080690384, 0.4010079503059387, 0.10055588185787201, -0.13812969624996185, -0.7643259167671204, 0.05067594349384308, 0.3045915961265564, -0.2847256362438202, 0.1806241124868393,...
I am trying to see the content in a `boot.img` file from an Android image. I googled and got http://javigon.com/category/development-tools/ to extract `system.img`. But it doesn't work for `boot.img`. When trying to do this for `boot.img`, it is showing the following: Invalid sparse file format at header magi Failed to read sparse file Is `simg2img` used only for extracting `system.img`? 1. If so, Is there any other method to extract `boot.img`? 2. If not, what is the problem for not extracting `boot.img`?
[ -0.004732427187263966, -0.0035563199780881405, -0.001081996364519, 0.01918601803481579, -0.02407299540936947, 0.007188315503299236, 0.007324996404349804, 0.022773664444684982, -0.016473006457090378, 0.02424885891377926, -0.011998518370091915, 0.008728167042136192, 0.00644141947850585, 0.01...
[ 0.2732153534889221, 0.19058138132095337, 0.18417595326900482, 0.005794471129775047, -0.10055657476186752, -0.10855288803577423, 0.5407028198242188, 0.0026238486170768738, -0.3501725196838379, -0.704586923122406, -0.03407035395503044, 0.37855300307273865, -0.30667901039123535, 0.36588877439...
Forgive me, but I am completely new to GIS. I am using QGIS. I have a raster layer (HDF format) that does not align with other layers. All layers are projected WGS84 (EPSG:4326) and still the problem persists. I know that the HDF layer is defined by rows and columns that need to be converted to longitudes and latitudes but I don't know how to do that. Also, the HDF file, for some reason, can't be exported as CSV file. ![enter image description here](http://i.stack.imgur.com/gTm0h.jpg)
[ -0.020318452268838882, 0.006516929715871811, 0.0037242420949041843, 0.028424229472875595, 0.0006297712679952383, 0.0034515634179115295, 0.010043038055300713, 0.0209090244024992, -0.015830472111701965, -0.02398538962006569, -0.001907329773530364, 0.003337469417601824, -0.007273420691490173, ...
[ 0.5238230228424072, 0.2656620442867279, 0.5239080786705017, 0.11287157237529755, -0.2407522201538086, -0.2550119459629059, 0.22709353268146515, 0.06789125502109528, -0.17155081033706665, -1.017269253730774, -0.1372680366039276, 0.4038304388523102, -0.0918259471654892, 0.3732704818248749, ...
I am using a WordPress plugin, SOLDPress. Two weeks ago, everything in this Real Estate plugins Search feature worked properly. This past week, the plugin has started failing with the following error 2013 Lost connection to MySQL server during query. The website is hosted by Go Daddy on one of their virtual machines. The exact same configuration on my development system, hosted by Host Papa works all the time. My plugin developer gave me the following SQL query to execute using phphMyAdmin, in the cPanel, and he was as surprised as I that it also failed. We felt that without all the other overhead it woudl work but I got the same error as above, 19 times out of 20. That means it succeeded once. If anyone could offer some troubleshooting tips and ideas and suggestions I thank you. SELECT SQL_CALC_FOUND_ROWS wp_posts.ID FROM wp_posts INNER JOIN wp_postmeta ON (wp_posts.ID = wp_postmeta.post_id) INNER JOIN wp_postmeta AS mt1 ON (wp_posts.ID = mt1.post_id) WHERE 1=1 AND wp_posts.post_type = 'sp_property' AND (wp_posts.post_status = 'publish' OR wp_posts.post_status = 'private') AND ( (wp_postmeta.meta_key = 'dfd_ListPrice' AND CAST(wp_postmeta.meta_value AS SIGNED) > '0') AND (mt1.meta_key = 'dfd_ListPrice' AND CAST(mt1.meta_value AS SIGNED) BETWEEN '0' AND '900000') ) GROUP BY wp_posts.ID ORDER BY wp_posts.post_date DESC LIMIT 0, 12
[ 0.003023957135155797, 0.00038200768176466227, 0.008035574108362198, 0.014436459168791771, -0.015952538698911667, 0.016091663390398026, 0.008633946999907494, 0.02055421471595764, -0.01640043780207634, 0.006269272416830063, -0.023532388731837273, 0.013801248744130135, 0.01802758313715458, 0....
[ 0.6516899466514587, 0.47504153847694397, 0.49113690853118896, -0.1812402606010437, 0.3253534436225891, 0.09558689594268799, 0.34768274426460266, 0.1391737312078476, -0.2898436188697815, -0.6644605398178101, 0.06169324368238449, 0.15017332136631012, -0.10635893791913986, 0.22513408958911896...
**Background:** I am proponent of functional programming who works at a VB.NET shop where the prevailing mental model is imperative programming. Being that foundation of our system is WinForms I can understand we're not going to get entirely away from imperative programming, but still I try to use FP (primarily via Linq) wherever possible because I believe in its merits. **Arguments & counter-arguments against FP** 1. One might notice that fluent Linq is less efficient than its imperative counterpart in that this style processes a sequence down to another sequence and repeats that. Generally, its going to take a few more passes than the imperative approach which can be better optimized to avoid repeat passes over a sequence. For this reason, the lead couldn't understand why I would choose a functional approach that is clearly "less efficient." * _Counter-argument_ : I argued that while it is sometimes less efficient in terms of CPU cycles, that I felt it is more humanly intelligible and easy to follow since each line does just one thing on its pass over the sequence. To me this feels like having an assembly line where each person at his station has just one job to do. I feel that the negligible trade off of efficiency is recompensed by code whose concerns are neatly separated. 2. The next argument against FP that I hear in my shop is that it's harder to debug -- which is true. It's not easy to step over Linq code. And I do sometimes have to unravel a method chain in order to better follow and dissect issues that I can't immediately spot. * _Counter-argument: For the most part though I don't have issue with this because I think the functional style is more declarative in how it reads and when an error is thrown within a functional chain, I can usually spot the issue immediately. **My Question** I've been trying to promote functional style in our shop and I don't feel like I'm making headway. I've done both styles of programming and have only recently dabbled in Haskell. Despite years of imperative experience, now that I'm making routine use of FP in JavaScript, it's grown on me. It rings a note of rightness in my core when I compare it to what I might have done if I stuck to an imperative style. I've retrained my brain toward functional thinking, toward functional composition. _What I can't understand is how hard its been to convince others of FP's merits._ For example, the developers in my shop do use Linq, but I think they generally use it in the context of dealing with domain data. I use it in a more general sense and prefer it anytime I'm dealing with sequences/lists or persistent data structures. I haven't been able to convince my teammates to expand their use of Linq. _What I'm trying to understand is what causes a developer to not like FP._ I would like to see an answer from someone who has a good deal of experience with FP but decided in favor of the imperative style. What drove the decision to stay with imperative instead of using functional? * * * Here's an additional example highlighting the differences between imperative & functional programming. I wrote the `SelectedRows` method of our grid in Linq as such: Public Property SelectedRows() As DataRow() Implements IDataSourceControl.SelectedRows Get Return Me.ugrBase.Selected.Rows. OfType(Of Infragistics.Win.UltraWinGrid.UltraGridRow)(). Select(Function(ugr) ugr.ListObject). OfType(Of DataRowView)(). Select(Function(drv) drv.Row). ToArray End Get However, this style of code makes some of our developers uncomfortable and so our lead rewrote it to the more familiar: Public Property SelectedRows() As DataRow() Implements IDataSourceControl.SelectedRows Get Dim plstRows As New List(Of DataRow) For Each bugrLoop As Infragistics.Win.UltraWinGrid.UltraGridRow In Me.ugrBase.Selected.Rows If bugrLoop.ListObject IsNot Nothing Then plstRows.Add(CType(bugrLoop.ListObject, DataRowView).Row) End If Next Return plstRows.ToArray() End Get
[ -0.005454786121845245, 0.01740024983882904, -0.01586253196001053, 0.008700745180249214, -0.006583884358406067, 0.007354468107223511, 0.007793571334332228, 0.013390765525400639, -0.011929012835025787, -0.016481276601552963, -0.019413422793149948, 0.018982242792844772, -0.017131788656115532, ...
[ 0.2512422502040863, 0.17590050399303436, 0.10990774631500244, -0.032109156250953674, -0.24055399000644684, 0.12111643701791763, 0.3655892610549927, 0.13406197726726532, -0.28598541021347046, -0.41805368661880493, -0.1973116397857666, 0.8066380023956299, -0.1633245050907135, -0.331375896930...
I have two counters, `\definenumber[words]` and `\definenumber[items]`. I have tried to use TeX conditionals to compare them, e.g.: \ifnum\getnumber[words]=\getnumber[items] This will print some text if the two counters have the same value. fi This does not work. How can I use TeX conditionals to compare two counters in ConTeXt?
[ 0.029907196760177612, 0.02169116400182247, -0.009720764122903347, 0.017059465870261192, -0.004166056867688894, -0.007677609566599131, 0.010698792524635792, -0.03613140434026718, -0.01526198536157608, -0.003110247664153576, -0.00972833950072527, -0.004713846370577812, 0.008020245470106602, ...
[ 0.28172919154167175, -0.13961352407932281, -0.0004037367179989815, -0.050215043127536774, -0.3642822206020355, 0.165322944521904, 0.39376869797706604, -0.23014304041862488, 0.3411603569984436, -0.7446513772010803, 0.23160453140735626, 0.7131701707839966, -0.06013733893632889, 0.11048950254...
I am regressing firm characteristics on some stock trading-related measures in a panel dataset. Firm size is a highly significant control variable, independent of the estimation method etc. My focus variables are related to firm size though, either by construction (e.g. $focus variable = x / firmsize$) or because of an economic relationship. As a consequence, I am finding myself in a classic multicollinearity situation: If firm size is put in as a control variable, my focus variables become insignificant. If firm size is left out, the focus variables are highly significant. Any of the usual advice (e.g. http://en.wikipedia.org/wiki/Multicollinearity) is not helping: I cannot obtain more data, I cannot run my regression on principal components because I need interpretable coefficients etc. I have little experience with this kind of problem but with some imagination, I came up with the following two ideas: 1. Running the regression with firm size as a control variables and additionally including interaction terms between each focus variable and firm size. 2. Trying to strip away the firm size effect from both the dependent variable and the focus variables, e.g. by first regressing firm size on the dependent/focus variable and then using the residuals as the dependent/focus variable in the actual regression. Would either or both idea make any sense? Any comment or alternative ideas would be very welcome!
[ -0.004476775415241718, 0.0043556648306548595, -0.006180483847856522, 0.008955534547567368, 0.020287685096263885, 0.016601139679551125, 0.008488882333040237, 0.006712809205055237, -0.005775859579443932, 0.008544549345970154, -0.006708875298500061, 0.01590624824166298, 0.0009660213254392147, ...
[ -0.11494457721710205, 0.26939818263053894, 0.22723691165447235, -0.11306869238615036, 0.05342746526002884, 0.1889370083808899, 0.07098590582609177, 0.16203227639198303, -0.22908151149749756, -0.5036999583244324, 0.08799044787883759, 0.3303212821483612, -0.27585917711257935, 0.4381842911243...
I'm running vim 7.0 on linux, with Vim LatexSuite. I'm trying to use the nifty-looking shortcut for adding items to an itemize environment. :h Alt-I <Alt-I> *ls_3_10_4* *ls_a_bE* *Alt-I* This mapping inserts an \item command at the current cursor location depending on which environment the cursor is enclosed in. ... But it's not working. Now `<Alt-I>` isn't a common mapping style on linux, `<M-I>` would be used normally, so I guess the mapping is just not set up correctly for linux. This is confirmed by: :verbose map! <M-I> No mapping found How can I fix this mapping? What's the function it usually maps to?
[ 0.01686004549264908, 0.007512037176638842, -0.016852812841534615, 0.009691848419606686, -0.0019326433539390564, 0.011052140966057777, 0.006405946798622608, 0.0005690981633961201, -0.012278389185667038, 0.024825986474752426, -0.017694707959890366, -0.003414616920053959, -0.012913030572235584,...
[ -0.03808598220348358, -0.16524355113506317, 1.1173679828643799, -0.0880994126200676, -0.03346064314246178, -0.060975342988967896, 0.35170048475265503, 0.021601084619760513, -0.36916178464889526, -0.8506973385810852, 0.04119831323623657, 0.7515220642089844, -0.4662255346775055, 0.2108695954...
I find METAPOST to be a wonderful software to create 2d charts and figures. Can you name me a few good alternatives to METAPOST for drawing 3d surfaces? The crucial feature I'm looking for is the ability to have all the fonts (the axis labels and all the numbers) in the LaTeX font, just like METAPOST so gracefully allows me to do.
[ -0.0018054347019642591, 0.005627406761050224, -0.003535816678777337, 0.012028869241476059, 0.012129945680499077, -0.015102040953934193, 0.00825536996126175, -0.0002920558908954263, -0.014745086431503296, -0.009432334452867508, -0.005337544716894627, 0.004204354714602232, 0.017414582893252373...
[ 0.35968396067619324, 0.10137277841567993, 0.39357927441596985, 0.2486260086297989, -0.3628898561000824, 0.3230340778827667, -0.21230964362621307, 0.2922121286392212, 0.03772984445095062, -0.8971561193466187, 0.4734911322593689, 0.4821661114692688, 0.05191567540168762, 0.09392903745174408, ...
For quite some time, I've been hearing the phrase "to be honest" almost every day. I've heard friends say it, characters on TV shows, and even an NPR reporter said it in an interview. Example: > "To be honest, I don't like spaghetti." Why do people have to specify when they are being honest? Or I've thought maybe the phrase is actually supposed to mean "to be blunt" or "let me tell it to you straight." How did the word "honest" get in the mix?
[ -0.009782624430954456, -0.0033378407824784517, -0.014577211812138557, 0.030681710690259933, -0.011000153608620167, -0.014266321435570717, 0.007440060377120972, -0.013199573382735252, -0.014999731443822384, -0.02053787373006344, -0.012926796451210976, 0.008987825363874435, -0.0019341709557920...
[ 0.9901711940765381, 0.12185526639223099, -0.12160686403512955, -0.08900346606969833, -0.5204729437828064, -0.2539054751396179, 0.6416069865226746, 0.5398787260055542, -0.14651766419410706, -0.18326933681964874, 0.3446131646633148, 0.595267117023468, 0.0513458326458931, 0.041098177433013916...
I am Ubuntu user; while I know that Ubuntu 12.04 is LTS and will be supported until April 2017, I'd like to know what the projected support lifetime of Debian Wheezy will be.
[ 0.004456363618373871, 0.00016841257456690073, -0.026693064719438553, -0.0009426003089174628, -0.015803417190909386, -0.01917315274477005, 0.014463620260357857, 0.025969194248318672, -0.022719530388712883, -0.07736317068338394, -0.04366428777575493, 0.022365611046552658, 0.0007719452842138708...
[ 0.5075947046279907, 0.33132800459861755, 0.3892991244792938, -0.19810068607330322, -0.4051670730113983, -0.5942959189414978, 0.21749021112918854, 0.3350757956504822, 0.07721954584121704, -0.6258623003959656, -0.2256683111190796, 0.6058896780014038, -0.3086073398590088, 0.22196371853351593,...
I don't know if this issue has been discussed before, if anyone know any related post, please let me know, thank you :) `Unprotect` the function `Sin` and create a new definition for it: Unprotect[Sin]; Sin[x_] := x^2; Because `x_` matches any expression, then I think (naively) the system definition of `Sin` would never be applied. {Sin[2], Sin[1.2], Sin[Sin]} (* {4, 1.44, Sin^2} *) Every thing works as expected except for `Plot`: Plot[Sin[x], {x, 0, Pi}] ![enter image description here](http://i.stack.imgur.com/6XmSH.jpg) I think it's related to the internal computation of `Plot`, I checked the mass output of `Trace` and found nothing useful. I know it's stupid to modify a system symbol this way, but could you please give me some explanation? Thanks in advance.
[ 0.0016791681991890073, 0.011099917814135551, -0.023861674591898918, 0.0052588507533073425, -0.0018046734621748328, -0.005370333790779114, 0.0062621599063277245, 0.008949697948992252, -0.01717677339911461, -0.003987691830843687, -0.00043381412979215384, -0.0030991071835160255, -0.015792939811...
[ 0.07318652421236038, 0.22817730903625488, 0.3273378312587738, -0.16339713335037231, 0.12559467554092407, -0.29803434014320374, 0.7075924873352051, -0.2581753432750702, 0.17798054218292236, -0.28642353415489197, -0.4203372895717621, 0.7520237565040588, -0.2902028262615204, 0.121890507638454...
For example, say I'm trying to find the command that displays the routing table (like `route`) or to run a `traceroute`, but I don't actually know the commands that accomplish this. My understanding is that I should be able to use `man -k [word]`, but it never seems to work: if I do `man -k route` on Scientific Linux 6.1, I get no results, even though I would expect it to return anything that includes the word "route" in it! Unfortunately, when I search for the answer to this, it's confounded by people explaining the `find` command for a file search. Thank you!
[ 0.0118185356259346, 0.004189460538327694, -0.016647402197122574, 0.00347869168035686, -0.032938823103904724, 0.008319483138620853, 0.005532583687454462, -0.002282223664224148, -0.014247728511691093, -0.009655071422457695, -0.006745017133653164, 0.0006417295662686229, -0.0036538923159241676, ...
[ 0.23092763125896454, 0.41207146644592285, -0.07721665501594543, -0.06982696056365967, 0.06734000891447067, -0.46449151635169983, 0.4090765118598938, 0.3768453001976013, -0.08373265713453293, -0.5027459859848022, -0.20502178370952606, 0.45514243841171265, -0.45804816484451294, 0.20740810036...
Does it apply for all non-elemental attacks, or is there more going on here?
[ -0.015942728146910667, 0.05222593620419502, -0.03374970704317093, 0.020519791170954704, 0.0029642365407198668, -0.013091625645756721, 0.01351542491465807, 0.025398319587111473, -0.05004154518246651, -0.033179473131895065, -0.006555248983204365, 0.019901758059859276, -0.015398030169308186, ...
[ 0.8004136681556702, 0.044737808406353, 0.016629638150334358, 0.10482782870531082, -0.11000225692987442, -0.483178973197937, 0.24054467678070068, -0.18124200403690338, -0.03403042256832123, 0.07052057236433029, 0.3613945543766022, 0.33746573328971863, -0.0563330240547657, 0.0192168094217777...
I linked my phone Samsung note2 with my PC , I find there is a new Update for the phone. the problem kies don't tell me wich version is avaible. I want to update my phone directly to kitkat 4.4 because 4.3 I think there is some problems with it. Kies give me just some information about the update but I want to know if the update shown on the picture is 4.4 or 4.3 please help ! ![](http://i.stack.imgur.com/rwENE.jpg)
[ -0.006734395865350962, -0.0030812558252364397, -0.00321876909583807, 0.007409965619444847, -0.03115883097052574, 0.006300350651144981, 0.006150921806693077, 0.011116499081254005, -0.01673653908073902, -0.003791636321693659, -0.014007901772856712, 0.003906279802322388, -0.0030606326181441545,...
[ 0.3921241760253906, 0.050631508231163025, 0.43364056944847107, -0.15943145751953125, -0.20416109263896942, 0.0914933905005455, 0.6536062359809875, -0.3765717148780823, -0.32530584931373596, -0.5393127799034119, 0.07318781316280365, 0.9041126370429993, -0.08803989738225937, 0.13956876099109...
I want to connect to a PostGIS table Using MS Access. The table has been created using the PostGIS Shapefile loader in pgAdmin. It has been encoded at LATIN1 (it wouldn't load with UTF-8 default). I have installed the ODBC driver from: http://www.postgresql.org/ftp/odbc/versions/msi/ (09_03_210) I have set up the link in MS Access (using "PostgreSQL Unicode" driver), and I can bring the table in. However in all of the fields it is displaying #Error rather than the data. I don't get this error with a table loaded with UTF-8. I'm not sure where to set the encoding variable to LATIN1 when I call the data. I have tried doing so in the "Data Source" Advanced options under "Connect Settings" SET client_encoding TO LATIN1 but it hasn't helped. Is there another problem here - am I setting the encoding variable in the wrong place? Any help appreciated.
[ -0.004877026658505201, 0.008029831573367119, 0.003751961747184396, 0.01321885734796524, 0.008094670251011848, 0.021427659317851067, 0.008316936902701855, 0.003660026704892516, -0.011843143962323666, -0.02914220280945301, 0.0047162375412881374, 0.0049909381195902824, 0.00758893508464098, 0....
[ -0.04012404754757881, 0.2593759596347809, 0.6875539422035217, -0.1640341579914093, 0.10200543701648712, -0.09463531523942947, 0.05314062908291817, -0.039651308208703995, -0.221944659948349, -1.1822034120559692, 0.040625542402267456, 0.17998619377613068, -0.496243417263031, 0.37597674131393...
I have a model to achieve Bayesian estimates the population size $N$ and probability of detection $\theta$ in a binomial distribution solely based on the observed number of observed objects $y$: $$ p(N,\theta|y)\propto \frac{ \text{Bin}(y|N,\theta)}{N} $$ for $ \left\{N|N\in\mathbb{Z}\land N\ge \max(y)\right\}\times(0,1) $. For simplicity, we assume that $N$ is fixed at the same, unknown value for each $y_i$. In this example, $y=53,57,66,67,73$. This model, when estimated in `rstan`, diverges from the results obtained from a grid approximation of the posterior. I'm trying to pin down why. (Interested readers might find that this question is a follow-on to my answer here.) # `rstan` Approximation For reference, this is the rstan code. raftery.model <- " data{ int I; int y[I]; } parameters{ real<lower=max(y)> N; simplex[2] theta; } transformed parameters{ } model{ vector[I] Pr_y; for(i in 1:I){ Pr_y[i] <- binomial_coefficient_log(N, y[i]) +multiply_log(y[i], theta[1]) +multiply_log((N-y[i]), theta[2]); } increment_log_prob(sum(Pr_y)); increment_log_prob(-log(N)); } " raft.data <- list(y=c(53,57,66,67,72), I=5) system.time(fit.test <- stan(model_code=raftery.model, data=raft.data,iter=10)) system.time(fit <- stan(fit=fit.test, data=raft.data,iter=10000,chains=5)) Note that I cast `theta` as a 2-simplex. This is just for simplicity. The quantity of interest is `theta[1]`; obviously `theta[2]` is superfluous information. Additionally, $N$ is a real value (`rstan` only accepts real-valued parameters because it is a gradient method), so I wrote a real-valued binomial distribution. ## Rstan results mean se_mean sd 2.5% 25% 50% 75% 97.5% n_eff Rhat N 1078.75 256.72 15159.79 94.44 148.28 230.61 461.63 4575.49 3487 1 theta[1] 0.29 0.00 0.19 0.01 0.14 0.27 0.42 0.67 2519 1 theta[2] 0.71 0.00 0.19 0.33 0.58 0.73 0.86 0.99 2519 1 lp__ -19.88 0.02 1.11 -22.89 -20.31 -19.54 -19.09 -18.82 3339 1 # Grid Approximation The grid approximation was produced as below. Memory constraints prevent me making a finer grid on my laptop. theta <- seq(0+1e-10,1-1e-10, len=1e3) N <- round(seq(72, 5000, len=1e3)); N[2]-N[1] grid <- expand.grid(N,theta) y <- c(53,57,66,67,72) raftery.prob <- function(x, z=y){ N <- x[1] theta <- x[2] exp(sum(dbinom(z, size=N, prob=theta, log=T)))/N } post <- matrix(apply(grid, 1, raftery.prob), nrow=length(N), ncol=length(theta),byrow=F) post.norm <- post/sum(post) I used the grid approximation to produce this display of the posterior density. We can see that the posterior is banana-shaped; this kind of posterior can be problematic for euclidian metric HMC. (The severity of the banana shape is actually suppressed here since $N$ is on the log scale.) If you think about the banana shape for a minute, you'll realize that it must lie on the line $\bar{y}=\theta N$. (Additionally, the grid approximation displayed in this graph is not normalized for reasons of clarity - else the banana is a little too narrow to clearly make out.) ![posterior over a grid](http://i.stack.imgur.com/kiR50.png) ## Grid approximation results do.call(cbind, lapply(c(0.025, .25, .5, .75, .975), function(quantile){ approx(y=N, x=cumsum(rowSums(post.norm))/sum(post.norm), xout=quantile) })) [,1] [,2] [,3] [,4] [,5] x 0.025 0.25 0.5 0.75 0.975 y 92.55068 144.7091 226.7845 443.6359 2475.398 # Discussion The 97.5% quantile for $N$ is much larger in my `rstan` model than it is for the grid approximation, but its quantiles are similar to the grid approximation otherwise. I interpret this as indicating that the two methods are generally in agreement. I do not know how to interpret the discrepancy in the 97.5% quantile, though. I've developed several possible explanations for what might be accounting for the divergence between the grid approximation and the results from `rstan` HMC-NUTS sampling, but I'm uncertain how to understand if one, both or neither explanation is correct. 1. Rstan is wrong and the grid is correct. The banana-shaped density is problematic for `rstan`, especially as $N$ drifts off towards $+\infty$, so these tail quantities are not trustworthy. We can see from the plot of the posterior over the grid that the tail is very sharp at larger values $N$. 2. Rstan is correct and the grid is wrong. The grid makes two approximations which may undermine the results. First, the grid is only a finite set of points over a subspace the posterior, so it is a rough approximation. Second, because it's a finite subspace, we're falsely declaring there to be 0 posterior probability over values $N$ larger than our largest grid value for $N$. Likewise, `rstan` is better at getting into the tails of the grid, so its tail quanitles are correct. * * * I needed more space to clarify a point from Juho's answer. If I understand correctly, we can integrate $\theta$ out of the posterior to obtain the beta- binomial distribution: $$ p(y|N,\alpha,\beta)={N\choose y} \frac{\text{Beta}(y+\alpha, N-y+\beta)}{\text{Beta}(\alpha,\beta)} $$ In our case, $\alpha=1$ and $\beta=1$ because we have a uniform prior on $\theta$. I believe that the posterior should then be $p(N|y)\propto N^{-1}\prod_{i=1}^K p(y_i|N,\alpha=1,\beta=1)$ where $K=\#(y)$ because $p(N)=N^{-1}$. But this appears to wildly diverge from Juho's answer. Where have I gone wrong?
[ -0.0034909609239548445, 0.012836357578635216, 0.0017525953007861972, 0.017349937930703163, -0.008193657733500004, -0.010095161385834217, 0.00577752199023962, -0.004639199934899807, -0.008898217231035233, -0.000733537133783102, -0.003618487622588873, 0.010577501729130745, -0.03577947616577148...
[ 0.13306526839733124, 0.24161776900291443, 0.38083386421203613, 0.022387653589248657, 0.20013131201267242, 0.4969441294670105, 0.17423798143863678, -0.3172362148761749, -0.10726839303970337, -0.5067152976989746, 0.05786019191145897, 0.3442643880844116, -0.40358519554138184, 0.35826671123504...
> **Possible Duplicate:** > "Neither Michael nor Albert is correct" or "Neither Michael nor Albert are > correct"? > Is "either you or [third-person]" followed by a singular verb or a plural > verb? If a noun phrase is made of two noun-like words that conjugate differently, then which conjugation do you use? Consider: 1) He nor I has... 2) He nor I have... "He" and "I" are connected with a conjunction. Between 1 and 2 which is correct? Are they both correct? Is neither correct?
[ 0.00861720647662878, 0.019290708005428314, -0.014271139167249203, 0.012225953862071037, -0.0009694923646748066, -0.008204220794141293, 0.009165529161691666, -0.007344387471675873, -0.014020229689776897, -0.017533890902996063, -0.006287151016294956, 0.004978707060217857, 0.0015901122242212296...
[ -0.1062321588397026, 0.14952890574932098, 0.0304703526198864, 0.024491578340530396, -0.28804856538772583, 0.42236489057540894, 0.7219815850257874, -0.15950988233089447, -0.3863297402858734, -0.48513808846473694, -0.470474511384964, 0.183091938495636, -0.29511740803718567, -0.09730117022991...
If I run pdflatex on the same .tex file, the resulting PDF files are not bitwise identical. How can I achieve that? Minimal example: \documentclass{article} \usepackage[pdftex]{hyperref} \hypersetup{ pdfcreationdate={D:20131010120000}, pdfmoddate={D:20131010120000} } \begin{document} foo \end{document} If I create two PDF files from this example, create hexdumps with _xxd_ and then compare these two hexdumps, this is the result: --- 1 2013-10-10 15:04:17.228898334 +0200 +++ 2 2013-10-10 15:04:28.152898213 +0200 @@ -537,11 +537,11 @@ 0002180: 2f53 697a 6520 3138 0a2f 5720 5b31 2032 /Size 18./W [1 2 0002190: 2031 5d0a 2f52 6f6f 7420 3135 2030 2052 1]./Root 15 0 R 00021a0: 0a2f 496e 666f 2031 3620 3020 520a 2f49 ./Info 16 0 R./I -00021b0: 4420 5b3c 3044 3938 3745 4533 3745 3642 D [<0D987EE37E6B -00021c0: 4344 4234 3838 3333 4237 4636 3434 3335 CDB48833B7F64435 -00021d0: 3134 3443 3e20 3c30 4439 3837 4545 3337 144C> <0D987EE37 -00021e0: 4536 4243 4442 3438 3833 3342 3746 3634 E6BCDB48833B7F64 -00021f0: 3433 3531 3434 433e 5d0a 2f4c 656e 6774 435144C>]./Lengt +00021b0: 4420 5b3c 3439 4644 3841 3445 3241 3439 D [<49FD8A4E2A49 +00021c0: 3434 4332 3244 4442 4536 4132 3534 3139 44C22DDBE6A25419 +00021d0: 3238 4334 3e20 3c34 3946 4438 4134 4532 28C4> <49FD8A4E2 +00021e0: 4134 3934 3443 3232 4444 4245 3641 3235 A4944C22DDBE6A25 +00021f0: 3431 3932 3843 343e 5d0a 2f4c 656e 6774 41928C4>]./Lengt 0002200: 6820 3632 2020 2020 2020 2020 0a2f 4669 h 62 ./Fi 0002210: 6c74 6572 202f 466c 6174 6544 6563 6f64 lter /FlateDecod 0002220: 650a 3e3e 0a73 7472 6561 6d0a 78da 6360 e.>>.stream.x.c` As you can see, there are still some differences, but I am not able to find out what this is. Any ideas?
[ -0.009546663612127304, 0.01852840557694435, 0.00022604758851230145, 0.026327569037675858, 0.013393545523285866, 0.007927719503641129, 0.00844519678503275, 0.025361720472574234, -0.012536901980638504, -0.052534736692905426, -0.010112429037690163, -0.0027197389863431454, 0.002469089115038514, ...
[ -0.0828537717461586, -0.026514392346143723, 0.6545765995979309, 0.07346826046705246, -0.16399885714054108, -0.07989498972892761, 0.24055789411067963, -0.3060458302497864, 0.01792418211698532, -0.97581547498703, 0.06099225580692291, 0.6136548519134521, -0.1523343026638031, 0.171237170696258...
I am applying a mixed model to predict tumor progression (y) using tumor volume (x) as the fixed effect and center ($i=1,...10$) as random intercept. The model can be written as: $y_{ij}=\alpha+\beta x_{ij}+b_{i}+\epsilon_{ij}$ I used the lme() function in R: `fit1 <- lme(PD ~ log(Volume), random = ~1 | CenterID, data=Data)` The result shows that the standard deviation of the random intercept is 0.079. Thus $b_{i}$ follows a normal distribution $N(0,0.079)$. In the meantime, I can extract the random intercept by applying ranef(fit1). This gives a list of $b_{i}$ corresponding to each center. Then I compute the standard of this vector. `sd(ranef(fit1)[[1]])` I would expect that it gives similar result as 0.079. However, it is far different. Can someone tell me why `sd(ranef(fit1)[[1]])` gives different result than the model output VarCorr(fit1)? What is exactly the relation between ranef(fit1) and VarCorr(fit1)? thanks
[ 0.011869999580085278, 0.004772696644067764, -0.007691610604524612, 0.005701462272554636, -0.01326822955161333, -0.025593217462301254, 0.005596640519797802, 0.0004635513760149479, -0.007963147014379501, -0.0034167806152254343, -0.010626360774040222, 0.010429086163640022, -0.001222430728375911...
[ 0.046744003891944885, 0.01472544763237238, 0.6002225875854492, -0.551689863204956, 0.1691100150346756, 0.4140913784503937, -0.11702628433704376, -0.3035893440246582, 0.20930388569831848, -0.5326100587844849, 0.22041678428649902, 0.18900637328624725, -0.2787279188632965, 0.2721492648124695,...
Chess: The Gathering has little to no documentation in its current state, and I have absolutely no idea what's going on. Is there a manual or tutorial for it somewhere?
[ -0.008652047254145145, 0.0036073753144592047, 0.011577948927879333, 0.030655916780233383, 0.04913415387272835, -0.02742352895438671, 0.008610235527157784, -0.032609716057777405, -0.029738718643784523, 0.004194617737084627, -0.02254437655210495, 0.014541815966367722, -0.007609858177602291, ...
[ 0.3698674142360687, 0.019398067146539688, -0.13628943264484406, 0.14767737686634064, 0.09479336440563202, -0.31086862087249756, -0.0981440544128418, 0.21721917390823364, -0.42273369431495667, -0.0005082034622319043, 0.046267393976449966, -0.1822764277458191, -0.09790506958961487, 0.2829357...
I'm writing a textbook and am using \marginpar to put a mnemonic on the margins (a triangle and a writing hand symbol from the marvosym package) for the reader to work out the derivation in the text. The mnemonic is correctly placed on odd-numbered pages but not on even-numbered ones. Here is a minimal working example--I apologize for the length, I didn't know how I could shorten it! \documentclass{book} \usepackage{marvosym} \newcommand{\dothis}{\marginpar[\Writinghand\ \Forward]{\Rewind\ \Writinghand}} \begin{document} \chapter{sample} The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog.\dothis The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. The quick brown fox jumps over the lazy dog. \dothis \end{document}
[ 0.012505554594099522, 0.014970298856496811, -0.022980347275733948, 0.034586794674396515, -0.02761993184685707, 0.015973838046193123, 0.009882243350148201, 0.006945527624338865, -0.011372990906238556, -0.022113334387540817, -0.02293434552848339, 0.0007050835993140936, 0.002651524031534791, ...
[ -0.09134665876626968, 0.20789644122123718, 0.33287692070007324, 0.16856889426708221, -0.036398787051439285, 0.02488960511982441, 0.09754085540771484, 0.12276333570480347, 0.21520638465881348, -0.4336410462856293, 0.007201865315437317, -0.02079308032989502, 0.09645994007587433, -0.240778148...
I need to have my document binded after being being printed, and so I need a 3.75 margin on every page on the side that will be binded. Is there any way of doing this?
[ 0.010073653422296047, 0.016231395304203033, -0.0210645142942667, 0.010631858371198177, 0.008879740722477436, 0.026510143652558327, 0.008056939579546452, -0.011603149585425854, -0.021356025710701942, 0.005408165976405144, 0.0018751997267827392, 0.01487184688448906, -0.013303461484611034, 0....
[ 0.5811727643013, 0.5125386714935303, 0.7718675136566162, 0.14444109797477722, 0.24930225312709808, -0.18536601960659027, 0.24748165905475616, -0.5262174606323242, 0.05074358731508255, -0.6317322850227356, 0.2714208960533142, 0.16983073949813843, 0.09703756868839264, -0.04934793710708618, ...
Most everyone has probably heard the phrase "tall, dark, and handsome" being used to describe the physical qualities of a perfect male romantic match. Where did this phrase come from, and who, if anyone, made it popular?
[ -0.00613036472350359, 0.03148243576288223, 0.005452245008200407, -0.006756410468369722, -0.0142375398427248, -0.006207433994859457, 0.014201772399246693, -0.024929242208600044, -0.03012845292687416, -0.06797824800014496, -0.019147787243127823, 0.024891795590519905, 0.028482627123594284, 0....
[ 0.5869245529174805, 0.14062033593654633, -0.0189573485404253, 0.2567054331302643, 0.31458771228790283, -0.1394178569316864, 0.255577027797699, 0.4046829640865326, -0.10206227004528046, 0.12859168648719788, -0.2804010510444641, 0.22321107983589172, 0.0012439346173778176, 0.19671174883842468...
I am new in Mapserver, now I want to create maplayer , so its required shapefile for this, and if yes how can i used it. Map file is given below eg2.map MAP NAME ENG1 STATUS ON SIZE 800 500 IMAGETYPE GIF IMAGECOLOR 240 240 240 SHAPEPATH "/www/geog5870/geog5870test/data" EXTENT 87000 8000 660000 660000 UNITS METERS WEB IMAGEPATH "/www/tmp/" IMAGEURL "/tmp/" END LAYER NAME "EnglishGOR" TYPE POLYGON STATUS DEFAULT DATA "England_gor_2001_area" CLASS NAME "English Government Office Regions" OUTLINECOLOR 60 60 60 COLOR 255 255 0 SYMBOL 0 END END END
[ -0.004087598528712988, 0.010605985298752785, 0.007287340238690376, -0.0015308447182178497, 0.007869048044085503, 0.025458132848143578, 0.005700547713786364, -0.018428929150104523, -0.012107913382351398, -0.026026464998722076, 0.011296934448182583, 0.014241929166018963, 0.01720476895570755, ...
[ -0.3269922733306885, 0.006296245846897364, 1.0760024785995483, 0.03214675188064575, -0.4357620179653168, 0.05111939087510109, 0.1283341646194458, -0.1781044900417328, -0.26483607292175293, -1.0199419260025024, -0.14187775552272797, 0.5565052032470703, 0.27980196475982666, 0.027284801006317...
The title says it all, basically I'm wondering if Set Items are more likely to drop from specific bosses, or are they just as likely to drop anywhere any other rare is likely to drop? Note: I know bosses have more chances to drop magic items, but I'm more curious if certain sets are supposed to drop from certain bosses or not.
[ 0.045754481106996536, 0.01742532104253769, 0.00962736364454031, 0.013969315215945244, 0.015640847384929657, 0.02530038170516491, 0.009235582314431667, 0.005839808378368616, -0.026032129302620888, 0.029784532263875008, -0.011947720311582088, 0.021808359771966934, 0.020334850996732712, 0.013...
[ 0.40482205152511597, -0.026811450719833374, -0.2829110324382782, 0.605889618396759, -0.5205948948860168, -0.642086923122406, 0.015348193235695362, 0.4106474816799164, -0.2881182134151459, -0.0547175295650959, -0.1274460405111313, 0.18445144593715668, 0.32721462845802307, 0.2951698005199432...
I have a list of IP address that I need to SSH to and log if I could get to them. I would use a Ping script, but DNS might have already reassigned the address. I do not care about the ones I can't get to. If the SSH works I need to log that IP address, I shouldn't have to worry about passwords or keys because I don't want to log in to the box's. I just want to see if I can SSH to them. So far i have: touch logfile_$(date "+%Y%m%d%T") IP_FILE="path/to/ip_address.txt" LOGFILE="path/to/logfile_$(date "+%Y%m%d%T")" if [[ ! -f ${IP_FILE} ]]; then echo "Cannot find IP address!" exit 1 fi for IP_ADDRESS in `cat $IP_FILE` ; do ssh $IP_ADDRESS >> $LOGFILE 2>&1 I am still new to scripting and any help would be great. I have looked at "How to check if I can log in to server via ssh?" posted by LanceBaynes but its not really what I need. I can't install anything on this boxes. Or I would have downloaded a nice scanner. Like Netcat or Nmap both of which are great options.
[ 0.01351393386721611, -0.00011679483577609062, -0.0005106383468955755, -0.0010423375060781837, -0.03155595809221268, 0.037633016705513, 0.007404591888189316, -0.002740312134847045, -0.02095503732562065, 0.004842807538807392, -0.013762347400188446, -0.0001815510622691363, -0.00622581085190177,...
[ 0.40865638852119446, 0.1313236951828003, 0.5292896032333374, -0.04289815574884415, 0.14854253828525543, 0.20854559540748596, 0.4141720235347748, 0.1320216953754425, -0.4057391285896301, -0.8856970071792603, 0.011888528242707253, 0.5535492897033691, -0.1566939800977707, 0.28989240527153015,...
I just found out about the Math Input Panel in Windows 7,and now there is commercial software using it for LaTex writing. For example Inlage. If you have a look of their first demo video, you know what I am talking about: http://www.inlage.com/videos I had a try with the Math Input Panel myself, and it seems you can basically write all math stuff, integration, super(sub) scripts, tensor, arrows (even with labels over them!), and matrix! The only big problem is there is NO commutative diagram. And it probably has trouble recognizing some math fonts, like \mathfrak or \mathcal. But in all, it really recognize handwritings pretty well. I don't have a writing pad, so I just tried writing with a mouse. I know there are people taking math notes using a tablet PC. Won't this tool drastically improve the quality of our note-taking? It basically changes all handwriting into TeX files! Maybe I am too late on this, are there more mature product for such purposes? I think handwriting math could actually be slower than typing LaTeX codes. But one good reason for doing handwriting is because sometimes I just don't like to make math writing into code writing (or something like programming). I would like to hear about your comments. Thank you!
[ -0.008249832317233086, -0.005337061360478401, -0.021091550588607788, 0.0062487367540597916, -0.017400288954377174, -0.010167702101171017, 0.008462655358016491, -0.005230090580880642, -0.01675693318247795, -0.008500427007675171, -0.010273769497871399, 0.009816128760576248, 0.00816529430449009...
[ 0.19940464198589325, 0.03140764683485031, 0.2699666917324066, 0.07729025930166245, -0.22410008311271667, -0.0957312285900116, -0.001058280118741095, 0.18560129404067993, -0.48989805579185486, -0.627808690071106, 0.2633208930492401, 0.50590980052948, -0.07294294238090515, -0.238669484853744...
I am using the nls procedure in R to fit a logistic growth model. In their SSlogis function, José Pinheiro and Douglas Bates chose the formulation Asym / (1 + exp((xmid-input) / scal)) for their model. As I am fairly inexperienced with the numerical properties of such models, I wonder: * Can somebody explain why the authors chose this formulation instead of possible alternatives? In particular, ecologists seem to prefer a model with initial population, carrying capacity and growth rate. Does the formulation above have favourable numerical properties? * It seems that when the model is misspecified and the data are actually fairly linear with time, this formulation often fails to converge. Could such a problem be avoided? * Is parameter orthogonality a key concern here or are other aspects of the model more important? * Is it trivial to extend this model to allow a flexible intercept? Would the following model provide sensible numerical properties? Intercept + (Asym - Intercept ) / (1 + exp((xmid-input) / scal)) I am, of course, open for alternatives as long as it allows for some flexibility in intercept, location where 50% of the growth has been achieved and asymptote.
[ -0.0022134927567094564, 0.014556997455656528, 0.0008175993571057916, 0.017181893810629845, -0.009449031203985214, -0.011676761321723461, 0.009773992002010345, 0.003693395759910345, -0.008924998342990875, -0.029002148658037186, -0.01618444174528122, 0.0027558228466659784, -0.01581062749028206...
[ 0.12304119765758514, -0.01599273644387722, -0.015501280315220356, -0.1340995728969574, 0.21936102211475372, 0.5683295726776123, -0.27125000953674316, -0.24503450095653534, -0.31503888964653015, -0.18195344507694244, 0.03542611375451088, 0.027348507195711136, -0.21381570398807526, 0.4154214...
Is it possible to detect if my visitors are coming to my website via a Google Adwords Ad? It would like to change the telephone number for those users only, that way I can measure the success of my adwords campaign with the number of calls.
[ -0.008546847850084305, 0.018119940534234047, -0.01186500582844019, 0.029646465554833412, -0.0016216130461543798, -0.00859713926911354, 0.012604126706719398, -0.0205787755548954, -0.026323281228542328, -0.03789430484175682, -0.004317348822951317, 0.014341284520924091, 0.012522654607892036, ...
[ 0.462106853723526, 0.17821066081523895, 0.5385193228721619, 0.041054125875234604, -0.014795400202274323, -0.11598622053861618, 0.45391136407852173, 0.17327429354190826, -0.2764923572540283, -0.3342267572879791, 0.3008715808391571, 0.2809087634086609, -0.0047605703584849834, 0.2779852747917...
I'm having a problem with the language in SWP 5.5. I am writting a text in spanish, so I am using the `\usepackage[spanish]{babel}` sentence at the beginning of the document and it works perfectly with all the format. However, the word 'Preface' still appears in english after the ToC in the pdf output. Any help is welcome.
[ -0.0071371146477758884, -0.0004512128361966461, -0.002383062383159995, 0.03394445776939392, -0.006919410545378923, 0.04011538624763489, 0.009975168853998184, 0.006101980339735746, -0.01803530752658844, -0.005289493128657341, -0.014551072381436825, 0.01572836935520172, -0.00010529565770411864...
[ 0.21511584520339966, 0.5342313051223755, 0.5646028518676758, -0.062053848057985306, 0.10186926275491714, -0.5019188523292542, 0.15499760210514069, 0.44471848011016846, 0.11685243248939514, -0.5749515295028687, -0.15760384500026703, 0.7010255455970764, -0.08322200924158096, -0.2461445778608...
I am part of a team tasked with performing exploratory analysis of a large data set containing neuro-imaging scans. For each scan I will likely calculate some variable that relates to brain function - for example functional connectivity. I may have 20 such statistical values per subject. The dataset has hundreds of subjects (1000 actually), so there is a lot of data. We are interested in finding relationships between various factors, such as age, health, socioeconomic status, there are 100s of measures taken for each subject. Lets say we want to data-mine this dataset to find what it is about a person's lifestyle that improves functional connectivity. Q1) What do you think the correct approach would be? I am thinking of splitting the data into a training set of data, in order to try to find patterns that relate to brain function, and a testing set, in order to test those models and hypotheses that are generated in the first phase. Anyone have any better ideas than this. Q2) If you think the proposed method of splitting the data set is a good idea, then which methods do you recommend for finding patterns? I have done a bit of searching and found articles on decision trees etc. My boss suggested just trying to correlate everything with everything and look for patterns but I suspect that this is a bad idea and I will likely miss many interactions and non-linear relationships. Q3) Which software do you think I should use? I am very familiar with MATLAB but don't mind learning R. Other than that, it would probably need to be free software as the department I work in is unlikely to want to pay for anything. Thanks
[ 0.010008344426751137, 0.0065882764756679535, -0.009018883109092712, 0.0047476585023105145, 0.020231405273079872, -0.0032579696271568537, 0.00687328539788723, 0.001754014752805233, -0.00731160631403327, 0.0048905787989497185, 0.00781651958823204, 0.009594685398042202, -0.001884564058855176, ...
[ 0.8194690346717834, 0.37119701504707336, 0.3290855884552002, 0.06874886155128479, 0.09304730594158173, 0.2585686147212982, 0.37467390298843384, -0.08334533870220184, -0.10354700684547424, -0.5764358043670654, 0.16659501194953918, 0.23047728836536407, 0.4598996043205261, 0.2549161911010742,...
I had a job interview today for a Job position as developer on an important site. They asked tonnes of programming language related questions, which I managed to answer without problems, but then they started asking question about how TCP/IP requests were made once I made a request on my PC to a web server. I did received those contents as a student, but I dont remember them well, because I'm working mostly in web development, my question is: As a software developer, mainly working on web applications, do I need to have **extensive** knowledge of TCP/IP and how routers manage requests or it's just black box knowledge to me?
[ 0.002491558901965618, 0.004128455184400082, -0.010702187195420265, -0.013154137879610062, -0.021353211253881454, 0.01907467469573021, 0.007589433807879686, -0.003810908179730177, -0.01609298214316368, -0.025104593485593796, 0.0010576109634712338, 0.019311990588903427, -0.00405096635222435, ...
[ 0.6562482714653015, 0.517636775970459, 0.2504124045372009, 0.10681036859750748, -0.04847754165530205, -0.22249405086040497, -0.018301935866475105, -0.013323407620191574, -0.2830335199832916, -0.3966045677661896, 0.24095073342323303, 0.23960725963115692, 0.18816320598125458, 0.0472738407552...
I want to do reverse geocoding from MySQL or PostGIS SQL. I have shapefile up to town level of India. So I have a geometry boundary for each town. What I want is to first find the nearest point from the current position and check that point is in which towns. I want to get the result in this format: Towns, District, State, Country I dont have knowledge of PostGIS SQL but I know MySQL. But if anyone can help me I will use the PostGIS SQL also. So I want fastest way to do reverse geocoding. Thank you.
[ 0.0002549622440710664, 0.01555743906646967, -0.011319804936647415, 0.01389341987669468, -0.017790479585528374, 0.02059771493077278, 0.009366273880004883, 0.020768368616700172, -0.01584150269627571, -0.004822432063519955, -0.0006534084677696228, 0.01299216691404581, 0.0008627197821624577, 0...
[ -0.03545914962887764, 0.37689676880836487, 0.4136979877948761, 0.06252096593379974, -0.08069714903831482, 0.4575812518596649, -0.004503630101680756, 0.15786220133304596, 0.1640123575925827, -0.33751380443573, 0.39910900592803955, 0.08058755099773407, 0.05640460550785065, 0.4542731642723083...
Is _That dog don't hunt_ an American slang expression? What does it mean exactly and where does it originate? If possible, please give some examples.
[ -0.021649107336997986, 0.014293366111814976, -0.005193724296987057, 0.028134988620877266, 0.03437225520610809, 0.0006223007221706212, 0.009108412079513073, 0.030869873240590096, -0.028414011001586914, 0.004398529417812824, 0.012554764747619629, 0.01988350786268711, 0.02242756448686123, -0....
[ 0.6310744881629944, 0.31268826127052307, -0.41606220602989197, -0.10346050560474396, 0.07292718440294266, -0.04757663235068321, 0.5626358389854431, 0.47704315185546875, -0.34351521730422974, -0.020080724731087685, -0.3488699495792389, 0.13254709541797638, -0.8767733573913574, 0.18993382155...
How can I completely remove Java 6 and install Java/JDK 7 on Linux Mint 13?
[ 0.06297297030687332, 0.016477663069963455, -0.06019780412316322, -0.012357774190604687, 0.019717447459697723, 0.024749351665377617, 0.020437749102711678, -0.045823171734809875, -0.02465224266052246, -0.07926065474748611, -0.04181508719921112, 0.07072052359580994, -0.042530596256256104, 0.0...
[ 0.3945847749710083, 0.20251935720443726, 0.04757343605160713, 0.023275509476661682, -0.04862753674387932, -0.4971751272678375, 0.5623193383216858, 0.009766495786607265, 0.10158877074718475, -0.46450555324554443, -0.03926856070756912, 0.5847048759460449, -0.24015578627586365, -0.02896751463...
Can anyone give me some pointers (code would be helpful) on how to go back to the extent a user was at after they have navigate away and clicked the back button? So I'm viewing a map at X extent I click on some link and page changes, on clicking the back browser button I want to be able to get back to the Map Extent I was just before the page changed. Any help much appreciated.
[ 0.004266703967005014, -0.0011786367977038026, 0.002267469186335802, -0.010866165161132812, 0.02097807452082634, -0.017567144706845284, 0.00825936533510685, 0.0161273330450058, -0.02757626213133335, -0.0011593009112402797, 0.006052388343960047, 0.014549988321959972, 0.00044425795204006135, ...
[ 0.265919029712677, 0.10043150931596756, 0.31365007162094116, 0.4269687235355377, -0.16251659393310547, -0.21150429546833038, 0.3123611509799957, 0.4805845618247986, -0.3460151255130768, -0.6979339718818665, 0.21282345056533813, 0.49924296140670776, 0.44622886180877686, 0.3311786353588104, ...
I have very much enjoyed William L. Burke's _Applied Differential Geometry_. Reading around on the web it seems that he discovered something which is called the (retarded) Burke Potential, but I have not been able to find any references to it. Can anyone point me at a reference or explain succinctly what it is?
[ -0.024119621142745018, 0.008525650948286057, -0.007217359729111195, 0.011101286858320236, -0.0012886046897619963, -0.03217396140098572, 0.006482576485723257, -0.026802444830536842, -0.014962485060095787, -0.019783567637205124, -0.007797427475452423, 0.013847817666828632, -0.00689223548397421...
[ 0.5225549340248108, 0.49101006984710693, -0.33424416184425354, -0.043536290526390076, -0.08731275051832199, -0.26682862639427185, 0.008661434054374695, 0.18502606451511383, -0.1283825933933258, -0.600975751876831, -0.10773531347513199, 0.16432632505893707, -0.07192686945199966, 0.450070172...
I don't believe there's a `is_new_year()` function in WordPress, as an alternative for `is_new_day()`. I'm trying to display the year only once for each set of posts, what would be the best way to do this? Should I just check the date manually with PHP in the loop or are there any better ways?
[ 0.010620640590786934, 0.011883181519806385, -0.012637315317988396, 0.019594881683588028, -0.02096211351454258, 0.017775889486074448, 0.007412446662783623, 0.020990461111068726, -0.02581881918013096, 0.0014705342473462224, -0.007928765378892422, 0.009447801858186722, 0.021464528515934944, 0...
[ 0.5125927925109863, 0.15404686331748962, 0.6149696111679077, 0.02263893000781536, -0.2256782352924347, -0.23317155241966248, 0.4535062313079834, 0.613351583480835, -0.47728076577186584, -0.5507763624191284, 0.40130504965782166, -0.1051659807562828, 0.18678626418113708, 0.35242587327957153,...
(I modified the question. i .e. I added hyperlink) My website uses Jobroller theme. The problem is that it is not compatible with WPML I am translating themes using codestyling localization [Link to the code. It is compressed][1] Could you give me some instructions on what I need to do from here? Jobroller does not seem to be compatible with codestyling localization. If I am using an incompatible them, is there a way to achieve compatibility? ( I have added some of the source codes) [1]: Code http://cafeattach.naver.net/a732bb081c2c2d9db35c320338dca1da782ad1f4/20130527_261_cafefile/24495556_1369658452761_pDfySc_zip/sample_source.zip?type=attachment
[ -0.0032025063410401344, 0.013059519231319427, -0.0006556311855092645, 0.019516080617904663, -0.009521649219095707, 0.024787675589323044, 0.007181535474956036, 0.017083045095205307, -0.013706689700484276, -0.022325728088617325, -0.01880093291401863, 0.009504931047558784, 0.0018388602184131742...
[ 0.2615300416946411, 0.2384277880191803, 0.4430485963821411, -0.0991961732506752, -0.13312900066375732, -0.1586262583732605, 0.23835410177707672, 0.5772224068641663, -0.5873684287071228, -0.7365803122520447, 0.04442298784852028, 0.06247860565781593, -0.16880016028881073, 0.04820306226611137...
I created a new custom post type 'Projects' and want the archive of all posts of this type to be available at mysite.com/projects. At the moment all project single posts are shown with a slug as follows mysite.com/projects/project- title, but when I go to mysite.com/projects I get a 404. Here is how I built the custom post type: /* Create the Project Custom Post Type ------------------------------------------*/ function create_post_type_project() { $labels = array( 'name' => __( 'Projects' ), 'singular_name' => __( 'Project' ), 'add_new' => __('Add New'), 'add_new_item' => __('Add New Project'), 'edit_item' => __('Edit Project'), 'new_item' => __('New Project'), 'view_item' => __('View Project'), 'search_items' => __('Search Project'), 'not_found' => __('No project found'), 'not_found_in_trash' => __('No project found in Trash'), 'parent_item_colon' => '' ); $args = array( 'labels' => $labels, 'public' => true, 'exclude_from_search' => true, 'publicly_queryable' => true, 'show_ui' => true, 'query_var' => true, 'capability_type' => 'post', 'hierarchical' => false, 'menu_position' => null, // Uncomment the following line to change the slug; // You must also save your permalink structure to prevent 404 errors 'rewrite' => array( 'slug' => 'projects' ), 'has_archive' => true, 'supports' => array('title','editor','thumbnail'), ); register_post_type(__( 'project' ),$args); }
[ 0.013041377998888493, 0.008022462949156761, 0.010992562398314476, 0.019159486517310143, 0.04025439918041229, 0.002917246427386999, 0.007520467042922974, 0.03152690827846527, -0.015392884612083435, -0.002156251110136509, 0.0026287587825208902, 0.01478767953813076, 0.007509130984544754, 0.01...
[ 0.874100923538208, 0.09718533605337143, 0.2583826184272766, 0.06262270361185074, -0.1533806025981903, 0.09022267162799835, 0.21769633889198303, -0.03685125336050987, -0.47290387749671936, -0.3345552682876587, 0.4657721519470215, 0.08387520909309387, 0.09687281399965286, 0.4370885193347931,...
Is there some automated way to test broken urls over a LaTex document for errors such as 404 etc. I noticed that some broken urls but the list of reference is too massive to go through one-by-one. Ideas how to solve this kind quality control problem fast? I have some ideas to solve this: 1. open up writing-process like in SO, with watchful peers, bugs tend become shallow 2. some at least partly automated processing of the document, perhaps something like apache-logging or something in that degree? (this is mere speculation and I have no idea how this could be done for a LaTex document to find and process broken material -- perhaps breaking up the large document into some blog series for peer-review where anyone can comment but it may become a larger problem of data-management) 3. other idea?
[ 0.01910359039902687, 0.013891022652387619, -0.003447724971920252, 0.010840913280844688, 0.01487734541296959, 0.005062376148998737, 0.006662547588348389, 0.017887277528643608, -0.014976520091295242, 0.002699329750612378, -0.01179315336048603, 0.005218948237597942, -0.010686248540878296, 0.0...
[ 0.3143877685070038, 0.11821809411048889, 0.07978261262178421, 0.5553542971611023, -0.106718510389328, -0.2742530107498169, 0.32044145464897156, 0.14739936590194702, -0.25916722416877747, -0.33859461545944214, 0.19540399312973022, 0.4884267747402191, -0.14180488884449005, -0.315525650978088...
Yesterday, I tested Debian Live 7.6 with non-free drivers using a USB stick. I used that because the completely-free version was not recognizing my wifi and video drivers out-of-the-box. Everything went fine, got everything working - audio, video, wifi, card reader, etc... so I decided to install. I restarted my laptop and began the installation process, which was successful. When starting the machine after the installation process had finished, the only thing I got was a black screen. The `tty`s where available and working, though. I was able to manage the issue upgrading to testing - that was my goal after installing the OS in the first place, so I didn't even take the time to properly troubleshoot the issue. So, I just changed the `sources.list` repositories to point to `testing` and issued an `aptitude safe-upgrade`. After that, everything was working. **Why the machine worked fine while using the Live USB stick, but then it failed to work when the OS was finally installed in my machine?** Does the Live distro use some other way to recognize the hardware, so it's able to handle more scenarios than usual? Does it do some kind of fail-safe run in order to ensure the system can load and be usable? I'm not trying to debug the problem here - I'm just trying to understand why the Live system have worked, but then when the OS was installed it's not. What have changed between the 2 runs?
[ -0.019211463630199432, -0.0007072305306792259, -0.011641399934887886, 0.00028915354050695896, -0.024389345198869705, -0.016873879358172417, 0.007045876234769821, 0.0034089998807758093, -0.0120621956884861, -0.02458593063056469, -0.019473660737276077, 0.010930037125945091, -0.0032881717197597...
[ 0.8267502188682556, 0.1334303915500641, -0.03414861112833023, -0.013552512042224407, 0.023084314540028572, -0.2524227797985077, 0.32585886120796204, 0.013779086992144585, -0.13337156176567078, -0.39071395993232727, -0.04858763888478279, 1.1926696300506592, -0.5205621719360352, -0.103408880...
I'm using MinionPro for the body of a document and Myriad for figure/table captions. Because I can't load packages MinionPro and Myriad at the same time, I load the Myriad fonts for the sans-serif fonts using a `\renewcommand{\sfdefault}{\Myriad-LF}` call in the preamble. Then I set the figure/table captions to use the sans-serif font using package `caption`. This works great, except that the en dashes (`\--`) and em dashes (`\---`) are not properly interpreted as long dashes; they're displayed as two or three hyphens. A minimal working example is below. Any ideas? \documentclass[12pt]{article} % Set fonts (Minion Pro, Myriad): \usepackage[lf]{MinionPro} % 'lf' option to use full-height (lining) figures % We can't use package Myriad here as it conflicts with MinionPro, so % instead we declare the Myriad fonts manually. % Set body sans serif (\textsf) text to Myriad with lining figures: \renewcommand{\sfdefault}{Myriad-LF} % Set math sans serif (\mathsf) alphabet to Myriad with lining figures: \SetMathAlphabet{\mathsf}{normal}{OML}{Myriad-LF}{m}{n} \SetMathAlphabet{\mathsf}{bold}{OML}{Myriad-LF}{b}{n} \usepackage[margin=10pt,font={small,sf},labelfont={sf,bf},labelsep=endash]{caption} \begin{document} Some text. \begin{figure} \centering \ldots \caption{The mass of the coal storage piles ranged from 8--20 megatonnes.} \label{fig:coal} \end{figure} \end{document} This outputs the following, with the incorrect dash highlighted. ![mwe image](http://i.stack.imgur.com/u99WH.png)
[ 0.027442488819360733, -0.0038279807195067406, -0.006085903383791447, 0.030278559774160385, 0.005531457252800465, 0.006098489742726088, 0.011961681768298149, 0.0037335273809731007, -0.01682433858513832, -0.009735251776874065, -0.026307201012969017, -0.011598790995776653, 0.006332197692245245,...
[ 0.23235610127449036, 0.018497634679079056, 0.7507133483886719, 0.15914848446846008, 0.09324108064174652, 0.14558273553848267, -0.06022731959819794, -0.274616539478302, 0.060370832681655884, -0.673036515712738, -0.3413195312023163, 0.5121628046035767, -0.08790940791368484, -0.25826409459114...
I am not great at using the `mail` command on Linux. How can I get better? I don't even know what program this is. Is it pine? I doubt it. It is in `/usr/bin/mail` and when I start it is tells me that it is `Version 8.1.2 01/15/2001`. When I type `list` it tells me Commands are: next, alias, print, type, Type, Print, visual, top, touch, preserve, delete, dp, dt, undelete, unset, mail, mbox, pipe, |, more, page, More, Page, unread, Unread, !, copy, chdir, cd, save, source, set, shell, version, group, write, from, file, folder, folders, ?, z, headers, help, =, Reply, Respond, reply, respond, edit, echo, quit, list, xit, exit, size, hold, if, else, endif, alternates, ignore, discard, retain, saveignore, savediscard, saveretain, core, #, inc, new If I type in `?` it tells me this: Mail Command Description ------------------------- -------------------------------------------- t [message list] type message(s). n goto and type next message. e [message list] edit message(s). f [message list] give head lines of messages. d [message list] delete message(s). s [message list] <file> append message(s) to file. u [message list] undelete message(s). R [message list] reply to message sender(s). r [message list] reply to message sender(s) and all recipients. p [message list] print message list. pre [message list] make messages go back to /var/mail. m <recipient list> mail to specific recipient(s). q quit, saving unresolved messages in mbox. x quit, do not remove system mailbox. h print out active message headers. ! shell escape. | [msglist] command pipe message(s) to shell command. pi [msglist] command pipe message(s) to shell command. cd [directory] chdir to directory or home if none given fi <file> switch to file (%=system inbox, %user=user's system inbox). + searches in your folder directory for the file. set variable[=value] set Mail variable. I type `z` and it shows me the end of the list of messages - but that command is not showing up in the `?` help page. My biggest questions are: 1. What program is this? 2. How can I learn it better? Are there some links to tutorials? 3. **MOST IMPORTANTLY How can I get back to the message list after I run another command (or refresh the message list, or move up and down in the message list??)** 4. What are the most command commands, or what are the most helpful tricks to using it? 5. What is the opposite command of `z` for scroll down the list page by page? **Clarification** : I'm asking about how to use the interactive program and not the scriptable command - i.e. what happens when you just type `mail` with no flags or parameters. I only seem to be able to find tutorials online of how to script with the program, not how to use it. I know it hasn't been updated since 2001, I would just like to get a little more efficient at using it and I'm having a hard time finding resources to do that.
[ -0.014579295180737972, -0.009705813601613045, -0.015208946540951729, -0.0030997642315924168, 0.014660785906016827, 0.0016772029921412468, 0.008098602294921875, 0.008883551694452763, -0.016166601330041885, 0.017892535775899887, -0.008071890100836754, -0.007818732410669327, -0.0006360568804666...
[ 0.09500135481357574, 0.05960036814212799, 0.32991206645965576, -0.16088427603244781, 0.12596262991428375, 0.10524124652147293, 0.4148082435131073, 0.3096047341823578, -0.34018850326538086, -0.7079427242279053, -0.08800230920314789, 0.7541364431381226, -0.2682899236679077, 0.119379818439483...
The word "humiliate" carries very negative connotations; is there a verb for "making someone humble" or "giving someone humility" in a positive way? For instance, "Having children of my own _humiliated_ me" isn't the best choice of words.
[ 0.003585333237424493, 0.014059356413781643, -0.010707472451031208, 0.04168058931827545, -0.012698628939688206, 0.0023294384591281414, 0.012052466161549091, 0.023675676435232162, -0.024851538240909576, 0.004205627832561731, -0.006552914157509804, 0.01410017628222704, 0.011085566133260727, -...
[ 0.07156269252300262, 0.29394879937171936, -0.35574084520339966, -0.18497715890407562, -0.6926620006561279, 0.24626211822032928, 0.8058229088783264, 0.49335977435112, -0.006685085594654083, -0.30158892273902893, -0.04012996703386307, 0.39334097504615784, -0.23058651387691498, 0.402063846588...
It seems to me it would be very useful to use Javascript for general server side scripting tasks as it has more or less the same features as Perl and Python. But AFAIK there are no generally available Javascript interpreters for the major machine architectures. I guess the other problem may be lack of libraries but surely these would come if the interpreters were there. Google's V8 maybe could be a starting point. Does anyone think we'll see this soon?
[ 0.006079400423914194, -0.001966421725228429, -0.013032980263233185, -0.009622227400541306, -0.019385501742362976, -0.004244667012244463, 0.005300641525536776, 0.0245964452624321, -0.01706065610051155, -0.011132030747830868, -0.006143980659544468, 0.012702099047601223, 0.0025751988869160414, ...
[ 0.476656049489975, 0.3086191415786743, -0.03737437725067139, 0.21571020781993866, -0.159499391913414, -0.08207543939352036, 0.31944698095321655, 0.3510389029979706, -0.14660276472568512, -0.5577566027641296, 0.207073375582695, 0.6756898164749146, -0.12015870213508606, -0.07659316807985306,...
My Question is not about how to use HTTP session or how values are being used in session. Question is more oriented about approach. In my current application there are few values being stored in the Session and we require those values every now and than, here are few values we fetching from session 1. Store 2. Language 3. Customer 4. few others I can fetch these values from session like request.getSession().getAttribute( "name" ) I need those values not directly in my Controller, but in other layer (say Facade) so I have 2 options 1. Pass each required values to Facade method by `request.getSession().getAttribute( "name" )`. 2. Create a method in each Facade Class like. Method in Facade private Object getSessionValue(final String key) { ServletRequestAttributes reqAttr = (ServletRequestAttributes) RequestContextHolder .currentRequestAttributes(); return reqAttr.getRequest().getAttribute(key); } Though both are working find in my case, but with one exception, I need to have this method on each Facade Class or need to fetch each object in Controller. Is there any other way or approach which might be more clean or my current approach is fine
[ 0.0032338101882487535, 0.014156118035316467, -0.0016500300262123346, -0.01006239838898182, -0.0038689710199832916, -0.013101665303111076, 0.00714206974953413, -0.006708970759063959, -0.011593318544328213, -0.010022900998592377, -0.0022544581443071365, 0.014932396821677685, -0.005002000369131...
[ 0.28604981303215027, -0.25347182154655457, 0.5079060196876526, 0.2197657972574234, -0.15908032655715942, -0.06224764883518219, 0.4276311993598938, -0.5131732821464539, -0.062292296439409256, -0.7770798802375793, -0.0884915366768837, 0.7236102819442749, -0.06176895648241043, -0.071645900607...
What are the most useful features of Google Analytics that are not known to common users? I know this is a typical question for many categories in stackexchange. I think Google Analytics deserves that question. There are so many things to learn about it. It would be very useful for many people if you can share your favorite hidden gems of Google Analytics.
[ 0.02480154298245907, 0.0040175109170377254, -0.0100769167765975, 0.01765715517103672, 0.022205883637070656, -0.00961946789175272, 0.007987054996192455, -0.020045680925250053, -0.034014880657196045, -0.03597720339894295, 0.005048389080911875, 0.00933690182864666, -0.00023704847262706608, 0....
[ 0.6992018818855286, 0.1294746696949005, -0.06487923860549927, 0.377475380897522, -0.016290223225951195, -0.0629829466342926, -0.000907358480617404, 0.5168241858482361, -0.28536155819892883, -0.33193156123161316, 0.27408960461616516, 0.6996831893920898, 0.1489970088005066, 0.350857377052307...
1. How can I see how apps are using _Internal Storage_? 2. How can I see percistent/constant usage? 3. How can I see how much applications on SD-card would use extra, if I moved them to the phone. I am referring to /data, Which Link2SD calls Internal (Storage). There is also "System" (/system) and "Cache" (/cache), which are also probably internal in the sense that they are not on the SD - but my problem is the Internal as defined above.
[ 0.005367062520235777, -0.0008899586391635239, 0.0006005078321322799, 0.012903479859232903, -0.0031936410814523697, -0.013410552404820919, 0.008285828866064548, 0.025997258722782135, -0.009729313664138317, 0.0038660550490021706, -0.013740279711782932, 0.0022089548874646425, 0.0132728451862931...
[ 0.3312685489654541, 0.29646968841552734, 0.2906912863254547, 0.10873492062091827, 0.17723383009433746, 0.211382195353508, -0.11245299130678177, -0.21211989223957062, -0.4736025333404541, -0.546855628490448, 0.004603206180036068, 0.5512441396713257, -0.0946313887834549, -0.176198810338974, ...
So I have a property field called 'offset', type=integer, range=0-359. The intent is to draw an svg symbol that it offset 5 units radially away from the feature coordinate at the angle (in degrees) specified in the 'offset' field. Basically if 'offset' is `45` and the feature coordinate is `0,0` then the symbol will be drawn at `x=5*acos(45) y=5*asin(45)` (sorry if the trig is wrong for the moment, you get the idea). If 'offset' is `90` and the feature is at `0,0` then the symbol will be at `x=5 y=0`. SO: the question is how to write this expression in the Data Defined Properties of the marker in the style dialog box? (After hitting Data Defined Properties in the SVG marker dialog, check the 'Offset' box and no others...) Ignoring the formulas for a minute, to get the syntax straight, I'm trying these on an object with 'offset' value = 5. I've tried: `"offset","offset"` -> not a valid expression `'"offset","offset"'` -> valid expression, but no actual offset `concat(tostring("offset"),',',tostring("offset")`) -> valid expression, but no actual offset `concat(tostring(offset),',',tostring(offset))` -> valid expression, but no actual offset Several of the attempts give 'Expression is invalid' under 'Output preview' with the following, or similar, under 'more info': Parser Error: syntax error, unexpected COMMA, expecting $end Eval Error: No root node! Parsing failed? As a reality check, the following expressions DO work as expected: '5,5' concat(tostring(5),',',tostring(5)) But the following gives the 'Expression is invalid' error: `5,5` So, apparently it does want a string? Can anyone provide the correct syntax? Thanks!
[ -0.0022266176529228687, 0.009535730816423893, -0.011921796016395092, 0.013285672292113304, -0.037166696041822433, 0.003343665972352028, 0.008625523187220097, 0.015692809596657753, -0.01129130832850933, 0.006295470520853996, -0.007437506690621376, 0.011062569916248322, 0.0017423850949853659, ...
[ 0.34440869092941284, 0.008813641034066677, 0.8880708813667297, 0.026912355795502663, -0.00419093482196331, 0.13572457432746887, -0.4499117136001587, -0.5400435924530029, -0.15848128497600555, -0.49216076731681824, 0.19114257395267487, 0.5886663794517517, 0.15016643702983856, 0.192621976137...
I'm a biologist trying to verify that the age, gender, and family group of my samples from a population of whales is representative of the population, in order to make inferences in the discussion. I've been in the field a long time and cannot remember what tests I use to compare goodness of fit in a known population. I am also confused if I should be using two different tests for the continuous variables (age) and the categorized ones (gender and family). My data are below. Could someone give me a clue? Samples Population Gender F 18 49 M 11 36 UK 3 3 Family J 13 26 K 7 20 L 3 42 UK 9 0 ![age distribution here](http://i.stack.imgur.com/6K1VK.png)
[ -0.01928277686238289, 0.01088304165750742, -0.009602892212569714, 0.02185840532183647, 0.006557968445122242, 0.014030929654836655, 0.00958559662103653, 0.006281848531216383, -0.014492698013782501, -0.009601401165127754, 0.008134478703141212, 0.006509169936180115, 0.013582105748355389, 0.00...
[ 0.6638581156730652, 0.29614582657814026, -0.062468208372592926, -0.05096274986863136, 0.03402796387672424, 0.5022516846656799, 0.18233883380889893, -0.22762702405452728, -0.35554927587509155, -0.33687904477119446, 0.2746771275997162, 0.18310140073299408, 0.06402294337749481, 0.441419512033...
I have two questions: Suppose we uncorrelate variables of a multivariate normal distribution using Cholesky transformation. Then: 1. What is the relation between Mahalanobis distances before and after this transformation? 2. Can I use this method to calculate the Mahalanobis distance between the means of two multivariate normal distributions with different means and different covariance matrices? In other word, can I calculate the Mahalanobis distance by uncorrelating each distribution and then using Euclidean distance?
[ 0.020611083135008812, 0.044575825333595276, -0.007257980760186911, 0.025083133950829506, -0.035719066858291626, -0.035979270935058594, 0.013633170165121555, -0.036374304443597794, -0.015354901552200317, -0.04784872755408287, -0.013051928952336311, 0.026974905282258987, -0.025567254051566124,...
[ 0.14465171098709106, -0.1090158224105835, 0.18894022703170776, 0.13574548065662384, -0.01941436342895031, 0.3697325587272644, 0.017713362351059914, -0.5472356081008911, -0.0851280689239502, -0.5159196257591248, 0.16506069898605347, 0.1713409572839737, -0.3369910418987274, 0.657791554927825...
I'm trying to get the equation shown in the picture. But i only manage to get the following coding. Can anyone help me to label it with numbering like the picture? I tried with `\begin{eqnarray}` or `\begin{equation}` but end up with many errors. \[ \left\{ \begin{array}{lr} \dot {x}_1 = f_1(x_1,\dots,x_n) \\ \hspace{13pt} \dots \\ \dot {x}_n = f_n(x_1,\dots,x_n) \\ \end{array} \right. \] ![HELP](http://i.stack.imgur.com/pgcUJ.png)
[ 0.0013525969116017222, 0.013749505393207073, -0.006826389580965042, 0.006204914301633835, -0.01867714151740074, -0.015299767255783081, 0.006768393330276012, 0.020528707653284073, -0.018964609131217003, 0.00480996910482645, -0.006260288879275322, 0.006630574353039265, -0.02209111489355564, ...
[ 0.20986023545265198, 0.3546185791492462, 0.2328493446111679, -0.1543310433626175, 0.014866386540234089, 0.1887352019548416, 0.312673419713974, -0.27880313992500305, -0.33268412947654724, -0.7134617567062378, 0.1612819880247116, 0.2843154966831207, -0.33019185066223145, 0.2478739321231842, ...
Is "Happy Birthday!" a complete sentence? And if it is, what role are the words happy and birthday playing? Where is the verb? Can "happy" **be** a verb? I know in a sentence like "Go get the milk" there is an implied "You" at the beginning that makes it complete. Is there an implied verb?
[ -0.023284191265702248, 0.009757203981280327, -0.0052886889316141605, 0.022350160405039787, 0.007940089330077171, 0.0107868742197752, 0.010906902141869068, 0.012080805376172066, -0.02562994882464409, -0.00376501539722085, -0.01894211396574974, 0.011804811656475067, 0.021308403462171555, -0....
[ -0.13578596711158752, 0.19038273394107819, 0.24921786785125732, -0.19128291308879852, -0.6442599892616272, -0.1725839525461197, 0.4935348331928253, 0.3385375142097473, -0.5256473422050476, -0.04715800657868385, -0.34765762090682983, 0.45852333307266235, 0.23453789949417114, 0.0400163941085...
Even after playing castle crashers for quite a bit and leveling up some of my characters, I am still having problems with the beefy versions of enemies that you run across every now and then. I'm talking about enemies like the one on the bottom right in the following image: ![Beefcake!](http://i.stack.imgur.com/7rKek.jpg) Seeing how they can take quite a beating, do not really seem to be stunned by melee attacks and can simply pick you up and throw you away, I spend most of my time running away from them and using my magic every now and then until they die. Not very heroic, and very difficult if there are many bad guys around to interfere. So my question is, is there something I'm missing here, or must I rely on ranged attacks to take them down without getting pummeled?
[ -0.010429790243506432, 0.010047363117337227, -0.006983848288655281, 0.0003721603425219655, -0.0021666414104402065, -0.007448531221598387, 0.0054547046311199665, 0.001768320333212614, -0.012579208239912987, 0.00798877514898777, -0.007311668246984482, 0.011908527463674545, -0.00323729589581489...
[ -0.0707024410367012, 0.046327222138643265, -0.25054049491882324, 0.09382271766662598, -0.30646902322769165, -0.09628555178642273, 0.6889548897743225, 0.308968186378479, -0.6246148347854614, -0.26611098647117615, 0.03859363868832588, 0.38368457555770874, 0.3027350902557373, 0.23025240004062...
> Two balls of mass $m$ each one are connected with mass-less rope with the > same length as the radius of earth. The system is in free fall. Prove that > the tension of the rope when the nearest (to the earth) ball's distance from > the earth surface is $R_E/2$ is: $T = \frac{32}{225} mg$ ![Illustration](http://i.stack.imgur.com/TgZjm.png) What I did is the following: $F_1$ is a gravitation force exerted on the nearest ball by the earth: $F_1=G \frac{M_Em}{(1.5R_E)^2}$ $F_2$ is a gravitation force exerted on the farthest ball by the earth: $F_2=G \frac{M_E m}{(2.5R_E)^2}$ $T=F_1-F_2=G \frac{M_E m}{(1.5R_E)^2}-G \frac{M_E m}{(2.5R_E)^2}=\frac{G M_E m}{R_E^2} \left (\frac{4}{9} - \frac{4}{25} \right)=\frac{64}{225} mg$ However, my answer is somehow twice bigger than what is expected. Where am I wrong? What am I missing?
[ 0.017845477908849716, 0.0010894339065998793, -0.017011510208249092, -0.0008514957735314965, -0.030168578028678894, -0.0011582085862755775, 0.007175731472671032, -0.0077631669119000435, -0.012017394416034222, -0.033229388296604156, -0.005914644338190556, 0.004862847272306681, -0.0202931016683...
[ 0.14900735020637512, -0.05780323967337608, 0.7721825838088989, 0.06663931906223297, -0.09851374477148056, 0.11440779268741608, 0.46704015135765076, -0.6844900250434875, -0.46841469407081604, -0.1640664041042328, 0.1750849187374115, 0.3006245195865631, -0.3654083013534546, -0.04239211976528...
I am currently developing a small plugin which queries an API for some XML data. The XML data is then used to display a page. I wanted to do most of the display and formatting related stuff in client-side Javascript, because it was what I was comfortable with, but because the API was on another server, I was receiving _Not Same Origin_ errors. So I decided to create a PHP file (a "wrapper" for the API) in my plugin's directory which requests the data, and then echos the result. Then all I had to do was request the PHP file from inside the Javascript file, and I got access to the XML data without having the Same Origin policy causing errors. This all worked fine, but recently I've been working towards making the plugin not hard-coded with our API auth key, but instead use a Settings page. I set it up, and verified that all of the settings saved correctly. **The problem** I have is that my "API wrapper" PHP file does not know what the function get_options is (the error says it is "Not Defined"), and all the solutions I've read have suggested using include_once('wp-content/options.php), but that is also shooting out an error about not being able to find the file (Failed to Open Stream: No such file/directory). In summary, I need a way to access get_options from inside a PHP file which is inside my plugin's directory, but is NOT the main plugin file.
[ -0.0060168164782226086, 0.021735135465860367, 0.004547539167106152, -0.004662409890443087, -0.00044277566485106945, -0.007670586928725243, 0.006635445170104504, 0.012093117460608482, -0.016378013417124748, -0.019073734059929848, -0.007267807610332966, 0.005425420124083757, -0.008235580287873...
[ 0.47707319259643555, 0.3371151089668274, 0.5025746822357178, -0.21088770031929016, -0.0755227580666542, -0.17257513105869293, 0.17945216596126556, -0.1401456743478775, -0.23527966439723969, -0.5928850769996643, 0.3137893080711365, 0.21785968542099, -0.3362996578216553, 0.12426268309354782,...
Last day my phone turned off unexpectedly and I found that it wont turns on until I connect it via usb to my PC! other problem is with SIM card: operators network do not recongnized and device will turns off after some mintues if SIM is placed in device even I connected the usb cable. after this problem I wiped all data and installed CyanogenMod 10.2. my previous ROM was: CyanogenMod 10.1 Device model: LG-Prada P940
[ -0.01062044408172369, 0.002535521984100342, -0.0021093508694320917, 0.015314229764044285, -0.040070630609989166, -0.02319713868200779, 0.009498406201601028, -0.009502534754574299, -0.018328431993722916, -0.019964050501585007, -0.019763238728046417, 0.016342023387551308, -0.000143858022056519...
[ 0.19680917263031006, 0.014399896375834942, 0.22217614948749542, 0.3371438682079315, 0.24534928798675537, -0.13259097933769226, 0.005092571023851633, 0.19487030804157257, -0.13804931938648224, -0.5013172030448914, -0.1360211819410324, 0.5102179646492004, -0.05514521896839142, 0.206166222691...
I don't think my wolves would eat an ocelot but I've had one before and it disappeared. There was nowhere for it to leave my house and it was tame.
[ 0.016416389495134354, 0.031802788376808167, -0.004751394037157297, 0.01848188042640686, -0.016490483656525612, 0.004263607785105705, 0.010164248757064342, 0.01815773732960224, -0.04506514221429825, -0.00984194502234459, -0.011052079498767853, 0.03366896137595177, 0.04699204862117767, -0.01...
[ 0.22435347735881805, 0.05661306530237198, -0.45957592129707336, -0.12349526584148407, -0.2217591106891632, 0.003136958461254835, 0.5065187215805054, 0.2864978313446045, 0.08401278406381607, -0.45562461018562317, 0.5948280692100525, 0.29718881845474243, -0.6297479867935181, 0.17491038143634...
For instance, should I say: > She woke to the rattling sound of the alarm clock. or > She awoke to the rattling sound of the alarm clock. (If both are correct, which one is more commonly used?)
[ -0.002671356312930584, 0.0013149890583008528, -0.021642781794071198, 0.033513471484184265, 0.05344327166676521, 0.004554602317512035, 0.019320478662848473, 0.014656975865364075, -0.012640602886676788, 0.00030347108258865774, -0.008128504268825054, 0.004343886394053698, -0.02523292601108551, ...
[ 0.24213379621505737, 0.19286417961120605, 0.10023611038923264, -0.15238770842552185, -0.33613258600234985, -0.017788676545023918, 0.4669431447982788, -0.30627185106277466, -0.48928502202033997, -0.6170986890792847, 0.225285604596138, 0.43197548389434814, 0.11065234988927841, 0.266855806112...
Once an early access game has been purchased on Steam, how do you find it in your library to install it? Do you have to enable "Beta participation" in Steam? Or do you have to wait a couple of days before the game shows up in your "Game Library|All Games"?
[ -0.03752675652503967, -0.005554067902266979, 0.02030118554830551, 0.004456293303519487, 0.05332321673631668, 0.007078539580106735, 0.010883474722504616, -0.035288285464048386, -0.02061426267027855, 0.02246987447142601, 0.0051476252265274525, 0.011103988625109196, -0.018364563584327698, 0.0...
[ 0.7128461003303528, -0.18693672120571136, -0.22080685198307037, 0.3616420328617096, 0.40636178851127625, -0.5543136596679688, -0.09468275308609009, 0.08844342082738876, 0.1137871965765953, -0.4553503096103668, -0.376350998878479, 0.2858833074569702, 0.22738751769065857, 0.26063576340675354...
I am writing a plugin for a customer and I need to use lightbox/thickbox/fancybox/etc in my admin piece. No matter what jquery plugin I use, and I'm using wp_enqueue_style and wp_enqueue_script, nothing seems to work. No errors or anything. I click a link to open a hidden DIV, and nothing happens. Anyone have an alternative or a solution? Thanks.
[ 0.009010884910821915, -0.003105651820078492, 0.007301905192434788, 0.01435497310012579, -0.009349043481051922, -0.01443133782595396, 0.009729335084557533, 0.010745142586529255, -0.021580468863248825, 0.002285536378622055, -0.021748678758740425, 0.016819605603814125, -0.014359052293002605, ...
[ 0.360258013010025, -0.0040068007074296474, 0.267849326133728, -0.09096713364124298, -0.035261232405900955, 0.2140539437532425, 0.4274497330188751, 0.4667762815952301, -0.007295145653188229, -0.7142393589019775, 0.21440249681472778, 0.6233385801315308, -0.1847357153892517, 0.299814909696579...