text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I try to build a multi-linear regression model, several explanatory variables have impact from period1 to period2. I need to do data transformation before modeling. Let's simplify the problem first: only one explanatory variable needs to be transformed. 2 data transformation methods are taken into consideration: * A carry-over rate: `co <- function(variable, i){ var1 <- variable for (p in (2:length(var1))){ var1[p] = var1[p-1] * i + var1[p] } return(var1) }` * S-curve: `sc <- function(variable, j, k){ var2 <- variable for (q in (1:length(var2))){ var2[q] = 1 + j * exp(k * var2[q]) } return(var2) }` My question is: **how could I get the best values for i, j & k ?** My initial idea is like this: I use `lm()` and `stepAIC()` (under `library(MASS)`) to try out all possible combination of these 3 parameters and keep the combination which leads to the highest R square and the lowest AIC (no matter if there are insignificant predictors/coefficients). (Here, `BoxTidwell()` (under `library(car)`) is not used since I was told these 2 methods of transformation should be applied in the same time for a variable.) Is there better methodology to figure out the parameters?
[ 0.024084873497486115, 0.03159106522798538, -0.011674543842673302, 0.004783163778483868, 0.005768658593297005, -0.016875995323061943, 0.007361984346061945, 0.011557461693882942, -0.01324581541121006, -0.0257282555103302, -0.006577971391379833, 0.011346136219799519, -0.01530828420072794, 0.0...
[ -0.15595604479312897, -0.045067813247442245, 0.28129154443740845, -0.17782039940357208, 0.023852406069636345, 0.28972506523132324, 0.1068480908870697, -0.49926236271858215, -0.2466839998960495, -0.5114192962646484, -0.02628796361386776, 0.3632486164569855, -0.11084776371717453, 0.390001535...
I want to create a TeX document for which the first (summary) page is `\documentclass[10pt]{article}`, but for which the rest of the document is `\documentclass[12pt]{article}`. Is there an easy way to do this?
[ -0.0012630103155970573, 0.02207804284989834, 0.002130463719367981, 0.017998188734054565, 0.004264540039002895, 0.014663477428257465, 0.010980489663779736, 0.03961240127682686, -0.019095757976174355, -0.028850415721535683, -0.010830423794686794, -0.003602579701691866, -0.027146343141794205, ...
[ 0.21514178812503815, 0.27193358540534973, 0.2652008831501007, 0.004531329032033682, 0.23034322261810303, -0.1990731805562973, 0.15259850025177002, -0.005841746926307678, 0.3671583831310272, -0.613598108291626, -0.029966115951538086, 0.8893945813179016, -0.29575470089912415, -0.158479839563...
I have a very dramatic and hypocondriac program, and it likes to spew alarming messages to syslog in cases when nothing really important happens. Besides, it sends them with priority "alert", so they float above the rest of messages of lower priority (anything lower than "err"), that I usually skip. The messages have a very specific format that I could filter or match, given the tools. Is it possible, using rsyslog, to change the priority of these messages so they become "warning"? I don't want to get rid of them, just put them in their proper place.
[ 0.002226481446996331, 0.018456775695085526, -0.008019184693694115, 0.011170675978064537, -0.013960091397166252, 0.022934947162866592, 0.007786664180457592, 0.024009104818105698, -0.014214381575584412, 0.007329270243644714, -0.01632416434586048, 0.008828792721033096, -0.010126634500920773, ...
[ 0.3461778461933136, 0.22252747416496277, 0.038377512246370316, -0.0618487223982811, -0.24079753458499908, -0.27983415126800537, 0.3972567617893219, 0.36738666892051697, -0.35201409459114075, -0.28574490547180176, 0.10423703491687775, 0.06746868789196014, -0.3355637490749359, 0.199807345867...
In a argument with my friend who lost her love, I came across her experience of life and what she said is : **Opposite** of **love** is **NOT** **Hate**. why, Because in love people have feeling and think about other person. And in Hate it is same with different intentions. Meaningfully Opposite of Love is not Hate. **Opposite of Love or Hate is Apathy.** Which is actually feelings-free. Is it right. Is she right. I am not looking for discussions?
[ -0.009402677416801453, 0.01463773101568222, -0.010623320005834103, 0.0198964960873127, -0.005793583579361439, 0.001264149439521134, 0.009320803917944431, 0.004297939129173756, -0.008158992044627666, -0.02491212822496891, -0.013190984725952148, -0.004100452177226543, 0.011549144983291626, 0...
[ 0.025323698297142982, 0.3085081875324249, 0.01785186491906643, -0.06660179048776627, -0.4799240231513977, 0.4434833228588104, 0.2886485159397125, -0.13111332058906555, -0.1554008573293686, -0.39615195989608765, 0.5471906661987305, 0.2914450466632843, -0.2685187757015228, 0.3903768956661224...
My understanding is that R squared cannot be negative as it is the square of R. However I ran a simple linear regression in SPSS with a single independent variable and a dependent variable. My SPSS output give me a negative value for R-squared. If I was to calculate this by hand from R then R squared would be positive. What has SPSS done to calculate this as negative? R=-.395 R squared =-.156 B (un-standardized)=-1261.611 Code I've used: DATASET ACTIVATE DataSet1. REGRESSION /MISSING LISTWISE /STATISTICS COEFF OUTS R ANOVA /CRITERIA=PIN(.05) POUT(.10) /NOORIGIN /DEPENDENT valueP /METHOD=ENTER ageP I get a negative value. Can anyone explain what this means? Thanks. ![Negative RSquared](http://i.stack.imgur.com/fCCDi.png) ![enter image description here](http://i.stack.imgur.com/h1cBS.png)
[ 0.03051425889134407, 0.005257419776171446, -0.016441555693745613, 0.023127108812332153, 0.006005218718200922, 0.0044638607650995255, 0.0080728679895401, -0.00237444369122386, -0.01017463393509388, -0.03960787504911423, -0.007995656691491604, 0.006945446133613586, -0.013578051701188087, 0.0...
[ 0.07169342041015625, 0.038614269345998764, 0.320822536945343, -0.060488052666187286, -0.21687541902065277, 0.2720686197280884, 0.208065927028656, -0.6200185418128967, -0.2925559878349304, -0.18983933329582214, 0.265182763338089, 0.5080111622810364, -0.36683622002601624, 0.4927097260951996,...
We have a model relationship between three random variables like this: $$ U = C + S $$ I have a ton of measurements of realizations of $U$, as well as a ton of realizations of $C$. But the interesting thing is the density of $S$. The measurements are not paired, so I can not take the differences $U-C$. One solution we tried was called _deconvolution_ , which does exactly what we want, i.e. extract the density of a "noise" variable $S$. This approach brought us fine results. Alternatively, we wanted to try a parametric approach, specifically the method of moments. **My first question:** Is the following algorithm correct, assuming gamma distributed $U$, $C$ and $S$? This should exactly reflect the method of moments, if I understood it correctly: 1. Fit a gamma distribution to $U$ and $C$, and store the parameters $k_U$, $k_C$, $\theta_U$ and $\theta_C$, using the wikipedia notation. 2. Compute the expected value and variance of $U$ and $C$, using $\mu = k\theta$ and $\sigma^2 = k\theta^2$. Denote them as $\mu_U$, $\sigma^2_U$, $\mu_C$ and $\sigma^2_C$. 3. Assuming independence, compute $\mu_S = \mu_U - \mu_C$ and $\sigma^2_S = \sigma^2_U - \sigma^2_C$. 4. Solve the equation system $\mu_S=k_S\theta_S$ and $\sigma_S^2 = k_S\theta_S^2$ for $\mu_S$ and $\theta_S$. Then I have the two parameters, and thus the gamma density of $S$. But the kernel density estimates of $U$ and $C$ are very different from the fitted gamma densities. So, I'd like to try this approach with Weibull distribution assumptions. This leads to **my second question:** How would I proceed when I assume Weibull distributed $U$, $C$ and $S$? The formula for the mean and variance look very ugly over at Wikipedia. Are there analytical solutions or do I have to go R and do some numerical optimizing there? Finally, if you are still there, **my third question:** Deconvolution fails as soon as $C$ and $S$ are correlated. With the method of moments, I think I can just see how the distribution of $S$ would change by assuming a correlation of, say, 0.5. I would do this by altering only the step 3. in my first question, and computing $\sigma^2_S = \sigma^2_U - \sigma^2_C - 2 Cov(U,C)$. Is that a valid approach?
[ 0.004775521345436573, 0.010579651221632957, -0.008159060031175613, 0.018964769318699837, -0.01505833026021719, -0.025542952120304108, 0.0067817168310284615, -0.02132364735007286, -0.015028579160571098, -0.016747454181313515, 0.001603426644578576, 0.010290184989571571, -0.011671219021081924, ...
[ 0.24139977991580963, -0.09739374369382858, 0.13669145107269287, 0.26824674010276794, -0.19142624735832214, 0.3038530945777893, -0.04195041581988335, -0.48899829387664795, -0.2301737368106842, -0.5821591019630432, 0.3902202844619751, 0.4373815357685089, -0.1658269315958023, 0.50991094112396...
I have finally faced a problem when all googling, this forum and any wittiness could not help me out. I have been working in a `\longtable` environment when in seemingly the most simple table I have ever made an unwanted extra spacing appeared with no visible for me reason. Here is the code: \usepackage{longtable} \usepackage{polyglossia} \setdefaultlanguage{latvian} \begin{longtable}{ccccc} \caption{Virknes slēguma bezdimensionālie lielumi} \\ %fixed \hline npk & $\omega$ & $i$ & $u_C$ & $u_L$ \\ \hline \endfirsthead \multicolumn{5}{c} {{\tablename\ \thetable{} (turpināta no iepriekšējas lappuses)}} \\ \\ \hline npk & $\omega$ & $i$ & $u_C$ & $u_L$ \\ \hline \endhead \hline \multicolumn{5}{l}{{Tabulas turpinājums nākamā lappusē}} \\ \endfoot \hline \endlastfoot %just data 1 & 1,71 & 0,891 & 0,522 & 1,52 \\ 2 & 1,68 & 0,923 & 0,551 & 1,55 \\ 3 & 1,62 & 0,995 & 0,614 & 1,61 \\ 4 & 1,45 & 1,32 & 0,911 & 1,91 \\ 5 & 1,30 & 1,86 & 1,43 & 2,42 \\ 6 & 1,21 & 2,51 & 2,07 & 3,05 \\ 7 & 1,16 & 3,32 & 2,87 & 3,84 \\ 8 & 1,10 & 4,94 & 4,49 & 5,43 \\ 9 & 1,01 & 12,8 & 12,6 & 12,9 \\ 10 & 0,984 & 12,3 & 12,5 & 12,1 \\ 11 & 0,955 & 8,47 & 8,87 & 8,09 \\ 12 & 0,941 & 6,98 & 7,42 & 6,57 \\ 13 & 0,926 & 5,87 & 6,33 & 5,43 \\ 14 & 1,04 & 9,07 & 8,71 & 9,45 \\ 15 & 0,868 & 3,41 & 3,93 & 2,96 \\ 16 & 0,752 & 1,71 & 2,28 & 1,29 \\ 17 & 0,665 & 1,19 & 1,79 & 0,790 \\ 18 & 0,578 & 0,867 & 1,50 & 0,501 \\ \end{longtable} I need to use Latvian language, I don't think, that it is important, but just prefered to show my problem as it is. Will appreciate any help and advice. ![unwanted gap between 4th and 5th column](http://i.stack.imgur.com/avw2t.png)
[ 0.02786802500486374, 0.003008349798619747, -0.0056003388017416, 0.02756168134510517, 0.004716571420431137, 0.0020587267354130745, 0.007517368532717228, 0.010681313462555408, -0.010990874841809273, -0.006954207085072994, -0.015232495963573456, -0.003952671308070421, -0.007702662609517574, 0...
[ 0.10867975652217865, -0.1651511788368225, 0.07534387707710266, -0.24250049889087677, 0.22220294177532196, -0.06584735959768295, 0.2821285128593445, 0.35314103960990906, -0.47313782572746277, -0.34899401664733887, 0.005860395263880491, 0.11935340613126755, -0.2606040835380554, 0.24127869307...
I want links to the acronym directory only for "PC" and not for the rest. How can I do this? \documentclass{article} \usepackage[english]{babel} \usepackage[latin1]{inputenc} \usepackage[T1]{fontenc} \usepackage{hyperref} \usepackage[acronym]{glossaries} \makeglossaries \newacronym{PC}{PC}{Personal Computer} \begin{document} \begin{enumerate} \item \gls{PC} \item \gls{PC} \item \acrshort{PC} \item \acrlong{PC} \end{enumerate} \printglossary[type=acronym] \end{document}
[ -0.004004005808383226, 0.009572722017765045, -0.0015778520610183477, 0.02870945632457733, -0.002803416922688484, -0.024336975067853928, 0.0076589882373809814, -0.005007311701774597, -0.012760595418512821, -0.006825691554695368, -0.005851792637258768, -0.01265336386859417, 0.01658905856311321...
[ 0.36620214581489563, 0.4903118312358856, 0.5152169466018677, -0.012172864750027657, 0.35126832127571106, 0.012639806605875492, 0.31721386313438416, 0.07928925007581711, -0.1565321385860443, -0.7874870896339417, -0.012684189714491367, 0.5748213529586792, -0.3983822762966156, 0.2329935133457...
I have some university rankings data where I am asked to analyze whether a higher %of one variable leads to a higher ranking in a particular area. What kind of statistical analysis should I use in excel and how should I report the findings?
[ 0.01485129538923502, 0.03609958663582802, 0.005486414302140474, 0.01825651526451111, -0.009178227744996548, 0.0026161246933043003, 0.01383802480995655, 0.00877759326249361, -0.026235423982143402, -0.02151501178741455, 0.0033936048857867718, 0.011706208810210228, -0.012050085701048374, 0.03...
[ 0.41126519441604614, 0.3200127184391022, 0.1477506309747696, 0.08370233327150345, -0.17023201286792755, -0.060757704079151154, 0.10451442003250122, 0.16771692037582397, -0.019218435510993004, -0.39039376378059387, 0.10105226188898087, 0.24008409678936005, 0.4556824266910553, 0.336816221475...
If a thug is holding a person hostage and using her as a human shield, is there any way to resolve the situation without killing anyone? I tried to aim for the arm or leg, but the shooter always ended up dead or killing the hostage. Is it possible to catch him alive somehow?
[ -0.028430746868252754, 0.034595195204019547, -0.005542038008570671, -0.008223381824791431, -0.03821700066328049, -0.009133566170930862, 0.01166217215359211, 0.00950103159993887, -0.03645310178399086, -0.06439565867185593, -0.002605058252811432, 0.03461102023720741, 0.004209928214550018, 0....
[ 0.28626081347465515, -0.12282565236091614, -0.27621200680732727, 0.2597540020942688, -0.006379101891070604, 0.441778302192688, 0.4500196576118469, -0.29666486382484436, -0.35507798194885254, -0.3573094308376312, 0.303514301776886, 0.3442833125591278, -0.31268101930618286, -0.39510497450828...
In Quake Live I see users crouching when using gaunt or shotgun. Is there any advantage to crouching? Or are these players doing it for no reason?
[ -0.018451189622282982, 0.014161468483507633, 0.007915335707366467, 0.012144290842115879, 0.008711493574082851, -0.04206235334277153, 0.016253339126706123, -0.05913100019097328, -0.019295373931527138, 0.0023494900669902563, 0.008457030169665813, 0.04589662328362465, -0.01641782931983471, 0....
[ 0.2884008586406708, -0.19014553725719452, -0.03780383989214897, -0.1473979651927948, -0.3470335602760315, -0.5099561214447021, 0.383156418800354, -0.41839146614074707, -0.11883730441331863, -0.373550683259964, 0.5234378576278687, 0.4852403700351715, -0.16153474152088165, -0.350915849208831...
I recently changed to a newer phone and currently using my old android phone as a sort of "ipod". However, it always shows the message that there is no sim card. I know this is just an annoyance but is there anyway to disable it? I am rooted running a custom sense 4 rom on ICS. Is there anything I could flash to remove this message? Or disable the cellular radio completely? ![Screenshot](http://i.stack.imgur.com/Iu31Hm.jpg) Screenshot (click for larger variant)
[ -0.0036191889084875584, -0.003749311901628971, -0.00001850910484790802, 0.012556102126836777, -0.021165885031223297, -0.0010162439430132508, 0.0036814555060118437, 0.007610897999256849, -0.014410463161766529, -0.007532862480729818, -0.020839542150497437, 0.0012774149654433131, 0.005422881804...
[ 0.4012361466884613, 0.0663885772228241, 0.5395932793617249, -0.2571292221546173, 0.14959736168384552, 0.017755454406142235, 0.635931134223938, -0.05665479972958565, -0.3615272641181946, -0.4140198230743408, 0.23463566601276398, 0.7104511857032776, -0.17010490596294403, 0.12889447808265686,...
When I put this code right above the closing of my body tag it works, but when I try to enqueue it, it breaks. I'm thinking maybe jQuery isn't getting called? The snippet is from here FUNCTIONS.PHP add_action( 'wp_enqueue_scripts', 'scroll_port' ); function scroll_port() { wp_register_script( 'scrolling', get_template_directory_uri() . '/js/scroll.js', array('jquery'), '', true ); wp_enqueue_script('scrolling'); } SCROLL.JS $('a[href*=#]:not([href=#])').click(function() { if (location.pathname.replace(/^\//,'') == this.pathname.replace(/^\//,'') || location.hostname == this.hostname) { var target = $(this.hash); target = target.length ? target : $('[name=' + this.hash.slice(1) +']'); if (target.length) { $('html,body').animate({ scrollTop: target.offset().top }, 1000); return false; } } });`
[ 0.00013892364222556353, 0.012775562703609467, 0.006115823518484831, 0.0054662227630615234, 0.005705421790480614, -0.015747951343655586, 0.0067946636117994785, 0.018578771501779556, -0.019053567200899124, -0.0024778335355222225, -0.01660584658384323, 0.019498586654663086, -0.01967461779713630...
[ 0.36219725012779236, -0.1487789899110794, 0.9692551493644714, -0.019683368504047394, 0.029316464439034462, 0.39924412965774536, 0.393399715423584, -0.27085697650909424, -0.14402657747268677, -0.5246660709381104, -0.0032107096631079912, 0.6919562816619873, 0.09563513100147247, -0.3192374408...
Suppose E1 is an expression in LaTeX, where one entry is x. Suppose E2 is another expression. How do you replace x by E2? Can I wrap E2 in a box of certain size and have LaTeX replace x with that box without asking LaTeX to look inside the box and try to rearrange its internal structure? For example: How do you replace an element of a given 3*3 matrix with a particular 2*2 matrix? The answers so far are very interesting in themselves but none worked for me. So let me re-state. Suppose in an environment I can put a literal $x$ without any problem but I cannot put a matrix or a multi-line equation in the same spot, as it conflicts with various commands etc. How can I make a matrix act like an $x$ to the rest of program? As if I have to build the matrix somewhere else, put it in a box, and sell it to LateX as a single item as if it has no line return.
[ 0.02355325035750866, 0.0258930791169405, -0.019285546615719795, 0.012248918414115906, 0.023883096873760223, 0.007184455171227455, 0.009382136166095734, -0.015268219634890556, -0.018266933038830757, -0.01865263096988201, -0.009947165846824646, 0.008965039625763893, -0.0036760494112968445, 0...
[ 0.15723809599876404, -0.13797730207443237, -0.03165612742304802, -0.09913542866706848, -0.0857405737042427, -0.12854266166687012, 0.2037009596824646, -0.42497289180755615, 0.14857929944992065, -0.6857681274414062, -0.3889789879322052, 0.29708585143089294, -0.5971825122833252, -0.2979879975...
Let's assume that another user started a bunzip process, and I have a script that I'd like to start running after that bunzip finishes. What's the best way to check from inside my script that the bunzip process has finished? A call to `ps` with the pid specified? A pgrep bunzip? Assume that I will throw my script into a crontab that checks every 5 minutes to see if it can run. Also assume that I don't want to stop the bunzip, as it's been running for over an hour and will likely take another hour to complete. My first reaction would be to use something like if `ps -p 12938` # bunzip is done, execute the code fi # exit But, I'm wondering if there is a better way. Also, I'm not sure how cross-unix ps -pid is.
[ 0.010474524460732937, 0.0013140137307345867, -0.014442646875977516, 0.017700830474495888, -0.012084179557859898, 0.012878109700977802, 0.0085072573274374, -0.016718745231628418, -0.017550934106111526, 0.029900765046477318, -0.011551421135663986, 0.0014109837356954813, 0.010620505549013615, ...
[ 0.4004686772823334, 0.13327330350875854, 0.44453150033950806, 0.20686304569244385, 0.07935725152492523, -0.25085312128067017, 0.2769639194011688, 0.14505064487457275, -0.2898097336292267, -0.3830624222755432, -0.10646528005599976, 0.4857271909713745, -0.22326523065567017, -0.03285159170627...
.htaccess is not working. * `/index.php/user` -- is working * `http://www.example.com/user` is not working Current .htaccess file: RewriteEngine on RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule .* index.php/$0 [PT,L]
[ -0.008705124258995056, 0.013262375257909298, 0.004224587231874466, 0.02224707044661045, 0.0028923149220645428, -0.006112067494541407, 0.00861344300210476, -0.030826328322291374, -0.019996268674731255, -0.021775677800178528, -0.024029787629842758, 0.001709499629214406, -0.00978997815400362, ...
[ 0.05244097858667374, 0.20149631798267365, 0.8264695405960083, -0.05950121954083443, 0.3378690183162689, -0.013024958781898022, 0.4426616430282593, -0.10736572742462158, 0.006737754680216312, -0.509300708770752, 0.01219504326581955, 0.6990810036659241, -0.2897106111049652, 0.254413843154907...
I've heard a lot of mention of "cursed charm tables". What exactly are these and why are they considered "cursed"? Possibly most important, how do I avoid getting a cursed charm table?
[ 0.006322601810097694, 0.013788565993309021, -0.006821888964623213, 0.029101433232426643, -0.00985525269061327, 0.007121110334992409, 0.009468978270888329, 0.029974667355418205, -0.02131403051316738, -0.043525580316782, -0.011853761039674282, 0.015148311853408813, -0.02109292335808277, -0.0...
[ 0.8454099297523499, 0.6936643719673157, -0.22293482720851898, 0.2093760222196579, -0.11449701339006424, -0.36194977164268494, 0.6326530575752258, 0.3005236089229584, -0.3309788405895233, -0.49960923194885254, 0.19706885516643524, 0.20937298238277435, -0.01677723228931427, 0.516988813877105...
I am new to QGIS and would like to merge selected data/attributes out of the attribute tables from overlaying vector objects on different vector layers without creating new additional layers. Is there any standard QGIS operation or QGIS plugin to accomplish this task. Basically to explain it better I have a vector layer with polygons representing buildings and its assigned attribute table. A second vector layer with polygons represents communal planning zones and its assigned attribute table. Both layers can not be joined by attributes as there have no corresponding keys. So I would like to perform a spatial query over both vector layers to find all buildings overlapping polygons of the zone layer and write back to the same building layer attribute table the corresponding zone code from the zone layer attribute table. How can this be done? I am appreciating any hint. Many thanks in advance
[ -0.0017961389385163784, 0.024303197860717773, 0.004006954841315746, 0.020578894764184952, 0.014019403606653214, -0.009863290935754776, 0.011357806622982025, 0.006103313062340021, -0.016236696392297745, 0.014656764455139637, -0.01687767170369625, 0.016756277531385422, -0.01028384454548359, ...
[ 0.5072214603424072, 0.2327750325202942, 0.3715990483760834, 0.07687681168317795, -0.04195823892951012, -0.0787755623459816, 0.1627373844385147, -0.20287463068962097, 0.05634301155805588, -1.1818671226501465, -0.003627593396231532, 0.22890006005764008, -0.20836496353149414, 0.12721018493175...
I've inherited the code maintenance of a complex web site for a customer that continuously requests enhancements for it. This application took years to develop and I'm facing increased difficulties to enhance it. It's organized but at the same time the .net code is mixed with ajax, javascripts and old school html that it takes me days to figure out how some pages work. First off, I not new to asp.net but I'm not familiar with the new MVC stuff but from what I've read it seems to be a step in a better direction. The current code is all in one big dll. The application code is divided into multiple folders representing the different departments and each department has it's own pages for handling general stuff like employee management, reports, budget and also their own information. For example, even though each department uses a different webpage for employee information handling, they want different fields and so it was simpler to create different pages than to use a single page that adapts to each department. But it is really a nightmare to maintain right now and I would like to create a parallel project where I could start a fresh project, create a better structure and from there start migrating the old code to this new environment and refactor it as I go. The idea is to migrate the old application to a new web site that has a similar look while maintaining both operating until everything is running in the new site. It may sound insane but it really is used extensively for hundreds of people everyday and it bugs me that I have to modify crappy code to make it work. How would you go about this issue? Thanks [edit] If found this link Things You Should Never Do on another post, very much to the point of my question.
[ -0.01361033134162426, 0.016795214265584946, -0.0025469413958489895, 0.013782700523734093, -0.0075515154749155045, 0.0076161399483680725, 0.00588276656344533, -0.005315943621098995, -0.012850684113800526, -0.014956032857298851, -0.0176544226706028, 0.015514735132455826, -0.005906473845243454,...
[ 0.4985547363758087, 0.3323047459125519, 0.37713128328323364, 0.1426466554403305, 0.04514127969741821, -0.27441078424453735, 0.07001838833093643, -0.011760584078729153, -0.23892784118652344, -0.9381876587867737, 0.22759169340133667, 0.4390419125556946, -0.06052476540207863, 0.11721994727849...
The following is a simple script, named `myscript`, that I wrote in order to run a command. dcmpath='$HOME/Data' dcmfile='IM1' dcm2nii $dcmpath/$dcmfile Unfortunately, bash does not recognize the path in front of `dcm2nii` and I cannot run the last command in the file. I would be thankful if someone could let me know what I am missing. What would be the best way of customizing the path and file name which is the argument of a command?
[ 0.002924501895904541, 0.009919527918100357, -0.017658665776252747, 0.003996466286480427, -0.0038046040572226048, -0.024184364825487137, 0.007166162133216858, 0.0019649644382297993, -0.016192849725484848, -0.0009532200638204813, -0.02164449915289879, -0.0009270636364817619, 0.0130279269069433...
[ 0.2593063414096832, 0.18636061251163483, 0.13662685453891754, -0.1737067997455597, -0.07140357792377472, -0.003262538695707917, 0.2538743317127228, -0.1948164999485016, -0.17883820831775665, -0.6975879073143005, 0.2838535010814667, 0.6131312847137451, -0.29077017307281494, 0.16270270943641...
I just started playing Thief: Deadly Shadows. I'm trying to change the difficulty on the first mission: Checking Inn - Cashing Out, but nothing works! **How do I change the difficulty level?**
[ 0.01393243856728077, 0.006940419785678387, -0.019660674035549164, 0.005076463334262371, 0.00893455371260643, -0.00643481221050024, 0.014783327467739582, -0.014380417764186859, -0.026740994304418564, -0.007580032106488943, -0.016727246344089508, 0.02036188915371895, -0.04665675759315491, 0....
[ -0.08180274814367294, -0.30961844325065613, 0.16731536388397217, -0.04801250994205475, -0.02423243038356304, 0.054629549384117126, 0.4841073453426361, 0.0863836258649826, -0.3936876058578491, -0.4421890377998352, 0.21363812685012817, 0.8244686722755432, 0.07130374014377594, -0.088345587253...
So, I'm trying to solve for the torque $\tau_A$ of a motor. I have attached a strong stick to the motor, like so: ![diagram of the problem](http://i.stack.imgur.com/ZUJCG.jpg) I apply a force $F$ on the stick which stops the motor. The distance from the outside edge of the cylinder to the end of the stick is $L$. The torque for the motor is $\tau_A=F(L+r)$, $r$ is the radius. My friends believe that the torque at point $B$ is $\tau_B=-FL$, but I believe even though the motor is not moving (due to the force), it still applies a torque at point $B$. It would be less than $\tau_A$ since it doesn't push around the point uniformly, but it should be $\tau_B=-FL+\tau_Ac$ ($c$ is a constant). Using their method, they got that $\tau_A=0$, which I believe happened because in calculating the torque at point $B$, they make $\tau_A=0$. Who is right? How do I calculate how much $\tau_A$ is applied about point $B$ (assuming I'm correct)?
[ -0.01590966433286667, 0.0002881910768337548, -0.013133782893419266, 0.004481873940676451, -0.020026080310344696, -0.0019986950792372227, 0.007191088981926441, -0.0030556311830878258, -0.01556357927620411, 0.008830324746668339, -0.016684921458363533, 0.006356186233460903, -0.01939285919070244...
[ 0.34113723039627075, 0.25294381380081177, 0.7786465883255005, -0.08510232716798782, -0.31454113125801086, 0.009203280322253704, 0.26628220081329346, -0.7473680973052979, -0.3524746000766754, 0.03427448868751526, 0.3783038854598999, 0.8967549204826355, -0.23172995448112488, 0.44794306159019...
When writing about LaTeX, I tend to use `\verb`, which could look like that: \documentclass{article} \usepackage{lipsum} \begin{document} Lorem ipsum dolor sit amet, consectetuer adipiscing elit. \verb+\documentclass{article}+ \lipsum[2] \end{document} Thus I get the following result: ![Output](http://i.stack.imgur.com/hWK6A.png) Is there another way to display LaTeX commands properly in continuous text or atleast some way to prevent `\verb` from going over the end of the line? (For longer code I obviously would use something like `listings`)
[ 0.0213887020945549, -0.0035690111108124256, -0.00896519050002098, 0.02044079639017582, -0.0026424378156661987, 0.008692596107721329, 0.009866554290056229, -0.003296148031949997, -0.010278582572937012, 0.03241153061389923, -0.00825487356632948, -0.009461179375648499, 0.010060335509479046, 0...
[ 0.03635714575648308, 0.04761794954538345, 0.5562887787818909, -0.2308707982301712, 0.05814545974135399, 0.21777136623859406, 0.270630806684494, -0.016446862369775772, -0.13133488595485687, -0.2837457060813904, -0.19093970954418182, 0.5527750253677368, -0.31357431411743164, 0.04858428984880...
My end goal is to get an excel (I'll probably just settle for .csv) file loaded from a silverlight app into my "%SCRATCHWORKSPACE%" for my gp service. I will be linear referencing the data onto a polyline. I have silverlight reading the csv file's contents and packing it into a string to send over to ArcGIS Server. Then I want to write this data to a csv file in the "%SCRATCHWORKSPACE%". But I don't seem to be able to use any standard python code to hit the "%SCRATCHWORKSPACE%". Any thoughts or critiques?
[ -0.0015410452615469694, 0.014455761760473251, -0.00760587677359581, 0.02798428386449814, 0.0007300756406039, 0.007544481195509434, 0.009371000342071056, 0.015809910371899605, -0.016150014474987984, -0.009782006964087486, 0.0038457131013274193, 0.013964694924652576, -0.0027277658227831125, ...
[ 0.28055012226104736, 0.6115959286689758, 0.2840379476547241, -0.16151179373264313, -0.12515661120414734, 0.26188984513282776, 0.2027495950460434, -0.06881581246852875, 0.20795920491218567, -0.7904832363128662, 0.2189026176929474, 0.34617018699645996, -0.11385541409254074, 0.110657505691051...
we use models with multiplicative interaction effects when relationship between independent variable and dependent variable are non-additive. My question is, Are all models with multiplicative interaction effects non- linear? and all models with additive interaction effects linear? Also, With non-linearity, the effect of independent variable on dependent variable depends on the value of independent variable, in effect, independent variable somehow interacts with itself. Does that mean that an independent variable(x1) interacts with itself(x1)? or does it mean that independent variables(x1) interact with other independent variables(x2, x3...) and not with itself? I am confused with concepts of of linearity, non-linearity, additivity and non-additivity (multiplicativity). Any help is much appreciated.
[ 0.007124701980501413, 0.01312519796192646, -0.004018841776996851, 0.029899992048740387, -0.0023295872379094362, -0.04216296970844269, 0.01133224181830883, -0.010655442252755165, -0.01802513748407364, -0.009435483254492283, -0.01884426921606064, 0.019183527678251266, -0.003937044180929661, ...
[ 0.0064831688068807125, -0.16175474226474762, 0.2712586224079132, 0.3587302565574646, -0.11101576685905457, 0.15976792573928833, -0.015284095890820026, -0.0926339328289032, -0.23004595935344696, -0.6776465177536011, 0.17081095278263092, 0.5635442137718201, -0.5318630337715149, 0.16017997264...
By suggestion in comments, I'll try to rephrase my question to better reflect my problem: I have to present users (few at first, as many as possible later) with a website. There they will be able to browse various data. The data will be served from a local "cache" db. Users should be always presented with data as fresh as possible. The source for most of the data will be external services with possible limits, long response times etc. The process of processing data from those sources to the schema of local db is also not straightforward (conversion, calculations, decoupling, recreating relations etc.) Is putting all that data mining and processing code ok (and doing all that work in the time between users click a link and receive data), or is my second idea - another, constantly running data-miner application - better? Or is there another option? Old text: I'm currently developing a website that will present both its own data (user profiles, their submited content) as well as data obtained from various external sources (e.g. via webservice calls). Varied sources will have varied limits of allowed calls, will respond with varied speed etc. No, the data user is presented with should be, by default, as fresh as possible with an option to request refresh (unless refreshed very recently). I started development in ASP.NET (for better or for worse), with some templates and some customization, and quite a lot of coding I ended up with a workable proof of concept. But then I started to think about large-scale operations. Displaying cached data is quick and easy, I didn't mess that up. Refreshing though is a problem. a) Requesting data for one user is fast, but it will eat through allowed api calls quickly. b) Requesting data for multiple users (for apis that allow that) saves potentialy great amount of api calls, but takes linearly longer (not 1:1, but still, it takes time to request, receive and parse tens of times more data than single request). There also comes the problem of exceeding the limits. Usually that would force the app to either cancel refresh, or wait few seconds for new allowance. But I don't really see the user happily waiting for several seconds for the page to load. I am currently concidering gutting the data-gathering part of the website and moving it into a single managing application, sharing a db with the website. The website would provide info which users need their data to be refreshed, the app would load-balance the external services to use them efficiently (not limited to "when a user opens a website"), and keep the data as fresh as possible - also possibly keeping part of the allowance for forced refreshes. Now, the main problem I see with this is cost of hosting not just website, but also a background running .NET application. But let's assume that the cost is not that relevant. What I actually fear is that I overcomplicate the solution. I might just put more of the website in Ajax and just wait for the request to complete in the background (when calls become available). So, does anyone has any simillar experience? I'm especially interested in the scenario where it is all scaled up, that is we can assume that there will always be deficit of available calls. PS I also worry about keeping website (which can be triggered independently by many users) from trying to refresh the same records twice etc. - I guess I should put some additional safeguards, lock or at least mark the records undergoing refresh etc. - a lot of code just to solve the problem that wouldn't exist with a single-backend....
[ 0.006802053190767765, 0.013735135085880756, 0.0023869522847235203, 0.007091981824487448, 0.006301228888332844, -0.019607318565249443, 0.005044036544859409, 0.0010168550070375204, -0.009437212720513344, -0.015145452693104744, -0.009514960460364819, 0.008606678806245327, -0.0058991676196455956...
[ 0.47090741991996765, 0.060938019305467606, 0.37152543663978577, 0.12337548285722733, -0.26356133818626404, -0.13509102165699005, 0.26393669843673706, 0.022123947739601135, 0.017177287489175797, -0.6078338027000427, -0.33492812514305115, 0.17875292897224426, 0.15636888146400452, 0.296156734...
In Sleeping Dogs, you can purchase improvements for your apartments. So far I've found a new bed, a sound system, a bird and AC in the Night Market for the North Point apartment and a furniture package at K-Bar for the Central Apartment.
[ 0.018218986690044403, 0.006856668274849653, -0.0028335487004369497, -0.0010614503407850862, -0.024438945576548576, 0.0034867858048528433, 0.01235959306359291, -0.0009506019996479154, -0.02625424973666668, 0.005458664149045944, -0.019641129299998283, 0.024826977401971817, -0.00526365451514720...
[ 0.8028415441513062, 0.3143608868122101, 0.3417055308818817, -0.003785863518714905, 0.4180721640586853, 0.22599221765995026, -0.0018045464530587196, -0.09336448460817337, -0.24061284959316254, -0.9636603593826294, 0.29521581530570984, 0.38218769431114197, -0.05166440084576607, 0.23749607801...
I'm brand new to creating gis map applications and I am trying to get a web project that uses function init(){ map = new OpenLayers.Map( 'map' ); layer = new OpenLayers.Layer.WMS( "OpenLayers WMS", "http://vmap0.tiles.osgeo.org/wms/vmap0", {layers: 'basic'} ); map.addLayer(layer); map.setCenter(new OpenLayers.LonLat(lon, lat), zoom); var featurecollection = { "type": "FeatureCollection", "features": [ {"geometry": { "type": "GeometryCollection", "geometries": [ { "type": "LineString", "coordinates": [[11.0878902207, 45.1602390564], [15.01953125, 48.1298828125]] }, { "type": "Polygon", "coordinates": [[[11.0878902207, 45.1602390564], [14.931640625, 40.9228515625], [0.8251953125, 41.0986328125], [7.63671875, 48.96484375], [11.0878902207, 45.1602390564]]] }, { "type":"Point", "coordinates":[15.87646484375, 44.1748046875] } ] }, "type": "Feature", "properties": {}} ] }; var geojson_format = new OpenLayers.Format.GeoJSON(); var vector_layer = new OpenLayers.Layer.Vector(); map.addLayer(vector_layer); vector_layer.addFeatures(geojson_format.read(featurecollection)); } I've included the OpenLayers.js with the img and theme folders under the WEB- IMF. When I run it only the text shows, not the map. Is there something I'm missing.
[ -0.008236387744545937, 0.004072069656103849, -0.00540725514292717, 0.010491684079170227, -0.00929329264909029, 0.00931618269532919, 0.008160327561199665, 0.011628231033682823, -0.011728642508387566, 0.0037193559110164642, 0.005133400205522776, 0.015376601368188858, 0.010599697008728981, 0....
[ 0.48076575994491577, -0.2853679060935974, 0.8550693392753601, -0.0630674809217453, -0.12036741524934769, 0.022495495155453682, 0.0876978412270546, -0.15082848072052002, -0.16212688386440277, -1.0110527276992798, -0.3407587707042694, 0.47642770409584045, -0.15656153857707977, 0.154558524489...
I have pages and posts on a WordPress site. The pages have various layouts. Parent pages are following a 2-column (maincontent + right sidebar) layout while child pages have a one-column layout (no sidebar). I want to display a JavaScript code on one-column layout, and a different JavaScript code on 2-column layout. Must also add that one-colum has a custom field "one-col" and 2-column is the default layout. Not sure if it's possible to add codes based on custom fields. Theme: WooThemes, Canvas. Really appreciate your help.
[ 0.013328894041478634, 0.01833600550889969, -0.01946388930082321, 0.0065307216718792915, -0.019659506157040596, -0.005006858147680759, 0.011846859008073807, -0.0009310794994235039, -0.014085497707128525, 0.005295053124427795, -0.015442462638020515, 0.005341897718608379, 0.010168205946683884, ...
[ 0.6300727128982544, -0.0032065999694168568, 0.8470088839530945, 0.037685927003622055, -0.07236327230930328, 0.4035091996192932, -0.07242066413164139, 0.018138952553272247, -0.12430037558078766, -0.8662205934524536, 0.0014186238404363394, -0.12929566204547882, -0.14538529515266418, 0.409134...
I´m trying to visualize a `glm` model with a binomial response variable, I want to put a line in the plots, but neither `lines` or `abline` work and I don´t know why when I have significant responses according to the model I used. Any idea why plotting a line does not work? > summary(model2) Call: glm(formula = Lövförekomst ~ Areal + Si, family = binomial) Deviance Residuals: Min 1Q Median 3Q Max -1.254 -1.048 -1.012 1.309 1.422 Coefficients: Estimate Std. Error z value Pr(>|z|) (Intercept) -0.2158411 0.0634165 -3.404 0.000665 *** Areal 0.0009178 0.0002495 3.678 0.000235 *** Si -0.0117521 0.0038870 -3.023 0.002499 ** --- Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 (Dispersion parameter for binomial family taken to be 1) Null deviance: 16662 on 12237 degrees of freedom Residual deviance: 16642 on 12235 degrees of freedom (33 observations deleted due to missingness) AIC: 16648 Number of Fisher Scoring iterations: 4 > mod <- glm(Lövförekomst~Areal, binomial) > med <- glm(Lövförekomst~Si, binomial) > x <- seq(0,9,0.01) > y <- predict(mod, list(Areal=x), type="response") > par(mfrow=c(1,2)) > plot(Areal, Lövförekomst) > x2 <- seq(0,9,0.01) > y2 <- predict(med, list(Si=x), type="response") > plot(Si, Lövförekomst) > lines(x, y) > lines(x2, y2)
[ 0.0015550930984318256, 0.0032441639341413975, -0.013540482148528099, 0.011294737458229065, 0.008884204551577568, -0.007581618148833513, 0.008760731667280197, 0.01923855021595955, -0.01534350961446762, -0.023105062544345856, -0.010543402284383774, 0.010335884056985378, -0.016954654827713966, ...
[ -0.15771333873271942, -0.24007049202919006, 0.5915285348892212, 0.05836253985762596, -0.11948081851005554, 0.2870727479457855, 0.1805899441242218, -0.5483980178833008, -0.40298572182655334, -0.2693825960159302, -0.055879220366477966, 0.6283749938011169, 0.07528333365917206, 0.1966323405504...
I want to download a folder from my google drive using terminal? Is there any way to do that? I tried this: ‍ `wget "https://drive.google.com/folderview?id=0B-Zc9K0k9q-WWUlqMXAyTG40MjA&usp=sharing"` But it is downloading this text file: `folderview?id=0B-Zc9K0k9q-WdEY5a1BCUDBaejQ&usp=sharing` Is there any way to download google drive folder from terminal?
[ -0.022873664274811745, -0.0020703331101685762, 0.00834369845688343, 0.01528895366936922, 0.016527552157640457, -0.010857201181352139, 0.005239950958639383, 0.009457566775381565, -0.019824005663394928, -0.011363803409039974, 0.00025432047550566494, -0.0029145036824047565, 0.00705924816429615,...
[ 0.31581825017929077, 0.0454246960580349, 0.4041350185871124, 0.038505930453538895, -0.08939110487699509, 0.03876267746090889, 0.08153676986694336, 0.31051871180534363, -0.20165224373340607, -0.7426561713218689, 0.21997559070587158, 0.5852966904640198, -0.2888288199901581, 0.291198551654815...
Which is correct: > One of the clearest analogies which helps us understand ... or > One of the clearest analogies which help us understand ... I think it's the first one because 'One of ...' is the subject and it's in the singular form, but it doesn't sound right. Which one is grammatically correct?
[ -0.011544940061867237, 0.023093832656741142, -0.028832580894231796, 0.012723385356366634, -0.021896954625844955, -0.01868327148258686, 0.01228767354041338, 0.023118402808904648, -0.017840469256043434, -0.006194498855620623, -0.009160037152469158, 0.013827997259795666, 0.007636560592800379, ...
[ -0.39422404766082764, 0.24921788275241852, 0.11023882776498795, -0.1060665175318718, -0.4499119222164154, 0.37919822335243225, 0.4766249358654022, 0.23594138026237488, -0.0727267786860466, -0.36372989416122437, -0.3348999321460724, 0.3907104730606079, -0.2827330231666565, -0.32571423053741...
This is more of a conceptual question. I have learned about Neural Nets, and I have some clue as to how Support Vector Machines work. I read somewhere however that given the appropriate kernel (is that right?), the SVM is identical to the Neural Net. Could someone who understands this please enlighten me as to how that's possible?
[ -0.004321039654314518, 0.008690637536346912, 0.008225088939070702, 0.020113054662942886, -0.02715807408094406, -0.0029876925982534885, 0.007724718190729618, 0.016280829906463623, -0.021558916196227074, -0.0123938312754035, -0.00442930031567812, 0.01968863606452942, -0.0013640018878504634, ...
[ 0.44864943623542786, -0.07168411463499069, 0.29498568177223206, 0.30380114912986755, 0.0315348356962204, 0.1621558666229248, -0.20637430250644684, 0.08670024573802948, -0.5708576440811157, -0.386465847492218, 0.2391650229692459, 0.7100561857223511, -0.18235467374324799, 0.3829140365123749,...
After dealing with DDD for months now, I'm still confused about the general purposes of domain services, factories and aggregate roots in relation to each other, e.g. where they overlap in their responsibility. Example: I need to 1) create a complex domain entity in a saga (process manager) which is followed by 2) a certain domain event that needs to be handled elsewhere whereas 3) the entity is clearly an aggregate root that marks a bounded context to some other entities. 1. The factory IS be responsible for creation of the entity/aggregate root 2. The service CAN create the entity, since he also throws the domain event 3. The service CAN act as a aggregate root (create 'subentity' of 'entity' with ID 4) 4. The aggregate root can create and manage 'subentities' When I introduce the concept of a aggregate root as well as a factory to my domain, a service seems no longer needed. However, if I'm not, the service can handle everything needed as well with the knowledge and dependencies he has. **Code Example** _based on the ubiquitous language of a car repair shop_ public class Car : AggregateRoot { private readonly IWheelRepository _wheels; private readonly IMessageBus _messageBus; public void AddWheel(Wheel wheel) { _wheels.Add(wheel); _messageBus.Raise(new WheelAddedEvent()); } } public static class CarFactory { public static Car CreateCar(string model, int amountofWheels); } _..or..._ public class Car { public ICollection<Wheel> Wheels { get; set; } } public interface ICarService { Car CreateCar(args); void DeleteCar(args); Car AddWheel(int carId, Wheel wheel); }
[ -0.006874619051814079, 0.01289600133895874, -0.006484461948275566, 0.010578231886029243, -0.012289993464946747, 0.013759506866335869, 0.008022741414606571, 0.034371815621852875, -0.01465567946434021, 0.014775251969695091, -0.011185846291482449, 0.01331708487123251, -0.008329004049301147, 0...
[ 0.47335416078567505, 0.2569461464881897, 0.25913235545158386, 0.10310978442430496, -0.033441781997680664, -0.08812730014324188, 0.31543928384780884, -0.25511452555656433, -0.21205590665340424, -0.45813044905662537, -0.05648581683635712, 0.3544303774833679, 0.022064389660954475, 0.713188290...
I often try to memorize command options by looking up for what they stand for. This time I couldn't find any source on that: why is `quota -s` a `\--human- readable` alternative, what does `s` stand for?
[ 0.012876652181148529, 0.011156508699059486, -0.01362527534365654, 0.00714560691267252, -0.017350034788250923, 0.01999828964471817, 0.009871549904346466, 0.0014404109679162502, -0.03200451284646988, 0.004928730893880129, -0.011582442559301853, -0.0019457739545032382, 0.0002721656928770244, ...
[ 0.2313624620437622, 0.08109147846698761, -0.26224997639656067, 0.05057818442583084, -0.19227731227874756, 0.16602733731269836, 0.13146543502807617, 0.1440422534942627, -0.2887040376663208, -0.4593002498149872, -0.3367186188697815, 0.21990765631198883, -0.24233472347259521, 0.00391709990799...
How might it be possible to alter some variable in the `env` of an already running process, for example through `/proc/PID/environ?` That "file" is `read-only`. Need to change or unset the DISPLAY variable of a long-running batch job without killing it.
[ 0.019284827634692192, 0.02338164485991001, -0.005770512390881777, 0.009690545499324799, -0.004369056783616543, -0.0029426622204482555, 0.011808982118964195, 0.020718561485409737, -0.019390873610973358, -0.010386761277914047, -0.01598336733877659, 0.017531482502818108, -0.003359433263540268, ...
[ 0.2975005805492401, -0.126466304063797, 0.16865679621696472, 0.2117750495672226, 0.3719053566455841, -0.051997922360897064, 0.04726947844028473, -0.09592319279909134, -0.2544313967227936, -0.5008615255355835, -0.07340168952941895, 0.5536795258522034, -0.3218246102333069, -0.059980113059282...
I am working my way (self-study) through E.T. Jaynes' book _Probability Theory - The Logic of Science_ ## Original Problem _Exercise 2.1_ says: "Is it possible to find a general formula for $p(C|A+B)$ analogous to [the formula $p(A+B|C)=p(A|C)+p(B|C)-p(AB|C)$] from the product and sum rules. If so, derive it; if not, explain why this cannot be done." ### Givens The rules I have to work with are: $p(AB | C) = p(A|C)p(B|AC) = p(B|C)p(A|BC)$ and $p(A|B)+p(\bar{A}|B)=1$ Where we can also use logical identities to manipulate propositions. For example: $A+B=\overline{\bar{A}\bar{B}}$ ### Assumption of Solvability I believe it must be possible because he does not introduce any other rules later and having a simple logical combination of propositions that was not easily expressible would defeat Jaynes' central thesis. However, I've been unable to derive the rule. ## My Attempt To keep myself from getting confused due to using the same variable names as the givens, I am solving the problem as: Derive a formula for $p(X|Y+Z)$ ### Introducing a tautology for conditioning My best attempt at solving it so far has been to introduce a proposition $W$ which is always true. Thus I can rewrite $Y+Z$ as $(Y+Z)W$ (since truth is the multiplicative identity). Then, I can write: $p(X|Y+Z)=p(X|(Y+Z)W)$ So, rewriting one of the givens as Bayes' rule: $p(A|BC)=\frac{p(B|AC)p(A|C)}{p(B|C)}$, I can write: $p(X|(Y+Z)W)=\frac{p(Y+Z|XW)p(X|W)}{p(Y+Z|W)}=\frac{p(Y+Z|X)p(X|W)}{p(Y+Z|W)}$ ### Why this doesn't work The term $p(Y+Z|X)$ is easy to deal with. (Its expansion is referred to in the problem definition.) However, I don't know what to do with $p(X|W)$ and $p(Y+Z|W)$. There is no logical transformation I can apply to get rid of the $W$, nor can I think of any way of applying the given rules to get there. ## Other places I've looked I've done a Google search, which turned up this forum page. But the author does the same thing I tried without seeing the difficulty I have with the resulting conditioning on the introduced tautology. I also searched stats.stackexchange.com for "Jaynes" and also for "Exercise 2.1" without finding any useful results.
[ -0.019704313948750496, 0.02611267939209938, -0.006397312507033348, 0.010104548186063766, -0.002370379399508238, 0.0022711222991347313, 0.006858665496110916, -0.01763463020324707, -0.018488869071006775, -0.02657122164964676, -0.012002946808934212, 0.009661461226642132, -0.014694946818053722, ...
[ 0.05000150203704834, 0.12275762856006622, 0.14302511513233185, -0.14450910687446594, -0.05426378175616264, 0.055975232273340225, 0.28814083337783813, -0.3905170261859894, -0.11889316141605377, -0.47588664293289185, 0.05981982871890068, 0.09557650983333588, -0.07088636606931686, 0.104264803...
When we measure the speed of a moving element we do it with the help of a reference frame. Now if we need to measure the speed of time, is it possible? Is time really has speed? Thanks in advance.
[ 0.009100592695176601, 0.0106434915214777, -0.0162211861461401, -0.010327828116714954, 0.01276585552841425, -0.012160610407590866, 0.0127341840416193, -0.008568793535232544, -0.020948009565472603, -0.008607028983533382, 0.026047494262456894, 0.01540519017726183, -0.008219270966947079, 0.006...
[ 0.09381081908941269, -0.19307251274585724, 0.19014018774032593, 0.5228703022003174, 0.0008626864291727543, 0.21166804432868958, -0.09824023395776749, 0.17051073908805847, -0.40532609820365906, -0.6066926717758179, 0.3871602714061737, 0.49886322021484375, 0.17929843068122864, 0.143450602889...
As I understand, past perfect is used to express an action that had happened before something else in the past. But in the paragraph I came across, there appeared past perfect progressive in the middle of nowhere. Please refer to the paragraph below. In the popular movie Good Will Hunting, a janitor working at MIT, one of the best universities in America, solves a complicated math problem that was written on a board by a professor. The janitor, Will Hunting, **had been studying** mathematics on his own. The professor, Gerald Lambeau, realizes that Will has a very special talent. In the following scene, he explains the situation to another professor, Sean. To me, it sounds more right to say "has been studying" because it is describing something that started in the past and extends to the present point("solves"). Can anybody explain how past perfect progressive is possible in this paragraph?
[ -0.0016887492965906858, 0.01239354070276022, -0.006077526602894068, -0.0022944952361285686, -0.009145818650722504, -0.004760395735502243, 0.006321336142718792, -0.004505098797380924, -0.010938124731183052, 0.0260334312915802, -0.006830775178968906, 0.009872759692370892, 0.011127380654215813,...
[ -0.1008247584104538, -0.03884800150990486, -0.039834924042224884, 0.11545512825250626, 0.15226082503795624, -0.23343217372894287, 0.4088577628135681, -0.0633864477276802, 0.07223577797412872, -0.43735605478286743, -0.5311159491539001, 0.2781337797641754, -0.3520505726337433, 0.265417098999...
If light was infinitely fast, we could just send a light signal from Earth to the planet. But I was wondering, if we made a perfectly non-elastic rope as long as the distance between earth and the far-away planet. Could somebody pulling the rope from the earth send an instant signal to a person holding the end of the rope on the far-away planet? Will the rope side on the other planet move at the exact same moment as the rope side on the Earth? Of course, we should ignore the obvious factors making this means of communication impossible -_-
[ 0.017841152846813202, 0.013903059996664524, -0.0159743819385767, 0.004467744845896959, -0.056239720433950424, 0.0003295165952295065, 0.011846058070659637, 0.03346189111471176, -0.02203788235783577, -0.0333365760743618, -0.013384026475250721, 0.023815179243683815, 0.011179568246006966, 0.01...
[ 0.3503577709197998, -0.3437739312648773, 0.5168178081512451, 0.2330155372619629, -0.42112529277801514, -0.22389914095401764, 0.2676424980163574, -0.007917230948805809, -0.6879674792289734, -0.4551364481449127, 0.16087688505649567, 0.3536771535873413, -0.464825302362442, 0.03641967475414276...
I tried using a different distribution (debian) but the rest of my application did not work. I've tried opkg install mysql-client and opkg install mysql- server however I get the error Unknown package for both.
[ 0.02318621054291725, -0.008307910524308681, -0.007688342593610287, 0.02224469929933548, -0.032767750322818756, 0.047030530869960785, 0.012097478844225407, -0.02752673253417015, -0.03762946277856827, -0.07702147960662842, -0.013706204481422901, 0.019842220470309258, -0.012723804451525211, -...
[ 0.4384143352508545, 0.2528345584869385, -0.1905759871006012, 0.06442149728536606, -0.30018720030784607, -0.23378689587116241, 0.40435123443603516, 0.11494407802820206, -0.1231180727481842, -0.6800879836082458, 0.30468684434890747, 0.6369523406028748, -0.5611975789070129, 0.1315889358520507...
I'm confused as to the correct formula for approximate degrees of freedom to use for Welch's t-test. Satterthwaite's (1946) formula is the most commonly cited formula, but Welch gave an alternative in 1947. I'm not sure which is preferable (or used by most statistical software). Satterthwaite's formula: $$\frac{\left(s_x^2/n_x +s_y^2/n_y\right)^2}{(s_x^2/n_x )^2/(n_x-1)+(s_y^2/n_y )^2/(n_y-1)}$$ Welch's formula: $$-2+ \frac{\left(s_x^2/n_x +s_y^2/n_y\right)^2}{(s_x^2/n_x )^2/(n_x+1)+(s_y^2/n_y )^2/(n_y+1)}$$ References: * Satterthwaite, F.E. (1946). "An Approximate Distribution of Estimates of Variance Components". _Biometrics Bulletin_ , 2, 6, pp. 110–114. * Welch, B.L. (1947). "The generalization of "Student's" problem when several different population variances are involved". _Biometrika_ , 34, 1/2, pp. 28–35.
[ 0.011908218264579773, 0.005685667507350445, -0.026773758232593536, -0.005674777086824179, -0.007457457948476076, 0.019660528749227524, 0.009125543758273125, -0.007557859644293785, -0.01408638246357441, 0.0014646286144852638, 0.001550149405375123, 0.008685526438057423, -0.03676467761397362, ...
[ -0.08375582098960876, -0.0790565237402916, 0.2531689703464508, -0.20520488917827606, 0.05609031021595001, 0.14347019791603088, 0.1837441474199295, -0.4642428457736969, -0.3275648057460785, -0.19751481711864471, 0.03441435098648071, 0.3476647734642029, -0.22003836929798126, 0.56828922033309...
I'm writing a LaTeX document for a hands-on workshop which uses shell/bash commands. I'm making use of the `listings` package to format the chunks of code. Some of the commands are rather long so I'm making use of `breaklines` and `breakatwhitespace`. Here's my MWC: \documentclass{book} \usepackage{listings} \usepackage{color} \usepackage{xcolor} \definecolor{darkgreen}{rgb}{0,0.9,0} \lstset{% columns=fullflexible, aboveskip=5pt, belowskip=10pt, basicstyle=\small\ttfamily, numbers=left, numberstyle=\tiny\color{black!85}, stepnumber=1, numbersep=13pt, backgroundcolor=\color{black!5}, showspaces=false, showstringspaces=false, showtabs=false, xleftmargin=20pt, xrightmargin=10pt, framesep=5pt, framerule=3pt, frame=leftline, rulecolor=\color{darkgreen}, tabsize=2, breaklines=true, breakatwhitespace=true, } \begin{document} \begin{lstlisting} cd ~/ time velveth run_25 25 -fmtAuto -create_binary -shortPaired -separate SRR022863_1.fastq.gz SRR022863_1.fastq.gz -shortPaired2 -separate SRR022852_1.fastq.gz SRR022852_2.fastq.gz time velvetg run_25 \end{lstlisting} \end{document} This is displayed as: ![Current Styling](http://i.stack.imgur.com/gd5IH.png) However, I'd like to the listing to include a line continuation character (`\`) at the ends of autobroken lines. Is there a way that listing can do this? This is what I'd like the listing to look like (with or without line numbers on the indented lines) without explicitly putting in the backslash: ![Required Styling](http://i.stack.imgur.com/qECco.png)
[ -0.0007151247700676322, 0.008028464391827583, -0.010593067854642868, 0.0141893420368433, -0.0006118868477642536, -0.002413788577541709, 0.009659167379140854, -0.008661006577312946, -0.010515008121728897, -0.005567421671003103, -0.012408189475536346, 0.0011173944221809506, 0.00145293248351663...
[ 0.6996789574623108, 0.4031566381454468, 0.34953802824020386, -0.11439503729343414, 0.20212522149085999, -0.0002852328179869801, 0.0475035198032856, -0.3747900724411011, -0.2415412813425064, -0.5817755460739136, 0.06470013409852982, 0.35427677631378174, -0.21298907697200775, 0.0363977625966...
I noticed something strange today. Immediately after I successfully assaulted a holding, my warscore **dropped** about 10 percentage points. I figured it may had something to do with my losses, so I reloaded and just let the siege wait it out untill I won — however, the result was still the same. Shouldn't I **earn** warscore for kicking ass? Secondly, I think in EU3, you would gain much more warscore for capturing capitals. Is this still the case? I've read that you get 75% more warscore for holding all your objective wargoals. As a bonus question, what do you find is the usual point at which an enemy accepts defeat? I heard a warscore of 80% is a good rule of thumb, but I've had a warscore of 97% and the enemy was still extremely reluctant (four minuses) and though "things are going my way", even though I had held a lot of his holdings for five years and won every single battle.
[ -0.019557658582925797, 0.02712995372712612, -0.0041463328525424, 0.009623692370951176, 0.011697156354784966, -0.025135979056358337, 0.00577762583270669, -0.01905002072453499, -0.016058271750807762, 0.03240259364247322, -0.0004475953755900264, 0.020796965807676315, -0.023331942036747932, 0....
[ 0.30288663506507874, 0.016496075317263603, 0.6729339361190796, 0.04296204447746277, -0.8073511719703674, 0.2753971815109253, 0.26907864212989807, -0.02378266677260399, -0.5646580457687378, -0.5237765312194824, 0.4185161292552948, 0.16128133237361908, 0.24608071148395538, 0.0574554763734340...
There is a lecture on how facts about the cosmos can be derived form rather simple observation. Is there an analogy for the small scale? Is there a technically simple observation that would imply, say, atoms? Anything smaller?
[ -0.009027144871652126, 0.020917406305670738, 0.005162826273590326, 0.019467683508992195, -0.012514479458332062, -0.00345193175598979, 0.0123458756133914, -0.003518481971696019, -0.02878878451883793, -0.0194262582808733, 0.0019596137572079897, 0.011572644114494324, 0.040557537227869034, -0....
[ 0.35885608196258545, -0.17562034726142883, -0.12109582126140594, 0.6546195149421692, 0.06276066601276398, 0.14299172163009644, -0.24392394721508026, 0.15420308709144592, -0.21642428636550903, -0.4348425567150116, -0.1394568383693695, 0.3249983489513397, 0.06915926933288574, 0.2418966740369...
Why does the dipole moment of an electric dipole of different charges depend on the distance from the origin? Physically, I don't understand why something that measures how much of a dipole something is should depend on how far away that thing is from the origin.
[ 0.0013456010492518544, 0.03056461364030838, -0.004983640741556883, 0.031669728457927704, -0.02300536073744297, -0.031040571630001068, 0.017451291903853416, 0.016328537836670876, -0.02428860403597355, -0.03181842714548111, -0.008582927286624908, 0.03798763453960419, 0.013084892183542252, 0....
[ 0.5257941484451294, -0.07732469588518143, 0.0213658157736063, 0.3399895429611206, -0.19263948500156403, -0.21037611365318298, -0.34252604842185974, -0.2180994302034378, -0.21056002378463745, -0.050597164779901505, 0.2719402313232422, 0.01983613893389702, -0.03509087860584259, 0.77414417266...
> 1: _He is nice_ except that _he is a little shy_. > > 2: Except that _he is a little shy_ , _he is nice_. Are these sentences both correct? If so, are these two sentences identical?
[ 0.012592646293342113, 0.025892358273267746, -0.006953111384063959, 0.03362947329878807, -0.002754895482212305, 0.044060397893190384, 0.011795575730502605, -0.01747378334403038, -0.019727688282728195, 0.004924552980810404, -0.01615053415298462, -0.0036362851969897747, 0.007288150954991579, ...
[ -0.4071664810180664, 0.28704410791397095, 0.06764469295740128, -0.029307877644896507, -0.34069177508354187, 0.1978238821029663, 0.5779266953468323, 0.09115477651357651, -0.021459287032485008, -0.7746466398239136, -0.3503868579864502, 0.27504774928092957, -0.14828236401081085, -0.5535427331...
## Scenario I want to write a tutorial on PSTricks, but because `latex-dvips-ps2pdf` (see the code shown in First Attempt section) makes my effort to prevent reader from copying the listing numbers no longer work \usepackage{accsupp} \newcommand*{\noaccsupp}[1]{\BeginAccSupp{ActualText={}}#1\EndAccSupp{}} and I cannot use `microtype` and the output is a bit strange as follow ![enter image description here](http://i.stack.imgur.com/tLRtD.png) As a result, this problem forced me to use `pdflatex` (see Second Attempt section). ## First Attempt (latex-dvips-ps2pdf) \documentclass[dvipsnames,border=12pt,preview]{standalone} \usepackage{pstricks} \usepackage{showexpl} \usepackage{accsupp} \newcommand*{\noaccsupp}[1]{\BeginAccSupp{ActualText={}}#1\EndAccSupp{}} \makeatletter \newlength{\parindent@save} \setlength{\parindent@save}{\parindent} \lstdefinestyle{PSTricks} { language={[LaTeX]TeX}, alsolanguage={PSTricks}, basicstyle=\small\ttfamily\null, keywordstyle=\color{blue}, backgroundcolor=\color{yellow!10}, numbers=left, numbersep=1em, numberstyle=\tiny\color{Red}\noaccsupp, frame=single, framesep=\fboxsep,% expands outward, cannot affect if frame=none framerule=\fboxrule,% expands outward, cannot affect if frame=none rulecolor=\color{red},% cannot affect if frame=none xleftmargin=\dimexpr\fboxsep+\fboxrule, xrightmargin=\dimexpr\fboxsep+\fboxrule, breaklines=true, breakindent=0pt, tabsize=5, columns=flexible, preset={\setlength{\parindent}{\parindent@save}\centering}, } \makeatother \begin{document} \LTXexample[pos=t,style=PSTricks] \begin{pspicture}[showgrid](3,3) \pscircle[linecolor=red](1.5,1.5){1.5} \end{pspicture} \endLTXexample \end{document} ## Second Attempt (pdflatex) Using `pdflatex` can prevent users from copying the listing numbers. However, `preset=\centering` is no longer work when I use `graphic.` \documentclass[dvipsnames,border=12pt,preview]{standalone} \usepackage{xcolor} \usepackage{showexpl} \usepackage{accsupp} \newcommand*{\noaccsupp}[1]{\BeginAccSupp{ActualText={}}#1\EndAccSupp{}} \makeatletter \newlength{\parindent@save} \setlength{\parindent@save}{\parindent} \lstdefinestyle{PSTricks} { language={[LaTeX]TeX}, alsolanguage={PSTricks}, basicstyle=\small\ttfamily\null, keywordstyle=\color{blue}, backgroundcolor=\color{yellow!10}, numbers=left, numbersep=1em, numberstyle=\tiny\color{Red}\noaccsupp, frame=single, framesep=\fboxsep,% expands outward, cannot affect if frame=none framerule=\fboxrule,% expands outward, cannot affect if frame=none rulecolor=\color{red},% cannot affect if frame=none xleftmargin=\dimexpr\fboxsep+\fboxrule, xrightmargin=\dimexpr\fboxsep+\fboxrule, breaklines=true, breakindent=0pt, tabsize=5, columns=flexible, preset={\setlength{\parindent}{\parindent@save}\centering}, } \makeatother \usepackage{filecontents} \begin{filecontents*}{diagram.tex} \documentclass[pstricks,border=12pt]{standalone} \begin{document} \begin{pspicture}[showgrid](3,3) \pscircle[linecolor=red](1.5,1.5){1.5} \end{pspicture} \end{document} \end{filecontents*} \immediate\write18{latex diagram.tex && dvips diagram.dvi && ps2pdf diagram.ps} \begin{document} \LTXinputExample[pos=t,style=PSTricks,graphic={[scale=1]"diagram"}]{diagram.tex} \end{document} **Original Output:** ![enter image description here](http://i.stack.imgur.com/UkxW8.png) **Expected Output:** ![enter image description here](http://i.stack.imgur.com/KwTj1.png) ## Question How to center the graphic of `\LTXexample` for the code shown in the Second Attempt section?
[ 0.0070724948309361935, 0.012195777148008347, 0.0023710455279797316, 0.02586732804775238, 0.012169057503342628, 0.008081966079771519, 0.008782342076301575, 0.001015226822346449, -0.0165039524435997, -0.003847454907372594, -0.01962311938405037, -0.0013038456672802567, -0.001929297111928463, ...
[ -0.06156253442168236, -0.08187325298786163, 0.44620949029922485, 0.15007495880126953, 0.24429404735565186, -0.014665164053440094, 0.3171107769012451, -0.36517924070358276, -0.06912770867347717, -0.7969562411308289, 0.031538307666778564, 0.38512086868286133, -0.465440034866333, 0.0126281660...
on my theme, i found this script add_image_size( 'slider', 464, 249, true ); add_image_size( 'fmenu', 306, 280, true ); add_image_size( 'teamthumb', 138, 207, true ); add_image_size( 'imlink', 286, 140, true ); add_image_size( 'albmlink', 274, 274, true ); add_image_size( 'fppost', 90, 90, true ); add_image_size( 'fslide', 520, 280, true ); i have disabled this, but when i upload another image, its generates 3 more images... like this: original_file.jpg origiinal_file-100x100.jpg origiinal_file-109x109.jpg i need only to upload the original file, there's a way to solve this?? I can't find where is this another 2 images sizes...
[ 0.0028266150038689375, -0.0027604501228779554, 0.0003463309840299189, 0.007195951417088509, 0.026803985238075256, -0.009949395433068275, 0.005145201925188303, 0.0009305310668423772, -0.009593775495886803, 0.0011560152051970363, -0.01650679111480713, 0.0019536507315933704, 0.00890009663999080...
[ -0.16009554266929626, -0.010773928835988045, 0.9923124313354492, 0.03794068843126297, -0.13142849504947662, 0.3585350513458252, 0.07020041346549988, -0.2255638688802719, -0.18146465718746185, -0.717657208442688, -0.3427900969982147, 0.5302435159683228, -0.3109697103500366, -0.2791722118854...
According to Davison and Hinkley's book "Bootstrap Methods and their Application", page 154, based on an `RxN` array where `R=999` is the number of simulations and `N` is the number of ordered samples (smallest to largest) in each row, we could construct: 1. **pointwise** test envelope: by extracting the 2.5th and 97.5th quantile of each column, and then connect the `N` (ordered) dots 2. **simultaneous** test envelope: "we first compute **columnwise ranks**. Then we calculate the proportion of rows in which either the minimum rank is less than or equal to `k`, or the maximum rank is greater than or equal to `R + 1 — k`, or both" Here `k = p(R+1) = 0.05*(999+1) = 50` and `R + 1 - k = 950`. I am confused about the construction of #2 "simultaneous test envelope" above (the explanation is unclear to me...) -- can it be directly made with the `R x N` matrix? Could anyone help me in explaining the procedure in detail? Thanks! Here is the plot that shows both the pointwise (dashed) and simultaneous (solid) envelopes from the book: ![enter image description here](http://i.stack.imgur.com/Z6xhY.png)
[ 0.011689558625221252, 0.0028627675492316484, -0.01342313177883625, 0.005733699072152376, -0.01464929711073637, 0.006476295180618763, 0.007727505639195442, -0.006252309773117304, -0.011269439943134785, -0.012106429785490036, -0.010703841224312782, -0.005488050170242786, -0.018813159316778183,...
[ -0.1495058387517929, -0.1268502175807953, 0.04156051576137543, 0.11647959053516388, 0.046673621982336044, 0.6088268160820007, 0.12144626677036285, -0.8314967751502991, -0.07405950874090195, -0.4780746400356293, 0.09688185155391693, 0.09678847342729568, -0.1790059208869934, 0.18515108525753...
Not all functions seem to work with `SetOptions`. e.g. SetOptions[Grid, BaseStyle -> Directive[Red]]; Grid[{{"hello", "world"}}] hello world the font _is not_ red. SetOptions[Row, BaseStyle -> Directive[Red]]; Row[{"hello", "world"}] hello world ...and the font _is_ red. ![enter image description here](http://i.stack.imgur.com/TWDM4.png) SetOptions[InputField, FieldSize -> 5]; InputField[Dynamic[x]] the input field size is _much_ larger than 5. But on the other hand InputField[Dynamic[x], Sequence @@ Options[InputField]] yields an input field with field size 5. ![enter image description here](http://i.stack.imgur.com/XAS9x.png) ...and so on. What is the easiest way to work out (i.e. make a list of ...) which functions can't be used with `SetOptions`?
[ 0.003938863053917885, 0.010016816668212414, 0.0027394266799092293, 0.014564551413059235, -0.0014613564126193523, -0.011728959158062935, 0.006959210615605116, -0.0024692192673683167, -0.01030644029378891, 0.007775604259222746, -0.017994839698076248, 0.007487909868359566, 0.0004831306869164109...
[ -0.03241490572690964, -0.17274798452854156, 0.47451767325401306, -0.2403387576341629, -0.018312007188796997, 0.1373773068189621, 0.2047768235206604, -0.18422828614711761, -0.1881188601255417, -0.631615936756134, -0.19882343709468842, 0.7289993762969971, -0.5341042280197144, -0.105836048722...
I think my question is pretty simple, but since I've been using LaTex for 2 days by now... Anyway, the thing is I'm having a problem adding a new section. I use the command `\section`, but in the previous section there is three subsections, the last one is filled up with lots of images and when I input a new section, the section begins right after the first image of the previous subsection and I don't know how to solve it. End of the code: %last picture of the subsection% \begin{figure}[h] \centering \includegraphics[width=1\textwidth]{28_google} \caption{$\sqrt[3]{28}$ \hspace{0.5mm} calculated by Google with 11 places of precision and 11 places after the comma.} \label{fig:28_google} \end{figure} \section {Conclusions} This section appears after the first image instead of the end of the subsection. If anyone could help, thank you! PS: I want the fourth section to start after the previous subsection (the one with images).
[ -0.0016864468343555927, 0.012131772004067898, -0.014085337519645691, 0.022538410499691963, 0.0019060634076595306, 0.001448431983590126, 0.00704880291596055, 0.0042647444643080235, -0.016482306644320488, 0.007142098154872656, -0.017597291618585587, -0.0015245031099766493, -0.01190075092017650...
[ 0.5071866512298584, 0.17729884386062622, 0.6796315908432007, -0.06029164046049118, 0.15522952377796173, 0.23796629905700684, 0.0044982656836509705, -0.05977129563689232, -0.3716016411781311, -0.45243126153945923, 0.09886321425437927, 0.22668452560901642, -0.31610891222953796, 0.24909687042...
I want to encourage users to register on my website, so as a registered member perk, the ads would be removed, so I was thinking of implementing AdSense in a div that slides out from the bottom of the page and has a Close button (for non-registered users). Would either of this be against Google's TOS?
[ -0.01545278262346983, 0.006168273743242025, 0.00327541446313262, 0.020385978743433952, -0.023276764899492264, 0.01686389558017254, 0.012087679468095303, 0.01086699590086937, -0.025505362078547478, 0.034113023430109024, -0.020362062379717827, 0.01337889302521944, 0.0036704347003251314, 0.03...
[ 0.669028103351593, -0.03142857924103737, 0.5587801933288574, -0.10044195502996445, -0.06219838559627533, -0.15302826464176178, 0.0034976874012500048, -0.14574658870697021, -0.11997807770967484, -0.15546034276485443, 0.3089304566383362, 0.2160131335258484, -0.31975266337394714, -0.026076154...
Need to brush up on my late-undergrad and early-grad physics and was wondering if anyone can recommend books or lecture notes (hard copy, or on-line) that also have solutions. Two that I have come across are: **Princeton Problems in Physics with Solutions** \- Nathan Newbury **University of Chicago Graduate Problems in Physics with Solutions** \- Jeremiah A. Cronin **Spacetime Physics** \- Taylor & Wheeler (favorite book on special relativity; has a lot of problems with solutions at the back; a lot of the problems really enforce the material and discuss paradoxes) If possible, please also provide a reason why you like the books as opposed to just listing them.
[ 0.002785841003060341, -0.0024513236712664366, -0.010031450539827347, 0.019809970632195473, -0.0003806161694228649, -0.00925089605152607, 0.00854790210723877, 0.002707070903852582, -0.016829604282975197, -0.004882515873759985, 0.010596404783427715, 0.010881108231842518, -0.006212015636265278,...
[ -0.20202863216400146, 0.25562500953674316, -0.0324455201625824, 0.2520763874053955, 0.008222680538892746, 0.040715981274843216, -0.06308970600366592, 0.3366900682449341, -0.31106600165367126, -0.6509470343589783, -0.240737646818161, 0.2733016312122345, -0.0516953319311142, -0.0673080608248...
Since upgrading to WP 3.9, my clients have been complaining that the new editor keeps _too much_ formatting. They'd like it so when you paste it keeps the formatting of bold, underline, paragraph breaks, etc. but ignores any color formatting. They tell me the old tinymce editor did this. Is there a way to do this? No look searching around the web and here. Thanks.
[ -0.00006182328070281073, -0.0003786800953093916, -0.009370904415845871, 0.026326095685362816, 0.013479367829859257, 0.003331175772473216, 0.008891241624951363, 0.01199389435350895, -0.024467289447784424, 0.016148807480931282, -0.022971531376242638, 0.010152988135814667, 0.005192536860704422,...
[ 0.4162085950374603, 0.169989213347435, 0.53688645362854, 0.12633271515369415, 0.08766170591115952, -0.1803242713212967, 0.1746959686279297, 0.2639889717102051, -0.42094680666923523, -0.7867911458015442, 0.08876679837703705, 0.8026911020278931, -0.1173204779624939, 0.041944317519664764, -...
**Problem** : For a $P,V,T$ system with constant mass,entropy can be considered a function of $U,V$, i.e. $S=S(U,V)$. The goal is to derive the rest of the thermodynamic functions via Legendre transformations. **Attempted answer** : I start with the 1st law of thermodynamics $dU = TdS - PdV$, which I solve for $dS$: \begin{equation}dS=(1/T)dU+(P/T)dV\end{equation} Let, $Y\equiv S,C_1\equiv 1/T,X_1\equiv U, C_2\equiv P/T, X_2\equiv V$. The first Legendre transformation and the associated differential are: \begin{equation}\widetilde{T1} = Y - C_1 X_1 = S - (1/T)U\end{equation} \begin{equation}d\widetilde{T1} = dS - (1/T)dU + (1/T^2)UdT = (P/T)dV +(U/T^2)dT\end{equation} Therefore $\widetilde{T1} = \widetilde{T1}(V,T)$. Does this seem correct to you ? 1. Do I stop here for $\widetilde{T_1}$ or is there more to do? 2. What are the natural variables for $\widetilde{T_1}$? I am a little concerned due to the fact that $T$ is intensive variable and if I recollect correctly, the natural variables are always extensive.
[ -0.003947280813008547, 0.016436956822872162, -0.013338655233383179, 0.01095874235033989, -0.025949113070964813, -0.012741532176733017, 0.007839212194085121, -0.009820498526096344, -0.011958085000514984, -0.021876299753785133, -0.010427627712488174, 0.010003024712204933, -0.040784433484077454...
[ -0.12087765336036682, -0.11863890290260315, 0.46835431456565857, 0.009983648546040058, 0.17680048942565918, 0.4249204099178314, 0.08190549910068512, -0.7122749090194702, -0.23033109307289124, -0.5326823592185974, -0.2748304009437561, 0.3054015338420868, -0.06481246650218964, 0.513055205345...
Is there a good plugin for rating blog posts that doesn't depend on PollDaddy or another external service? Ideally I'd like a plugin similar in appearance to Polldaddy Polls & Ratings, but without the PollDaddy dependency (I'd prefer to have the rating data in the WP database). _EDIT: just to make things clear: I'm not interested in polls, only in rating for blog posts_ * * * What I tried so far: * GD Star rating: too complex, and doesn't work at all for me (*) * WP-PostRatings: ~~doesn't work at all for me (*)~~ _(my mistake... I didn't realize I had to modify the theme. This one works fine)_ * kk Star ratings: doesn't look too good with my theme (Garland revisited) * Post ratings: simple, good looking, but there's a bug with my theme; ~~I'll try to fix the CSS if I can't find anything better~~ _(OK, I'm really, really bad at CSS...)_ * Rating-Widget: really nice, but depends on external site _(*) perhaps this is because I have an unusual setup (Windows/IIS server, SQL Server database with WP database abstraction plugin, multi-site with domain mapping)_
[ -0.006716263014823198, 0.011790351942181587, 0.005900188349187374, 0.02182609960436821, 0.0034146669786423445, 0.014217907562851906, 0.00772142643108964, -0.013843245804309845, -0.01836799830198288, 0.0002561332657933235, -0.012503759935498238, 0.017664384096860886, 0.0025765900500118732, ...
[ 0.4497154653072357, -0.04310812056064606, 0.2885725796222687, 0.35717007517814636, -0.14840134978294373, -0.10694176703691483, 0.22709424793720245, 0.06926410645246506, -0.08985395729541779, -0.4033818244934082, 0.41088688373565674, 0.5682075023651123, -0.12758176028728485, -0.011360326781...
I've been reading up about the repository pattern, with a view to implementing it in my own application. Almost all examples I've found on the internet use some kind of existing framework rather than showing how to implement it 'from scratch'. Here's my first thoughts of how I might implement it - I was wondering if anyone could advise me on whether this is correct? I have two tables, named CONTAINERS and BITS. Each CONTAINER can contain any number of BITs. I represent them as two classes: class Container{ private $bits; private $id; //...and a property for each column in the table... public function __construct(){ $this->bits = array(); } public function addBit($bit){ $this->bits[] = $bit; } //...getters and setters... } class Bit{ //some properties, methods etc... } Each class will have a property for each column in its respective table. I then have a couple of 'repositories' which handle things to do with saving/retrieving these objects from the database: //repository to control saving/retrieving Containers from the database class ContainerRepository{ //inject the bit repository for use later public function __construct($bitRepo){ $this->bitRepo = $bitRepo; } public function getById($id){ //talk directly to Oracle here to all column data into the object //get all the bits in the container $bits = $this->bitRepo->getByContainerId($id); foreach($bits as $bit){ $container->addBit($bit); } //return an instance of Container } public function persist($container){ //talk directly to Oracle here to save it to the database //if its ID is NULL, create a new container in database, otherwise update the existing one //use BitRepository to save each of the Bits inside the Container $bitRepo = $this->bitRepo; foreach($container->bits as $bit){ $bitRepo->persist($bit); } } } //repository to control saving/retrieving Bits from the database class BitRepository{ public function getById($id){} public function getByContainerId($containerId){} public function persist($bit){} } Therefore, the code I would use to get an instance of Container from the database would be: $bitRepo = new BitRepository(); $containerRepo = new ContainerRepository($bitRepo); $container = $containerRepo->getById($id); Or to create a new one and save to the database: $bitRepo = new BitRepository(); $containerRepo = new ContainerRepository($bitRepo); $container = new Container(); $container->setSomeProperty(1); $bit = new Bit(); $container->addBit($bit); $containerRepo->persist($container); Can someone advise me as to whether I have implemented this pattern correctly? Thanks!
[ 0.010505003854632378, 0.02528519183397293, -0.010237645357847214, 0.017988745123147964, 0.013212419115006924, -0.0074167107231915, 0.006098932586610317, -0.007773816119879484, -0.01561732031404972, -0.008154172450304031, -0.012002510949969292, 0.008867273107171059, -0.006903531961143017, -...
[ 0.36134395003318787, -0.015943652018904686, -0.034856460988521576, 0.29912030696868896, 0.1561068594455719, 0.47861096262931824, -0.21502891182899475, 0.006426944397389889, -0.11412975192070007, -0.5898913145065308, 0.10771873593330383, 0.1383284479379654, 0.03352511301636696, 0.4930873513...
Actually I had the haptic feedback option enabled by default when I bought the Samsung Galaxy Y mobile. But it seems that somebody has changed the Settings. So I would like to know how to enable haptic feedback in Samsung Galaxy Y? But I need it badly. Anybody please help out. Thanks in advance.
[ 0.007424537092447281, -0.021696284413337708, -0.025161944329738617, 0.030865689739584923, 0.022595444694161415, -0.0035185576416552067, 0.010816846042871475, 0.01269068755209446, -0.032004307955503464, -0.04354831203818321, -0.016592731699347496, 0.006463482044637203, -0.030987989157438278, ...
[ 0.6179934144020081, 0.2444174587726593, 0.656278133392334, 0.27967143058776855, -0.16159817576408386, -0.07202395051717758, 0.23604370653629303, 0.43090131878852844, -0.1940739005804062, -0.30341169238090515, 0.0444304533302784, 0.4945766031742096, -0.10945918411016464, 0.05710713565349579...
I'm writing a paper in IEEE journal format. I have a BIG matrix as follows: \begin{align} \begin{bmatrix} \begin{pmatrix} E^\top KA &0 \\ 0 &V^\top E^\top U \hat{K} U^\top AV \end{pmatrix} + \begin{pmatrix} A^\top KE &0 \\ 0 &V^\top A^\top U \hat{K} U^\top EV \end{pmatrix} &\begin{pmatrix} E^\top KB \\ V^\top E^\top U \hat{K} U^\top B \end{pmatrix} \\ \begin{pmatrix} B^\top KE &B^\top U \hat{K}U^\top EV \end{pmatrix} &0 \end{bmatrix} + \cr \begin{bmatrix} \begin{pmatrix} C^\top C &-C^\top CV \\ -V^\top C^\top C &V^\top C^\top CV \end{pmatrix} &0 \\ 0 &-\gamma^2 I \end{bmatrix} = Q_2 \preceq 0 \end{align} How can i make it fit in the two column paper format? I am relatively new to LaTeX. Thanks in advance.
[ 0.005566657055169344, 0.018961764872074127, -0.00631968816742301, 0.0175640769302845, 0.0026751321274787188, 0.02490866556763649, 0.006794069893658161, -0.0036337701603770256, -0.009941760450601578, 0.0012016696855425835, -0.006363751832395792, -0.001550557790324092, -0.02459256537258625, ...
[ 0.09094439446926117, 0.4724434018135071, 0.7560023665428162, -0.19905978441238403, 0.09259086847305298, 0.2488918900489807, -0.28881874680519104, -0.44860905408859253, -0.10415112972259521, -0.2766245901584625, 0.03469858318567276, 0.21150580048561096, 0.0010854860302060843, 0.381486266851...
I recently posted this answer to How to highlight all words of the form [0-9][A-Za-z0-9]* immediately following an equal sign?. It works, but it suffers from (a lot) of code duplication. ## What I have at the moment More specifically, I have to use a line of the form \lst@DefSaveDef{`<char>}\jubobs@<char>{\jubobs@<char>\foo} no fewer than **52 times** , for `<char>` in `A`-`Z` and `a`-`z`. For information, `\lst@DefSaveDef` is a `listings` internal macro with the following syntax: \lst@DefSaveDef{<charcode>}<some-macro>{<replacement-text>} Using my newfound understanding of `\expandafter`, I've managed to somewhat reduce the code by defining a macro that takes a letter as argument and performs the same thing as the line show above: \makeatletter \newcommand\addtoletterdef[1] {% \expandafter\expandafter\expandafter\lst@DefSaveDef% \expandafter\expandafter\expandafter{% \expandafter\expandafter\expandafter`% \expandafter\expandafter\expandafter#1% \expandafter\expandafter\expandafter}% \expandafter\csname jubobs@#1\expandafter\endcsname% \expandafter{\csname jubobs@#1\endcsname\foo}% } \makeatother However, that didn't help much in reducing code duplication: I still have to invoke `\addtoletterdef` 52 times... `;...(` ## What I want I would like to define a macro similar to my `\addtoletterdef` but that would accept a charcode instead of a character. That way, it would be easy to loop through charcodes 65 to 90 (A-Z) and 97 to 122 (a-z) and perform all those 52 operations without any code duplication. I've got a feeling that the answer lies in using the `\begingroup\lccode` trick, but that trick is very new to me, and I'm far from mastering it. **How can I redefine`\addtoletterdef` to accept a charcode instead of a character?** * * * MWE: \documentclass{article} \usepackage{listings} % dummy macro \def\foo{foo} \makeatletter \newcommand\addtoletterdef[1] {% \expandafter\expandafter\expandafter\lst@DefSaveDef% \expandafter\expandafter\expandafter{% \expandafter\expandafter\expandafter`% \expandafter\expandafter\expandafter#1% \expandafter\expandafter\expandafter}% \expandafter\csname jubobs@#1\expandafter\endcsname% \expandafter{\csname jubobs@#1\endcsname\foo}% } \makeatother \lstdefinestyle{mycode} { language=C, SelectCharTable= \addtoletterdef{A} % ... \addtoletterdef{Z} \addtoletterdef{a} % ... \addtoletterdef{z} } \begin{document} \begin{lstlisting}[style=mycode] a1 = 6.12234Z a2 = Z1324124 \end{lstlisting} \end{document}
[ 0.006363963708281517, 0.008698596619069576, -0.02404721826314926, 0.0014833607710897923, -0.0050514014437794685, 0.016127431765198708, 0.006332266144454479, 0.00822461023926735, -0.013656861148774624, -0.02013222500681877, -0.007020613178610802, 0.0008659020531922579, -0.007642109412699938, ...
[ 0.3260740041732788, 0.21729230880737305, 0.4534951448440552, -0.4202408492565155, -0.2689622938632965, 0.40681958198547363, 0.5320714116096497, 0.12121320515871048, -0.17200808227062225, -0.7670522928237915, -0.07393032312393188, 0.2723821699619293, -0.10330947488546371, -0.000848379160743...
The short question: Has anyone used _Mathematica_ in conjunction with Hadoop and does _Mathematica_ 's built in parallelization play well with Hadoop? The long version: So I have a _Mathematica_ program that I would like to do the following: 1. I have a kernel does some initial computations and produces some sets of equations which it outputs as context files to a bucket of some kind. The way it does this is essentially through searching a binary tree until it either a) finds a solution, b) finds a contradiction and thus prunes that branch, or c) it can't solve it for some reason. 2. I have several remote kernels running which monitor this directory and pick out context files (which are essentially sets of equations) to try and solve. If they succeed, they throw the solutions that they've found into a bucket for solutions. If they produce more equations to be solved I want them to put them into context files and put them back in the original bucket. If they fail for some reason (which is for all intents and purposes saying that the algorithm I just used to try and solve them did not work), I want them to save the context they are working on as is and put it in a separate bucket that I somehow mark "hard". 3. I want to have certain kernels which are marked to look into the "hard" bucket and try more intensive algorithms for solving them. I would like for them to do this in an "intelligent" way, whatever that ends up being. 4. I produce new sets of context files for computation by recursing further down the tree. I would like to (somehow) treat my bucket as a priority queue so that context files generated at greater depth are given priority over those closer to the root. 5. When all is said and done and I have (hopefully) produced all of the sets of solutions that I can to this system of equations, I want to have a kernel that goes through the sets of solutions and computes when they are equivalent. We (since this is certainly not a one person effort) have been looking at using the Parallelization capabilities built into Mathematica for this task. Some of the advantages to this are that when I initialize remote kernels, _Mathematica_ is supposed to have a means of making sure that the context running in that kernel has certain appropriate definitions. There are a few apparent problems we have identified: One is in handling the file distribution. Ideally we would like to make sure that two kernels are not trying to solve the equations in the same context at the same time. What also happens is that at some point all of the remote kernels are doing disk reads and writes from the same directory, which would probably be bad. Additionally, as it stands right now, the way we can think of to do this with the built in _Mathematica_ parallelization requires that all communication go through the original kernel which spawned the process. We would like to decentralize the algorithm to make it as modular as possible. Finally, not all of the problems that this software is used to solve are beyond the realm of a single kernel, however as it stands doing things in the current version using only a single kernel still requires treating the program as parallelized. I am basically familiar with Hadoop, its DFS, and the MapReduce paradigm that it uses. As I see it, steps 1-4 above could be considered as the map step of an algorithm and step 5 could be a reduce step. Additionally, the HDFS seems like it would provide a solution to the file system problems. The potential problem with hadoop is how to implement access through Mathematica. I have run across the HadoopLink (https://github.com/shadanan/HadoopLink) project, and it seems like the goal of the project could help provide some of the framework we desire. However, some of us have already done work on implementing a solution using the _Mathematica_ parallelization functions, and it is highly desirable to not have to abandon this code, especially since it's what's likely to be optimized for doing parallelization with _Mathematica_. My questions are as follows: 1. As above, has anyone had any experience with trying to get these two things to work together, and if so, was it worth it? 2. Does anyone have any experience with the HadoopLink project, and is it compatible with the _Mathematica_ parallelization? I'm emailing the github project owner, but there are also three other forks out there for the project. 3. Would this be killing a fly with a bazooka and using a 40 lb sledge where a 12oz claw hammer would do? Part of the reason for investigating this is that we anticipate having access to a moderate (few dozen) machines on which we can run kernels. On the one hand we would like to not get stuck thinking too small and having to implement something new all over again as we scale up, and on the other hand we don't want to waste time because we were trying to anticipate problems it was never reasonable to expect in the first place. 4. Is there a better way to do this? At this point, things are rather exploratory and I have no problems with taking new suggestions. Thanks!
[ 0.0060666585341095924, 0.009715375490486622, 0.0036052498035132885, 0.022691424936056137, 0.017736800014972687, 0.0034461289178580046, 0.008029766380786896, -0.012256067246198654, -0.016998454928398132, -0.040821004658937454, -0.004718022421002388, 0.013796437531709671, -0.010564119555056095...
[ -0.07450340688228607, -0.08102910220623016, -0.041658807545900345, 0.18705478310585022, -0.31370973587036133, -0.032642945647239685, 0.05880898982286453, 0.17717988789081573, -0.2768888473510742, -0.30059948563575745, 0.11312758922576904, 0.5839515924453735, -0.27378910779953003, -0.062695...
I'm looking for some inspiration on the best way to run a php script away from the template. I know that the wordpress core files won't have loaded, so I am wondering what workgflow you follow when doing this? cheers
[ 0.020419076085090637, 0.02666730433702469, 0.015621406026184559, 0.02874143049120903, -0.03949752077460289, 0.008953414857387543, 0.008507413789629936, -0.008200960233807564, -0.03260411322116852, -0.017126539722085, -0.025112511590123177, 0.012273631989955902, 0.0023761617485433817, 0.018...
[ 0.421623170375824, 0.4592934846878052, 0.6892278790473938, -0.016532596200704575, -0.5708605647087097, -0.2931954860687256, 0.5237939357757568, 0.3640756905078888, -0.22474271059036255, -0.40053996443748474, 0.2005898654460907, 0.5488665103912354, 0.5701983571052551, 0.11828664690256119, ...
I am a newb in writing this and since this is dangerous stuff to fidlle with (if you are not sure what you are doing) can you please verify it? I use wordpress multisite, this is why I chose to allow blogs.dir I want only my posts and categories indexed. That's it:) User-agent: * Disallow: /cron/ Disallow: /lo/ Disallow: /portfolio.html Disallow: /wp-admin/ Disallow: /wp-content/ Disallow: /wp-includes/ Disallow: /wp-login.php Allow: /wp-content/uploads/ Allow: /wp-content/blogs.dir/ Disallow: /author/ Disallow: /tag/ Disallow: /archives/ Disallow: /2010/* Disallow: /2011/* Disallow: /2012/* Disallow: /about/ Disallow: /trackback Disallow: */trackback Disallow: /comments Disallow: */comments Disallow: /feed Disallow: */feed Disallow: /*.php$ Disallow: /*.js$ Disallow: /*.inc$ Disallow: /*.css$ Disallow: /*.gz$ Disallow: /*.wmv$ Disallow: /*.cgi$ Disallow: /*.xhtml$ Disallow: /*.xlsx $ Disallow: /*.doc$ Disallow: /*.pdf$ Disallow: /*.zip$ # Google Image User-agent: Googlebot-Image Disallow: Allow: /* Ty, take care;)
[ -0.0008831527084112167, 0.010373665951192379, -0.004540733061730862, 0.018872573971748352, -0.00898829661309719, 0.009881366044282913, 0.007618192117661238, 0.024289924651384354, -0.015892695635557175, -0.00374020729213953, -0.012562533840537071, 0.005069204606115818, -0.007667035795748234, ...
[ 0.26465606689453125, -0.0018496779957786202, 0.756497859954834, -0.01938362419605255, -0.23322099447250366, -0.3743126392364502, -0.04969058558344841, -0.32509052753448486, -0.09512760490179062, -0.435814768075943, 0.15418657660484314, 0.8817360401153564, -0.09036193788051605, -0.136454656...
I have a question regarding mixed effect models. Is it possible that the degrees of freedom differ when I use the same model with different dependent variables? Or does that indicate that something went wrong? I analyzed my data using the MIXED procedure in SPSS. There are two random factors (participants and stimuli) and one fixed (continuous) factor (PD) in my design. The number of degrees of freedom is 1201,436 for my first dependent variable "perf". When I use the same model and only choose a different dependent variable ("wr") the number of degrees of freedom is 1194,166. The number of data sets and the model dimensions are identical for both dependent variables. The information criteria differ (e.g., AIC for "perf" = 3975 and AIC for "wr" = 9942). The tests of fixed effects are significant for both dependent variables. Does it make sense that the degrees of freedom differ?
[ 0.02024376578629017, 0.013017712160944939, -0.02488355152308941, 0.01598295569419861, -0.001876726862974465, -0.011916765943169594, 0.010119406506419182, -0.018402036279439926, -0.013400018215179443, -0.030427981168031693, -0.002670634537935257, 0.01454920694231987, -0.023677602410316467, ...
[ 0.06061386689543724, -0.13266946375370026, 0.5304383039474487, 0.14669252932071686, -0.48909488320350647, 0.6461728811264038, 0.2531796991825104, -0.6265220642089844, -0.5074796080589294, -0.3057374656200409, 0.1821591556072235, 0.11495700478553772, -0.2471543848514557, 0.44128018617630005...
I have asked this question in http://tex.stackexchange.com/ but then i got a reply that it is off topic there and webmasters stack exchange site would be useful. This is a continuation from my previous question: http://tex.stackexchange.com/questions/164670/convert-tex-file-to-html-in- miktex?noredirect=1#comment378543_164670 As I have realized that LaTeX cannot be put directly in WordPress pages, I tried to convert into HTML and use that in WordPress, and I am facing the following problem. I would like to write some mathematical notes on a WordPress site. I am expecting the output to be something like http://crazyproject.wordpress.com/2010/04/24/stabilizer-commutes-with- conjugation-2/ But instead I am getting for the same content output as ![enter image description here](http://i.stack.imgur.com/UpScX.png) I do not really understand what is the problem. I think all the mathematical symbols in the Crazy Project link are actually images, whereas mine are just LaTeX code converted into HTML. I am waiting for some better idea to convert my ugly looking statement to change to something like that in Crazy Project notes. Another problem I am facing is: I read that to write something in WordPress with LaTeX, I should just write `$latex {code }$`, but then it is showing the error "formula does not parse" in red colour. What am I supposed to do with this kind of problem? Am I not allowed to just copy the content of LaTeX code I have written before and paste it in WordPress? If I am writing all those just by typing, I am not getting any problem, but if I copy the code then I get the "formula does not parse" error.
[ 0.002992208581417799, 0.00021181997726671398, -0.001181459752842784, 0.01539185456931591, 0.03243072330951691, 0.006244994699954987, 0.007641866337507963, 0.01669396087527275, -0.01844816841185093, -0.015415452420711517, -0.003829195164144039, 0.006563510745763779, -0.005071884021162987, 0...
[ 0.2215622216463089, 0.10671399533748627, 0.5108646154403687, -0.23826688528060913, -0.05096223205327988, 0.09238074719905853, -0.13501901924610138, -0.12864452600479126, -0.24914833903312683, -0.4939649701118469, -0.04337944835424423, 0.49030429124832153, -0.32992836833000183, 0.0215238686...
As a supplement to this question as to whether particles can be observers, supposing that the answer is yes. One could suppose a setup where particle A is observing particle B, but what to stop us switching viewpoints around here and supposing particle B is observing particle A? I find this is an intriguing possibility considering the importance of symmetry in Physics.
[ 0.0055378517135977745, 0.03229750692844391, -0.0000010198209565714933, 0.012831749394536018, 0.01605367101728916, -0.019916949793696404, 0.008258930407464504, -0.00910887960344553, -0.019793519750237465, 0.008760171011090279, -0.01049294788390398, 0.027032405138015747, -0.008910122327506542,...
[ 0.7407545447349548, -0.02636026404798031, -0.1541777104139328, 0.2294759303331375, -0.29441705346107483, 0.03768410533666611, 0.22163543105125427, -0.05750451236963272, -0.5641558766365051, -0.5001298785209656, 0.12254759669303894, 0.2669156491756439, -0.24635732173919678, 0.43375316262245...
One of the advantages of older version control systems such as CVS and SVN in enterprise development is that anyone can connect to source control and see all the projects that the company has. This can make it easier to get a high level view of what kid of development is happening outside your sprint and also keeps everything in one place and easy to find. However, distributed version control systems (Git, specifically) use the repository as their base unit. They work best with one project (or several closely related projects) per repository. This makes repository management more difficult in most enterprise environments where it is not unusual to have more than 25-50 projects to support. As far as I have been able to determine, you have to keep a list somewhere else of all the repos you have. There is software available, like GitHub, that help, but that still is an extra step beyond a single connection string and listing the contents of the repository. What is the best way to deal with the complexity of multiple repositories?
[ 0.007658353075385094, 0.002530961763113737, -0.018031081184744835, 0.006497509311884642, 0.024224717170000076, 0.001557561568915844, 0.0070690857246518135, -0.01151931844651699, -0.011825033463537693, -0.017035426571965218, -0.006145327351987362, 0.01572347804903984, 0.015036661177873611, ...
[ 0.32210344076156616, 0.041517093777656555, 0.04616249352693558, 0.48044198751449585, 0.030867984518408775, -0.4601011276245117, -0.07267899811267853, 0.12236076593399048, -0.4787512421607971, -0.5592918395996094, -0.3144308030605316, 0.35290706157684326, 0.11629406362771988, 0.372579693794...
What is a car's _interior packaging_? This is from Wikipedia article on Crossovers: > Using the unibody construction typical of passenger vehicles, the crossover > combines SUV design features such as tall **interior packaging** , high > H-point seating, high ground-clearance or all-wheel-drive capability — with > design features from an automobile such as a passenger vehicle's platform, > independent rear suspension, car-like handling and fuel economy.
[ -0.03796576336026192, 0.005127740558236837, 0.007483760826289654, 0.022398708388209343, 0.00332166557200253, -0.014730031602084637, 0.010577425360679626, -0.003614172339439392, -0.016239697113633156, -0.02128785103559494, -0.011527017690241337, 0.013522683642804623, 0.011317423544824123, 0...
[ 0.40579402446746826, 0.18003077805042267, 0.3478568196296692, 0.4647108018398285, -0.27323031425476074, 0.15769760310649872, 0.0017630434595048428, -0.27739471197128296, -0.20844222605228424, -0.3590089678764343, 0.08437246084213257, 0.5570614337921143, 0.5057653784751892, 0.07804621756076...
I'm writing a paper using Aldous Huxley's The Perennial Philosophy. Huxley doesn't use a single word when he quotes, his book is full of complete quotes that aren't a part of a sentence or a paragraph, and he uses the quotes as supporting evidence to what he has said. For example, take this as a paragraph from the book. End paragraph. He who suffers for love does not suffer, for all suffering is forgot. -Eckhart How do I quote this quote in my paper?
[ -0.0015821220586076379, 0.017627935856580734, -0.015904104337096214, 0.017239941284060478, -0.027819577604532242, -0.012075818143785, 0.01113760843873024, -0.007857924327254295, -0.014036139473319054, -0.02628844417631626, -0.008488865569233894, -0.009256608784198761, -0.01604422554373741, ...
[ 0.3696569502353668, 0.3545870780944824, -0.04857196658849716, 0.08362872153520584, -0.16682204604148865, 0.1230587512254715, 0.3474101126194, -0.05626692250370979, 0.4152265787124634, -0.4994843602180481, 0.14794683456420898, 0.06685514748096466, -0.48015642166137695, 0.2825317680835724, ...
> **Possible Duplicate:** > “The thing is, is that…” > Why do some people say “the reason is is that,” with “is” twice in a row? For the past few years I've been noticing a curious phenomenon. People are quite commonly saying things like "... the problem _is is_ that we don't have access to ...". It appears that the first "is" has been grafted onto the semantic unit of "the problem" and is no longer being perceived as being present, so the speaker has to add the second "is". I first thought that this might be related to situations where a repeated "is" is valid, such as ... the question is "Is this our only option?" ... but it seems to occur mostly in constructs of the form ... the [x] is that ... instead. I seem to be very sensitive to this and hear it several times a day, often on local or national news broadcasts. In fact, it's now jarring (in a good way) when I hear someone NOT repeating "is". 1. Is there a name for this phenomenon? 2. Is it as rampant as it appears to me?
[ -0.0014116952661424875, -0.00028138456400483847, -0.011552945710718632, 0.014861632138490677, -0.0059402501210570335, -0.007308669853955507, 0.005807904526591301, 0.014519311487674713, -0.006344961002469063, -0.005069626495242119, -0.0015929981600493193, 0.0026088061276823282, -0.00014630984...
[ 0.12381412833929062, 0.2139839231967926, 0.1464458703994751, 0.042774684727191925, -0.11313100904226303, 0.19427737593650818, 0.5874242186546326, 0.4021807014942169, -0.7675897479057312, -0.493242084980011, -0.15734745562076569, 0.3406132161617279, -0.2530149519443512, 0.38152986764907837,...
I'm a junior developer at a company that has asked me to establish academic goals for the near future. I didn't realize how hard of a question this was until I could only come up with one answer, off the top of my head: Learn more Design Patterns What subjects have you learned, after you finished school, that have helped you significantly?
[ 0.0016155119519680738, 0.013046758249402046, -0.011791015975177288, -0.003312984947115183, -0.00493590859696269, 0.00879113283008337, 0.006121705286204815, 0.012678427621722221, -0.020318787544965744, -0.0204360019415617, 0.007214644458144903, 0.013245650567114353, 0.018109293654561043, 0....
[ 0.8515236973762512, 0.15587694942951202, -0.20191848278045654, -0.050023894757032394, 0.20653115212917328, 0.12535421550273895, 0.10914113372564316, 0.24925127625465393, -0.29815584421157837, -0.5054817199707031, 0.16752614080905914, 0.06343895196914673, 0.4372531771659851, 0.2768450677394...
According to Ampere's Ciruital Law: ![enter image description here](http://i.stack.imgur.com/lBBiE.png) Now consider two straight wires, each carrying current `I`, one of infinite length and another of finite length `l`. If you need to find out magnetic field because of each, at a point (X) whose perpendicular distance from wire is `d`. You get magnetic field as $\frac{\mu I}{2 \pi d}$. Same for both. But, Magnetic field due to infinitely long wire is : $\frac{\mu I}{2 \pi d}$ Magnetic field due to wire of finite length `l` : $\frac{\mu I (\sin(P)+\sin(Q)) }{2 \pi d}$, where P & Q are the angles subtended at the point by the ends of the wire. Why are we getting wrong value for using Ampere's circuit law?
[ 0.007817009463906288, 0.012174638919532299, -0.007575375027954578, -0.00308137945830822, -0.01498956698924303, 0.0024599507451057434, 0.007261955179274082, 0.004782667849212885, -0.015541605651378632, 0.02202349156141281, -0.0021525141783058643, 0.006796966306865215, -0.008733402006328106, ...
[ 0.5492140650749207, -0.5238351225852966, 0.6631755828857422, 0.41337019205093384, -0.10417630523443222, 0.3420598804950714, -0.08250100910663605, -0.7648836970329285, -0.6771265268325806, 0.05863909423351288, 0.4995173215866089, 0.6669885516166687, -0.33293449878692627, 0.3196713626384735,...
I am currently using both `matlab2tikz` and `pgfplots` with external graphics since I am not completely convinced by either yet. Is it possible to add a colorbar with any of the two afterwards (meaning the original MATLAB figure is not open anymore, and therefore I can not check for the minimum and maximum values anymore)? If I run `matlab2tikz` on a MATLAB figure with the colorbar open, `matlab2tikz` will place the correct colorbar automatically, if added afterwards it will always only give a scale from 0-1 as `point meta min` and `point meta max` are of course not specified. Is it possible to modify `matlab2tikz` such that `point meta min` and `point meta max` are always saved to the .tikz file? If i use `pgfplots` with external graphics, I will always have to assign `point meta min` and `point meta max` manually. Is there any way around this when using pgfplots? P.S. I am aware that it is possible to export the entire dataset from MATLAB and import it to `pgfplots`. However I would like to avoid this as the datasets are rather big (including quite some meshes) and the total amount of pictures in the final report will be >100.
[ 0.007386842742562294, 0.011896122246980667, -0.005325604695826769, 0.010296058841049671, -0.014115613885223866, -0.005172244738787413, 0.008491020649671555, 0.01560950931161642, -0.017272498458623886, -0.022058147937059402, -0.02138037607073784, 0.013737357221543789, -0.00923082884401083, ...
[ -0.0052898903377354145, -0.24432536959648132, 0.5884197950363159, -0.11394693702459335, -0.24755236506462097, 0.3850039541721344, 0.2506665885448456, -0.012855855748057365, -0.4611953794956207, -0.6427443027496338, 0.13155795633792877, 0.6738637089729309, -0.21317225694656372, -0.074572227...
With binaries, I can always do `which foo` to know which binary I am invoking with `$ foo`, but what about man pages? Is there a way to know the path of the man page will be loaded by default? Similarly, what about dynamic libraries? Is there any to find out the path of the `.so` file that will be loaded for a given library name? ## Update: I just learned about pkg-config, but I didn't see an option to ask it to output paths. Also, it looks like it requires having `.pc` files that specify paths. Still is this something that could help with this problem?
[ 0.004626601003110409, 0.007073144894093275, 0.0010901072528213263, 0.010317182168364525, -0.024956047534942627, 0.004426637198776007, 0.0068799350410699844, -0.004326567053794861, -0.014972741715610027, -0.0062113795429468155, -0.000416987226344645, 0.0009854835225269198, 0.01062201336026191...
[ 0.06501679122447968, -0.24993689358234406, 0.05501626804471016, 0.23082877695560455, -0.34109121561050415, -0.26462364196777344, 0.18959444761276245, -0.06335904449224472, -0.25497758388519287, -0.8526358008384705, 0.0879821702837944, 0.7134313583374023, -0.6707334518432617, -0.15546858310...
I use `multibib` which creates two families of citation commands: `pri` and `sec`. I would like to get rid of bibliography headings for the `\bibliographypri` and `\bibliographysec` commands while keeping the Bibliography heading for the main `\bibliography` at the end of the document. I also use `tocbibind`, as advised here, to include Bibliography in the TOC and have a correct PDF bookmark. I renew `thebibliography` environment to remove the internal `\chapter` code before `\newcites` and then restore it to get the main bibliography right. In the following MWE, using `tocbibind` modifies `\bibliography` in a way that it adds an empty PDF bookmark at the chapter level, which I don't want. Moreover, citations that should be on the same pages as sections _Primary_ and _Secondary_ are two pages farther than expected. \documentclass{book} \usepackage[resetlabels]{multibib} \usepackage[nottoc]{tocbibind} \makeatletter \newenvironment{thebibliographynohead}[1] {%\chapter*{\bibname}% %\@mkboth{\MakeUppercase\bibname}{\MakeUppercase\bibname}% \list{\@biblabel{\@arabic\c@enumiv}}% {\settowidth\labelwidth{\@biblabel{#1}}% \leftmargin\labelwidth \advance\leftmargin\labelsep \@openbib@code \usecounter{enumiv}% \let\p@enumiv\@empty \renewcommand\theenumiv{\@arabic\c@enumiv}}% \sloppy \clubpenalty4000 \@clubpenalty \clubpenalty \widowpenalty4000% \sfcode`\.\@m} {\def\@noitemerr {\@latex@warning{Empty `thebibliography' environment}}% \endlist} \makeatother \let\thebibliographyold\thebibliography \let\thebibliography\thebibliographynohead \newcites{pri,sec}{{},{}} \let\thebibliography\thebibliographyold \usepackage{bookmark} \begin{document} \frontmatter \chapter{Abstract} \chapter{Sources} \section*{Primary} \addcontentsline{toc}{section}{Primary} \nocitepri{greenwade93} \bibliographystylepri{plain} \bibliographypri{bibliography} \clearpage \section*{Secondary} \addcontentsline{toc}{section}{Secondary} \nocitesec{greenwade93} \bibliographystylesec{plain} \bibliographysec{bibliography} \clearpage \cleardoublepage \pdfbookmark{\contentsname}{Contents} \tableofcontents %% Skip from TOC, only PDF bookmark \mainmatter \pagenumbering{arabic} \chapter{Chapter A} \section{Section A} Ref.~\cite{greenwade93}. \backmatter %\addcontentsline{toc}{chapter}{Bibliography} \bibliographystyle{plain} \bibliography{bibliography} \chapter{Acronyms} \end{document}
[ 0.023757271468639374, 0.007574463728815317, -0.010856516659259796, 0.0230039544403553, 0.007215801626443863, 0.009033034555613995, 0.008611653000116348, -0.008224792778491974, -0.013553624041378498, 0.005376154091209173, -0.011631704866886139, 0.006030576769262552, -0.01637711189687252, 0....
[ 0.0906989574432373, 0.1587488055229187, 0.6707489490509033, 0.0724361315369606, -0.06694994121789932, -0.3549758493900299, 0.130880668759346, -0.1262393742799759, -0.029120219871401787, -0.7700770497322083, -0.2797035872936249, 0.7797005772590637, -0.40509548783302307, 0.16133277118206024,...
I am trying to get skylanders off of eBay but I am not sure which one to buy because some of them don't have online codes and I don't know if it matters.
[ 0.02054278366267681, 0.008864925242960453, -0.010110579431056976, 0.034741032868623734, -0.008604873903095722, 0.05391506478190422, 0.01024385541677475, 0.027482321485877037, -0.027515850961208344, -0.038694217801094055, -0.006216799840331078, 0.02085501328110695, -0.02198113687336445, -0....
[ 0.8563929200172424, 0.06501930207014084, -0.13498885929584503, 0.454659104347229, -0.043452370911836624, 0.4043758511543274, -0.0936410203576088, 0.5376745462417603, 0.0746951550245285, -0.33757367730140686, 0.3919183611869812, 0.1086873784661293, 0.2134782075881958, 0.9121414422988892, ...
I was looking for a way to attach a custom meta box to the edit screen of a page at a specific template. I found and tried this here and from my actual point of view this can't work that way. Let me explain why I think so: To determine, whether I'm on a screen using a specific template I need reference the global $post. But since this function to create the meta box is hooked into the admin_init the global $post will not yet exist. So it has to put out an Error message, doesn't it? After those thoughts I tried to tie my meta_box_init function to edit_post, save_post and such. But the codex says, even they need $post to be set. So how could I achieve my page template specific meta box, really? Is there a clean solution?
[ -0.01305130310356617, -0.00020789832342416048, 0.011673413217067719, 0.017991866916418076, 0.005086506716907024, -0.003001469187438488, 0.006958060432225466, 0.0053351158276200294, -0.014977144077420235, -0.0007118773646652699, -0.01052780169993639, 0.010868081822991371, 0.002482433803379535...
[ 0.5510852932929993, 0.18593859672546387, 0.42588871717453003, -0.038557201623916626, -0.08829015493392944, 0.05365056172013283, 0.1011526808142662, 0.039165861904621124, -0.23099955916404724, -0.8006731867790222, 0.6625127792358398, 0.5060009360313416, -0.6025846600532532, 0.38676837086677...
The Higgs mechanism allows massless fields to acquire mass through their coupling to a scalar field. But if the masses cannot be predicted because the couplings have to be fixed, what really is the utility of the Higgs mechanism? Instead of saying "Here are _a priori_ couplings; the Higgs mechanism generates mass.", I could just as well say "Here are _a priori_ masses. Period.". I understand that the Higgs mechanism is crucial to electroweak unification, but I have the same question there. Why does electromagnetism and the weak force have to be unified? Even if the couplings of the photon, Z and W bosons became related on unification, this is still at the cost of introducing new parameters - so it's not really clear to me that something has been explained or tidied up. Do either the Higgs mechanism or electroweak unification tell us something new? Do either make any predictions that don't come at the cost of extra parameters? (I'm not really challenging anything here; I'm sure the answer to both questions is 'yes' - I just want to fill the gaps in my understanding as I study the Standard Model)
[ 0.003871542401611805, 0.014659928157925606, 0.004634974058717489, 0.01922755129635334, -0.022157471626996994, -0.013930758461356163, 0.009289691224694252, -0.03165869787335396, -0.016259286552667618, -0.0051021883264184, -0.013268283568322659, 0.014777803793549538, -0.003821867285296321, 0...
[ 0.2239120900630951, -0.01627342775464058, 0.3705407679080963, -0.08065861463546753, -0.2599145174026489, -0.15015123784542084, -0.08715333789587021, -0.3783739507198334, -0.5237710475921631, -0.2343887835741043, -0.06773649901151657, 0.1419236958026886, -0.37990424036979675, 0.762795209884...
I am just starting a Monk build in Diablo 3, and notice that a dual-wielding setup or a weapon/shield combo is far more favorable than using a Daibo, due to the ability to equip more prefixes and faster attack speed. Are there any reasons to use a Daibo instead, besides the Spirit regeneration bonus?
[ -0.03269438073039055, 0.03086955100297928, -0.003971263766288757, -0.007064655423164368, -0.0258198119699955, -0.015151292085647583, 0.014052088372409344, 0.017756521701812744, -0.013661595992743969, -0.015081644058227539, -0.009495235979557037, 0.01677083410322666, -0.01182357408106327, 0...
[ 0.31633153557777405, 0.09830798953771591, 0.2737292945384979, -0.15461555123329163, -0.638694703578949, -0.39823728799819946, 0.7010615468025208, -0.6975126266479492, -0.18664591014385223, -0.30032217502593994, 0.13417021930217743, 0.6565523147583008, 0.5292295813560486, -0.268550336360931...
I have a Linux (Ubuntu 12.04) PC connected to the internet with a Greenpacket WiMax USB modem. I want to share the Internet connection with another computer running Windows 7 Home Premium, connected to the Linux PC over a LAN. Is this possible? How? Is the reverse possible instead (connecting the internet to the Windows computer and sharing it with Linux)?
[ -0.012857579626142979, 0.00014012763858772814, -0.011906024999916553, 0.027348848059773445, -0.026070773601531982, -0.03430471569299698, 0.01230954471975565, -0.004249376244843006, -0.024139808490872383, -0.020915551111102104, 0.005907198414206505, 0.012702619656920433, 0.006007462739944458,...
[ 0.5428181290626526, 0.23259778320789337, 0.37156665325164795, 0.13044194877147675, 0.19099532067775726, 0.07532869279384613, -0.10612645745277405, 0.22030124068260193, 0.012157070450484753, -0.6867933869361877, -0.03720693290233612, 0.39574095606803894, -0.10386427491903305, 0.301040083169...
The de-facto plotting program for LaTeX is, AFAIA, gnuplot. Is there such a program for drawing graphs/diagrams? I am aware of graphviz and at first glance it appears very useful, but it is missing certain basic features such as specifying multiple rank orders (arrange left to right, top to bottom... see here) or subscripts/superscripts. It apparently allows for html formatting which in theory provides for subscripts/superscripts, but it does not seem to work. dot2tex attempts to make graphviz more LaTeX friendly, but the generated graphs are not as sharp, C style comments, which are valid in graphviz files, appear to conflict with LaTeX files, even when included as a `doc` file, and I, personally, can't figure out how to embed the graph as a figure as opposed to a dangling graph. Is there a way to address some of these shortcomings: * no sub/superscripts * no figures * no comments If not, is there a way to output the file to a more LaTeX friendly format, much like gnuplot does with its latex format, such that I can use LaTeX's math mode? **EDIT** here are two pictures to better describe what I mean by graphs and/or diagrams ![Diagram 1](http://i.stack.imgur.com/dS0ZE.png) ![Diagram 2](http://i.stack.imgur.com/8HBss.png)
[ -0.022658001631498337, 0.0024794244673103094, -0.0030416431836783886, 0.009207649156451225, 0.0023838793858885765, -0.0020187769550830126, 0.007351364009082317, 0.0191800519824028, -0.01835867017507553, 0.005806738510727882, -0.01188749261200428, 0.009837382473051548, 0.002445777878165245, ...
[ 0.3758256137371063, -0.24164383113384247, 0.747657299041748, 0.19696608185768127, -0.24637140333652496, -0.33313632011413574, -0.2956576347351074, -0.03585151210427284, -0.28980758786201477, -0.3948765993118286, 0.42314669489860535, 0.3240205645561218, -0.6309592127799988, -0.0618391409516...
How can I get money on the Xbox 360 Minecraft game to purchase skins at the Minecraft store? I purchased a Minecraft Mojang prepaid card and would like to use it on the Xbox 360, is this possible?
[ 0.024048207327723503, -0.007448929361999035, 0.003183653112500906, 0.016206687316298485, -0.002171018859371543, -0.023759502917528152, 0.015241524204611778, -0.04671427980065346, -0.04087859392166138, -0.059662457555532455, -0.003199246944859624, 0.02805175445973873, -0.009659445844590664, ...
[ 0.7681695818901062, 0.03645461052656174, 0.14692184329032898, 0.5819072723388672, 0.07083585858345032, 0.18458926677703857, 0.007173560559749603, 0.1793157309293747, -0.2616152763366699, -0.37039488554000854, 0.4786149561405182, 0.83570796251297, 0.09536755830049515, -0.2794760763645172, ...
I am looking for good and recent references to constructing twistor space for curved spacetime. This could be a general spacetime, or specific ones (say maximally symmetric spaces different from Minkowski). This could be in he context of the twistor correspondence, or the twistor transform of field equations, either subject generalized to curved spacetime. The references I am familiar with are the standard ones from about 30-40 years ago, where most constructions involve flat spacetime. Some generalizations are mentioned, but my impression is that the community had not settled at the time on a single approach. Many things happened since, and one of the things I am hoping to get is some understanding of the landscape of current approaches to the subject.
[ 0.012062443420290947, 0.012088415212929249, -0.011445955373346806, 0.02746180258691311, 0.021687638014554977, -0.009472105652093887, 0.010344745591282845, -0.01193370670080185, -0.0166462492197752, -0.0030457088723778725, -0.0035827974788844585, 0.02289555035531521, 0.0025501581840217113, ...
[ 0.11537723988294601, 0.11356432735919952, 0.16138191521167755, -0.0459817573428154, -0.08924873173236847, 0.1816961020231247, -0.14822226762771606, 0.18658317625522614, -0.3040107190608978, -0.7480446100234985, 0.2748020589351654, 0.17897231876850128, 0.04824988543987274, 0.507845520973205...
I'm examining some genomic coverage data which is basically a long list (a few million values) of integers, each saying how well (or "deep") this position in the genome is covered. I would like to look for "valleys" in this data, that is, regions which are significantly "lower" than their surrounding environment. Note that the size of the valleys I'm looking for may range from 50 bases to a few thousands. What kind of paradigms would you recommend using to find those valleys? **UPDATE** Some graphical examples for the data: ![alt text](http://i.stack.imgur.com/5UEOg.png) ![alt text](http://i.stack.imgur.com/tlaV1.png) **UPDATE 2** Defining what is a valley is of course one of the question I'm struggling with. These are obvious ones for me: ![alt text](http://i.stack.imgur.com/MaR6O.jpg) ![alt text](http://i.stack.imgur.com/GP9UZ.jpg) but there some more complex situations. In general, there are 3 criteria I consider: 1\. The (average? maximal?) coverage in the window with respect to the global average. 2\. The (...) coverage in the window with respect to its immediate surrounding. 3\. How large is the window: if I see very low coverage for a short span it is interesting, if I see very low coverage for a long span it's also interesting, if I see mildly low coverage for a short span it's **not** really interesting, but if I see mildly low coverage for a long span - it is.. So it's a combination of the length of the sapn and it's coverage. The longer it is, the higher I let the coverage be and still consider it a valley. Thanks, Dave
[ -0.0063009727746248245, 0.014137795194983482, -0.009603842161595821, 0.009161906316876411, 0.008700777776539326, -0.002720186021178961, 0.006133050657808781, -0.004238619469106197, -0.012904960662126541, -0.001121464534662664, 0.0027814358472824097, 0.010833910666406155, -0.00964830443263053...
[ 0.6322855353355408, 0.02269688807427883, 0.24528640508651733, 0.10443826764822006, 0.31917521357536316, 0.09020525962114334, 0.6520594954490662, 0.4131567180156708, -0.08543221652507782, -0.6999198198318481, -0.22990331053733826, 0.04096267744898796, 0.4113437235355377, 0.3499494194984436,...
Taito's _Groove Coaster_ for iOS has an achievement called "Use Item" which you get for simply using an item on a level. Items are expendable powerups that make levels slightly easier. Items are available for purchase in the Upgrade menu of the game for $0.99 for 10. Is there any way to acquire an item, thereby allowing the achievement, without buying some in the Upgrade shop?
[ -0.0013439947506412864, 0.006667078007012606, 0.0008430348825640976, -0.0012368799652904272, 0.021368827670812607, -0.01962961070239544, 0.010379779152572155, -0.017816787585616112, -0.01593800261616707, 0.037570033222436905, -0.02494806796312332, 0.010483894497156143, 0.0064339605160057545,...
[ 0.002202237257733941, -0.26520881056785583, 0.3201300799846649, 0.4085179269313812, -0.17361114919185638, -0.09197407215833664, 0.19846656918525696, -0.09532439708709717, -0.01755869947373867, -0.03951355442404747, 0.060434333980083466, 0.40048009157180786, 0.19819629192352295, -0.29956659...
This is a rather simple question but I cannot seem to find the answer anywhere. I have a classification tree (Exact details irrelevant). My question is, say at the very top of the tree the condition is number < 5. If the condition is true (say, my number = 10), then do I go left or right down the tree? What is the rule for the directions on which you proceed down the tree?
[ -0.013775310479104519, 0.018522433936595917, -0.01785171404480934, -0.00008102582796709612, 0.003473630640655756, 0.013651498593389988, 0.009737455286085606, 0.005372214131057262, -0.019854580983519554, 0.0003993050486315042, -0.008462189696729183, 0.001539540709927678, 0.0013002394698560238...
[ -0.19080083072185516, 0.13517612218856812, 0.4135049283504486, -0.002994627458974719, 0.010040395893156528, 0.16584624350070953, 0.36094358563423157, -0.26872098445892334, -0.5034799575805664, -0.4097886383533478, 0.1063525378704071, 0.26428303122520447, 0.06496502459049225, 0.196459874510...
I want to be able to reference a table value in my text (this is because I often update my tables, and then list the specific values in the text). Here is an example table I would use: % Example Table \documentclass{minimal} \begin{filecontents*}{scientists.csv} name,surname,age Albert,Einstein,133 Marie,Curie,145 Thomas,Edison,165 \end{filecontents*} % Read in Table \documentclass{article} \usepackage{pgfplotstable} \begin{document} \pgfplotstabletypeset[ col sep=comma, string type, columns/name/.style={column name=Name, column type={|l}}, columns/surname/.style={column name=Surname, column type={|l}}, columns/age/.style={column name=Age, column type={|c|}}, every head row/.style={before row=\hline,after row=\hline}, every last row/.style={after row=\hline}, ]{scientists.csv} \end{document} I may want to be able to reference a given scientists age in the text by a reference of his/her name (ie,:) Albert Einstein is \ref{albert} years old. Ideally, this would still be using pgfplotstable because it is how I currently read in many tables. Thanks,
[ 0.018235253170132637, 0.01375911757349968, -0.013836260885000229, 0.017578259110450745, 0.017777474597096443, 0.012350508943200111, 0.006590981502085924, 0.011968288570642471, -0.014739971607923508, -0.021895524114370346, 0.01003511343151331, -0.008348358795046806, -0.002091251779347658, 0...
[ 0.23606276512145996, 0.51273512840271, 0.2063826322555542, 0.3938632011413574, 0.34032005071640015, 0.2720962464809418, 0.194193497300148, -0.01720513589680195, -0.4981268644332886, -0.44427329301834106, -0.20711158215999603, -0.33259153366088867, -0.12533222138881683, 0.5230821967124939, ...
Non-parametric ANOVA – a hot topic that is unanswered. There are many questions on this topic online. However, they all seem to end in a debate and no definite answer or clear explanation (that I can relate to my data set). The only good answers I can find seem to relate to one treatment, that was repeated over time. How do I handle two treatments? More importantly, should this be handled as a linear model? I cannot meet assumptions of homogeneity, normality, equal sample size, and transformations were not helpful. Data setup: * Soil cores growing plants. * Subject – about 33 samples. A core will consist of a treatment combo ( _n_ = 5 to 6; repeats) * `Treatment01` - clay, sand, loamy sand * `Treatment02` - water, fertiliser * Time - an sample is analysed from each core sample, every week, for 3 months. * Variable - Nutrients (mg/kg soil), moisture, pH, mass of soil etc. For simplicity, let's do carbon. Here are my failed attempts on R. This problem rests in the statistical theory. `Anova(lm(Carbon ~ Treatment01 * Treatment02, data), type = "3")` or `model1 <\- lm(Carbon ~ Treatment01 * Treatment02)` Then something like... print(lsmeans(model1, list(pairwise ~ Treatment01)), adjust = c("tukey")) print(lsmeans(model1, list(pairwise ~ Treatment02)), adjust = c("tukey")) print(lsmeans(model1, list(pairwise ~ Treatment01 | Treatment02)), adjust = c("tukey")) print(lsmeans(model1, list(pairwise ~ Treatment02 | Treatment01)), adjust = c("tukey")) Any suggestions? What are the disadvantages in pulling the data set apart to do a series of Friedman or Kruskal–Wallis tests (depending on how much I pull this data set apart?)
[ 0.01884298585355282, 0.021807517856359482, -0.004654870368540287, 0.017885081470012665, -0.015247134491801262, -0.004781678784638643, 0.004625669680535793, 0.02357299253344536, -0.010776749812066555, -0.00896446593105793, -0.006261266767978668, 0.014409257099032402, -0.009904414415359497, ...
[ 0.26883020997047424, -0.3673321008682251, -0.1399995982646942, 0.14901955425739288, -0.24196569621562958, 0.46311038732528687, 0.2516104578971863, -0.26069894433021545, -0.15385469794273376, -0.43031781911849976, 0.05024698004126549, 0.2959529757499695, -0.2881760597229004, 0.3149255812168...
I'm interested primarly on bifins (one fin for each foot) with a monofin movement, e.g. undulatory. I'd like to know how they work and where I can find articles/books about it.
[ 0.022440074011683464, 0.04792213439941406, -0.015970595180988312, 0.02765735797584057, 0.05227778106927872, -0.0288265198469162, 0.013560761697590351, 0.03774413838982582, -0.025834914296865463, -0.011870001442730427, 0.003568840678781271, 0.029092708602547646, -0.041538432240486145, -0.01...
[ 0.418212354183197, -0.04142308235168457, 0.025892209261655807, -0.06311658024787903, -0.10049982368946075, 0.3213431239128113, 0.31927743554115295, 0.19604824483394623, -0.8031063675880432, -0.4303763806819916, 0.49611911177635193, -0.007923148572444916, 0.06193310767412186, 0.324534028768...
Is there any proper wordpress way of including the JS and CSS through one merged URL? So that instead of loading in x number of separate calls for each JS asset, it can be all merged into one URL? Samething for CSS. I know this is easy enough to do using straight up PHP, but I wanted to check if wordpress has some nifty way of combining requests.
[ 0.022223930805921555, 0.013652694411575794, 0.0056396592408418655, 0.024214891716837883, 0.00839888397604227, 0.0022080058697611094, 0.009104302152991295, -0.014926341362297535, -0.03104325570166111, 0.0059797316789627075, -0.010470353066921234, 0.017307208850979805, -0.02194572053849697, ...
[ 0.5210104584693909, 0.061414074152708054, 0.12441366910934448, 0.3236521780490875, -0.17469839751720428, -0.11372379213571548, -0.026065444573760033, 0.16729098558425903, -0.2876381278038025, -0.4580759108066559, 0.17989881336688995, 0.20787034928798676, -0.07676203548908234, 0.15184777975...
I have searched online, and can't find a concrete answer, would this be breaching the Adsense rules? I have a Wordpress site with a youtube video custom post type. When a query is run, it will display all posts in full view, which is basically a youtube video, a few lines of text and hopefully an Adsense text ad at the bottom. I was then planning on having a couple ads on the sidebar. The problem is i'm not sure if that is breaking the Adsense rules, because the query will show multiple posts (more than 3). However, there isn't a "page" that would have more than 3 on it. Would the query results be a "page" and subject to 3 total ads?
[ -0.009635350666940212, 0.00220666010864079, 0.00032395171001553535, 0.02124142274260521, -0.03416164219379425, 0.0032861116342246532, 0.009368626400828362, 0.04364429786801338, -0.018925286829471588, 0.011271354742348194, -0.023022329434752464, 0.01763610914349556, 0.005882480181753635, 0....
[ 0.7087317109107971, 0.13691887259483337, 0.5571938753128052, 0.06785783916711807, -0.20523041486740112, -0.04166310280561447, 0.2968329191207886, 0.13339968025684357, -0.2716343104839325, -0.5813627243041992, 0.3939092457294464, 0.31222498416900635, -0.13500817120075226, 0.3998647034168243...
I have the latest version of the YouTube app for Android (v5.1.10). At some point, the app began overlaying the uploader's user image in the upper-right- hand corner of the video. I can't click on it (it acts as if I clicked anywhere else on the video) and I can't remove it. It never moves (unless I switch to/from full screen) or goes away, and sometimes blocks important portions of the video. Is there any way to turn this annoying non-feature off?
[ -0.013370849192142487, -0.004656407982110977, -0.01181389857083559, 0.00517059164121747, -0.02525673247873783, 0.0036817772779613733, 0.005209473893046379, 0.026192903518676758, -0.014591355808079243, 0.02488868311047554, -0.02935251221060753, 0.010859153233468533, 0.020206710323691368, 0....
[ 0.7383350729942322, -0.009210282936692238, 1.0778114795684814, -0.003730407217517495, 0.13664479553699493, -0.0064843385480344296, 0.09736371785402298, 0.2788275182247162, -0.20668897032737732, -0.6431485414505005, 0.10347671806812286, 0.46217668056488037, -0.3578680753707886, 0.3310643732...
I'd like to create a certain kind of voting/recommendation system. I'm sure there must be a name for what I'm trying to do, but I'm not sure. Basically, I start with a distribution of binary vectors of length 512. Each vector determines a visual image (that is, an algorithm is applied to the vector to generate a larger image). I select 9 vectors from the distribution at random, and display the corresponding images to a user in a 3x3 grid. The user clicks on the one he thinks is "best." After doing this, the distribution should update, and display 9 more images such that the probability of a vector being chosen is directly proportional to its generated image being "best". Any help on where to look for algorithms/methods to do this kind of thing? I know that I could build up a simple genetic algorithm to do this, but it seems that there might be more direct Bayesian approaches (if I add in some assumptions about how the vectors are related).
[ -0.00391707569360733, 0.011021246202290058, -0.005917144939303398, 0.00631957221776247, -0.018686238676309586, -0.012202372774481773, 0.004512421786785126, 0.0006655396427959204, -0.011725543066859245, 0.01281760260462761, -0.01440283004194498, 0.003793012350797653, 0.00652251485735178, 0....
[ 0.1509704887866974, -0.33171501755714417, 0.432886004447937, 0.28833892941474915, -0.039250243455171585, 0.6165735125541687, -0.29017454385757446, -0.08084817230701447, -0.18428604304790497, -0.7230499386787415, 0.13545121252536774, 0.13039791584014893, -0.4591636657714844, 0.0847469270229...
I used a solution from @Jake from this post Insertion of Perpendicular Symbol at Intersection of Two Lines, but the alignment is not quite right. I have my suspicions why, but I don't understand the `pgf` code enough to figure it out. Here is my MWE: \documentclass{article} \usepackage{tikz} \usetikzlibrary{shapes,snakes,backgrounds,arrows} \usetikzlibrary{calc} \tikzset{ right angle quadrant/.code={ \pgfmathsetmacro\quadranta{{1,1,-1,-1}[#1-1]} % Arrays for selecting quadrant \pgfmathsetmacro\quadrantb{{1,-1,-1,1}[#1-1]}}, right angle quadrant=1, % Make sure it is set, even if not called explicitly right angle length/.code={\def\rightanglelength{#1}}, % Length of symbol right angle length=2ex, % Make sure it is set... right angle symbol/.style n args={3}{ insert path={ let \p0 = ($(#1)!(#3)!(#2)$) in % Intersection let \p1 = ($(\p0)!\quadranta*\rightanglelength!(#3)$), % Point on base line \p2 = ($(\p0)!\quadrantb*\rightanglelength!(#2)$) in % Point on perpendicular line let \p3 = ($(\p1)+(\p2)-(\p0)$) in % Corner point of symbol (\p1) -- (\p3) -- (\p2) } } } \begin{document} \begin{tikzpicture}[line width=1pt,>=triangle 45,x=1.0cm,y=1.0cm,point/.style={name={#1}},sharp corners] \clip (-0.1,-0.1) rectangle (6.1,6.1); \draw [step=0.5cm,dotted,line width=0.35pt] (0,0) grid (6,6); % Parallel Line Graph % Axis \draw [line width=0.75pt] (2,0) --(2,6) (0,2) --(6,2); % Lines & Right Angle \node [point=A] at (5.5,5) {}; \node [point=B] at (1,0.5) {}; \node [point=C] at (5.5,0.5) {}; \node [point=D] at (0.5,5) {}; \draw [<->] (A) -- (B); \draw [<->] (C) -- (D); \draw [right angle quadrant=4,right angle symbol={A}{B}{D}] ($(A)!(D)!(B)$); \end{tikzpicture}% \end{document} I imagine the solution is within the macro, but can someone help me understand where?
[ -0.01024126447737217, 0.008959894068539143, -0.003518221667036414, 0.025059029459953308, -0.015046417713165283, 0.005090741906315088, 0.006775846239179373, 0.002218246227130294, -0.011575039476156235, -0.01523485779762268, -0.010314279235899448, 0.0024924525059759617, -0.0063740573823452, ...
[ 0.1573483943939209, 0.20861110091209412, 0.7049219608306885, 0.0884893387556076, 0.04170271009206772, 0.3370267152786255, 0.16546830534934998, -0.3384850025177002, -0.13725781440734863, -0.6590113639831543, 0.3311217427253723, 0.15970706939697266, -0.2083602398633957, 0.24090057611465454, ...
I try to password protect my custom page such that before any content is shown the user must enter a password. More specific, in my code below, anything inside the content-div-container should be password protected: <?php /* Template Name: custom_page */ ?> <?php get_header(); ?> <div id="content"> <div id="main"> <ul class="post"> <li><b>LATEST POSTS</li> <?php $args = array('category' => 5, 'post_type' => 'post'); $postslist = get_posts( $args ); foreach ($postslist as $post) : setup_postdata($post); ?> <li id="post-<?php the_ID(); ?>"> <h2><a href="<?php the_permalink(); ?>"><?php the_title(); ?></a></h2> <?php the_excerpt(); ?> </li> <?php endforeach; ?> </ul> </div><!-- end content --> <?php get_sidebar(); ?> <?php get_footer(); ?> I tried to use the approach explained here Password protecting a page but cannot make it work, since I do not know how to wrap my php code into the suggested solution.
[ 0.00533007038757205, -0.0017295179422944784, 0.0005360302748158574, 0.007039734628051519, -0.0007569482550024986, 0.004477381240576506, 0.006575130857527256, -0.009563946165144444, -0.012858721427619457, 0.0022515489254146814, -0.0027879702392965555, -0.0026150187477469444, 0.002051083836704...
[ 0.397208571434021, 0.27027446031570435, 0.3631316125392914, -0.2522631883621216, -0.044966068118810654, 0.16805654764175415, 0.4512418210506439, -0.20487265288829803, 0.024448076263070107, -0.7598991394042969, -0.06987930089235306, 0.6065899133682251, -0.17804676294326782, 0.10497139394283...
Probably 90% of my time on my phone is spent typing. So I like to be able to do it quickly. Also, because I take complex notes and write code on my phone, I would like to be able to have the full ASCII character set available to me. I also like to work in the dark, so backlighting is important. I've found that I can thumb-type much more quickly on physical keyboards, and find the "slider" form factor ideal for holding the phone while thumb-typing on it. There don't appear to be any good new slider phones (in the US market), so I'm considering Bluetooth options. For example, I've found this keyboard, available in S3 and S4 designs, which can apparently be modded to fit a Nexus 4. The clever fellow who wrote the mod post likes it, and seems to have done quite a bit of research on other options, finding them seriously lacking in quality. However, I have some potential problems with this product: * There are some pretty useful characters like `{` and `}` missing from the layout. If they had added a dedicated row of keys for the digits, they probably could have completed the ASCII set and still kept all of the multimedia keys. * I don't know if the Fn-Arrow shortcut used for quick (read: acceptably fast) text navigation will work with this keyboard, since there are Fn-functions mapped to the arrow keys. * There are also some reports in the reviews that the depicted keyboard is not the one delivered; I need to do some more research to sift through this. It may just be that the two different colours of S3 keyboards have different layouts (as depicted on the product page).
[ 0.00392506318166852, 0.002013660967350006, -0.019615784287452698, -0.004077923484146595, -0.016989106312394142, 0.004253011662513018, 0.008257711306214333, 0.004977836739271879, -0.011766336858272552, -0.02037673071026802, -0.01451418362557888, 0.0011815810576081276, 0.013059960678219795, ...
[ 0.19134509563446045, 0.13572268187999725, 0.262763112783432, 0.004455320071429014, 0.3259061574935913, 0.16141656041145325, 0.14358732104301453, 0.026576153934001923, -0.220571830868721, -0.5171963572502136, -0.04009613022208214, 0.6116334795951843, 0.027157198637723923, -0.362653017044067...
I want a chapter tittle and a table to fit in just one page. It looks possible as it is enough space, but this is the output: ![enter image description here](http://i.stack.imgur.com/Dsnpo.png) The code used is: \afterpage{ \chapter{Esquemáticos y PCB} \clearpage % flush out other floats waiting to be typeset \begin{table} \centering \begin{tabular}{@{}cp{10cm}@{}} \toprule Nº referencia & Descripción \\ \cmidrule(l){1-1}\cmidrule(l){2-2} 3113067-01 & Esquemático del primer prototipo de placa (página 1)\\ 3113067-02 & Esquemático del primer prototipo de placa (página 2)\\ 3113067-03 & Esquemático del primer prototipo de placa (página 3)\\ 3113067-04 & Cara TOP de la PCB del primer prototipo de placa\\ 3113067-05 & Cara BOTTOM de la PCB del primer prototipo de placa\\ 3113067-06 & Vías de la PCB del primer prototipo de placa\\ 3113067-07 & Esquemático del diseño final de placa (página 1)\\ 3113067-08 & Esquemático del diseño final de placa (página 2)\\ 3113067-09 & Esquemático del diseño final de placa (página 3)\\ 3113067-10 & Cara TOP de la PCB del diseño final de placa\\ 3113067-11 & Cara BOTTOM de la PCB del diseño final de placa\\ 3113067-12 & Vías de la PCB del diseño final de placa\\ 3113067-13 & Esquemático de la placa de la tarjeta SD \\ 3113067-14 & Cara TOP de la PCB de la placa de la tarjeta SD\\ 3113067-15 & Cara BOTTOM de la PCB de la placa de la tarjeta SD\\ 3113067-16 & Vías de la PCB de la placa de la tarjeta SD\\ \bottomrule \end{tabular} \caption{Referencia y descripción de los planos del Apéndice A} \label{planostable} \end{table} \clearpage} % prevent other material from being placed on this page NOTE: I posted before for this same problem for figures and I tried aswking this there too once the image problem was solved, but I got the question closed, so I open a new one as I understand this is a different question.
[ 0.018458109349012375, 0.023582547903060913, -0.0030282828956842422, 0.02053973451256752, 0.001430355361662805, 0.014593386091291904, 0.0049696872010827065, 0.0023549688048660755, -0.01807890459895134, -0.007782959379255772, -0.015151155181229115, 0.008457774296402931, -0.001722807646729052, ...
[ 0.27945834398269653, 0.0006785429432056844, 0.8653322458267212, 0.13351912796497345, 0.19645987451076508, -0.06901271641254425, 0.023482492193579674, -0.16136856377124786, -0.06652656197547913, -0.590944766998291, 0.019169388338923454, 0.3723631799221039, -0.433370977640152, -0.06853831559...
A chain has N segments which can be oriented in either the x or y directions. For each segment oriented along y, there is an energy penalty of $\epsilon$. We also know the end segment is at $(L_x, L_y)$. How can we define the entropy S as a function of E,N,Lx, and Ly? I know $S=k*ln(\Omega)$, but is can you actually solve for the possible arrangements of the chain? I'm assume overlapping and crossing cannot be done.
[ -0.0026160869747400284, 0.017197486013174057, -0.018235420808196068, 0.023296542465686798, -0.017589101567864418, -0.014388044364750385, 0.00772132957354188, 0.016411788761615753, -0.01626541092991829, 0.0053101335652172565, -0.014966432005167007, 0.009472515434026718, -0.021800756454467773,...
[ 0.15267792344093323, -0.1597636491060257, 0.30663177371025085, -0.04572228342294693, -0.19850780069828033, 0.01923992484807968, 0.29123547673225403, -0.4748557507991791, -0.24827870726585388, -0.3689536154270172, -0.10002646595239639, 0.14959976077079773, -0.40245184302330017, 0.6680291891...
If I have 2 pens and I want to say all of them are green, I can say "Both of them are green" but if I have 3 pens should I use "Triple of them are green" or "All of them are green"?
[ 0.03032415173947811, 0.026089996099472046, -0.025439206510782242, 0.028662661090493202, -0.0426039844751358, 0.03290840983390808, 0.018133005127310753, 0.006669184658676386, -0.019231803715229034, -0.054249752312898636, -0.014250978827476501, -0.0008953883079811931, -0.0017089697066694498, ...
[ 0.6938758492469788, 0.18603312969207764, -0.06717082113027573, -0.3144707977771759, -0.2863454520702362, 0.4865221083164215, 0.14894694089889526, -0.6038007140159607, -0.16190709173679352, -0.582470178604126, 0.19954296946525574, 0.5043084025382996, -0.2414589673280716, -0.0634153112769126...