text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I'm trying to create a search page for a custom post type. This is the code I'm using: $s = $_GET['s']; $args = array( 'post_type' => 'companies', 's' => $s ); query_posts($args); if ( have_posts() ) : ....results..... endif The problem is that adding the "s" parameter in query_posts() gives me zero results no matter what the "s" value is. It seems like the whole search mechanism is not working at all. Any suggestions?
[ 0.010443026199936867, 0.0010972109157592058, 0.0012483918108046055, 0.011829981580376625, 0.022691134363412857, 0.009756292216479778, 0.006994991563260555, 0.014508802443742752, -0.00833029393106699, -0.015420685522258282, 0.0018218244658783078, 0.0017699790187180042, -0.007374941371381283, ...
[ 0.2637534737586975, 0.12265443056821823, 0.5290324091911316, -0.09955062717199326, -0.19801919162273407, 0.03435347601771355, 0.4647187292575836, -0.18830715119838715, -0.23912133276462555, -0.8146801590919495, -0.1709253191947937, 0.2597759962081909, -0.5696550607681274, 0.415916621685028...
I know I can use `:sp` or `:vsp` to open files in vim in splitted panes. But how can I increase/decrease the size of each pane?
[ -0.007699558045715094, 0.010379686020314693, -0.005250724498182535, 0.031240083277225494, 0.020045550540089607, 0.03380410745739937, 0.01509840041399002, 0.0062804571352899075, -0.027129797264933586, 0.0115354647859931, -0.0040399436838924885, 0.026477297767996788, -0.025493070483207703, 0...
[ 0.5166804194450378, 0.21077241003513336, 0.2393544316291809, 0.012743653729557991, 0.03662625327706337, 0.21691958606243134, 0.08204162120819092, -0.125099778175354, -0.5188168287277222, -0.519515335559845, -0.15544144809246063, 0.242719829082489, 0.023119328543543816, 0.39701998233795166,...
I have a study in which participants facebook profiles were rated by a group of students. The participants were then randomly allocated into one of three groups: the **uncertain condition** (where they could have been assigned either a liked-best or liked-least rating but they are not told which one), the **Liked-Best condition** (where they assigned the highest rating) and the **Liked-Least condition** (where they are assigned the lowest rating). Once they saw their ratings they completed a measure of mood and then in turn were able to rate how attractive they found the facebook profiles of those who rated them. After this they completed a measure of attachment style (ECR-S), self-esteem (RSES), the mood measure for a second time, and a measure of how much they had thought about the rating group in the last 15 minutes. The **dependent variables** in the study are **attraction levels** and **reported thoughts** of the participants to the rating group of students, as well as an **overall measure of mood** (mood across both time points summed and averaged). **Please bear with me as I am a complete novice at statistics, but the analysis of the data has me extremely confused.** My **IV** in the study is **Condition** , in that I want to see whether there was a significant difference between groups on the dependent variables. However I also have the measures of attachment style and self-esteem, the data for which my supervisor is planning on using in future studies but also wants me to include in mine. The problem is, I do not know where they fit in! Originally I thought that any effect of uncertainty on the DVs may be mediated by attachment style and self-esteem, BUT because the measurements were taken after the ratings had been given, they aren't exactly IVs as they may well have been affected by the ratings given. Please any help on how to analyse this would be much appreciated. I have been told I should use separate ANOVAs, but I do not feel this is justified due to the familywise inflation of the Type I error rate. I was thinking of using MANOVA to analyse Attraction and Reported thoughts (correlated at r = .34, p < .01) by Condition, with a one-way ANOVA for mood by condition. But then I am not sure how to deal with attachment style and self-esteem. Please Help!
[ 0.00458869244903326, 0.012150997295975685, 0.004707407671958208, 0.02319370023906231, 0.006029129959642887, -0.006024688482284546, 0.008340001106262207, -0.029316943138837814, -0.010700277984142303, 0.037174537777900696, -0.022085493430495262, 0.009671840816736221, -0.0011124606244266033, ...
[ 0.2018442451953888, -0.05798915773630142, 0.5264473557472229, -0.04603709653019905, -0.30046993494033813, 0.5502099990844727, 0.1911551058292389, 0.017654741182923317, 0.031132491305470467, -0.7027977108955383, 0.12610222399234772, 0.3104077875614166, 0.3318557143211365, -0.387269705533981...
If pymc.numpy.random.seed(0) guarantee the same random number sequence to initialize a stochastic variable (say a Uniform distribution), why does its posterior samples (from trace plot) don't have the same values for multiple runs with the same seed=0 ? Is there any internal random seed encoded in the pymc module ? Or, this is cause by the assigned "probability α" ? ( See http://en.wikipedia.org/wiki/Metropolis%E2%80%93Hastings_algorithm#Intuition ) Have anyone encountered this problem ? Is this related to the older version of pymc that I'm using ? Currently, I'm using pymc version 2.2 because failed to install version 3 to my computer with windows 7 platform.
[ -0.000051866285502910614, 0.005312682129442692, 0.00612262636423111, 0.027262546122074127, -0.022890528663992882, -0.008518781512975693, 0.01014459878206253, 0.007863403297960758, -0.014637480489909649, -0.0380726158618927, -0.02458849735558033, 0.015619768761098385, 0.001044127973727882, ...
[ 0.02531265839934349, -0.4831094443798065, -0.22608809173107147, 0.19904986023902893, 0.0849858745932579, 0.10604781657457352, -0.1736629754304886, -0.24876131117343903, -0.24212245643138885, -0.24512793123722076, 0.2360362857580185, 0.2292575240135193, -0.5149205923080444, 0.19324252009391...
I am just messing around, trying to figure out how stuff works and right now I have a couple questions about HTML, JS and CSS engines. I know there are two major JavaScript engines out there - V8 and JavaScriptCore (WebKit's JS engine as far as I know). Is that correct? And what are the main HTML + CSS renderers out there? Let's say I want to build a web browser using V8 (I saw it has some documentation and stuff + I like the way it works), what are the best options for me? Partially another question. Is there any bare browser that uses V8 and runs on Ubuntu at least? P.S. I am a Ubuntu user and prefer C++.
[ 0.00506901741027832, -0.0042377375066280365, -0.009910192340612411, -0.0020738723687827587, -0.01440814696252346, -0.005758045241236687, 0.006210098508745432, 0.007903625257313251, -0.01901765912771225, -0.032358791679143906, -0.005574248265475035, 0.008651580661535263, -0.000530646881088614...
[ 0.7908357977867126, 0.405751496553421, -0.1810016632080078, -0.03388741984963417, -0.6044341921806335, 0.14886485040187836, 0.20828312635421753, 0.20315779745578766, -0.4929347634315491, -0.740531325340271, 0.41031450033187866, 0.8513773679733276, -0.17695297300815582, 0.15431906282901764,...
I’m wondering if an article is used with the word _overkill_ : * Something seems like an overkill (to me). * Something seems like overkill (to me). Which is grammatical?
[ -0.0034641700331121683, 0.01003144308924675, -0.004314861260354519, 0.05542314797639847, -0.0463554821908474, -0.02687794528901577, 0.014676186256110668, -0.02801106870174408, -0.009640020318329334, 0.0073757413774728775, 0.004893392790108919, 0.02418733574450016, 0.0165577195584774, 0.009...
[ 0.3084714412689209, 0.553814709186554, 0.053304824978113174, -0.21301910281181335, -0.2797010540962219, 0.197001650929451, 0.6242151260375977, 0.2746109366416931, -0.21992219984531403, -0.2423270046710968, -0.056829702109098434, 0.20251786708831787, -0.2504023313522339, -0.0150733552873134...
1. Since the unit of distance is defined in terms of speed of light in vacuum, if the speed of light in vacuum were to change by "magic", what experiment would be able to detect that? 2. Is there a theory which says that the speed of light is not a function of time (in the same medium)? edit. people voting for close, please see my comment
[ -0.00651145912706852, 0.011264181695878506, -0.008383078500628471, 0.010218982584774494, 0.016903912648558617, -0.043407000601291656, 0.009961266070604324, -0.008861549198627472, -0.016382412984967232, -0.014578117057681084, -0.009368844330310822, 0.01677485555410385, 0.0010783320758491755, ...
[ 0.054687753319740295, -0.04273151978850365, 0.40642064809799194, 0.4051433503627777, -0.034197110682725906, -0.11923196166753769, -0.06077463552355766, -0.5555554628372192, -0.18351568281650543, -0.3018234968185425, 0.023969747126102448, 0.6421983242034912, -0.0007024988299235702, 0.190488...
Trying to remove the WYSIWYG Editor on all post types except the default pages and posts. Shouldn't this work // Remove WYSIWYG Editor function remove_wysiwyg( $hook ) { if ( $hook != 'post-new.php' || $hook != 'post.php' ) add_filter('user_can_richedit', '__return_false'); } add_action( 'init', 'remove_wysiwyg' );
[ 0.005252103321254253, 0.01303982175886631, 0.0015457009430974722, 0.027983933687210083, 0.0019925411324948072, 0.009770672768354416, 0.007928053848445415, 0.012679827399551868, -0.014287268742918968, 0.026444323360919952, -0.026158999651670456, 0.017669839784502983, 0.0004481932264752686, ...
[ -0.1803511083126068, 0.011088252067565918, 0.3916686475276947, -0.3570953905582428, -0.07884979993104935, 0.12486319988965988, 0.39158526062965393, -0.1029239371418953, 0.0738513320684433, -0.45686864852905273, -0.3129936456680298, 0.4956420361995697, -0.8991311192512512, 0.152311056852340...
I have seen 5 different type of error rates. They include: * **Componentwise Error Rate** * **Experimentwise Error Rate** * **False Discovery Rate** * **Strong Familywise Error Rate** * **Simultaneous Confidence Intervals** Suppose $H_0 = H_{01} \cap \cdots \cap H_{0K}$ Here are some questions that I have: 1. So basically if the componentwise error rate is $\alpha$ then the probability of rejecting a single hypothesis in a single test is $\alpha$? 2. If the experimentwise error rate is $\alpha$ then the probability of rejecting any of the $H_{0i}$ when all of the $H_{0i}$ are true is $\alpha$? 3. What is the difference between the FDR and the experimentwise error rate? Doesn't the FDR control the experimentwise error rate? 4. Is the strong familywise error rate basically the strictest error rate? So the FDR allows for some slack if we have more correct rejections but the strong familywise error rate does not? 5. Simultaneous confidence intervals must cover their true parameter with probability $1-\alpha$. But a single confidence interval will cover its true parameter with probability greater than $1-\alpha$?
[ -0.0016239304095506668, -0.002359966281801462, -0.012358436360955238, 0.020414644852280617, 0.02594827115535736, 0.010274175554513931, 0.007929841056466103, -0.003715783590450883, -0.015270383097231388, 0.01384761556982994, -0.0077511342242360115, 0.004496478009968996, -0.0037253000773489475...
[ 0.05045457184314728, 0.17036211490631104, -0.2586296796798706, 0.05845842510461807, -0.13491351902484894, 0.2613663673400879, 0.27730822563171387, -0.5057961344718933, -0.08387427031993866, -0.440434992313385, 0.2500605285167694, 0.3655940890312195, -0.23514387011528015, 0.5633701682090759...
How can I fill with a color in one side of the diagonal of the following graph (only one triangle area) in R? I have used `abline(0,1)` for drawing the diagonal line. However, as we can see the the diagonal line is not perfect in the left-bottom corner. Could someone help me to draw a perfect diagonal line that starts exactly from the left-bottom corner, please. ![enter image description here](http://i.stack.imgur.com/0QxuF.jpg)
[ 0.016370080411434174, 0.013528335839509964, -0.01725952886044979, 0.011712010018527508, -0.04755028709769249, -0.010016261599957943, 0.007658323738723993, 0.011562172323465347, -0.017912600189447403, -0.015324571169912815, -0.01453716866672039, -0.007749977987259626, -0.006623929366469383, ...
[ 0.2640429735183716, -0.11334611475467682, 0.2590475082397461, 0.009226354770362377, 0.16774767637252808, 0.09810689091682434, 0.22737033665180206, -0.29547205567359924, -0.3681059777736664, -0.822139322757721, 0.07675755769014359, 0.3762504458427429, -0.12825600802898407, 0.059924382716417...
I use this sleep 900; <command> on my shell. Just wanted to know is there is some alternate/better way that you use?
[ 0.0034228817094117403, 0.008163281716406345, -0.04305160790681839, 0.01587378792464733, -0.029947275295853615, 0.020126141607761383, 0.012731299735605717, -0.08558017015457153, -0.03367585688829422, -0.04308070242404938, -0.009609009139239788, -0.008440575562417507, -0.006039714906364679, ...
[ 0.678009569644928, 0.11166387051343918, 0.24381870031356812, 0.00432099774479866, 0.2863895297050476, 0.09434350579977036, 0.27843308448791504, 0.4180605411529541, -0.138760045170784, -0.8181381821632385, 0.153091162443161, 0.4043956696987152, -0.0789388045668602, -0.3070249557495117, -0...
I have a Dell Inspiron Duo - The tablet netbook, the 1090. It's beautiful, and I want to get the most out of it. I don't mind digging deep into a distro and installing patches and such to get multitouch working, but that's me. I have to think of others that also would like this goodness. So my question, which distro has the best multi touch / tablet support out the box. > Imagine I have to recommend something to someone about to go on an Arctic > expedition, who has a touch screen laptop, and needs FOSS installed because > of cutbacks :)
[ -0.002351190662011504, -0.01054807472974062, -0.012676674872636795, 0.003123408416286111, -0.03958463668823242, -0.008309759199619293, 0.005734422709792852, 0.018854601308703423, -0.020047813653945923, -0.03989788517355919, -0.0008046422153711319, 0.005546222440898418, -0.002245111158117652,...
[ 0.2910592555999756, 0.11155745387077332, 0.32427549362182617, 0.45298776030540466, 0.11257621645927429, 0.3172450065612793, -0.09734862297773361, 0.21446575224399567, -0.14969885349273682, -0.5960373282432556, 0.18922749161720276, 0.6213268637657166, 0.13951842486858368, 0.0472231395542621...
I have been studying scattering theory in Sakurai's quantum mechanics. The phase shift in scattering theory has been a major conceptual and computational stumbling block for me. How (if at all) does the phase shift relate to the scattering amplitude? What does it help you calculate? Also, any literature or book references that might be more accessible than Sakurai would be greatly appreciated.
[ -0.011772062629461288, 0.014226024970412254, 0.0038147219456732273, 0.041436437517404556, 0.041026100516319275, 0.0032799800392240286, 0.009856486693024635, -0.01458401046693325, -0.02840357832610607, -0.020305834710597992, -0.015384682454168797, 0.017977561801671982, -0.045434530824422836, ...
[ 0.25600945949554443, 0.021640265360474586, -0.3157634437084198, -0.09941720962524414, -0.3291091322898865, -0.21904240548610687, -0.23281331360340118, -0.025211604312062263, -0.0770508274435997, -0.523570716381073, 0.16764304041862488, -0.05081113055348396, 0.173996701836586, 0.54728990793...
The problem is to define when a variable might be considered as a latent variable. I am interested in how to describe a latent variable, and what are the properties of latent variables. My twofold question is: * When you try to explain what a latent variable is, what do you consider as the main differences between a manifest and a latent variable? * When does factor analysis or item response modeling seem more appropriate ? **Example.** On the one hand, if you want to measure fish weight without any instrument, you can devise items to measure it. In this case, do you rely on a latent variable model? On the other hand, social level is sometimes measured directly through answers to a survey with a linear model (or other models) applied to items such as _highest diploma_ , _numbers of books at home_ , _numbers of electronics devices_ , but not by considering a latent variable. But why can't we use a latent variable model in this case? Thanks in advance.
[ 0.005046853795647621, 0.011513850651681423, -0.007968632504343987, 0.006248921155929565, -0.004850252531468868, 0.0022845289204269648, 0.008172315545380116, 0.0023232402745634317, -0.016066856682300568, -0.03272487223148346, 0.00120770581997931, 0.015571657568216324, -0.02023935131728649, ...
[ 0.29734107851982117, -0.4055825173854828, -0.2645275294780731, 0.18569275736808777, -0.2675963342189789, 0.3497733175754547, -0.0837085172533989, 0.025240102782845497, -0.39271581172943115, -0.827568769454956, 0.5394372940063477, 0.33750373125076294, -0.05496076121926308, 0.113633424043655...
Is it possible to make an insulator have two poles like a magnet? If yes, how to make it?
[ 0.038713131099939346, 0.05229822173714638, -0.01682780496776104, 0.01291795913130045, 0.0302411001175642, 0.005422962363809347, 0.015485352836549282, -0.014374924823641777, -0.02173822745680809, 0.061764393001794815, -0.017253737896680832, 0.03811369463801384, -0.02729858085513115, -0.0000...
[ 0.5667749643325806, -0.17393632233142853, 0.44180405139923096, 0.3788929283618927, 0.04183608666062355, 0.02497766725718975, -0.25681599974632263, -0.20882073044776917, -0.32901307940483093, -0.056205008178949356, 0.38139352202415466, 0.331098347902298, -0.09216246753931046, 0.212296694517...
I live in China. Just purchased a Samsung Galaxy Note 10.1 - 2014 Edition (SM-P600) tablet. Since I have other devices that came with Google Play Store, I had no idea that Samsung's wouldn't. I was wondering, is there a way to get Google Play installed and working on this device? Is it possible to purchase the Samsung/Google version and replace the Samsung/Baidu one that it came with? Or, is there some other workaround for this specific issue? I live in China, and so I don't have many options to buy products designed for Westerners. Hope someone can offer some assistance. Thanks! ~Anthony
[ -0.01879420131444931, -0.010858085006475449, -0.004660516045987606, 0.013014202006161213, 0.010354576632380486, 0.012322474271059036, 0.007184668909758329, 0.03136281669139862, -0.015922270715236664, -0.047175534069538116, -0.007600302342325449, 0.01243361085653305, -0.003687196644023061, ...
[ 0.8464592695236206, 0.35948920249938965, 0.3049193024635315, -0.20105986297130585, 0.20684313774108887, 0.27061495184898376, 0.45840221643447876, 0.30758675932884216, -0.1418450027704239, -0.5886961817741394, 0.13063517212867737, -0.22032465040683746, -0.07723690569400787, 0.03283619135618...
I am trying to change `$GRAILS_HOME` from `/usr/grails-1.3.7` to `/usr/grails-2.0.0` If I just do `export GRAILS_HOME=/usr/grails-2.0.0` it doesn't persist when I log out and back in. I found a line in `/etc/profile` using `grep -R "GRAILS_HOME" /etc/` which reads like GRAILS_HOME=/usr/grails-1.3.7 ... export PATH ... GRAILS_HOME ... So I changed that and logged out. The change does not seem to have taken effect, although my edits to the file remain. I cannot reboot the machine as it is serving stuff for my buddy and he is only giving me some space. But I do want the environment variable changed permanently and everywhere. What do I need to do?
[ -0.004996380768716335, 0.005304867867380381, -0.0018916592234745622, 0.017496023327112198, -0.02536000683903694, 0.015934869647026062, 0.0074857864528894424, -0.012379341758787632, -0.015265632420778275, -0.014497997239232063, -0.00045334198512136936, 0.006524041295051575, 0.0027245848905295...
[ 0.057004526257514954, 0.4137457311153412, 0.4210335910320282, -0.4857628345489502, -0.1015177071094513, -0.04804351553320885, 0.5720653533935547, 0.30119889974594116, -0.22166171669960022, -0.46560296416282654, -0.17244672775268555, 0.41133594512939453, -0.2552204132080078, 0.2933126688003...
I am a new user of LaTeX. I want to underline the answers to mathematics problems using a `\rightharpoondown` symbol. If possible, I would like a `\rightharpoondown` of variable size, as in the image below. How can I achieve this? ![enter image description here](http://i.stack.imgur.com/2nFaF.jpg)
[ 0.023060403764247894, 0.00839055236428976, -0.010645520873367786, 0.015325881540775299, 0.0032726721838116646, -0.005278232041746378, 0.008594903163611889, 0.017259830608963966, -0.01840386912226677, -0.003534377785399556, -0.0007351453532464802, 0.0005902430275455117, -0.012880642898380756,...
[ 0.31881558895111084, -0.06251005083322525, 0.39735180139541626, -0.1645369529724121, 0.04661610722541809, 0.12629707157611847, 0.03246967867016792, -0.1344844400882721, -0.26294276118278503, -0.5864928960800171, -0.10440969467163086, 0.5126762986183167, -0.1291244924068451, 0.2628993690013...
What is the chance Android 4+ phone or tablet with physical HDMI-out connector will be able to output FullHD resolution if the device itself does not have FullHD screen? Are there any rules for this? What if the device's screen resolution is even smaller than 720p (like 800x480)?
[ -0.02514210343360901, -0.007905772887170315, 0.006859404034912586, 0.011812309734523296, -0.06213882938027382, -0.003145399736240506, 0.011716288514435291, 0.013559510000050068, -0.015973582863807678, -0.05470436438918114, -0.013028091751039028, 0.01014979463070631, 0.028999602422118187, 0...
[ 0.2196773886680603, -0.08595330268144608, 0.6918238997459412, 0.21682016551494598, 0.007682720199227333, -0.20039567351341248, 0.4222041666507721, 0.027422990649938583, -0.41048088669776917, -0.3788819909095764, -0.10406147688627243, 0.6384208798408508, -0.2146165817975998, -0.078910857439...
Please tell me if the following sentences have the same meaning or if there is any difference between them. > 1. I can't do this task. > 2. I didn't finish this task. >
[ -0.002557197120040655, 0.030366135761141777, -0.008212968707084656, 0.028078192844986916, -0.00677831145003438, -0.005867263302206993, 0.014695647172629833, -0.0038306971546262503, -0.01754557341337204, 0.10381349921226501, -0.04900268465280533, 0.005263937637209892, -0.001175588695332408, ...
[ 0.23700499534606934, 0.021377306431531906, 0.16698327660560608, -0.11983264982700348, -0.4660957455635071, 0.2887250483036041, 1.0050584077835083, -0.10683943331241608, -0.15836167335510254, -0.5449803471565247, -0.07812978327274323, 0.18259447813034058, -0.16547131538391113, 0.12211851775...
I'm reviewing the cookies served up from my website and one of them is a first-party cookie `__qca` which seems to be set by Quantcast. I'm sure this is just used for tracking, but I can't find this cookie discussed anywhere on Quantcast's website or generally on the web at all? Can anyone provide a reference or further information on what this cookie actually does? Thanks.
[ 0.01377030834555626, 0.0049859569407999516, 0.0069472938776016235, 0.009625189937651157, 0.026820773258805275, -0.0037515831645578146, 0.005964193027466536, 0.021036582067608833, -0.01572350040078163, 0.002526944736018777, -0.01448734663426876, 0.0020242384634912014, 0.009511471726000309, ...
[ 0.7327291965484619, 0.4363883137702942, 0.13404826819896698, 0.3991957902908325, 0.008445898070931435, -0.42822179198265076, 0.03695589676499367, 0.6782655715942383, -0.32672977447509766, -0.4315987825393677, 0.3317639231681824, 0.46309006214141846, 0.02016383782029152, 0.22905869781970978...
I want to know why logistic regression is called a linear model. It uses a sigmoid function, which is not linear. So why is logistic regression a linear model?
[ 0.017826048657298088, 0.011303110979497433, 0.005082620773464441, 0.009775088168680668, 0.01411282829940319, -0.033619314432144165, 0.019523780792951584, 0.0427328497171402, -0.031023776158690453, -0.07558758556842804, -0.026633556932210922, 0.012402130290865898, 0.007636321242898703, 0.00...
[ 0.24424952268600464, 0.0207260362803936, 0.03313741460442543, 0.3438703715801239, -0.0021149117965251207, -0.12936446070671082, -0.380427747964859, 0.2044776976108551, -0.20452654361724854, -0.2535187005996704, 0.28511670231819153, 0.37097597122192383, -0.28977304697036743, 0.5847854018211...
In this answer on Stack Overflow, the term "several" is used as an indeterminate number, the actual value of which is literally in the quintillions: > Zero is one of several values that can be represented exactly. To my ear, this is an exceedingly strange use of "several", which led me to believe that the writer was confused. I realize that trying to truly pin down "several" is probably a hopeless task, but I'm curious if anyone else would use it for such an immensely vast quantity, and if there's regional variation in the usage. So: How many is "several"? Would anyone else use "several" for "quintillions"? This question addresses similar issues, but doesn't seem to have the answer I'm interested in (much of the discussion mentions lower bounds for "several", but not upper bounds).
[ 0.007240127772092819, 0.006338861305266619, -0.013963498175144196, 0.0003706644056364894, -0.0001859969925135374, -0.0021590846590697765, 0.006279318127781153, 0.0036045180168002844, -0.00850037019699812, -0.008273448795080185, -0.0023006435949355364, 0.0030625578947365284, 0.006597222294658...
[ 0.08586690574884415, 0.23738449811935425, -0.004511865321546793, 0.17426592111587524, -0.3007911741733551, 0.06832177937030792, -0.01747271977365017, -0.3196234703063965, -0.4250684082508087, -0.16219981014728546, 0.1081400066614151, 0.19192899763584137, -0.02066017873585224, 0.40970373153...
This always happens to me. Someone tells me to access some folder on Android (in this case `/sdcard/Android/data/com.android.lmt/`) and I can't find it. I'm using the "File Manager" app and asking it to show hidden files and folders but I still can't find the folder. This happens a lot. ![Screenshot](http://i.imgur.com/xcENdycm.jpg) Screenshot (click image for larger variant)
[ 0.0015872506191954017, -0.008337648585438728, -0.008298316039144993, 0.00015209402772597969, 0.01585516519844532, -0.005284591112285852, 0.006746880244463682, 0.022421492263674736, -0.015002354979515076, -0.014687643386423588, -0.010347617790102959, 0.0022969478741288185, 0.01170386560261249...
[ 0.34767842292785645, 0.08514847606420517, 0.32139459252357483, 0.034377165138721466, 0.21056988835334778, -0.021821338683366776, 0.4600786566734314, 0.1898140162229538, -0.7033060193061829, -0.4871009588241577, 0.10714352875947952, 0.2453344762325287, -0.47427526116371155, 0.10295148938894...
I have 3 classes: a JButton extension, FButton; a JTextField; and a JPanel extension, FPanel, that would contain one instance of both. The FButton is supposed to open a file, and make the file name and directory for the file available. The JTextField is meant to display the file name and directory provided by FButton. I figured FPanel should be the one containing the file name and directory. If I make FButton and JTextField private classes of FPanel, problem solved with public variables. How about if I want to separate them into a public FPanel class, a package visible FButton class, and a package visible JTextField class, within the same package? **How can I make the file variables available to all classes in the package?** Should I be doing it that way?
[ -0.03284987062215805, 0.020876575261354446, 0.008274274878203869, -0.006665720604360104, -0.011557580903172493, -0.009022300131618977, 0.012014592066407204, 0.04812650382518768, -0.016131695359945297, -0.013781709596514702, -0.031008444726467133, 0.01211935468018055, 0.037882283329963684, ...
[ 0.03281881660223007, -0.25763702392578125, 0.14260442554950714, -0.2335028499364853, -0.1222725436091423, 0.4763219654560089, -0.2370840609073639, 0.05224165692925453, -0.023190464824438095, -0.5494778156280518, -0.3915867805480957, 0.33561939001083374, -0.2559012174606323, -0.029860019683...
In Emacs/Xemacs, there is a rectangle feature, such that you can delete, copy, etc., "rectangles" of text. Now suppose I have two text files, each containing one column of numbers. Is it possible to use Mathematica to "paste" the columns of text, side by side (perhaps separated by at least one space)? For example, suppose I have two text files: (* file 1 *) 1 20 300 (* file 2 *) 4 50 600 I would like to be able to obtain a single text file like this: (* file 3 *) 1 4 20 50 300 600 Do you have any ideas of how I can do this efficiently? The ideas I have (like stripping and joining row-by-row, and adding the appropriate number of spaces) are really complicated, and I feel like there must be an easier way. Thanks!
[ 0.010977812111377716, 0.010737837292253971, -0.015078194439411163, 0.007661774754524231, 0.004489103797823191, -0.002386217936873436, 0.007400943897664547, -0.012059655971825123, -0.014406895264983177, -0.00313686765730381, -0.016527293249964714, 0.004446370992809534, 0.003677056636661291, ...
[ 0.06203276664018631, 0.2606090009212494, 0.3957418203353882, 0.25982269644737244, 0.13853886723518372, 0.32683366537094116, -0.057895172387361526, -0.22885648906230927, -0.2727726697921753, -0.5215401649475098, 0.14198412001132965, 0.6056017875671387, -0.2014036327600479, -0.05157103762030...
How to force "lpstat" utility to show output messages in English under localized Mac OS X (version 10.6)? Example: lpstat -v Shows in Dutch: apparaat voor Samsung_SCX_4300_Series: usb://..... I need to force it to show the message in English in order to be able to parse the output by the application using regexp ("device for" prefix): device for Samsung_SCX_4300_Series: usb:// I have tried running it with "LANG=en_US.UTF-8" prefix, "LC_MESSAGES=C" prefix, etc. but none of those appear to have any effect under Mac. English language is listed under system preferences on the second position.
[ 0.02402387373149395, -0.004245941527187824, -0.011003541760146618, 0.013796783983707428, -0.01719778962433338, 0.028615090996026993, 0.010420906357467175, 0.006670959759503603, -0.019191619008779526, -0.0028120672795921564, -0.008217323571443558, 0.0009324546554125845, 0.0019452203996479511,...
[ 0.07936622202396393, -0.3491872549057007, 0.6557729840278625, -0.1295369267463684, 0.018384171649813652, -0.007969919592142105, 0.35196131467819214, -0.5150235891342163, -0.335412859916687, -0.46363356709480286, -0.04632445424795151, 0.8318106532096863, -0.5785315036773682, -0.337678134441...
I use the wrapfigure package for placing images in my text at the right side. Now my problem is, that my image is really small and my caption is larger then the image. Instead of a line break at the end of the line it adds the chars below each other. This looks like the following: ![Error on wrapfigure package](http://i.stack.imgur.com/29yk4.png) So is it possible to add a line break after "Abbildung 4.3:" with the content of \caption? My code looks like this: \begin{wrapfigure}{R}{2,76cm} \includegraphics[]{./bilder/mockup-endtimeError} \caption{Error Start- Endzeit} \label{mockup-endtime-error} \end{wrapfigure} BR & Thanks, mybecks
[ -0.0012973519042134285, 0.003533169161528349, -0.010357937775552273, 0.024151090532541275, 0.0017034094780683517, 0.007873543538153172, 0.007159234955906868, 0.008739718236029148, -0.012644513510167599, -0.000044259242713451385, -0.024068545550107956, 0.003669622354209423, -0.009032574482262...
[ 0.032484207302331924, 0.22897931933403015, 0.2854716181755066, -0.3358782231807709, -0.07564231753349304, 0.07215632498264313, 0.7281319499015808, -0.04022665694355965, -0.4395982027053833, -0.2161674201488495, -0.1965002864599228, 0.596691906452179, -0.16819386184215546, -0.16286540031433...
In inelastic collisions, the kinetic energy of the system is not conserved but the momentum is. Kinetic energy is: $0.5 \times \text{mass} \times \text{velocity}^2$. Momentum is: $\text{mass}\times\text{velocity}$. I think that, considering that mass is constant: * if Ke must be different also the velocity of the centre of mass of the system _must be different_ , after the collision. On the other hand: * if the momentum of the system is conserved, the velocity of the centre of mass of the system _cannot be different_. So, how can there be a change in kinetic energy of the system if there is no change in momentum? $mv = m_1v_1$
[ 0.016156913712620735, 0.0014289754908531904, -0.016154920682311058, 0.01227091159671545, -0.008003386668860912, -0.009695095010101795, 0.008266505785286427, -0.023784857243299484, -0.008576592430472374, -0.019962487742304802, -0.0030094238463789225, 0.02041282132267952, -0.009975776076316833...
[ -0.527694582939148, 0.009347563609480858, 0.5461837649345398, 0.014067159965634346, -0.17712616920471191, 0.11691843718290329, -0.001603517564944923, -0.1216350868344307, -0.3476853370666504, -0.33280766010284424, -0.1970689296722412, 0.08172168582677841, -0.4517473578453064, 0.42353737354...
I am not sure how to markup accepted payment methods for an offer. Here are the schema: Offer PaymentMethod Here is my basic example, is this correct? <div itemscope="" itemtype="http://schema.org/Product"> <span itemprop="name">Product Name</span> <span itemprop="description">Product Description</span> <div itemprop="offers" itemscope="" itemtype="http://schema.org/Offer"> <span itemprop="price">$19.95</span> <meta itemprop="priceCurrency" content="USD" /> <meta itemprop="availability" content="in_stock" /> <a href="/buy-now" itemprop="url">Buy Now</a> <meta itemprop="acceptedPaymentMethod" content="http://purl.org/goodrelations/v1#PayPal" /> <meta itemprop="acceptedPaymentMethod" content=" http://purl.org/goodrelations/v1#PaymentMethodCreditCard" /> </div> </div>
[ -0.008491439744830132, 0.011583948507905006, 0.0011137574911117554, 0.017231924459338188, -0.02049553021788597, -0.010296657681465149, 0.007519846316426992, -0.01119623240083456, -0.017880946397781372, 0.017209842801094055, -0.009511529467999935, 0.010193144902586937, -0.015370670706033707, ...
[ 0.18642914295196533, 0.10486985743045807, 0.5860448479652405, -0.1672884076833725, -0.1659359633922577, 0.2688121497631073, -0.1363571733236313, -0.6122731566429138, -0.059432193636894226, -0.6082628965377808, 0.31482189893722534, 0.49992117285728455, 0.3264915645122528, -0.033165134489536...
In the MWE below, are all general methods, of achieving arbitrarily shifting of symbols up or down or left or right in math mode, represented? If yes, that is the answer. If no, what good methods are missed? Specifically, are there any ways to get around the related problems listed below? 1. A tabular stack does the work almost everywhere there are ordered rows and columns of symbols, but a scalebox won't compile inside it. 2. A raisebox seems to work well in some cases, but it won't compile when placing math mode font size commands inside it, and in similar instances. 3. Also one cannot place a raisebox inside a raisebox: the second one causes the first one to be ignored. 4. Horizontal spacing via `\;` , `\`, ``, `\:` often leads to cluttering of some formulas. Is there any way to set the spacing between all symbols in math mode to be a definite size by default? This can be done by doing all work inside tabular stack and setting a parameter, but is there a parameter of this sort in the ordinary math environment? The formula in the MWE is for illustration: \documentclass[12pt]{article} \usepackage{amsfonts,mathtools,amssymb} \usepackage{tabstackengine} \usepackage{tikz} \begin{document} \[% \setstackgap{L}{25pt}\setstacktabulargap{2pt}\TABbinary% \underset{\raisebox{-7pt}{$\textstyle 1$}}{\underbrace{\raisebox{4pt}{\;qbqb\;}}}% \raisebox{-9pt}{\hspace{7pt}${\tabularCenterstack{rcl}{\textstyle = \\ \textstyle +}}$\hspace{7pt}}% \underset{\raisebox{-7pt}{$\textstyle 1$}}{\underbrace{\raisebox{4pt}{\;qqbb\;}}}% \raisebox{-9pt}{\hspace{7pt}${\tabularCenterstack{rcl}{\textstyle = \\ \textstyle +}}$\hspace{7pt}}% \underset{\raisebox{-7pt}{$\textstyle 1$}}{\underbrace{\raisebox{4pt}{\;qb\;}}}% \mathrel{\raisebox{-21pt}{$\;=\;\;\;3$}}% \] \[% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qbqb\;}}}% \;\stackrel{\raisebox{2pt}{=}}{\raisebox{-19pt}{\scalebox{0.9}{\,+}}}\;% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qqbb\;}}}% \;\stackrel{\raisebox{2pt}{=}}{\raisebox{-19pt}{\scalebox{0.9}{\,+}}}\;% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qb\;}}}% \mathrel{\raisebox{-19pt}{\scalebox{0.9}{$\;=\;\;\;3$}}}% \] \scalebox{1}{\begin{tikzpicture}[scale=1]% \node at (2.75,-1.2) {$+$};% \node at (2.75,-1.9) {$d$};% \node at (0,0) {$% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qbqb\;}}}% \;\stackrel{\raisebox{2pt}{=}}{\raisebox{-19pt}{\scalebox{0.9}{\,+}}}\;% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qqbb\;}}}% \;\stackrel{\raisebox{2pt}{=}}{\raisebox{-19pt}{\scalebox{0.9}{\,+}}}\;% \underset{\raisebox{-5pt}{\scalebox{0.9}{$1$}}}{\underbrace{\raisebox{4pt}{\;qb\;}}}% \mathrel{\raisebox{-19pt}{\scalebox{0.9}{$\;+\;\;\;p$}}}% $};\end{tikzpicture}} \end{document} Of course a formula can also be placed in a `standalone` document, `output`, and then included as graphics in the `tikz` environment above where other symbols are drawn over it, and so on, avoiding all issues with macros not compiling one inside the other, but that s a last resort. 5. Better: is there a way to do this in one document, say code halfway through outputs to a `pdf`. This is then is read by the compiler (it would be in the same folder as the ultimate output `pdf` and the `.tex` file) as it continues down the document code and treated as graphics by the remaining code in the compiler, which outputs a second `pdf` (the actual paper). **EDIT:** to clarify, has this been implemented anywhere / can it be done?
[ 0.0019587932620197535, 0.007432172074913979, -0.010404936969280243, 0.007986191660165787, 0.010531069710850716, -0.006862396374344826, 0.008138800039887428, 0.013373833149671555, -0.015964236110448837, 0.00030001357663422823, -0.01474759355187416, 0.001293042441830039, -0.005601249635219574,...
[ 0.22838309407234192, -0.027513153851032257, 0.46960750222206116, 0.23461419343948364, -0.2461630403995514, -0.03002181649208069, 0.24478785693645477, -0.6471686363220215, -0.282534122467041, -0.5648680925369263, 0.1282057762145996, 0.46698611974716187, -0.2952214181423187, 0.07039395719766...
Have very recently started a new position. One of my key challenges is a C# WinForms application (let's call it AppX) that causes a lot of non-planned work. The main reason for this is that it needs to be deployed to several hundred branch offices. Currently, this is done via GP update or bespoke application. Both of these methods are unreliable and means we can't effectively deploy version upgrades. Bit of background to what the software needs to do and some of the technical issues / background:- 1. AppX needs to produce results in real time 2. AppX needs to connect to a SQL server (express or full edition) which is physically located in the branch office 3. AppX needs to ensure that after a version upgrade all remote sites are on the same version 4. AppX can't be deployed via a thin client solution such as Citrix or Remote Desktop services 5. The headquarters network can see the entire branch office network and connect into their SQL servers My initial thoughts were ClickOnce deployment -- however there seems to be some suggestions that there may be some issues with conflicts of the .Net framework with other software Deploy the solution as a web app at the headquarters which connects back into the remote SQL server - technically I'm not sure how I would go about this I guess that this is something that others may have come across so I would be interested to know if there are any other suggestions on how to approach this Thanks
[ -0.018201427534222603, 0.004600760526955128, -0.006655183620750904, 0.018315061926841736, -0.0018372138729318976, -0.01516906451433897, 0.00631877314299345, 0.022571483626961708, -0.009209413081407547, -0.017233338207006454, -0.010709665715694427, 0.008130839094519615, -0.0001483620144426822...
[ 0.6220172047615051, 0.33322426676750183, 0.35762396454811096, -0.1478557288646698, -0.07258497178554535, 0.038811180740594864, 0.17058852314949036, 0.07955057173967361, -0.44277647137641907, -0.6118040084838867, -0.27311185002326965, 0.4767366647720337, 0.1520552784204483, 0.09471805393695...
I try to create the speech recognition system based on 5-state HMM + Multivariate Gaussian function. I use my own feature vector derived from MFCC (Mel-frequency cepstral coefficients). The problem is that after training only a few entries of the HMM transition matrix are left greater then zero. It should be mentioned that only states 2 -- 4 can emit observation. States 1 and 5 help connecting model with other models to make a chain. The initial matrix: $$\left[ \begin{array}{ccccc} 0.0 & 1.0 & 0.0 & 0.0 & 0.0 \\\ 0.0 & 0.6 & 0.4 & 0.0 & 0.0 \\\ 0.0 & 0.0 & 0.6 & 0.4 & 0.0 \\\ 0.0 & 0.0 & 0.0 & 0.7 & 0.3 \\\ 0.0 & 0.0 & 0.0 & 0.0 & 0.0 \end{array} \right]$$ Matrix after model training on phoneme "a" samples is looking like that: $$\left[ \begin{array}{ccccc} 0.0 & 1.0 & 0.0 & 0.0 & 0.0 \\\ 0.0 & 0.0 & 1.0 & 0.0 & 0.0 \\\ 0.0 & 0.0 & 0.8910256 & 0.1089744 & 0.0 \\\ 0.0 & 0.0 & 0.0 & 0.0 & 1.0 \\\ 0.0 & 0.0 & 0.0 & 0.0 & 0.0 \end{array} \right]$$ The training corpus consists of 187 sequences with about 10 vectors in average in every sequence. The performance of such model is quite poor. So what might be the main reasons for such sparse transition matrix after model training? Thank you for any help in advance.
[ 0.0005252690752968192, 0.005352601408958435, -0.013319028541445732, 0.0033876991365104914, 0.0003311967011541128, -0.012233955785632133, 0.007238006219267845, -0.009273339994251728, -0.015181506983935833, -0.003735550679266453, -0.007963930256664753, 0.009817193262279034, -0.0004091882146894...
[ 0.28156206011772156, 0.22750328481197357, 0.7799698710441589, -0.15494339168071747, -0.004878440871834755, 0.2185327559709549, 0.31905055046081543, -0.4459424316883087, 0.1300787627696991, -0.46722713112831116, 0.1988137662410736, 0.6927464604377747, -0.33501559495925903, 0.422040164470672...
I've been playing World of Tanks for a little while now and have got to the KV-1 which, now that I have the top gun, seems to be miles ahead of every other tank I've played. Is this what people call the 'Russian Bias'? Are the higher Russian tanks as good as the KV-1 or is this the best it gets?
[ -0.001382459537126124, 0.013949757441878319, -0.028013592585921288, 0.0016822899924591184, 0.0013978538336232305, 0.015597639605402946, 0.007785489782691002, 0.008235437795519829, -0.014777435921132565, 0.005260620731860399, -0.0035140286199748516, 0.015570521354675293, -0.020004892721772194...
[ 0.0472339391708374, -0.014656144194304943, 0.04133247584104538, -0.06012318655848503, -0.31044188141822815, -0.031393297016620636, 0.22905333340168, 0.07163425534963608, -0.25518909096717834, -0.36714616417884827, 0.11379464715719223, 0.8460181951522827, 0.31090572476387024, 0.118860952556...
I use find -type f -print0 | xargs -0 -n 10 grep -i searchstring to search recursively for a string. But it also gives me error messages like "permission denied". How can I avoid such error messages and only see results?
[ 0.005579597782343626, 0.0036671864800155163, -0.02123153768479824, -0.005434312857687473, -0.01523592695593834, -0.00420202175155282, 0.009997669607400894, 0.011905615217983723, -0.023028165102005005, -0.01667887531220913, 0.009932013228535652, -0.0016612352337688208, -0.016689280048012733, ...
[ -0.22787770628929138, 0.2540675103664398, 0.3227432370185852, 0.11756754666566849, -0.13057447969913483, -0.06396552175283432, 0.37970679998397827, -0.0018035173416137695, -0.04639294743537903, -0.304291695356369, -0.04137243703007698, 0.33380040526390076, -0.10351080447435379, 0.297654181...
I came across the phrase _Gotten to like him_ in one of the English movies. Is it appropriate and grammatical to say the following: > Sarah has gotten to like him.
[ -0.0170024074614048, 0.00904090516269207, -0.01178696658462286, 0.02984447032213211, 0.04233454167842865, 0.011854741722345352, 0.015453651547431946, -0.03903237357735634, -0.01706852577626705, -0.015013930387794971, 0.006023087538778782, 0.02649184875190258, 0.037277087569236755, 0.015969...
[ 0.28673237562179565, 0.1295102834701538, 0.12767282128334045, -0.25963687896728516, 0.06946321576833725, 0.1819794625043869, 0.27109429240226746, 0.5608440637588501, -0.8148928880691528, -0.2760448157787323, 0.04595452919602394, -0.17106789350509644, 0.22821560502052307, 0.1078764423727989...
I have several large data sets which follow the following pattern: A position is measured, a force is applied until a new equilibrium is found. I'd like to find a fit for the position, at least at the plateaus, and preferably of the inter lying section, which in this case approaches a line. I tried fitting the data with `Clip`, and with `Piecewise`. nlm = NonlinearModelFit[v40s1000h,Piecewise[{{a, x < A}, {b, x > B}}], {a, b, A, B, c, d}, x] This creates a decent fit only if I specify the values for A and B, but then I have to estimate those values for each data set manually. It also doesn't really work to just add `NMinimize`, or add the piecewise part for the middle bit. Is There anything else I can try? ![Setting the A to 200 and B to 250 works decently.](http://i.stack.imgur.com/Dsbuf.png) s000.tinyupload.com/?file_id=35616536027018518052 << file
[ 0.0025033350102603436, 0.012613438069820404, -0.005881074350327253, 0.006324027199298143, 0.004652593284845352, 0.009846510365605354, 0.006634421646595001, -0.0010569895384833217, -0.008297396823763847, 0.0030708888079971075, 0.00047817861195653677, 0.009515544399619102, -0.00248342915438115...
[ -0.2770666480064392, -0.5168728828430176, 0.5281686186790466, 0.05461633577942848, -0.051791273057460785, 0.5134120583534241, -0.20596864819526672, -0.31400448083877563, -0.4064507484436035, -0.44192299246788025, 0.2378586083650589, 0.6316248178482056, 0.06027454137802124, 0.17529429495334...
Let's say I use a misunderstood word in direct speech. After that, which is correct to use from the following? 1. _I can explain, if you don't catch the meaning._ 2. _I can explain, if you haven't caught the meaning._ 3. _I can explain, if you didn't catch the meaning._
[ 0.02635105699300766, 0.023479614406824112, -0.020933672785758972, 0.013079335913062096, -0.005139724351465702, -0.02961871214210987, 0.011576344259083271, -0.010363233275711536, -0.017598992213606834, -0.004237334709614515, -0.012642420828342438, 0.0063356817699968815, 0.0022395115811377764,...
[ 0.01676064543426037, 0.35054197907447815, -0.10117845982313156, -0.017801666632294655, -0.5889873504638672, 0.02387564443051815, 0.8572887778282166, -0.13759422302246094, 0.1470346301794052, -0.39185580611228943, -0.24724555015563965, 0.3953353762626648, -0.44713014364242554, 0.11590538918...
_I've read the following question:_ Negative probabilities in quantum physics _and I'm not sure I understand all the details about my actual question. I think mine is more direct._ It is known that the Wigner function can become negative in certain region of phase-space. Some people claim that the negativity of this quasi-probability distribution signifies that the system behaves quantum mechanically (as opposed to classical physics, when probabilities are always positive). Apparently, there are still some controversies about this point. Please read the answers from the previously cited post: Negative probabilities in quantum physics **I would like to know whether there is an equivalence between the negativity of the Wigner distribution and some quantum behaviours or not.** Is it still a question under debate / actual research or not ? My main concern is that there are more and more experimental studies of the Wigner function (or other tomography captures) reporting negativity of the Wigner function. I would like to understand what did these studies actually probe. As an extra question (that I could eventually switch to an other question): What is the _quantum behaviour_ the negativity of the Wigner function may probe ? _Having not a lot of time at the moment, I would prefer an explicit answer rather than a bunch of (perhaps contradictory) papers regarding this subject. But I would satisfy myself with what you want to share of course :-)_
[ -0.003582082688808441, 0.013494856655597687, 0.005722008645534515, 0.017951756715774536, -0.015206106938421726, -0.01799742877483368, 0.006976306438446045, -0.0007493497105315328, -0.009451758116483688, 0.01065298542380333, -0.013567853718996048, 0.01486942172050476, -0.0035728751681745052, ...
[ 0.17781126499176025, 0.028859971091151237, 0.08477380871772766, -0.08716228604316711, -0.2793417274951935, -0.29671987891197205, 0.26544591784477234, 0.03341684862971306, -0.3562900424003601, -0.34533801674842834, -0.07640436291694641, 0.13487643003463745, -0.48685017228126526, 0.593632102...
I tried downloading the blue image in myheritage.com by right clicking. But somehow I always get the picture with parachute no matter where I right click in the blue area. I downloaded the complete page in firefox and then opened images the blue background was missing. How did they manage to do that? Would like to know the method. PS: I went through FAQ of webmasters.SE, nothing was said regarding the explicit mentioning of a third party domain name in the question so I am going ahead and posting it. Edit: How is the image above 955M profiles is also not available to download when we right click?
[ -0.012018658220767975, -0.00914792064577341, 0.0018349499441683292, 0.005804333835840225, 0.007765947841107845, -0.025585971772670746, 0.007682884112000465, 0.020402804017066956, -0.019944749772548676, 0.002533096820116043, -0.010676342993974686, 0.015053056180477142, -0.009111864492297173, ...
[ 0.6546539664268494, -0.09927527606487274, 0.1515677124261856, 0.2080354541540146, -0.23592478036880493, -0.2147151231765747, 0.5031639337539673, 0.19858814775943756, -0.3414697051048279, -0.505428671836853, -0.06959037482738495, 0.3739848732948303, -0.19294323027133942, -0.1709949225187301...
I can't log in through my android Gmail application and have to log in using the browser. I am kind of confused. The response I got from the app was "Can't establish a reliable data connection to the server".
[ -0.03451927378773689, -0.004592457786202431, -0.002946197986602783, 0.0075605339370667934, -0.00659822765737772, -0.004090727306902409, 0.014472872950136662, 0.018764188513159752, -0.02827625349164009, -0.025055576115846634, -0.029059259220957756, 0.02529488503932953, 0.006810469087213278, ...
[ 0.18455597758293152, 0.3211379945278168, 0.25517627596855164, -0.10759057104587555, 0.13134720921516418, -0.034527696669101715, 0.9676858186721802, 0.4927111566066742, 0.05579790100455284, -0.6753866672515869, 0.11264760792255402, 0.2743719816207886, -0.24622459709644318, 0.363135248422622...
I have a world map made up of 6 shp files. All the files and the data frame are set to WGS84. The way Arc 10 and 10.1 decides to display the files, it splits Russia into 2. The majority of Russia can be seen to the East, however, one small bit is displayed at the western extent of the map. This is a single polygon. Is there a way to change the way ArcGIS treats this data so the the whole of Russia is displayed to the east? (I will eventually need to print the data). ![enter image description here](http://i.stack.imgur.com/KZNQo.jpg) EDIT: I'm starting to think that in order to do this I might need to create a new coordinate system (a copy of wgs84) and offset the centre of it so that it sits to the east. Thoughts?
[ -0.0010502112563699484, 0.01666100136935711, -0.013965998776257038, 0.01969575509428978, 0.010750954039394855, -0.00327725475654006, 0.008753083646297455, 0.031401652842760086, -0.01814313977956772, 0.01262364536523819, -0.0019419388845562935, 0.015493806451559067, -0.026937108486890793, 0...
[ 0.3326375484466553, -0.07740451395511627, 0.6366492509841919, 0.056736622005701065, -0.11968302726745605, 0.21062515676021576, 0.3666583001613617, 0.040528926998376846, -0.27933794260025024, -0.872828483581543, -0.05179766193032265, -0.010257111862301826, -0.24071915447711945, 0.5271150469...
I'm trying to learn how to use Latex so I've got a little bit of problems in understanding what is the mistake that I've made. In this case I'm compiling a Table containing multiple datas about currents, voltages and resistences of a circuit and the programme gives back a houndred mistakes about **Missing $** or **Extra {** , then there is a **fatal mistake at lines 1** , but by putting everything under the **%** comment sign I've discovered that the problems still lies in the way I've encoded the lines of the table... > **Error in main.tex (line 29): Missing $ inserted.** $ l.29 {\Large R_ B}& > {\Large 2.2 M\Omega}& {\Large 2.237 M\Omega}& {\Large 2.2... I've inserted a > begin-math/end-math symbol since I think you left one out. Proceed, with > fingers crossed. > > LaTeX Font Info: External font `cmex10' loaded for size (Font) > <17.28> on input line 29. ! Extra }, or forgotten $. l.29 {\Large R_B} & > {\Large 2.2 M\Omega}& {\Large 2.237 M\Omega}& {\Large 2.2... > > I've deleted a group-closing symbol because it seems to be spurious, as in > `$x}$'. But perhaps the } is legitimate and you forgot something else, as > in`\hbox{$x}'. In such cases the way to recover is to insert both the > forgotten and the deleted material, e.g., by typing `I$}'. > > **Error in main.tex (line 29): Missing } inserted.** } l.29 {\Large R_B}& > {\Large 2.2 M\Omega}& {\Large 2.237 M\Omega}& {\Large 2.2... > > I've put in what seems to be necessary to fix the current column of the > current alignment. Try to go on, since this might almost work. > > ! Extra }, or forgotten $. } > > **Error in main.tex:** ==> Fatal error occurred, no output PDF file > produced!* These are the Errors. Here you can find the code that I've written. The troubled linesare the ones are the ones presenting numeric Result...they should all contain the same mistake because I've copy/paste them since I've to create 6 Tables, all with the same Layout. \documentclass[12pt,a4paper]{report} \usepackage[italian]{babel} \title{Relazione - Circuiti con BJT e Punti di Bias} \author{Mengops} \date{Laboratory of April, 14th 2014} \begin{document} \begin{tabular}{cccc} \hline \multicolumn{4}{c}{\Large Tabella Circuito 1}\\ \hline {\Large }& {\Large Teorico}& {\Large Reale}& {\Large Spice}\\ \hline {\Large R_{B}}& {\Large 2.2 M\Omega}& {\Large 2.237 M\Omega}& {\Large 2.2 M\Omega}\\ \hline {\Large R_{C}}& {\Large 5.6 k\Omega}& {\Large 5.555 k\Omega}& {\Large 56 k\Omega}\\ \hline {\Large I_{B}}& {\Large 5 \mu A}& {\Large k\Omega}& {\Large 5.15622 \mu A}\\ \hline {\Large I_{C}}& {\Large 1 \mu A}& {\Large k\Omega}& {\Large 1.082 \mu A}\\ \hline {\Large V_{RC}}& {\Large V}& {\Large 7.079 V}& {\Large 6.0594 V}\\ \hline {\Large V_{RB}}& {\Large 11.3 V}& {\Large 11.388 V}& 11.344 V}\\ \hline {\Large V_{BE}}& {\Large 0.7 V}& {\Large 0.615 V}& {\Large 0.656 V}\\ \hline {\Large V_{CE}}& {\Large V}& {\Large 4.896 V}& {\Large 5.94 V}\\ \hline \end{tabular} \end{document}
[ 0.011344143189489841, 0.014254191890358925, -0.014586333185434341, 0.009211390279233456, 0.01356138288974762, 0.0008679465390741825, 0.006683433428406715, -0.002816710388287902, -0.012892307713627815, -0.010584541596472263, 0.006287388503551483, 0.0009470083750784397, 0.0008506132289767265, ...
[ 0.542324960231781, 0.2588951885700226, 0.19605021178722382, 0.06817451119422913, 0.05382959544658661, 0.4219270348548889, 0.15037813782691956, 0.0003967030788771808, -0.1685841977596283, -0.5545984506607056, 0.10709616541862488, 0.15744417905807495, -0.22778435051441193, 0.3875330090522766...
Once you've finished Diablo II you are able to play hardcore mode. So far, I am aware of these elements when playing hardcore: * When you die, your character is permanently killed and unplayable. * Your name appears in red on b.net. * Your character will wear a robe once killed and you can still participate in b.net conversations etc. In games with similar modes, you will be rewarded whilst playing. For example: * Better items drop far more frequently. * Access to a couple more side-quests. * Ability to play against special boss's (for example Uber Diablo/Baal/Mephisto if you play hell mode + b.net). * A small selection of new skills / spells. * etc Are there any perks when playing hardcore such as an increased chance of finding uniques and set items? Or is the sole benefit bragging rights?
[ 0.0010429567191749811, 0.02718476764857769, -0.003934236243367195, -0.010617085732519627, -0.00679089454934001, -0.0033199104946106672, 0.006897030398249626, -0.007246753200888634, -0.01407511718571186, 0.026191163808107376, -0.017352644354104996, 0.007923371158540249, 0.010079935193061829, ...
[ 0.42882847785949707, -0.2895703613758087, 0.28892576694488525, 0.009361812844872475, -0.5333749651908875, -0.3064017593860626, 0.8037390112876892, -0.2670949399471283, -0.5315833687782288, -0.5699549317359924, -0.07171943038702011, 0.4221566617488861, -0.00962288398295641, 0.04940532520413...
Context/Resume: * I'm changing the android platform to create two partitions in a SD Card. I need to do this as early as possible. I'm currently trying in `init.rc` * It would be nice to obfuscate the access to one of the partitions. If I could keep it hidden would be better And...the long story: I'm trying to create new partitions on an SD card in a device, and I need to be done as early as possible. I thought that the init.rc should be the best location for this, so I tried to add a script call to perform the task, but I'm unable to create these partitions (or get the reason it fails). First of all, is this premise valid? Should I be able to do this? I call the script by: service myscript /system/bin/logwrapper /system/bin/myscript.sh disabled oneshot at init-time. And the content´s of the .sh file is fdisk /dev/sdcard < mykeys.input where "mykeys.input" is the sequence of commands used to perform the task of creating the partitions. Is this the recommended way of doing this?
[ 0.0074335746467113495, 0.009663937613368034, -0.01206907071173191, 0.014469766989350319, -0.033251263201236725, 0.0171479694545269, 0.008290885016322136, 0.026676874607801437, -0.017400460317730904, 0.007450344040989876, -0.01293402723968029, 0.0077436985448002815, -0.0033621378242969513, ...
[ 0.3680823743343353, 0.021547939628362656, 0.6140112280845642, -0.0450032614171505, 0.40707796812057495, 0.04662506282329559, 0.06671184301376343, -0.03177719563245773, -0.4464859366416931, -0.9168634414672852, -0.06345045566558838, 0.5931581854820251, 0.21733638644218445, -0.08873720467090...
I've imported vector layers from map info pro to both SQL server and post gis using easy loader. In both cases I've included style column in the exported table. When I create new store in geoserver and see the list of layers , each layer has the mi_style column in the list of features at the bottom of the screen . But when I publish the layer it still have the default style of geoserver . How can I set it to read the style from the column ?
[ 0.009626807644963264, 0.0015732425963506103, -0.002511331345885992, 0.025371350347995758, -0.010792454704642296, 0.022332172840833664, 0.010017321445047855, -0.013260371051728725, -0.01692228764295578, 0.0020236424170434475, -0.000201464252313599, 0.020777493715286255, 0.016384383663535118, ...
[ 0.3145952820777893, 0.011844358406960964, 0.595819890499115, -0.048455581068992615, 0.046569641679525375, 0.07447553426027298, -0.004692746326327324, 0.07425311207771301, -0.05237139016389847, -0.9484871625900269, -0.009007954970002174, 0.6060143709182739, -0.15086987614631653, 0.122657507...
I like to think of myself as an expert in the area of Final Fantasy VII. As it was on steam for $8.00 recently, I bought a copy for my girlfriend. I've been answering her questions with relative ease, but she asked me something that took me completely by surprise. Does it matter which equipment slot you place your materia in? It's something that I guess I never really gave thought to, and there isn't any information I can find online. Thanks in advance!
[ -0.002700011944398284, -0.0022112340666353703, -0.0018815892981365323, 0.0009708378929644823, 0.01570764370262623, 0.0017130088526755571, 0.003911470528692007, 0.0025750568602234125, -0.014822544530034065, 0.015449880622327328, -0.0023491380270570517, 0.014583677053451538, 0.0008946650777943...
[ 0.5620625019073486, 0.03886425122618675, 0.20310205221176147, 0.4783853590488434, -0.21516183018684387, -0.045518726110458374, -0.020083358511328697, 0.30160194635391235, -0.07353291660547256, -0.34903502464294434, 0.45405837893486023, 0.24628876149654388, 0.6087286472320557, 0.69478583335...
Example 7.2 in David Griffiths E & M book (3rd edition) has a side view of 2 concentric cylinders, with smaller radius $a$ and larger radius $b$. The region in between $a$ and $b$ has conductivity $\sigma$. "If they are maintained at a potential difference $\textit{V}$, what current flows from one cylinder to the other for a given length $L$?" The E field is pointing radially outward along $\textit{s}$. My question is: what direction is the current? Do electrons flow in the opposite direction of an E field? If so, does that mean the current is flowing radially inward, along $\textit{-s}$, from $b$ to $a$?
[ -0.005278067663311958, 0.010722656734287739, -0.010699529200792313, 0.019084980711340904, 0.006084710359573364, -0.025924019515514374, 0.008985641412436962, -0.006377696059644222, -0.010414748452603817, 0.003248053602874279, -0.008674058131873608, 0.00816495344042778, -0.0316222719848156, ...
[ 0.33474254608154297, 0.17841604351997375, 0.6033591628074646, -0.049912042915821075, 0.08842546492815018, -0.21481551229953766, -0.2483687698841095, -0.49282607436180115, -0.23295988142490387, -0.23718251287937164, 0.22453635931015015, 0.19349849224090576, -0.43546584248542786, 0.552838742...
I'm used to _modeless_ editors. Only the past year I've been using vim/gvim, which has a _modal_ approach. I'm used to tabs in all editors, since before vim all of them are used to it. In gvim, you don't necessarily need to use tabs: you can just use buffers. My question is: what are the advantages/disadvantages between these two approaches (buffers vs tabs)? Why do these both options exist?
[ 0.0065165613777935505, 0.015737079083919525, -0.0019204930867999792, 0.012344435788691044, -0.016977548599243164, 0.005937275942414999, 0.008293427526950836, -0.019903745502233505, -0.015883853659033775, -0.007482703309506178, -0.005420724395662546, 0.008558583445847034, -0.00285425572656095...
[ 0.10023277252912521, 0.11954305320978165, 0.46307894587516785, -0.1111004576086998, -0.15706823766231537, 0.025279242545366287, -0.04899094998836517, 0.2529616951942444, -0.28125372529029846, -0.9306945204734802, 0.13874582946300507, 0.5431379675865173, -0.39270949363708496, 0.153373703360...
\documentclass[12pt,a4paper]{article} \usepackage[version=3]{mhchem} \begin{document} \ce{K_a \times K_b = K_w} \end{document} ![enter image description here](http://i.stack.imgur.com/epFeS.png) In `\ce{}` all text is supposed to be regularized. See how the `K_a` is regularized but then the `K_b` is italicised and the ensuing `K_w` is again subject to `\ce`'s regularization. Why does this interruption in formatting happen? I mean, it probably has to do with the `\times`, but then why does italicised formatting resume after that?
[ 0.001126964925788343, 0.007739860564470291, -0.011133024469017982, 0.016630293801426888, -0.005792777519673109, 0.009991275146603584, 0.0072491783648729324, -0.004594963043928146, -0.012000110931694508, 0.007434300612658262, -0.011465951800346375, 0.00259688263759017, 0.005806056782603264, ...
[ -0.12410363554954529, 0.04185429960489273, 0.5613792538642883, 0.07921196520328522, 0.012033587321639061, -0.15870274603366852, 0.22462807595729828, -0.2836642563343048, -0.46372318267822266, -0.6055896282196045, -0.4685898721218109, 0.41554129123687744, -0.29247066378593445, -0.0087822685...
My Samsung Galaxy Ace has problems with its Wi-fi connection. It's very weird because it shows that its connected to a network but i still cannot connect to the internet. Can someone please help me out with this? thanks a lot!
[ -0.011007084511220455, -0.0061557712033391, 0.006648472975939512, 0.02103441022336483, -0.029427213594317436, 0.00739371357485652, 0.008663143962621689, 0.030116604641079903, -0.025447893887758255, 0.013799844309687614, -0.015230178833007812, 0.01262371614575386, -0.02091941609978676, 0.00...
[ 0.17990897595882416, 0.22074739634990692, 0.48870643973350525, 0.3711753785610199, 0.07522420585155487, -0.10390721261501312, 0.5426178574562073, 0.13572748005390167, -0.37395015358924866, -0.6059108972549438, 0.08717641979455948, 0.3011610209941864, -0.29806727170944214, 0.257006913423538...
I am performing this simple experiment: I have one variety of grass and 8 different fungi (say #1 to #8). I am going to put 10-20 grass plants in each one of 18 containers, and then I will put each one of the fungi in 2 containers and let 2 with no fungus ("negative control"). After some time I will count how many plants have been killed by each one of the fungi in each one of two replicates for that fungus. I want to compare the results for each one of the fungi with respect to the negative control. I think I can do that using the Fisher exact test, but I don't know if I have to combine the counts of the two replicates into a single count (like using a single container with all the plants together) or if I can keep those count separated. Also, I would like to know if I have to run the test doing pairwise comparisons (controls vs #n) or if I should run a single test considering all the results at the same time. May I do that also with the Fisher exact test? Do I have to make a multiple comparison adjustment? Thank you for any help. I am very confused about this.
[ 0.010150757618248463, 0.016856271773576736, -0.013404492288827896, 0.019614923745393753, 0.019442712888121605, -0.010215183719992638, 0.008141607977449894, -0.005242181941866875, -0.0165998637676239, -0.015825314447283745, -0.007885166443884373, 0.010975981131196022, -0.01880774274468422, ...
[ 0.6572685241699219, 0.1991610825061798, 0.08387133479118347, 0.16711443662643433, 0.060230765491724014, 0.5488117933273315, 0.027332866564393044, -0.13760139048099518, -0.4991764426231384, -0.6648631691932678, 0.4574778079986572, 0.24468165636062622, -0.3978639841079712, 0.4652106463909149...
For example > /data/data/com.android.providers.contacts/databases/contacts.db I cannot see it both in Windows 7 and Ubuntu 11.04.
[ -0.006094834301620722, 0.008659428916871548, -0.010083484463393688, 0.025941478088498116, -0.009638863615691662, 0.00015454583626706153, 0.017565850168466568, 0.0295390747487545, -0.025112247094511986, -0.052628837525844574, -0.02638811618089676, 0.03254112973809242, -0.028294922783970833, ...
[ 0.0839432030916214, 0.03195522725582123, 0.44683608412742615, -0.3081504702568054, 0.021495800465345383, -0.11096339672803879, 0.38506263494491577, 0.15119832754135132, -0.0776781216263771, -0.7560248970985413, -0.4177325367927551, 0.4299115538597107, -0.3603014349937439, 0.203635603189468...
I want to have a pressure plate beside my door on the inside. When I step on it I want it to toggle my door open or closed. So if I walk out my door should stay open until I (or someone else) walks back inside.
[ 0.008923622779548168, 0.03259745612740517, -0.001143994159065187, 0.0123617984354496, -0.032825883477926254, 0.0015102113829925656, 0.012147421948611736, -0.017824320122599602, -0.020305396988987923, 0.03614002466201782, -0.01852666400372982, -0.002530238125473261, 0.02816854603588581, 0.0...
[ 0.18823707103729248, 0.35647913813591003, 0.4076004922389984, -0.1290971040725708, 0.18757526576519012, 0.30613449215888977, 0.26302799582481384, -0.36646798253059387, -0.4941486120223999, -0.7078255414962769, 0.012600698508322239, 0.037290289998054504, -0.031246621161699295, 0.08352603018...
I tried defining a new command `\Quest` as follows but the error Lonely \item--perhaps a missing list environment appears. \def\solnversion{0} %%set this to 1 to get solutions \newcommand\Quest[3]{\item\ifnum\solnversion=1 #3\else #2\fi} But if the `\item` command is removed, the error Something's wrong--perhaps a missing \item appears instead. How should the code be corrected?
[ 0.011118927039206028, 0.016320673748850822, -0.004592479206621647, 0.009151481091976166, -0.004333812743425369, 0.0079596396535635, 0.008388449437916279, 0.004546306561678648, -0.013533185236155987, 0.0073406267911195755, -0.015797778964042664, 0.006702587474137545, -0.007679433096200228, ...
[ -0.23419822752475739, 0.03803848475217819, 0.34252575039863586, -0.13944847881793976, -0.027404753491282463, 0.0009030451183207333, 0.6142798066139221, -0.2720465660095215, -0.22614067792892456, -0.5290270447731018, 0.008961102925240993, 0.49367085099220276, -0.5011814832687378, 0.38031560...
The following code retrieves only custom post types with the custom taxonomy "Slider." I would like to change their order to ASC. The code: <?php // Retrive custom post type with a custom taxonomy assigned to it $posts = new WP_Query('post_type=page_content&page_sections=Slider (Front Page)') ?> <?php while ( $posts->have_posts() ) : $posts->the_post(); ?> <?php the_content(); ?> <?php endwhile; ?> <?php wp_reset_query(); ?> Not sure if I should use an array (not sure how anyways). Any suggestions?
[ 0.01283260714262724, 0.009556259959936142, 0.00393674336373806, 0.024605542421340942, 0.02463676780462265, 0.003988834097981453, 0.00895924773067236, 0.00837934110313654, -0.012619039043784142, 0.0018604989163577557, -0.010475877672433853, 0.004169904161244631, 0.001818820834159851, 0.0281...
[ 0.13489355146884918, 0.03792104870080948, 0.8125178813934326, 0.017171265557408333, -0.2165941447019577, 0.2090035080909729, 0.1256961077451706, -0.1987929493188858, -0.33446311950683594, -0.5858331322669983, -0.23962707817554474, 0.23383019864559174, -0.35515013337135315, 0.25011384487152...
When the system emits an Intent for which you have multiple apps that can handle the intent, a dialog opens to tell you which application you want to use with that intent. Is there a way to see what these associations are, and clear them? If there is no app or interface to do that, do you know of an API to access these associations? EDIT: When I said associations, I was thiking of the "late-bound" associations. For example, when I open a link and I have multiple browsers, an intent is fired saying "view link", and I choose "always use that application"; what can I do if, after some time, I would prefer to use another application for that intent?
[ 0.004342994652688503, 0.009149642661213875, 0.006130200810730457, 0.014171669259667397, -0.013996230438351631, 0.004159173928201199, 0.007341733202338219, 0.012122157029807568, -0.017488587647676468, 0.013949920423328876, -0.0040079024620354176, 0.01482115127146244, 0.018615838140249252, 0...
[ 0.311606764793396, 0.05080540105700493, 0.4190073013305664, 0.23010587692260742, 0.18638156354427338, -0.23353593051433563, 0.38632673025131226, 0.07881525158882141, -0.38544347882270813, -0.5022475719451904, -0.07311027497053146, 0.5258826613426208, -0.4648202955722809, 0.0815657898783683...
I'd like to make a document that would leave ample room for the reader to fill in answers on the sheet. (nothing fancy and interactive, just whitespace for writing on the printed version) What packages/classes exist for achieving this? I suppose I could just use massive tables and lots of `\bigskip`s but that seems suboptimal... Here's an example of the sort of thing I want to achieve, I could add lots of space manually, but that doesn't seem ideal... \documentclass{article} \begin{document} \begin{table} \begin{tabular}{|l|r|r|r|r|r|r|} \hline & A & B & C & D & E & F \\ \hline Spelling & & & & & & \\ \hline Grammar & & & & & & \\ \hline Use of literature & & & & & & \\ \hline Strength of Argument & & & & & & \\ \hline \end{tabular} \end{table} \end{document}
[ -0.005841006990522146, 0.002370845526456833, -0.006483909208327532, 0.017206946387887, -0.004101553000509739, 0.010989569127559662, 0.005600046366453171, 0.021279316395521164, -0.015030553564429283, -0.0013005367945879698, -0.003750039730221033, 0.002850928343832493, 0.003909687511622906, ...
[ 0.22601650655269623, 0.17960526049137115, 0.0884498804807663, 0.2350350320339203, 0.20976145565509796, -0.1498984545469284, 0.25973108410835266, -0.10899631679058075, -0.15810781717300415, -0.7714244723320007, -0.01805190183222294, 0.5669428706169128, -0.005793026648461819, -0.170638710260...
For those familiar with the Google Chrome browser, I'm looking for the feature which basically treats the address bar as a search bar when not navigating to a site. I would find this handy since it's annoying to have to navigate to Google.com first (which is the only option on the stock browser) or using the search widget. Is there a browser for that?
[ -0.009747427888214588, -0.0019805715419352055, -0.001382390153594315, 0.0029331345576792955, -0.04228828474879265, 0.0024029554333537817, 0.008448691107332706, -0.006571614183485508, -0.022158028557896614, 0.0223050806671381, -0.014537137001752853, 0.0027132935356348753, -0.00966441351920366...
[ 0.5512375235557556, 0.11576585471630096, 0.3913436532020569, 0.16785591840744019, 0.22825326025485992, -0.0310980211943388, 0.20271286368370056, 0.3327828347682953, -0.27767977118492126, -0.5591334700584412, 0.24934665858745575, 0.467069149017334, 0.12632372975349426, 0.030699223279953003,...
I registered 2 taxonomies: `festival_year` and `art_genre`. I would like users to search for a specific year and genre but am not sure the best way to do this. I have tried various plugins but none have worked. I was able to create separate files, like `taxonomy-art_genre-consumable-crafts.php` with the following code: <?php $festivalquery = wp_parse_args($query_string); $festivalquery['tax_query'] = array( 'showposts' => 10 , 'orderby' => 'title', 'order' => 'ASC', 'paged' => $paged, 'relation' => 'AND', array( 'taxonomy' => 'festival_year', 'terms' => array('2010-artists'), 'field' => 'slug', ), array( 'taxonomy' => 'art_genre', 'terms' => array('consumable-crafts'), 'field' => 'slug', ), ); query_posts($festivalquery); ?> which gives me the query I'm looking for. However, there are 11 genres (and of course there will be multiple years). Instead of creating separate page templates, what is the best way to accomplish this? The test site: http://www.beta.cornhillartsfestival.com/artists/ Thank you so much!
[ 0.015536913648247719, 0.01472515519708395, 0.009621978737413883, 0.019756242632865906, 0.014380216598510742, -0.005289343651384115, 0.008543015457689762, 0.024524081498384476, -0.017334016039967537, 0.011649693362414837, -0.006293831393122673, 0.01624280959367752, 0.0019682857673615217, 0....
[ 0.9129775166511536, 0.22311004996299744, 0.4373926818370819, -0.043049756437540054, -0.4641263782978058, 0.04730570688843727, 0.00589505210518837, 0.004269284196197987, -0.4948349893093109, -0.0009176351595669985, 0.4080541133880615, -0.09804453700780869, 0.3643026053905487, 0.537686467170...
I have very little internal memory (user memory 150mb), so after the auto update of Play Service and the Play Store, I took a back up (using ES File Explorer). I have root access, so I pasted those back ups in the `/system/app` folder. After that I restarted my phone, and that's when the problem started. The Play Store opens but can't see a thing, and Play Services crashes every time. What I have done: * I installed 3 version of the Play Store. No positive result. * Installed the latest version of Play Services. No positive result. * Searched many blogs. They say it might happen as not com.android.gapps, but I can see the service running. However, after installing Play Services, both Play Services and gapps are crashing. What can I do?
[ -0.02131405472755432, -0.002396028023213148, 0.0006937272846698761, 0.011684685945510864, 0.008797218091785908, -0.0033081085421144962, 0.005153602920472622, 0.00947378110140562, -0.014484383165836334, -0.009137271903455257, -0.018053535372018814, 0.012505419552326202, 0.005282850004732609, ...
[ 0.4312085807323456, 0.24354292452335358, 0.43246084451675415, -0.2800533175468445, 0.34477779269218445, 0.049386441707611084, 0.7754570841789246, 0.04292239248752594, -0.15742801129817963, -0.8197802901268005, 0.1213185265660286, 0.8874391913414001, -0.42671695351600647, 0.2370963394641876...
I have data from an experiment that is testing how the order of two studying methods (visual or auditory) affects word recall. For analysis a multi-factor anova with a repeated measure is appropriate, but I am not sure if I am structuring my data correctly. This is the command I'm using: `aov(recalled_items~task*order)+Error(subject/task)+(order))` Here is an example of the data structure: Subject Task Order Recalled Items A Visual First 13 A Auditory Second 22 B Visual First 14 B Auditory Second 28 C Visual Second 10 C Auditory First 15 D Visual Second 14 D Auditory First 29 * Does R know to compare Visual 1 and Visual 2 recall values and Auditory 1 and Auditory 2 recall values? I am worried that because of the way I structured my data R is just comparing Visual 1 and Auditory 1 and as a result I am getting no effect.
[ -0.005784090608358383, 0.022865762934088707, -0.024635296314954758, 0.023921441286802292, 0.004614995792508125, 0.003239216050133109, 0.009328916668891907, 0.016446225345134735, -0.010720761492848396, 0.04382750391960144, -0.019189583137631416, 0.009291790425777435, 0.006397188641130924, 0...
[ -0.23967385292053223, 0.20152106881141663, 0.11736001819372177, -0.06879168748855591, -0.3295679986476898, 0.6071277856826782, 0.2338334023952484, -0.4020105004310608, -0.2087506651878357, -0.5841279625892639, -0.17649993300437927, 0.4859407842159271, 0.08234919607639313, 0.121994331479072...
How do I `InstallR` on OS X to use my already installed and configured R? Simply using InstallR["RHomeLocation" -> "/Library/Frameworks/R.framework/Resources"] results in $Failed InstallR::nopaclet: Could not find RLink runtime installed. Please use RLinkResourcesInstall to install it How do I connect _Mathematica_ to my installed R? Do I have to install a separate R for _Mathematica_?
[ 0.02083212323486805, 0.0009339869720861316, 0.006891445256769657, 0.0076318709179759026, 0.025058235973119736, -0.017729105427861214, 0.009168833494186401, 0.001862280652858317, -0.015479041263461113, -0.016754982993006706, -0.0036120200529694557, -0.001095297047868371, -0.001783253625035286...
[ 0.09351888298988342, 0.17384207248687744, 0.577972948551178, -0.02609354816377163, 0.00887732207775116, -0.0992601066827774, 0.5568235516548157, -0.48010554909706116, 0.021599050611257553, -0.5958204865455627, -0.12688618898391724, 0.7437042593955994, -0.3876919448375702, -0.01234613917768...
## Background: This question came to mind after I ignorantly came to the conclusion that I was driven to go into game development, because it is the only place I can do creative development without the hinderance of clients wanting things tweaked, or security needing to be extreme, or deadlines pushing away research. I'm not talking about all cases, I'm talking about "in the end", where I've found my "final" kind of job. Game development I speak of here is indie development, where original intuitive design is important. As far as I know, this form of game development is the only way a person with only a bachelors can do research level development [AI, original UI (like voice commands / NLP), advanced graphical development]. With a phD I understand i could do any research, but I'm still unsure that is the path I wish to take. ## Question: What type of jobs or industry can allow someone with only a bachelors degree to do virtually unhindered creative research and development? This is subjective, so I'm mainly looking for opinions and RL accounts.
[ -0.020766258239746094, 0.004915307275950909, -0.005889951251447201, -0.00894039124250412, -0.013780297711491585, -0.00024941470474004745, 0.005341438576579094, 0.01171506941318512, -0.01019919291138649, -0.009545616805553436, 0.00011411291779950261, 0.012282337993383408, 0.01764363795518875,...
[ 0.6289535760879517, -0.03397664427757263, 0.18804021179676056, 0.25327807664871216, 0.25457751750946045, 0.25658753514289856, -0.1769343465566635, 0.25951242446899414, -0.12790192663669586, -0.23718270659446716, 0.011662832461297512, 0.640749454498291, 0.09873804450035095, 0.07621436566114...
I got puzzled when putting the following pieces together. I checked it online but couldn't find a convincing explanation. > In large supermarkets, management must _decide (on)_ what to put on sale. Is the "on" preposition necessary here? What difference would it make if we removed it?
[ 0.005020912270992994, 0.005737997125834227, 0.004226456396281719, 0.04833007976412773, -0.002084557432681322, 0.02442004904150963, 0.009532609954476357, 0.03151366114616394, -0.020808378234505653, 0.015100091695785522, -0.03084542043507099, 0.022321315482258797, 0.012496746145188808, 0.024...
[ 0.8009341955184937, -0.018974587321281433, -0.09429724514484406, -0.1524084508419037, -0.09084147214889526, 0.3040866553783417, 0.014708670787513256, 0.30414092540740967, -0.15848016738891602, -0.33640700578689575, 0.019504012539982796, 0.3920799493789673, 0.0038403524085879326, 0.32886680...
I need to access an USB camera connected to my (rooted) Android. I'm on a Samsung Galaxy S4 with Android 4.2.2. I want to access it with a normal app, so I need the device to be world readable. I can connect via shell and type root@android:/ # su root@android:/ # cd /dev root@android:/dev # chmod 666 video4 and it works. But each time I disconnect the camera, and reconnect it, I have to do it again. Via shell I can remount the root file system r/w mount -o rw,remount -t rootfs rootfs / and then edit the file `/ueventd.rc` (actually it's `/ueventd.qcom.rc` ) and add /dev/video* 0666 system camera But even if I put 666 I only get root@android:/dev # ls -la video4 crw-rw---- system camera 81, 18 2013-08-08 10:14 video4 Is there a way to grant a 666 permission permanently? So that each time I connect the camera it gets the right file permissions? Thanks
[ -0.00609154487028718, -0.0059430659748613834, -0.005688799545168877, 0.005068700760602951, -0.021884415298700333, 0.014483338221907616, 0.008566179312765598, 0.0008849237929098308, -0.015758970752358437, -0.017781861126422882, -0.018469469621777534, 0.0023800963535904884, -0.0131925018504261...
[ 0.1737809032201767, 0.23687079548835754, 0.5251500010490417, 0.09964209049940109, 0.208439439535141, 0.1133401021361351, 0.6026944518089294, 0.051589976996183395, -0.20326481759548187, -0.5791934132575989, -0.21725696325302124, 0.5512862801551819, -0.18564994633197784, 0.19319282472133636,...
I have a facebook app page, and a separate facebook business profile page. The business page was created, but not by me, some time before the app and its page were created. **Is there any way to connect the two pages, or import the content and friends from one to the other**? The older profile page has some content; a set of friends and wall posts that I don't want to lose. It was created before I had a chance to set up an app page. Since the app was created more recently, it does not have any content posted to it. I intended the app page to eventually hold some advertising info for my main website itself (non-canvas, just using fb for the connect api etc). The idea being that as people sign up on my site through facebook's OAuth, I could use the graph api to post to their wall. The wall posts are working as expected but naturally they are directing users to the facebook app page, which has no content, friends etc. I'd prefer to be directed to the original business page, where the party is really happening. Now it seems that the two pages are completely separate; what would I need to do to direct the users to the business page?
[ -0.008321625180542469, 0.003978610038757324, 0.010095600038766861, 0.014105336740612984, 0.009945614263415337, 0.024458669126033783, 0.008692475967109203, 0.017537934705615044, -0.01914343796670437, -0.02043376863002777, -0.009978905320167542, 0.02013617753982544, -0.00044147809967398643, ...
[ 0.5447956323623657, 0.3639455735683441, 0.4080439805984497, -0.11280633509159088, -0.05441303923726082, 0.28400930762290955, -0.031510937958955765, -0.020397666841745377, -0.34215399622917175, -0.49487772583961487, 0.24350346624851227, 0.34163305163383484, -0.1302047222852707, -0.027086753...
I have recently updated my website. In fact I have completely rewritten it using wordpress. It was originally constructed using ASP. In the process, I have recycled a a large amount of my original website. This resulted in a large amount of broken links. Google webmaster tools report more than 30000 broken links and most of them are internal. These mainly result from there being reported links to urls not used anymore. That is links to scrapped, obsolete url from the previous version that does not exist any more. From SEO perspective I have read that if 404 errors are from internal links, it is best to delete the links. When I click on a broken link from the list displayed at health-> crawl errors, google shows me where this broken link appears at 'linked from' tab. When I click on links from 'linked from' , the purported broken link is not displayed on the 'linked from' page. That is the users of my website has no chance of trying to load this broken link. The broken link still shows up in page source. i.e. when i try view page source from chrome. The broken link is usually used in some javascript To clafiry, let www.myhomepage/broken_link be the broken link reported by google webmaster tools. The google 'linked from' tab shows that this url was linked from www.myhomepage/some_other_page. When I view the page source, the broken link is usually in [script type="text/javascript"]var _bpfbRootUrl="www.myhomepage/broken_link";[/script] Again, i read that it is advisable to just delete internal broken links, but there are just so many of them and it would be very time consuming to manually delete each and every of them. What would be the best way to deal with this situation? I would like to avoid using custom 404 page. 302 does not seem like an option either, parsing url using regular expression to not redirect valid pages seem very complicated too Google webmaster tools report that index on my page had been sharply declining since it was rewritten with wordpress and dropped to 10% of original. Thank You.
[ 0.0033427197486162186, 0.009417916648089886, -0.006848314311355352, 0.023268375545740128, 0.014032678678631783, -0.009450443089008331, 0.007312421686947346, -0.001831153640523553, -0.013146731071174145, -0.02167697623372078, -0.0039815716445446014, 0.01555081270635128, 0.004330066032707691, ...
[ 0.4081459939479828, 0.4543282985687256, 0.3128456771373749, 0.0682448297739029, -0.08148954063653946, -0.0007122729439288378, 0.6545716524124146, 0.14265139400959015, -0.28963151574134827, -0.8565881848335266, -0.11339090764522552, 0.24499043822288513, -0.14569957554340363, 0.2150592654943...
When we say, > It is **not** fair. or > It is **unfair**. I'm not sure enough to say whether both of the sentences have the same meaning or not though superficially, there is no difference between them but if we say, > Something is **not** possible. or > Something is **impossible**. then there is **really an observable difference** in my first language. Hence, the question - is there a difference between the preceding two sentences **in English**?
[ -0.012214841321110725, 0.01803980953991413, -0.004372136201709509, 0.02536514215171337, -0.009060014970600605, -0.015081332065165043, 0.009335056878626347, 0.006957785226404667, -0.0070076738484203815, 0.00045815418707206845, -0.019070737063884735, 0.004196001682430506, 0.011177262291312218,...
[ 0.035810407251119614, 0.1692674160003662, 0.037932850420475006, 0.09276694059371948, -0.5653260350227356, 0.3011740446090698, 0.6484097242355347, -0.010833080857992172, -0.13931594789028168, -0.9693722128868103, 0.04680310934782028, 0.33398276567459106, 0.04020466282963753, -0.056016292423...
I need one (or maybe two) LaTeX symbols that I don't find in the comprehensive list of symbols. The symbol is like the normal plus-minus sign `\pm`, but with parenthesis around the minus sign. Is such a symbol available somewhere? And if not, what is the best way to construct it from `(`, `\pm` and `)`? Preferably I would also want the opposite: `\mp` with parenthesis around the plus. Any ideas?
[ 0.002130392473191023, 0.01441999338567257, -0.01299250964075327, 0.018943000584840775, 0.0056783161126077175, -0.002342689549550414, 0.007702081464231014, -0.012227719649672508, -0.016529114916920662, 0.013623724691569805, -0.004979992751032114, -0.0073420158587396145, -0.0028127050027251244...
[ 0.22748452425003052, 0.3678152859210968, 0.3320835828781128, 0.1567612737417221, 0.17676132917404175, 0.6294065117835999, 0.019579073414206505, 0.19219854474067688, -0.17646482586860657, -0.5682489275932312, 0.025453532114624977, 0.4210049510002136, -0.10262373089790344, 0.0225561149418354...
I have a Linux Mint 11 instalation (until some months ago I had Ubuntu 11.04) and since yesterday I have different problems. Let's explaine in steps: * I had updated my system (I can't say what files were installed) * restarted and the system entered in "low graphic mode" but no options from menu are working. I have made a video (bad quality but to get the idea) here. * I have made Bootable USB stick with different distributions like Puppy, DSL, Fedora, Ubuntu also tried with Live CD (Ubuntu) but I get errors and the system remains in "standby". Here I have made pictures of the errors: Linux Mint editing the grub Fedora live USB Ubuntu Live CD * conclusion until now is that I can't manage to boot with LiveCD or USB stick although with Puppy Linux it worked but I couldn't mount the ext4 drives but the ntfs drives were mounted corectly. * the tty doesn't work that means that I can't write something like: sudo dpkg-reconfigure xserver-xorg * I have dual boot with WinXP. Entered here and installed Ext2explore and Ext2fsd at least to back up my files from /home. When entering on /root partition with Ext2fsd is asking me if I want to format it. I didn't continued but as final resort can I format the /root and try again to install from scratch a new distribution? My /home partition will remain untouched hopefully. Someone knows a reason why I can't boot or why even the Live CD gives errors? I have used more than one live cd and more than one bootable usb stick. I don't see on my system file /etc/X11/xorg.conf but is xorg.conf.failsafe. Here you have an archive where I managed to save xorg.0.log and other files: link. If you can please tell me what other files are relevant to save to place here maybe it's shows clues why this is happening... Thanks and sorry for the mistakes in writting this topic...
[ 0.004267159383744001, 0.003958639223128557, 0.005134573206305504, 0.004982842598110437, -0.021782904863357544, -0.024019703269004822, 0.005744437221437693, -0.004063262138515711, -0.011903800070285797, -0.005580445751547813, -0.03248211741447449, 0.008505879901349545, -0.009738674387335777, ...
[ 0.38987329602241516, 0.3129364848136902, 0.10603415220975876, -0.05482747405767441, -0.15015068650245667, 0.06335397809743881, 0.32967695593833923, 0.5222615003585815, -0.1315614879131317, -0.6388424038887024, 0.04562004655599594, 0.6678901314735413, -0.23303839564323425, 0.150323122739791...
This is my sentence: > The A & P’s regular customers are depicted as "scared little pigs in a > chute." Daily, they live the same monotonous existence. Microsoft word insists that there should be a comma after "depicted". I don't think it makes sense. What say you?
[ -0.037026386708021164, 0.02963467687368393, -0.009334304369986057, 0.03406917303800583, -0.00904759019613266, -0.003732280805706978, 0.010678901337087154, -0.012535116635262966, -0.01613253727555275, -0.02457592822611332, -0.035618651658296585, 0.004133459646254778, 0.048132266849279404, 0...
[ 0.6137375831604004, 0.2171250730752945, 0.3973798453807831, 0.003063524141907692, -0.033258773386478424, 0.23851589858531952, -0.08134149014949799, 0.5436567068099976, -0.3019358813762665, -0.2079039067029953, -0.13537341356277466, 0.3535624146461487, -0.4383593201637268, 0.197539806365966...
I've attached a picture of the time series I'm talking about. The top is the original series, the bottom is the differenced series. Each data point is a 5 minute average reading from a strain gauge. This strain gauge is placed on a machine. The noisy areas correspond to areas where the machine is turned on, the clean areas are when the machine is turned off. If you look at the area circled in red, there are anomalous steps in the reading that I would like to be able to detect automatically. I'm completely stumped on how I might be able to do this - any ideas? ![enter image description here](http://i.stack.imgur.com/FbGXD.png)
[ -0.0036850073374807835, 0.009370893239974976, -0.006908251903951168, 0.008990335278213024, 0.0018795686773955822, -0.009099197573959827, 0.006460199132561684, -0.005838803946971893, -0.010528615675866604, 0.019058074802160263, -0.011200232431292534, 0.005576442927122116, 0.009948479011654854...
[ 0.5950527787208557, -0.08592084050178528, 0.4340515732765198, 0.5192471742630005, 0.18031641840934753, 0.31970933079719543, -0.2738061547279358, -0.256817102432251, -0.12618817389011383, -0.5700327157974243, 0.3757045865058899, 0.1724419891834259, -0.001129364361986518, 0.422576367855072, ...
I have a plugin that creates a custom post type. I am able to control the rewite rules and have also created a plugin settings for the archive url and taxonomy urls slugs. However if a user enter "category" for the taxonomy slug then it collides with a defauly wordpress naming convention and I get 404. I can deal with this by putting warning telling the user that they need to head on over to their permalink setting and change the "Category base" value in. Permalink Settings. I see some plugins have added what looks like custom post type fields in the Permalink Settings interface. Woo Commercefor example has "Product permalink base" section also has added some additional fields in "Optional" section of the Permalink Settings interface. How are they doing this? I have looked around and cannot fine any documentation on this.
[ -0.001640257891267538, 0.01381523534655571, 0.0008155102841556072, 0.019831592217087746, 0.0352618545293808, 0.010465936735272408, 0.008972300216555595, 0.009685400873422623, -0.012614341452717781, -0.008751560933887959, -0.012667804956436157, 0.006609364412724972, -0.011524925008416176, 0...
[ 0.6773911714553833, 0.22231312096118927, 0.193026602268219, -0.01354132778942585, -0.2490990310907364, -0.2367694079875946, 0.3981988728046417, -0.03285341337323189, -0.13257989287376404, -0.6146491765975952, 0.1489308774471283, 0.4186757802963257, -0.4081299602985382, 0.5251067280769348, ...
Assume you're building a bulletproof CMS for your company. Cost, time, and technology are no issue. You've got developers ready to go with the knowledge and experience to crank out outstanding code. Now, you have to answer the age old CMS question: what steps do you take to ensure that the code entered into the CMS by the user is acceptable for your site? I'm thinking beyond just technically escaping a string to make it able to go into a database--there are considerations for w3 standards, filtering for inappropriate content, spelling, and such conundrums as Microsoft Word code scrubbing. Certainly, there are concerns that I've missed...
[ 0.02219488099217415, 0.01352628506720066, -0.008752113208174706, 0.006178770214319229, -0.0114469388499856, 0.01925649493932724, 0.00586679857224226, -0.0021380335092544556, -0.014916591346263885, -0.03248143941164017, -0.0016313036903738976, 0.011282465420663357, -0.0019316801335662603, 0...
[ 0.4039023816585541, 0.47471868991851807, -0.13662254810333252, 0.21721552312374115, 0.2270132303237915, -0.38477659225463867, 0.20293979346752167, -0.026265228167176247, -0.10480041056871414, -0.466459721326828, -0.1572403460741043, 0.43594443798065186, -0.29852384328842163, -0.11736886203...
When designing a website that is targetting large number of consummers, what design tips would you give any designer to make the website popular and memorable amongst its users?
[ 0.007948082871735096, 0.01591246761381626, -0.00006675800977973267, 0.02025921270251274, -0.024676304310560226, -0.017317987978458405, 0.014471862465143204, 0.011066010221838951, -0.018482202664017677, -0.022859396412968636, -0.012047218158841133, 0.007105190306901932, 0.019833169877529144, ...
[ 0.7825703024864197, 0.05332629382610321, 0.006356042344123125, 0.3172852396965027, 0.27493053674697876, -0.13944582641124725, 0.09369052946567535, 0.16833777725696564, -0.4176456928253174, -0.2191399484872818, 0.21791410446166992, 0.40982896089553833, 0.04090292751789093, -0.01823459193110...
The stochastic difference equation is $$ {{y}_{t+1}}=\frac{1}{a}{{y}_{t}}-\frac{b}{a}{{x}_{t}}+{{\varsigma }_{t+1}} $$ where $ {a}>1; $ $ {{x}_{t}}=(1-\rho )\overline{x}+\rho {{x}_{t-1}}+{{\varepsilon }_{t}}, 0<\rho<1; $ $ {\varsigma }_{t+1} $ is a forecast error, that is, $ {{y}_{t+1}}={{E}_{t}}{{y}_{t+1}}+{{\varsigma }_{t+1}} $. $ {{\varsigma }_{t}} $ is serially uncorrelated, and uncorrelated with ${\varepsilon }_{t}$. $ {{E}_{t}} {{\varsigma }_{t+1}}=0 $. How to calculate the variance and the first order autocorrelation of $ {y}_{t} $? The reference answers are, $ \begin{align} & \sigma _{y}^{2}=\frac{{{b}^{2}}(a+\rho )}{({{a}^{2}}-1)(a-\rho )}\sigma _{x}^{2}+\frac{{{a}^{2}}}{({{a}^{2}}-1)}\sigma _{\varsigma }^{2} \\\ & {{\rho }_{y}}(1)=\frac{1}{a}\left[ 1+\frac{{{b}^{2}}\rho ({{a}^{2}}-1)\sigma _{x}^{2}}{{{b}^{2}}(a+\rho )\sigma _{x}^{2}+{{a}^{2}}(a-\rho )\sigma _{\varsigma }^{2}} \right] \\\ \end{align} $ My answers are different from the above ones, could someone give me a detailed derivation? Thanks a lot.
[ 0.0006610740092583001, 0.007576385047286749, -0.002051331102848053, 0.006591808050870895, 0.008269406855106354, -0.020991530269384384, 0.003582301549613476, -0.0027077901177108288, -0.006735595408827066, 0.008735287003219128, -0.005711571779102087, -0.0006763922865502536, -0.0230946950614452...
[ -0.18996818363666534, -0.1808149218559265, 0.5910333395004272, -0.13744081556797028, -0.11563259363174438, 0.44068199396133423, -0.07937999814748764, -0.4763248860836029, 0.12322388589382172, -0.040618523955345154, 0.3452880382537842, 0.5592831373214722, -0.49244460463523865, 0.08252030611...
I'm looking for Debian package of Tomee plus but I have not found anything. Also, I would like to know if it is possible to download the package( on windows) then be installed on Debian or I have to download and install directly on Debian?
[ -0.02146974578499794, -0.018297893926501274, -0.014484419487416744, 0.01828412339091301, -0.007938497699797153, 0.00006667941488558426, 0.011885206215083599, 0.018634609878063202, -0.040308695286512375, -0.03462669253349304, -0.016299724578857422, 0.017865106463432312, -0.029410777613520622,...
[ 0.6649789810180664, -0.10161685943603516, -0.298990935087204, 0.15511442720890045, -0.19427958130836487, -0.30401885509490967, 0.0806494653224945, 0.05902278795838356, -0.1384698450565338, -0.5692390203475952, 0.09246715158224106, 0.7841920256614685, -0.14056845009326935, 0.092275485396385...
I am estimating a structural model with the following data generating process: $$x=\begin{cases}\begin{array}{c}y-\theta\\\0\end{array} & \begin{array}{c}if\ y>\bar{y}(\lambda_1)\\\if\ y\leq\bar{y}(\lambda_1)\end{array}\end{cases}$$ Further, y is a random variable with distribution $F(\cdot|\lambda_2)$, and $\bar{y}(\cdot)$ is an increasing function. The parameters I want to estimate are $\theta$, $\lambda_1$, and $\lambda_2$. I thought that this was a standard Type 1 censored model, and so wrote the likelihood as follows: $$L=\prod f(x+\theta|\lambda_2)^{I(x>0))}*F(\bar{y}(\lambda_1)|\lambda_2)^{1-I(x>0)}$$ However, after estimating the model, I do not think this is correct. You can see that the parameter $\lambda_1$ only contributes to the likelihood when the data is censored (ie $x=0$). Hence, since $\bar{y}(\cdot)$ is an increasing function, the value of $\lambda_1$ that maximizes the likelihood is positive infinity, which is essentially what my estimate for $\lambda_1$ was. However, $\lambda_1$ is clearly identfied, since it controls the selection equation. I then thought that instead the likelihood should be this: $$L=\prod \left(\frac{f(x+\theta|\lambda_2)}{1-F(\bar{y}(\lambda_1)|\lambda_2)}\right) ^{I(x>0))}*F(\bar{y}(\lambda_1)|\lambda_2)^{1-I(x>0)}$$ The reasoning is pretty straight forward; when $x>0$, y is truncated from below at $\bar{y}(\lambda_1)$. However, this turns out also not to work, for the same reason as before: the likelihood is strictly increasing in the parameter $\lambda_1$. I am almost positive that $\lambda_1$ is identified; if I was simply estimating whether or not x=0, then the likelihood would be: $$L=\prod \left(1-F(\bar{y}(\lambda_1)|\lambda_2)\right)^{I(x>0))}*F(\bar{y}(\lambda_1)|\lambda_2)^{1-I(x>0)}$$ which would have a unique maximizer in $\lambda_1$. So what should the correct likelihood be? Thanks!
[ -0.009301922284066677, 0.019311444833874702, 0.0010658579412847757, 0.01288942713290453, -0.00027886126190423965, -0.016976334154605865, 0.004733298439532518, -0.006171294953674078, -0.005611073225736618, -0.0027237858157604933, -0.014082614332437515, 0.0040625520050525665, -0.01741254888474...
[ 0.09697593003511429, 0.12252280116081238, 0.3248644173145294, -0.10373303294181824, -0.05421268194913864, 0.2967927157878876, -0.18454709649085999, -0.4804369807243347, 0.11104217916727066, -0.5488287210464478, 0.10288190096616745, 0.40217486023902893, -0.2481006532907486, 0.32155776023864...
In the philosophy of science, there are three terms which are used to describe three different related notions. In both Polish (pl) and German (de), these three terms are unique such that there is no risk of equivocation. Briefly: * metoda (pl); Methode (de) an ordered sequence of actions chosen to economically and efficiently achieve a desired end * metodyka (pl); Methodik (de) a set of methods chosen for the purpose of achieving a common end or related ends (e.g. the cluster of methods used in molecular biology) * metodologia (pl); Methodologie (de) the science whose proper objects of study are the previous two In English, metoda is _method_ , while both metodyka and metodologia are _methodology_. In English, we have the word _methodic_ , but this is an adjective, and if I were to invent a word such as _methodics_ or the noun form _methodic_ , it would appear, given the Greek etymology of similar terms (concerning the -ikos ending, e.g. mathematikos, logikos, physikos) and the meanings of words with similar endings in English, that the word would suggest a field of study. Oddly enough, _methodology_ , etymologically speaking, does concern a field of study and is thus the wrong word to describe the second term above. Is anyone aware of an alternative that would work? If no such alternative exists, any ideas about what new word could be coined to better suit the intended meaning? EDIT: If anyone is able to read Polish, here is a PDF with definitions supplied by Herbut.
[ 0.011061743833124638, 0.002412125002592802, -0.008704555220901966, 0.009155932813882828, -0.008505747653543949, -0.002900324761867523, 0.009652851149439812, -0.023486651480197906, -0.011008085682988167, 0.012024875730276108, -0.008781813085079193, 0.007165238261222839, 0.006778989918529987, ...
[ 0.5158978700637817, 0.07142260670661926, -0.3594229221343994, 0.2673700153827667, -0.1295529007911682, 0.28855568170547485, 0.5543006062507629, -0.2963828444480896, 0.13263213634490967, -0.1911616176366806, -0.3048570156097412, 0.17195990681648254, -0.14101211726665497, 0.4049721956253052,...
I have a `SpatialPointsDataFrame` with some additional data. I would like to extract those points inside a polygon and at the same time, preserve `SPDF` object and its corresponding data. So far I've had little luck and resorted to matching and merging through a common ID, but this works only becaues I have gridded data with individual IDS. Here's a quick example, I'm looking for points inside the red square. library(sp) set.seed(357) pts <- data.frame(x = rnorm(100), y = rnorm(100), var1 = runif(100), var2 = sample(letters, 100, replace = TRUE)) coordinates(pts) <- ~ x + y class(pts) plot(pts) axis(1); axis(2) ply <- matrix(c(-1,-1, 1,-1, 1,1, -1,1, -1,-1), ncol = 2, byrow = TRUE) ply <- SpatialPolygons(list(Polygons(list(Polygon(ply)), ID = 1))) ply <- SpatialPolygonsDataFrame(Sr = ply, data = data.frame(polyvar = 357)) plot(ply, add = TRUE, border = "red") The most obvious approach would be to use `over`, but this returns the data from the polygon. > over(pts, ply) polyvar 1 NA 2 357 3 357 4 NA 5 357 6 357
[ -0.013053549453616142, 0.012844184413552284, -0.007103668991476297, 0.016356922686100006, 0.005493924021720886, 0.019940922036767006, 0.006527361460030079, 0.016983674839138985, -0.014648087322711945, 0.014356203377246857, 0.004064091481268406, 0.014548632316291332, -0.013123026117682457, ...
[ 0.13796934485435486, 0.03919099643826485, 0.7563831806182861, 0.11891557276248932, -0.18932229280471802, 0.2974480390548706, -0.13458487391471863, -0.6438376903533936, -0.23862408101558685, -0.8868675827980042, 0.30918243527412415, -0.0023842831142246723, 0.022208936512470245, 0.2923593223...
How do I make a WP page (w/ dynamic content) auto-refresh every minute? With PHP, I'd use a "Refresh:" header, and in HTML I could use a HTTP-EQUIV tag or something. However, WP generates headers itself, and I can't tweak anything until the body, no? I searched in vain for a plug-in that does this.
[ -0.002384474501013756, 0.00701840128749609, -0.009390444494783878, 0.016598764806985855, -0.007007965352386236, -0.02310311235487461, 0.009439618326723576, 0.010841245763003826, -0.025409545749425888, -0.003522058017551899, -0.011869420297443867, 0.004534948151558638, -0.001715264399535954, ...
[ 0.5851987600326538, -0.16856975853443146, 0.4249337911605835, -0.0006814193329773843, -0.3860906958580017, -0.09087881445884705, 0.3956424295902252, -0.284457266330719, -0.2526380121707916, -0.5015624165534973, 0.09855309873819351, 0.3763197660446167, -0.14328444004058838, 0.03574845939874...
I'm looking for a list of natural phenomenon which inspired science and technology like how bat inspired the radar concept....Any idea ? Thx for helping, Stephane
[ 0.004762687720358372, 0.011771184392273426, 0.010929657146334648, 0.03581269457936287, 0.006765297614037991, -0.013546807691454887, 0.01346714049577713, 0.04477493464946747, -0.040340062230825424, 0.013331516645848751, -0.0023136218078434467, 0.025516659021377563, -0.011008741334080696, 0....
[ 0.7847166657447815, 0.03170783817768097, 0.1252581924200058, 0.6983951926231384, -0.2738499045372009, -0.5354659557342529, 0.2907198965549469, 0.8778353333473206, -0.17249295115470886, -0.29478368163108826, 0.08760890364646912, 0.3122744858264923, 0.24022288620471954, 0.11680713295936584, ...
I have to draw a border around my tables, so I'm using `\fbox{}` (is there a better way?). It works when using a `threeparttable` but with the normal `tabular` latex throws an error stating Missing \endgroup inserted I have closed the curly brackets. What am I missing? I've found the post booktabs and tiny horizontal space where it seems to work. See the code below: \begin{table}[h] \fbox{ \centering\caption{\textbf{blabla 1}} \begin{tabular*}{0.9\columnwidth}{@{\extracolsep{\fill}}lc} \toprule \textbf{SF-36 Parameter} & \textbf{Mittelwert (SD)}\tabularnewline \midrule \midrule Körperliche Funktionsfähigkeit & 56,5 (33,6)\tabularnewline \midrule Psychisches Wohlbefinden & 70,7 (17,8)\tabularnewline \bottomrule \end{tabular*} } \end{table}
[ 0.01885177195072174, -0.0006897915154695511, -0.0089540621265769, 0.026457952335476875, 0.011216488666832447, 0.014465097337961197, 0.007521852385252714, 0.015313559211790562, -0.016185902059078217, 0.0022486583329737186, -0.021739330142736435, 0.0026880046352744102, -0.010542097501456738, ...
[ 0.117938332259655, 0.22089451551437378, 0.2965109348297119, -0.4068373441696167, 0.0449417307972908, -0.08754060417413712, 0.24204568564891815, -0.08066051453351974, -0.4692918658256531, -0.438821405172348, 0.07688101381063461, 0.22191576659679413, -0.33374691009521484, 0.05702416226267814...
I am following this guide to become Thane of all the holds in Skyrim. The Jarl Siddgeir does not ask me for a drink, I completed Dengeir's quests. I completed the all the quests Dengeir gave me plus most of the quests for the people of Falkreath. I do not have any more dialogue options from either the Jarl or Dengeir. Is this bugged? I have not completed much of the main story line quests or the civil war quests. Is it possible to become Thane of all the holds without completing the main quests or the civil war quests?
[ 0.02255829982459545, 0.010274030268192291, -0.0038938450161367655, -0.015283970162272453, -0.026822861284017563, -0.014517008326947689, 0.00940030999481678, -0.003366617253050208, -0.01864473521709442, -0.0071043092757463455, -0.021697696298360825, 0.01645423099398613, -0.023960258811712265,...
[ 0.37822672724723816, 0.08805403113365173, 0.454787015914917, 0.04699740558862686, -0.3024590015411377, -0.29318785667419434, 0.2338908314704895, 0.12873628735542297, -0.36284834146499634, -0.2454882711172104, 0.0019191630417481065, 0.1526736319065094, -0.16170856356620789, 0.43238669633865...
I'm trying to create an English document using `pdflatex` and want to insert a few cyrillic and arabic glyphs. something like the letter Я is cyrillic, whereas ش is arabic I tried to follow advice found Inserting some Arabic script in an English article, but it seems that `xetex` and `pdflatex` are mutually exclusive. my feeble attempts until now look like: \documentclass[letterpaper, 12pt]{article} \usepackage[utf8]{inputenc} \usepackage[T2A,LAE]{fontenc} \usepackage[russian,USenglish,arabic]{babel} \begin{document} the letter Я is Cyrillic, whereas ﺵ is arabic \end{document} any hints how to achieve this?
[ -0.006664206273853779, 0.029130499809980392, -0.010580078698694706, 0.035345714539289474, -0.014473752118647099, 0.014636158011853695, 0.012411670759320259, 0.005233696196228266, -0.02067285031080246, -0.04953564330935478, -0.01717243529856205, -0.009099066257476807, -0.02804867923259735, ...
[ 0.3394773006439209, 0.4289441704750061, 0.10240523517131805, 0.056606389582157135, -0.12267244607210159, 0.175923153758049, 0.34829139709472656, 0.11747438460588455, 0.10864239931106567, -0.5773170590400696, -0.1361067146062851, 0.20695063471794128, -0.2356531322002411, 0.09041766077280045...
In QGIS, whatever might be the source, I couldn't zoom beyond the scale of 1:2256. Whether it's a raster or vector, result is the same. Does anyone know how to fix this ? Edit: Thanks to Chris. I've uninstalled "Tile Map Scale Plugin" which fixed the problem (at least for me).
[ -0.008378323167562485, -0.01045234501361847, -0.004424580838531256, 0.006855466403067112, 0.004707080777734518, 0.018663259223103523, 0.007205838803201914, 0.0007964480319060385, -0.021354807540774345, -0.021382616832852364, 0.0015389756299555302, 0.012684402987360954, -0.027569247409701347,...
[ 0.17594069242477417, -0.15499944984912872, 0.2969307601451874, 0.30277520418167114, -0.20146699249744415, -0.15650808811187744, 0.4135320782661438, 0.3997906744480133, -0.27006790041923523, -0.917988121509552, -0.05635365843772888, 0.44261038303375244, 0.20606783032417297, -0.2050844728946...
I'd like to produce a document comparing all of the different bb (blackboard bold) fonts available in LaTeX. symbols-a4.pdf gives is what I will assume is a comprehensive list: ![enter image description here](http://i.stack.imgur.com/6Fq1e.png) The problem is that three different packages override the `\mathbb` command. I tried to 'save' each version as shown in the following MWE, but to no avail: \documentclass{article} \usepackage{bbold} \let\bboldbb\mathbb \usepackage{amsfonts} \let\amsbb\mathbb \begin{document} $\amsbb{N} \bboldbb{N}$ \end{document} Using `\def` in place of `\let` doesn't work either. How can I use multiple versions of `\mathbb` at once? (And what is wrong with my MWE?)
[ 0.00635154452174902, -0.007171346805989742, -0.01804562285542488, 0.015115847811102867, 0.004696249030530453, 0.007005807012319565, 0.007899053394794464, 0.012416507117450237, -0.01896529085934162, -0.0120590440928936, -0.016206884756684303, 0.0013518678024411201, 0.0034223953261971474, 0....
[ 0.39064833521842957, 0.1717202514410019, 0.3674685060977936, -0.2548396587371826, 0.14979031682014465, 0.05576583370566368, 0.20729109644889832, -0.4225105047225952, -0.3379828929901123, -0.6343313455581665, 0.17823657393455505, 0.8744348883628845, -0.43477725982666016, 0.07689425349235535...
I am working of a simple redirect plugin that match the given slug and redirect to the assigned url, which i supposed to working fine. But recently i am adding the functionality to update the stat for the visited slug and i found that my update query called twice, i checked and found that `template_redirect` hook called twice. My stat update query is applied on single time only. Here is my code: add_action("template_redirect", array($this, "redirection_section"), 11); function redirection_section() { if( is_404() ) $this->uni_page_redirect(); } function uni_page_redirect() { include_once dirname(__FILE__) . '/uni_links.php'; $Links = new UNI_Links(); $slug = ltrim(preg_replace('#/$#', '', urldecode($_SERVER['REQUEST_URI'])), '/'); if( $Links->is_valid_slug($slug) ) { // Initiate with Slug do_action("ulr_init_slug", $slug); $Links->generate_links($slug); exit(); } return false; } Now this is how i am updating my query: function slug_init_callback($slug_name) { global $wpdb; $table_slugs = ULR_TABLE_SLUGS; // Debug to check invoks file_put_contents(PLUGIN_BASE . '/debug.txt', "call \r\n", FILE_APPEND); $sql = $wpdb->prepare("update {$table_slugs} set stat= stat+1 where slug_name=%s", $slug_name); $wpdb->query($sql); } Your help is highly appreciated. Thanks
[ 0.00645638071000576, 0.020979810506105423, 0.004712388850748539, 0.020418815314769745, -0.004939175210893154, 0.0028991650324314833, 0.01043031644076109, -0.00624080840498209, -0.017677364870905876, -0.021226633340120316, -0.018305007368326187, 0.015346016734838486, -0.012569993734359741, ...
[ 0.461052805185318, -0.0725494995713234, 0.603351354598999, -0.21953150629997253, -0.22323156893253326, 0.31844109296798706, -0.006368862930685282, -0.3543950021266937, -0.319530189037323, -0.7326540946960449, 0.4098871946334839, 0.5734081864356995, -0.3972737789154053, -0.09761983156204224...
> **Possible Duplicate:** > Incorrect bookmarks and page number in table of contents I've created a table of contents with links to each chapter and section. I use this package \usepackage[colorlinks=true,urlcolor=red,citecolor=blue,linkcolor=blue]{hyperref} and I use this for bibliography \addcontentsline{toc}{chapter}{\bibname} \begin{thebibliography}{99} \bibitem{} bla bla bla \end{thebibliography} If I click on the item in table of contents, the bibliography's link is wrong: it is linked to the previous page. How I can fix this problem?
[ 0.016611266881227493, 0.010116933844983578, -0.00998331606388092, 0.01970953494310379, 0.014646166935563087, 0.008175451308488846, 0.007804269436746836, -0.014990872703492641, -0.018524739891290665, -0.02664746344089508, -0.007394698448479176, -0.008238324895501137, -0.011117691174149513, ...
[ 0.03425748273730278, 0.3296501040458679, 0.6304087042808533, 0.15913252532482147, -0.0618453249335289, 0.013547523878514767, 0.2518925368785858, 0.09491067379713058, -0.5545424818992615, -0.7403972744941711, -0.15711019933223724, -0.007849611341953278, -0.6055729389190674, 0.22280482947826...
I am running a rare events logistic regression on a binary dependent variable. I have 538 observations and only 10 events (so 528 values of 0 and 10 of 1), which is why I chose to use a rare events logistic regression. When I run the regression, one of the independent variables in the model has a huge coefficient (around 25,000,000) and is found to be significant. The range on the independent variable is 0 to 1. Is this a problem? Could anyone explain why this is happening? When I run the same model with just a logistic regression this variable is insignificant. I'm not sure what is happening. Any advice would be appreciated.
[ 0.033495642244815826, 0.015659503638744354, -0.024538762867450714, -0.0013352155219763517, -0.03233326971530914, -0.01902313157916069, 0.00987221673130989, 0.01643691584467888, -0.0069397687911987305, -0.014876731671392918, -0.00674480851739645, 0.015418164432048798, -0.0010832531843334436, ...
[ 0.20105886459350586, 0.19658827781677246, 0.05991838499903679, 0.10151106119155884, -0.14109379053115845, 0.48737427592277527, 0.618830680847168, -0.003445201786234975, -0.3010856807231903, -0.5917074084281921, 0.29890894889831543, 0.2656606137752533, -0.2766363024711609, 0.558701336383819...
I'm currently studying an article by Maslov, in particular the first section about higher corrections to Fermi-liquid behavior of interacting electron systems. Unfortunately, I've hit a snag when trying to understand an argument concerning the (retarded) self-energy $\Sigma^R(ε,k)$. Maslov states that in a Fermi liquid, the real part and the imaginary part of the self-energy $\Sigma^R(ε,k)$ are given by $$ \mathop{\text{Re}}\Sigma^R(ε,k) = -Aε + B\xi_k + \dots $$ $$ -\mathop{\text{Im}}\Sigma^R(ε,k) = C(ε^2 + \pi^2T^2) + \dots $$ (equations 2.4a and 2.4b). These equations seem reasonable: when plugged into the fermion propagator, $$ G^R(ε,k) = \frac1{ε + i\delta - \xi_k - \Sigma^R(ε,k)} $$ the real part slightly modifies the dispersion relation $ε = \xi_k$ slightly and the imaginary part slightly broadens the peak. That's what I'd call a Fermi liquid: the bare electron peaks are smeared out a bit, but everything else stays as usual. Now, Maslov goes on to derive higher-order corrections to the imaginary part of the self-energy, for instance of the form $$ \mathop{\text{Im}}\Sigma^R(ε) = Cε^2 + D|ε|^3 + \dots .$$ First, I do not quite understand how to interpret this expansion. > How am I to understand the expansions in orders of $ε$? I suppose that $ε$ > is small, but in relation to what? The Fermi level seems to be given by > $ε=0$. Second, he states that this expansion is to be understood "on the mass-shell". > I take it that "on the mass shell" means to set $\xi_k=ε$? But what does the > expansion mean, then? Maybe I am supposed to expand in orders of > $(ε-\xi_k)$? Now the question that is the most important to me. Maslov argues that the real part of the self-energy can be obtained via the Kramers-Kronig relation from the imaginary part of self-energy. My problem is that the corresponding integrals diverge. > How can $$ \mathop{\text{Re}}\Sigma^R(ε,k) = > \mathcal{P}\frac1{\pi}\int_{-\infty}^{\infty} d\omega > \frac{\mathop{\text{Im}}\Sigma^R(\omega,k)}{\omega-ε} $$ be understood for > non-integrable functions like $\mathop{\text{Im}}\Sigma^R(ε,k) = ε^2$? It probably has to do with $ε$ being small, but I don't really understand what is going on. * * * I should probably mention my motivation for these questions: I have calculated the imaginary part of the self-energy for the one-dimensional Luttinger liquid $\xi_k=|k|$ as $$ \mathop{\text{Im}}\Sigma^R(ε,k) = (|ε|-|k|)θ(|ε|-|k|)\mathop{\text{sgn}}(ε) $$ and would like to make the connection to Maslov's interpretation and results. In particular, I want to calculate the imaginary part of the self-energy with the Kramers-Kronig relations.
[ 0.004175481386482716, 0.007007826119661331, 0.002411337336525321, 0.01578628085553646, -0.005716189742088318, -0.03236323967576027, 0.008971326984465122, -0.006863263435661793, -0.017984915524721146, -0.009950670413672924, -0.013219515793025494, 0.012341402471065521, -0.013303825631737709, ...
[ 0.24594120681285858, 0.17486940324306488, 0.14220795035362244, -0.19376039505004883, -0.0512283630669117, 0.2858540117740631, 0.22781619429588318, -0.6837000250816345, 0.17926651239395142, -0.3526521623134613, -0.03746618330478668, 0.5061447024345398, -0.393282026052475, 0.9199649095535278...
It was wondering: when installing something, there's an easy way of double clicking an install executable file, and on the other hand, there is a way of building it from source. The latter one, downloading a source bundle, is really cumbersome. But what is the fundamental difference between these two methods?
[ -0.020456723868846893, 0.0074086724780499935, -0.0019658247474581003, -0.002234565792605281, 0.007609704043716192, -0.01820414513349533, 0.011781466193497181, -0.008867166005074978, -0.019932866096496582, -0.01631416566669941, -0.021260002627968788, -0.006829883903264999, 0.01547801122069358...
[ 0.6084715127944946, -0.09952667355537415, 0.15258310735225677, 0.3782540559768677, 0.017645902931690216, -0.10416761040687561, -0.019525568932294846, 0.08665687590837479, -0.24884825944900513, -0.33765238523483276, -0.13766682147979736, 0.581673264503479, 0.12948846817016602, 0.15496850013...
I need to classify a list of events: each events is either A or NOT A As it turns out, if the nth event is A, then the (n+1)th event is more likely to be A. How would I go about adding this probabilistic component into a supervised learning model? Thanks!
[ -0.02166799083352089, 0.03123944066464901, -0.026786046102643013, 0.016043704003095627, -0.004118624143302441, -0.015242455527186394, 0.01158431638032198, 0.031078210100531578, -0.025703992694616318, 0.020940842106938362, -0.021515168249607086, 0.0042144921608269215, -0.024206949397921562, ...
[ -0.03840279206633568, -0.028353864327073097, -0.2319195717573166, 0.15006956458091736, -0.3031260371208191, -0.15787741541862488, 0.29488587379455566, -0.13794498145580292, -0.39382338523864746, -0.32616713643074036, 0.0026337397284805775, 0.4074881076812744, -0.12311708182096481, 0.298088...
When I am questing and/or exploring in diablo 3, I sometimes miss the _New Lore_ popup in the bottom right corner of my screen. After a short period of time, it disappears. Can I read or listen to this lore or is it lost till my next play through ?
[ -0.0005155188264325261, 0.028382685035467148, 0.00025449972599744797, -0.02759752795100212, 0.03538643941283226, 0.0023188763298094273, 0.009024711325764656, -0.010364620946347713, -0.018643351271748543, 0.051520802080631256, -0.021773047745227814, 0.009941421449184418, -0.003721113549545407...
[ 0.3789178133010864, -0.07862480729818344, 0.3976987600326538, 0.08749546855688095, -0.21544963121414185, -0.48155930638313293, 0.7048229575157166, 0.3657630681991577, -0.19357368350028992, -0.5993700623512268, -0.007736047264188528, 0.4640466272830963, 0.44723770022392273, 0.16015039384365...
Could anyone give me some information on who invented profile maximum likelihood estimation or who first use profile maximum likelihood estimation and the short history of profile maximum likelihood estimation? I would like to know about any paper on this.
[ 0.06918487697839737, 0.009538456797599792, -0.008310625329613686, 0.006168318446725607, 0.01136523112654686, 0.00572527851909399, 0.014952381141483784, -0.019167225807905197, -0.04271665960550308, 0.01831379532814026, 0.018662292510271072, 0.034690774977207184, -0.016189802438020706, -0.00...
[ 0.6522294878959656, 0.04976234585046768, -0.004544100258499384, 0.38046857714653015, -0.09270323067903519, -0.2978774309158325, -0.13760972023010254, 0.569567859172821, 0.05349785462021828, -0.4379737079143524, 0.3965100049972534, 0.2156813144683838, 0.3180112838745117, 0.49876752495765686...
When using a construction like the following, the height of the block changes between frames. This change distracts the audience from the content. How to fix the height of the block, so that it doesn't "flicker"? \documentclass{beamer} \title{MWE of flickering boxes} \author{\textsc{Jon Doe}} \begin{document} \frame{\titlepage} \section{Flickering boxheight} \begin{frame}\frametitle{The upper box flickers} \begin{block}{\only<1-2>{Examples:} \only<3>{Pausetext:}\only<4>{Examples again:}}<1-4> \only<1-2,4>{\begin{itemize} \item<1-2,4> Example 1 \item<2-,4> Example 2 \end{itemize}} \only<3>{Oneliner} \end{block} \begin{block}{Another Box}<3-> Your Adverrtisement could be here \end{block} \end{frame} \end{document}
[ -0.01384043414145708, -0.000657881610095501, 0.005340282339602709, 0.020462986081838608, 0.01387024112045765, -0.0024505876936018467, 0.007976260036230087, 0.0218522772192955, -0.012232326902449131, 0.013144422322511673, -0.019094085320830345, 0.013283955864608288, 0.006367402151226997, 0....
[ 0.5534838438034058, 0.02871697023510933, 0.3350878357887268, 0.14541879296302795, 0.001421682769432664, -0.1114397644996643, 0.5131693482398987, -0.5126692652702332, -0.2001369744539261, -0.4708840847015381, -0.056403107941150665, 0.15760438144207, -0.26137685775756836, 0.04729859530925751...
If I know the current post id that i have in the variable $pid I use $terms = get_the_terms($pid, 'custom_category'); How do I get just the term id/term_taxonomy_id if I var_dump $terms I see what I want...but I have no idea how the heck to return just the id, not an array, just the id. Bare in mind...I less than 1/2 know what I'm doing...just face rolling keyboard to get what I want...but learning more everyday. Cheers, Brian
[ 0.012617519125342369, 0.0033266437239944935, -0.010766427963972092, 0.00752711296081543, 0.0030832563061267138, 0.005419133231043816, 0.007523390464484692, 0.019728563725948334, -0.02003178372979164, -0.028630485758185387, -0.009827550500631332, 0.002851588884368539, 0.010363609530031681, ...
[ 0.15174725651741028, 0.12658564746379852, 0.6146885752677917, -0.07825576514005661, 0.03870471194386482, 0.07287003844976425, 0.2801404893398285, -0.3753666281700134, -0.0629936158657074, -0.6580711007118225, 0.09389188885688782, 0.6263290047645569, -0.03111308068037033, 0.2412049770355224...
In field theory, the energy momentum defined as the functional derivative wrt the metric $T_{\mu\nu}=\frac{2}{\sqrt{-g}}\frac{\delta S}{\delta g^{\mu\nu}}$ (up to a sign depending on conventions)For a theory in flat space, this has the advantage to give you directly an improved energy momentum tensor as the metric is symmetric, but it hold also for dynamical metrics. Now, my problem is the following: If you have in your Lagrangian a term like $A^{\mu\nu}\nabla_\mu V^\mu$, where $V^\mu$ is a vector field (not needed to be a gauge field, for instance say it represents the velocity in hydrodynamics) and $A^{\mu\nu}$ an arbitrary tensor a priori depending on anything (metric, $V$, or any other field), you will have a Christoffel symbol. If you take a functional derivative, the expression will contain terms like $\frac{\delta \Gamma^\rho_{\mu\nu}(x)}{\delta g^{\alpha\beta}(y)}\supset g^{\rho\sigma} \partial_\mu(g_{\nu\alpha}g_{\sigma\beta}\delta(x-y))$+permutations those terms will not combine again to form a christoffel symbol. How can one take the functional derivative of something covariant ($\nabla V$) with respect to something covariant (the metric) and end up with something not covariant? For flat space cases, everything is well since those terms will vanish, but for curved space, we loose general covariance/diff invariance. What am I missing here?
[ 0.012712766416370869, -0.00027042930014431477, -0.0029597945977002382, 0.001766827655956149, -0.0044467817060649395, -0.011246094480156898, 0.007736783474683762, -0.01633797213435173, -0.0083343256264925, -0.0483601912856102, -0.005466570612043142, 0.018143940716981888, -0.001987095456570387...
[ 0.037009820342063904, -0.6388726234436035, 0.41652706265449524, -0.04693673551082611, -0.26316362619400024, -0.21157380938529968, -0.552539050579071, -0.40711304545402527, -0.1674625128507614, -0.3883526921272278, -0.42039310932159424, 0.879386305809021, -0.26231813430786133, 0.46901875734...
**Update** Experimenting with descriptive statistics, I determined that if rank the scaled weighted average quantity for each set, computed as follows: $$Mean(Weighted Quantity) * \frac {Median ( Weighted Quantity)} {Range(Weighted Quantity)}$$ the results match my intuition. Although I've solved my immediate problem, I'd still like an explanation for what I'm doing in more formal terms. I'd like to start learning more about these types of problems. * * * **Update 2** _**Context, per request:**_ The data are inputs into a parametric (aka variance-covariance) interest rate value at risk (IR VaR) computation _(the associated correlation matrices are not considered here)_ * Each matrix is per security holding. * Each row is per key rate tenor _(i.e. 3 months, 1 year, 5 years, etc.)_ The number of tenors varies. For each tenor _(row)_ : * f is dv01 - or the $ change in market value of the holding corresponding to a 1bp shift in the yield curve at the row's tenor * v is the spot rate at that tenor, interpolated from the yield curve * q is the volatility of the the spot rate at that tenor The scaled weighted average volatility is a measure of, or proxy for, the degree of non-linearity of the security's pricing model. It accurately ranks holdings (based on test data) in order of the percentage variance of the holding's parametric IR VaR result from its Monte Carlo IR VaR result. * * * I wish to rank sets of data by some form of average quantity as weighted by a factor adjusted by a value. For example, consider the two sets of data: > **Set One:** \begin{matrix} f[1]_1 & v[1]_1 & q[1]_1\cr f[2]_1 & v[2]_1 & > q[2]_1\cr f[3]_1 & v[3]_1 & q[3]_1 \end{matrix} and > **Set Two:** \begin{matrix} f[1]_2 & v[1]_2 & q[1]_2\cr f[2]_2 & v[2]_2 & > q[2]_2\cr f[3]_2 & v[3]_2 & q[3]_2\cr f[4]_2 & v[4]_2 & q[4]_2\cr f[5]_2 & > v[5]_2 & q[5]_2 \end{matrix} My initial thought was to rank them by weighted arithmetic mean of quantity, _i.e._ $$\frac {\sum (f * v * q)}{\sum(f * v)}$$ But the results don't fit my intuition _(which may be faulty.)_ How can I tell whether I need to normalize before ranking? Which normalization methods should I consider?
[ -0.008595413528382778, 0.02866288088262081, -0.005727587267756462, 0.008303866721689701, -0.019088558852672577, -0.005367209203541279, 0.00570192513987422, -0.023170575499534607, -0.012076223269104958, 0.008574754931032658, -0.0032629931811243296, 0.01558318454772234, -0.004782833158969879, ...
[ 0.1831955909729004, -0.0009230849682353437, 0.5503920912742615, -0.08614737540483475, -0.034765902906656265, 0.12077755481004715, 0.256119042634964, -0.17534814774990082, 0.0002956483804155141, -0.5939739942550659, -0.0490308478474617, 0.5771278738975525, -0.15817494690418243, 0.0482898429...