text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
For Example I have simple linestring 'LINESTRING(25 50, 100 125, 150 190)' and I need to divide it to segments of length 10. In postgis i know there is a function ST_Segmentize but what about MS SQL Server?
[ 0.014145512133836746, 0.011429309844970703, -0.011115035973489285, 0.02156684175133705, 0.02928386814892292, 0.029832573607563972, 0.015300101600587368, 0.03718085214495659, -0.018204210326075554, -0.014673972502350807, -0.0038526663556694984, 0.009803170338273048, -0.016527311876416206, 0...
[ 0.17408441007137299, -0.08727732300758362, 0.41130882501602173, 0.08062633872032166, -0.1738484799861908, -0.06837314367294312, 0.012888947501778603, -0.27272728085517883, -0.13350681960582733, -0.5758250951766968, 0.2954372763633728, 0.12658128142356873, -0.09192921966314316, -0.009122721...
**Background on Two-Stage Designs** In clinical trials, we are often interested in the response rate $p$ for an experimental treatment. In a typical trial, we might expose $n$ patients to the treatment and observe the total number of responses $X$ in order to test the null hypothesis $H_0: p \leq p_0$ against a one-sided alternative $H_1: p > p_0$, where $p_0$ could be the response rate for a standard-of-care treatment. In order to limit patient exposure to an ineffective treatment, two-stage designs are popular: in the first stage, $n_1$ patients receive treatment and the number of responses $X_1$ is observed. If there are $r_1$ or fewer responses, the trial ends. If $X_1 > r_1$, a second stage is carried out: additional patients are enrolled until a total of $n$ patients have received treatment, and the total number of responses $X$ is observed. If there are more than $r$ total responses, the null hypothesis is rejected; otherwise, the null is not rejected and the new treatment is considered ineffective (or at least not superior to the standard of care). The parameters $n_1, r_1, n,$ and $r$ are chosen by the investigator before carrying out the study. **The Problem** While working on designing such a study, I have come across a phenomenon that I would like to better understand. Intuitively it seems that increasing the cut-off $r_1$ should decrease the type-1 error of the test. One might reason that for larger values of $r_1$, it is less likely that the study will proceed to the second stage, and thus less likely that one will ultimately reject the null. However, it appears that this is not always the case. For example, with the parameters $n_1 = 20, n = 30, r = 25,$ and $p_0 = 0.6$, I found the type-1 error to be the same (up to 9 digits) for $r = 10$ as for $r = 15$ (I got $\alpha = 0.001510074$). In fact, I get the same number for all values of $r_1 \leq 15$. See my code below. **My Question** So my question is, does it make sense that changing the value of $r_1$ does not affect the type-1 error in some cases? Can you help me refine my intuition to understand why this is true? **Code** Here is a quick function I have written to compute the type-1 error (as well as the power, for a given alternative, and the expected sample size) for two- stage designs. ##### Function to compute the type-1 error, power, and expected sample size for a ##### two-stage design. The parameters p0 and p1 are the response rates under the ##### null and alternative, respectively. twoStage = function(n, n1, r1, r, p0, p1){ # Type 1 error = P(reject H0 | p = p0) # = P(X1 > r1 && X > r | p = p0)] n2 = n - n1 x1 = (r1 + 1):n1 alpha = sum(dbinom(x1, n1, p0)*(1 - pbinom(r-x1, n2, p0))) # Power = P(reject H0 | p = p1) # = P(X1 > r1 && X > r | p = p1)] # Same as above, but with p = p1 power = sum(dbinom(x1, n1, p1)*(1 - pbinom(r-x1, n2, p1))) # Expected number of subjects enrolled under H0 expected.n = n1 + (1 - pbinom(r1, n1, p0))*n2 return(cbind(n = n, n1 = n1, r1 = r1, r = r, p0 = p0, p1 = p1, alpha = alpha, power = power, expected.n = expected.n)) } ##### Example: > twoStage(n=30, n1=20, r1=5, r=25, p0=0.6, p1=0.85)[,"alpha"] alpha 0.001510074 > twoStage(n=30, n1=20, r1=10, r=25, p0=0.6, p1=0.85)[,"alpha"] alpha 0.001510074
[ 0.02361169084906578, 0.007748657837510109, -0.00657112430781126, 0.009871428832411766, 0.005625096149742603, 0.005444203037768602, 0.008917206898331642, -0.015840886160731316, -0.013312401250004768, -0.015030820854008198, -0.009494992904365063, -0.0021641438361257315, -0.017201034352183342, ...
[ 0.45472654700279236, 0.02469608001410961, -0.20717711746692657, -0.017443649470806122, 0.15683989226818085, 0.5960353016853333, 0.14426028728485107, -0.3268541693687439, 0.0864245593547821, -0.7059880495071411, -0.05704912543296814, 0.42312172055244446, -0.22632184624671936, 0.152731642127...
> **Possible Duplicate:** > How can I manually install a package on MiKTeX (Windows) I am writing this because I need a more efficient way to use the `acmsmall` class. The problem is when I add the `.cls` and `.bst` files within my folder everything works just fine, but when I remove them from the folder the message `acmsmall.cls not found` appears. Is there a way to add the class into my MiKTeX or TeXnicCenter?
[ 0.017968211323022842, -0.014685573987662792, -0.012110594660043716, 0.016702445223927498, 0.025507275015115738, 0.012420233339071274, 0.008006650023162365, -0.007405548356473446, -0.016707556322216988, 0.005615385714918375, -0.01828419417142868, 0.007401837036013603, -0.01387306209653616, ...
[ 0.24032726883888245, 0.24462227523326874, 0.3025871515274048, 0.12020928412675858, 0.19355595111846924, -0.09215813875198364, 0.49289965629577637, 0.3362521529197693, -0.13073323667049408, -0.9059298038482666, 0.15412719547748566, 0.804783046245575, -0.417768269777298, 0.11130693554878235,...
I'm wondering what the arguments for/against Falsy values are. On what principles would you decide to add or exclude them from a language? Are there any problems you could see them causing off-hand? For users of languages that support Falsy values: * Where specifically have you used them to your advantage? * Where have you had unpleasant run-ins with them? * Are there any rules or finer points in your language/project/team about where it's appropriate or inappropriate to use them? For users of other languages: * Have you ever seen a situation where you've thought "I wish I could use a Falsy value here"? I'm tagging the question `haskell` and `python` because AFAIK those two represent opposite ends of the spectrum (Haskell demanding `Bool`s when you use `if`, and Python treating `None` and some "empty" values as Falsy), but feel free to talk about your experience from other languages. Just mention where on the spectrum they stand.
[ 0.008665340021252632, 0.031768567860126495, -0.01066146232187748, 0.028371695429086685, -0.003290147054940462, 0.007959110662341118, 0.005673435982316732, 0.024290362372994423, -0.018747758120298386, -0.030567754060029984, -0.011523582972586155, 0.008571562357246876, -0.03683020919561386, ...
[ 0.22137312591075897, 0.514866292476654, -0.1557728797197342, -0.11439137905836105, -0.1691698431968689, 0.15285301208496094, 0.396345853805542, 0.17589716613292694, -0.32368335127830505, -0.2714364230632782, -0.08139560371637344, 0.36439424753189087, -0.1521541029214859, 0.0161690637469291...
I'm looking for some theme or plugin examples that showcase ways in which to enhance the usability of editing shortcodes and/or custom html objects that have been inserted into the editor. I know that its possible to apply runtime styles to elements (nextpage and image rollovers are a couple examples) and I'm looking for some examples of scripts that seek to extend this to other elements (custom classed divs for example). I've never seen an example of shortcode editing that went beyond the basic bracket+properties+content. Ideally, I'd like to see an example where someone has abstracted shortcodes into their html equivalent, with rollover and right click events to modify element attrributes. Just looking for any examples of work in this field.
[ 0.004149602260440588, 0.004855740815401077, -0.003362539689987898, 0.017030514776706696, -0.009547008201479912, -0.0008392654126510024, 0.007818955928087234, 0.025015391409397125, -0.018656300380825996, 0.01613088883459568, -0.01307473611086607, 0.004022349137812853, 0.0008682481711730361, ...
[ 0.5880585312843323, 0.24180017411708832, -0.16224028170108795, 0.031097112223505974, -0.3803097903728485, -0.20370858907699585, -0.09791299700737, 0.09406574070453644, -0.26542219519615173, -0.7068780660629272, 0.09011375904083252, 0.31407254934310913, -0.14629021286964417, 0.0796375051140...
We're in the middle of a historical time. Two creatures will be separated from each other. Waffles will be torn in two. Meta Stack Overflow will be split. This banner is currently being shown on Meta Stack Overflow and on Stack Overflow itself: ![enter image description here](http://i.stack.imgur.com/wOfr5.png) > The MSO/MSE Split is soon underway. Please bear with us while this is > completed. The first sentence caught my eye, specifically the last few words: _"is soon underway"_. I'm not sure this is grammatically correct - my internal "something-isn't-right" meter doesn't like it. If I had to guess I'd say it should read _"is underway soon"_ , but that doesn't sound quite right either. I can't put a finger on what, exactly, is wrong with the sentence, nor what should be done to fix it. Can anyone explain this to me?
[ -0.014142699539661407, 0.006613859441131353, 0.015449315309524536, 0.020305819809436798, 0.022867154330015182, -0.0005959707777947187, 0.0055018458515405655, 0.008033670485019684, -0.010117512196302414, 0.021259650588035583, -0.010165420360863209, 0.02026764303445816, 0.0030998014844954014, ...
[ 0.6103634238243103, -0.22354145348072052, 0.41462114453315735, 0.13433583080768585, 0.26165348291397095, -0.043029025197029114, 0.3177322745323181, 0.08123494684696198, -0.23281274735927582, -0.6933648586273193, -0.225864976644516, -0.17165182530879974, -0.05879946053028107, 0.308029413223...
I am trying to make the margin narrower and use the command `\addtolength{\hoffset}{-0.5cm} and \addtolength{\voffset}{-0.5cm}` I have used the package `calc`, but there is still an error message saying 'Please type another file name for output' I have tried putting the addtolength command in the preamble and after the `\begin{document}` but none of them work. I am wondering how can I fix this? So here is my code: \documentclass[10pt,a4paper]{article} \usepackage{amsmath} \usepackage{amssymb} \usepackage{calc} \title{AAA} \addtolength{\hoffset}{-0.5cm} \addtolength{\voffset}{-0.5cm} \begin{document} \maketitle some text \end{document} Helps are really appreciated. Thanks!
[ 0.0015444508753716946, 0.0005309592233970761, -0.013146087527275085, 0.022930730134248734, -0.015696264803409576, 0.017421267926692963, 0.00901571661233902, -0.009414320811629295, -0.013887908309698105, -0.01574072800576687, -0.005170802585780621, 0.00686626648530364, -0.014951109886169434, ...
[ 0.44612935185432434, 0.3427655100822449, 0.6282408833503723, -0.22205476462841034, -0.05822909250855446, -0.008365524932742119, 0.1651352494955063, -0.1275147944688797, 0.09025894850492477, -0.6505733132362366, 0.35187575221061707, 0.5446279644966125, -0.2720045745372772, 0.137950986623764...
Is it possible to unpack files created with linux dd tool on Windows? I have tried that tool http://www.chrysocome.net/dd, but it crashes under Windows 7. Any ideas?
[ -0.051733046770095825, -0.008593405596911907, -0.03316286578774452, 0.024591993540525436, -0.020699264481663704, -0.005145081784576178, 0.012428710237145424, 0.007214637473225594, -0.049740541726350784, -0.025238459929823875, -0.0062622614204883575, 0.018773574382066727, 0.021318035200238228...
[ 0.5588896870613098, -0.041001565754413605, 0.09304360300302505, 0.32357674837112427, 0.35555291175842285, -0.17887412011623383, -0.04842822253704071, 0.3560645282268524, -0.4696449935436249, -0.5055735111236572, 0.03015819936990738, 0.5378237366676331, -0.19768035411834717, 0.0460417792201...
Why is this, that in string theory the maximum amount of supersymmetry is $\cal{N} = 2$, whereas in supergravity one can have up to $\cal{N} = 8$ ?
[ -0.008615676313638687, 0.028113752603530884, -0.003476140322163701, 0.00515268137678504, 0.023636406287550926, -0.036669015884399414, 0.014209403656423092, -0.023046744987368584, -0.013790853321552277, -0.05123688653111458, -0.022597836330533028, 0.008102447725832462, -0.03099413402378559, ...
[ 0.1936754733324051, -0.08975427597761154, 0.2598572373390198, 0.10504183918237686, 0.10160248726606369, -0.20010791718959808, 0.19760087132453918, -0.019950103014707565, -0.4782131612300873, -0.012831793166697025, -0.37481483817100525, 0.3639959990978241, -0.45292845368385315, 0.5975542664...
I've built a kernel with loadable module support for various reasons, one of them the possibility to compile modules and load them without rebooting. This is supposed to be useful when I need a module that I had not enabled in the kernel config. Now, with drivers like nouveau, it's as easy as going to the source directory, and running `make M=drivers/gpu/drm/nouveau`. How can I build an updated `iptables` module without compiling a whole kernel and rebooting? Is it even possible?
[ 0.026068154722452164, 0.005992577876895666, -0.004157538525760174, 0.0038634096272289753, -0.04532679170370102, 0.0029772853013128042, 0.008434247225522995, -0.012215867638587952, -0.013894187286496162, 0.01652519777417183, -0.006972402334213257, 0.009071161970496178, 0.0009297747164964676, ...
[ 0.37701523303985596, 0.10252514481544495, 0.06633581221103668, 0.19788746535778046, -0.11221987754106522, -0.25768154859542847, 0.3723273277282715, -0.10043495148420334, -0.31915923953056335, -0.6423869132995605, 0.07690601795911789, 0.7530953884124756, -0.3862645924091339, 0.2271460592746...
I try to understand Multilayer Perceptrons for function regression. As a starting reference I am using the matlab neuronal network toolbox. There I mentioned that the tool maps the input and output values to the range of [-1 1]. Why do they do this? I thought that the inputs need to be decorrelated and free of mean, but they don't do this.
[ -0.00047606902080588043, 0.02195354364812374, -0.0038758134469389915, 0.0015760769601911306, -0.03355108201503754, -0.0037349658086895943, 0.010255620814859867, 0.01346561685204506, -0.01858591102063656, -0.02266436628997326, -0.010136205703020096, 0.01041178684681654, 0.007986199110746384, ...
[ 0.21231906116008759, -0.16990475356578827, 0.17912039160728455, 0.063874751329422, -0.3709951937198639, 0.08706926554441452, 0.021224090829491615, -0.1352914422750473, -0.20274323225021362, -0.6823205351829529, 0.4960017502307892, 0.3156304955482483, -0.1617034524679184, 0.5311959981918335...
For example: http://commons.wikimedia.org/wiki/File:Epithalamus.png Could I take an image file like that, splice it apart, and use it in an application that would be sold?
[ 0.014983322471380234, 0.0006161099299788475, -0.005881156772375107, 0.04756711795926094, 0.016390172764658928, 0.023001132532954216, 0.014385518617928028, 0.00569785526022315, -0.038856931030750275, -0.021614035591483116, -0.010165375657379627, 0.028801321983337402, -0.00908667966723442, -...
[ 0.9385849237442017, -0.21838049590587616, -0.2779485881328583, 0.32922640442848206, 0.12406399846076965, -0.17334187030792236, -0.1743762195110321, -0.20688053965568542, -0.42875391244888306, -0.3294271230697632, 0.3017447292804718, 0.09975773096084595, -0.3694622218608856, 0.2577654123306...
I want to insert small images in sidenotes. Is it possible (everything seems to)? How could I do it?
[ 0.026887329295277596, 0.0026482194662094116, -0.03011753410100937, 0.03691806644201279, 0.0019198224181309342, -0.01773293875157833, 0.011109930463135242, 0.046539876610040665, -0.021779168397188187, 0.010402439162135124, -0.03063984401524067, 0.00760784512385726, -0.010862905532121658, 0....
[ 0.38132932782173157, 0.40671390295028687, 0.16824349761009216, 0.24371063709259033, 0.08848875015974045, -0.011537856422364712, 0.18607692420482635, 0.44716736674308777, -0.20490527153015137, -0.4474732577800751, 0.13656198978424072, -0.03564954549074173, -0.2324737012386322, 0.49915313720...
How to set up the citation square brakets and numbers in upright in theorems while keeping other stuffs slanted? (This style is suggest by my professor.) The usual appearance is the following > _Theorem. There is no biggest prime number (see [1, Theorem 30 at page > 100])._ I would like to achieve > _Theorem. There is no biggest prime number_ ( _see_ [1, _Theorem_ 30 _at > page_ 100]). I could manually set `\textup` to numbers to upright but can not set the square brackets to upright.
[ 0.02131444402039051, 0.004130902700126171, -0.00674903579056263, 0.02253316342830658, 0.026445195078849792, 0.0026014819741249084, 0.007716228719800711, -0.020583491772413254, -0.02324143424630165, -0.03006792813539505, -0.003093174658715725, 0.005641439463943243, -0.02628317102789879, 0.0...
[ -0.2861361503601074, 0.5002391934394836, 0.583897054195404, 0.0250762477517128, -0.5164442658424377, 0.1876198947429657, 0.16637901961803436, -0.3231430649757385, -0.28618672490119934, -0.6773141026496887, -0.0031619311776012182, 0.7413958311080933, -0.19741332530975342, -0.009612197056412...
I got some files that say "epub+zip" undrr Properties and I wanted to know how can I open it? OS: 4.1.1
[ -0.014773829840123653, 0.03582008183002472, -0.011966729536652565, 0.038711242377758026, -0.0934557169675827, 0.07969708740711212, 0.01703479513525963, -0.021671926602721214, -0.024439288303256035, -0.03016560897231102, -0.028282329440116882, 0.03364119306206703, 0.017235862091183662, -0.0...
[ 0.18381048738956451, -0.07145035266876221, 0.4837258756160736, 0.14735841751098633, 0.15956644713878632, -0.3071043789386749, 0.36117374897003174, -0.03318566083908081, -0.40879589319229126, -0.5563920736312866, -0.2741529643535614, 0.4296935498714447, -0.5343523621559143, -0.0230568237602...
This line: ST_AsText(ST_Transform(the_geom, 4326)) Returns the "coordinates" as utm coordinates. Although, I thought ST_Transform was supposed to convert the coordinates from utm to lat/lon. I found a tutorial where the behavior I need is shown in an example: http://workshops.boundlessgeo.com/postgis-intro/projection.html Is there any reason why it would not be doing this?
[ -0.00932926032692194, 0.009979989379644394, -0.008459145203232765, 0.016953857615590096, 0.013153987936675549, 0.003669960191473365, 0.009655819274485111, 0.011980078183114529, -0.01369291078299284, 0.0010131922317668796, -0.0029328076634556055, 0.005528000649064779, -0.013018890284001827, ...
[ -0.04319748654961586, -0.3454715609550476, 0.7430936098098755, 0.07604105770587921, -0.10351353138685226, 0.12781907618045807, 0.25701799988746643, 0.13679838180541992, -0.16727851331233978, -0.6499390602111816, -0.42856481671333313, 0.05876334756612778, -0.4070034921169281, 0.092953450977...
> **Possible Duplicate:** > Do you work contract projects in addition to your full-time job? I currently work as a C++ developer full time (not a contract). I'm seeking extra opportunities outside of work. So far, I've been doing small gigs for people on sites like Craigslist or ODesk.com on the side for extra cash but they are just never lucrative enough. My first question is ... is it possible to find a contract type position that will allow me to work outside of normal hours (i.e. evenings and weekends)? How would you even find a job like that? I'm also worried about potential conflicts of interest. If the contract was in a completely different field, is there any reason that I have to report this to my employer? Should I get the help of a recruiter to find these type of jobs? If it's too difficult to find something like that ... where do you guys go to find extra work outside of your full time job? I know a lot of people do their own projects but I just don't have any of my own ideas for an iPhone app or anything like that. :P I have had the most luck with Craigslist in the past but there has to be better sites out there somewhere. Anyone willing to reveal the secret place? ;) Thanks!
[ 0.0067330943420529366, 0.011095013469457626, 0.0012632329016923904, 0.0003650481812655926, -0.015097644180059433, -0.003335230750963092, 0.005671054590493441, -0.00615297257900238, -0.016376007348299026, 0.01020386815071106, -0.002997515257447958, 0.016913991421461105, 0.018860314041376114, ...
[ 0.9263401031494141, 0.07025869935750961, 0.02163751609623432, 0.09172255545854568, 0.2517135739326477, 0.2904825210571289, 0.3197311758995056, 0.1574779599905014, -0.5155917406082153, -0.46224498748779297, 0.27014297246932983, 0.46234387159347534, 0.09177181869745255, -0.022872090339660645...
I don't know why but the `citations` start from `0` in the `beamer` class. But I don't have that problem with `article class` Here is the `MWE` \documentclass{beamer} \usepackage[latin1]{inputenc} \usepackage{hyperref} \usepackage{color} \usepackage{textpos} \usepackage{framed} \usetheme{CambridgeUS} \usepackage[british,UKenglish,USenglish,english,american]{babel} \begin{document} \begin{frame} Policy3\cite{one}. \end{frame} \bibliographystyle{ieeetr} \bibliography{thesis.bib} \end{document} @misc{one, title = {{AMD} BIOS and Kernel Developer's Guide for AMD Athlon\textsuperscript{\texttrademark} 64 }, howpublished = {\url{http://support.amd.com/us/Processor_TechDocs/26094.PDF}}, note = {Accessed: 21/2/2013}, } Result > Policy3[0]. It should be > Policy3[1].
[ -0.0006717017386108637, 0.00474043283611536, 0.00155140261631459, 0.021249493584036827, 0.008273058570921421, 0.0019275597296655178, 0.007455042097717524, 0.00856104213744402, -0.007480144500732422, -0.014184588566422462, -0.001788813853636384, -0.0044977907091379166, -0.0004203040152788162,...
[ -0.0704474225640297, 0.5025620460510254, 0.6439138650894165, -0.05133998766541481, -0.19725456833839417, -0.06059453636407852, 0.30233481526374817, -0.12329915910959244, -0.2894715964794159, -0.46331509947776794, -0.3652648329734802, 0.25673696398735046, -0.17847242951393127, 0.00072386767...
I have ubuntu and I want to know where can I find the log file of iptables? I find out `/var/log/messages`, but I am not sure this is a correct log file or not. And I also want to know when this log file is changed? I add one rule to prevent my machine to respond to ping message but when I ping my machine I didn't see any changes to `/var/log/messages`
[ 0.02899312600493431, 0.006491581443697214, 0.0009699520887807012, 0.005036907736212015, 0.007844510488212109, 0.02303815260529518, 0.00955027062445879, 0.03561090677976608, -0.022829381749033928, -0.008634907193481922, -0.004019600339233875, 0.006384979467839003, -0.012385653331875801, 0.0...
[ 0.09739420562982559, 0.12222746759653091, 0.3961641788482666, 0.024664748460054398, -0.46073850989341736, -0.5259355306625366, -0.027714578434824944, 0.260111540555954, -0.23089000582695007, -0.8670018911361694, 0.1530541628599167, 0.5074297189712524, -0.46935009956359863, 0.31380707025527...
I am unable to use the `eulervm` package with `maketitle`. If I comment out the `\usepackage{eulervm}` the LaTeX renders fine, but as soon as I include it, I get the following > ! Math formula deleted: Insufficient symbol fonts. \endtabular ->\crcr > \egroup \egroup $ \egroup l.16 \maketitle Are these not compatible? Is there any way to make them so? Or is there another command that would substitute for maketitle? Minimal Working Example as asked for in comments. \documentclass[a4paper,10pt]{article} \usepackage{eulervm} \title{The Cave} \author{Batman, Robin} \date{} \begin{document} \maketitle \end{document}
[ 0.01983458176255226, 0.008347101509571075, -0.009482957422733307, 0.016379063948988914, -0.010899767279624939, 0.014076540246605873, 0.00842844694852829, 0.0047711594961583614, -0.013954324647784233, -0.022880040109157562, -0.013523878529667854, 0.010207025334239006, -0.009396286681294441, ...
[ 0.24160391092300415, 0.3510564863681793, 0.316017210483551, -0.3090344965457916, 0.17449186742305756, 0.010949810966849327, 0.5502557754516602, -0.18691855669021606, 0.12020120769739151, -0.577296793460846, -0.17586183547973633, 0.5043514966964722, -0.37427467107772827, -0.1115277856588363...
If you cut a thin slit in thin opaque material and then put it into water and pull it out, the meniscus will be formed in the slit. For my research I need to know if it is analytically possible to define the curvature of meniscus( is it hyperbola, parabola, circle or something else). I need the function of curvature so I will be able to make calculations of how the light passes through the meniscus. I know there are numerical methods to make an approximate function, but I want to know if there is any other way.
[ 0.013289405964314938, -0.005766966845840216, 0.0008180940058082342, 0.016515878960490227, -0.027650076895952225, -0.02382662519812584, 0.008179724216461182, 0.004370969720184803, -0.01966448687016964, -0.029559627175331116, 0.007381764240562916, 0.008147135376930237, -0.014336707070469856, ...
[ 0.583337664604187, -0.08039512485265732, 0.17725613713264465, 0.22576786577701569, -0.17374922335147858, 0.34708842635154724, -0.014212424866855145, 0.07040445506572723, -0.37764886021614075, -1.025911808013916, 0.29916590452194214, 0.22023354470729828, -0.08590976148843765, 0.816513836383...
I'm writing a program that displays various system information (on a CentOS system). For example, the processor type and speed (from `/proc/cpuinfo`), the last boot time (calculated from `/proc/uptime`), the IP address (from `ifconfig` output), and a list of installed printers (from `lpstat` output). Currently, several pieces of data are obtained from the `dmidecode` program: * The platform type (`dmidecode -s system-product-name`) * The BIOS version (`dmidecode -s bios-version`) * The amount of physical memory (`dmidecode -t17 | grep Size`) These are only available if my program is run as root (because otherwise the `dmidecode` subprocess fails with a `/dev/mem: Permission denied` error). Is there an alternative way to get this information, that a normal user can access?
[ -0.0033298684284090996, -0.0030720029026269913, -0.011898281052708626, 0.0017209129873663187, -0.00942030269652605, 0.002518261782824993, 0.00802028551697731, -0.00022668391466140747, -0.012134825810790062, -0.00013868045061826706, -0.0027022650465369225, 0.0008014905033633113, 0.00995324179...
[ 0.5576123595237732, 0.2837181091308594, 0.3721030354499817, 0.10204190760850906, 0.34014570713043213, 0.34742215275764465, -0.16698239743709564, -0.24927765130996704, 0.16504110395908356, -0.7633052468299866, 0.08892429620027542, 0.43802207708358765, 0.05602717027068138, 0.0843362361192703...
I'm finishing my dissertation and want/need to provide the source code of some analysis scripts I've used. I did all the version control with `git`. Because I want to publish the results generated with these scripts, I don't want to just put them in a public repository on GitHub. Is it possible to provide an "URL-only" access to a GitHub repo, meaning it is not publicly listed but only visible to persons who have the URL? If not with GitHub, is this possible with another service? I would like to have some kind of GUI to see the code, which is why I don't want to make a zip file with just the source code in it. Edit: To make that point more clear: There should not be any login or registration involved! Type in the URL printed in the dissertation and boom, there's the source code!
[ 0.015133648179471493, 0.003163897432386875, 0.005259714089334011, 0.022731654345989227, 0.0031414139084517956, -0.01182652823626995, 0.005258640274405479, -0.0007717821281403303, -0.014842931181192398, -0.01275133527815342, -0.007856839336454868, 0.011329076252877712, -0.0067833359353244305,...
[ 0.4608749747276306, 0.43861186504364014, 0.27277326583862305, 0.10545690357685089, 0.04012446478009224, -0.37677186727523804, 0.03184027597308159, 0.22023336589336395, -0.16940069198608398, -0.29402589797973633, -0.06084660813212395, 0.1502341628074646, 0.17059080302715302, 0.3349541425704...
I'm trying to use the ClockworkMod recovery image and ROM Manager to flash a custom ROM on my HTC Wildfire. I downloaded CyanogenMod 7.1, put it on the sdcard and followed the instructions. I selected Backup current ROM, Wipe User Data and Wipe System Data, then it reboots into recovery mode but I never get past the screen with the red triangle and exclamation mark (http://twitpic.com/7yzq9w) Any idea on what I'm doing wrong? **Edit** : forgot to mention that pressing the power + volume up buttons brings me to a menu with options for reboot, apply update.zip, wipe cache partition and wipe / factory reset. Apply update.zip does not work (signature cannot be verified) and wouldn't be desirable since I'd first like to make a backup of my current ROM.
[ 0.007663613185286522, -0.0006158333271741867, -0.010855624452233315, 0.021411936730146408, -0.046392835676670074, -0.014833039604127407, 0.00756430346518755, -0.0052820611745119095, -0.015737716108560562, -0.009117784909904003, -0.021786721423268318, -0.00037575699388980865, 0.00085719802882...
[ 0.38600945472717285, 0.286899209022522, 0.3810447156429291, 0.11063222587108612, 0.16053979098796844, 0.02407034859061241, 0.5475905537605286, -0.040151964873075485, -0.25773847103118896, -0.15639181435108185, -0.24185271561145782, 0.745051383972168, 0.04996171221137047, -0.061222977936267...
I have completed all the sidequests for the gratitude crystals (and found all the ones in the wild), but realized I was still five short (75). In trying to discover the issue, I realized the game had not recognized that I had completed the plow needed request from Kina, when in fact I had got Guld from Eldin Volcano and brought him to Kina (and received the crystals). But now Kina is again talking to Link about needing someone to plow the field, and Guld is back in Eldin volcano saying he wishes he could go to the sky. However, I cannot get Fi to appear once again in Kina's presence to set Guld on my dowsing function, so I cannot get Guld back up to the sky. This has basically voided my ability to complete the Gratitude Crystals sidequest and get the Tycoon wallet, which seriously pisses me off. Does anyone know how to fix this issue?
[ 0.007927805185317993, -0.0002141381846740842, -0.003844473510980606, -0.004253854975104332, -0.006554639432579279, 0.006026848219335079, 0.010699305683374405, 0.004089973866939545, -0.01999855414032936, 0.025811966508626938, -0.013435683213174343, 0.01439545676112175, -0.02593025378882885, ...
[ 0.06652452051639557, -0.0049807969480752945, 0.4496976435184479, 0.11681388318538666, -0.7248086929321289, 0.11988218873739243, 0.36749333143234253, -0.7216461300849915, -0.2705746591091156, -0.46977517008781433, 0.15623515844345093, 0.1019127368927002, 0.14990974962711334, 0.1590445637702...
The quasi-MWE below uses the `memoir` class and works fine. It produces a document with the following structure: * ToC * Book * Part1 * Chapter1 * Part2 * Chapter2 * Index Here, cross-references in the ToC and in the Index all work correctly. But I want to reset Chapter numbers back to "1" in each Part (and ultimately, in each Book, if there are many books). To do so, I uncomment the `\counterwithin` instructions in the preamble. Now, depending on the value assigned to `hyperref`'s `hypertexnames` option, I get the following behaviour: * if `true`, the ToC link to Chapter "2" in Part 2 (which is now renumbered to "1") does not works, but the Index link to AAAAA on page "5" does; * if `false`, the ToC link to Chapter "2" in Part 2 (which is now renumbered to "1") works, but the Index link to AAAAA on page "5" does not. Am I doing something wrong? * * * \documentclass[a4paper,12pt]{memoir} %\counterwithin*{chapter}{part} % Restarts chapter count within Part. \usepackage{hyperref} \hypersetup{ hypertexnames=true, linktocpage=false, colorlinks=true, } \indexintoc \makeindex[main] \begin{document} \frontmatter{} \tableofcontents* \mainmatter{} \book{BookTitleA} \part{PartTitleA} \chapter{TitleA} AAAAA\index[main]{AAAAA}.\\ \part{PartTitleB} \chapter{TitleB} BBBBB. \backmatter \clearpage \printindex[main] \end{document}
[ 0.01520930603146553, 0.019224360585212708, -0.021640270948410034, 0.01795087195932865, 0.024734044447541237, 0.015395027585327625, 0.009117835201323032, -0.00250720651820302, -0.019049063324928284, 0.010704380460083485, -0.009470507502555847, 0.006536558270454407, -0.01036473922431469, 0.0...
[ 0.2699049711227417, 0.2047080546617508, 0.7567648887634277, 0.06371334940195084, -0.1416134238243103, -0.2830802798271179, 0.5479934811592102, -0.2611505389213562, -0.06253144890069962, -0.0785011351108551, -0.3020765483379364, 0.6338372230529785, -0.10557539016008377, 0.41439664363861084,...
I am working on a custom Wordpress template for the archive. I have the archive set up by category but now I want to organize by category and month but I can't figure out how to access the month argument from the permalink. How would I get access to that argument.
[ 0.025445496663451195, 0.003996210638433695, -0.007938103750348091, 0.0358886793255806, 0.047246355563402176, 0.014061578549444675, 0.012961228378117085, 0.007370364852249622, -0.03007138893008232, -0.005903041455894709, -0.009946002624928951, -0.005797104444354773, -0.015623732469975948, -...
[ 0.38221317529678345, 0.3729519844055176, 0.41004085540771484, 0.018931392580270767, -0.11135813593864441, 0.3768300414085388, -0.03288416191935539, 0.09480728209018707, -0.35993051528930664, -0.5082817077636719, 0.2861082851886749, 0.10704417526721954, 0.13571611046791077, 0.47464001178741...
# Background I have a **registration form** that validates with AJAX using jQuery Validation Engine. The form was created using Theme My Login and contains **reCaptcha** which I implemented using WP-reCAPTCHA. Just like the rest of the fields, I needed to **validate reCaptcha using AJAX** as well, that is, without refreshing the whole page. So far I managed to make it work when the Captcha didn't validate. But **when it did validate** , ie. the input was correct, the page refreshed but with an **error message** which says **'That reCAPTCHA response was incorrect'**. Thus, the form was not saved. I tried using `console.log()` to check the output and the result is 'valid'. So I can't really tell where I went wrong. Currently I'm testing this locally using _XAMPP_. # The Code That I Used **The Javascript:** (function($){ // Validate Captcha $.fn.validateCaptcha = function() { challengeField = $('input#recaptcha_challenge_field').val(); responseField = $('input#recaptcha_response_field').val(); $.ajax({ type: 'POST', url: the_ajax_script.ajaxurl, data: 'action=validate_captcha&recaptcha_challenge_field=' + challengeField + '&recaptcha_response_field=' + responseField, async: false, success:function(result) { //console.log( result ); if(result != 'Valid') { $('#captchaError').html('<p class="error">The security code you entered did not match. Please try again.</p>'); $captchaFlag = 'Invalid'; Recaptcha.reload(); } else { $('#memberInformation span').css({'color':'green'}); $('#memberInformation span').html(html.message).show(3000); $('#captchaError').html('<p>Success!</p>'); $captchaFlag = 'Valid'; dataString = form.serializeArray(); getSearchMembers(dataString); } } }).responseText; } $('#registerform').submit(function() { var form = $(this); if (form.validationEngine('validate')) { $('#memberInformation span').html(''); form.validateCaptcha(); } else { $('#memberInformation span').css('color','#ff0000').html('Please fill out required fields').show(3000); } return false; }); })(jQuery); **The PHP:** function validate_Captha() { $privatekey = "xxxxxxxxxx"; //<!----- private key here $resp = recaptcha_check_answer ($privatekey, $_SERVER["REMOTE_ADDR"], $_POST["recaptcha_challenge_field"], $_POST["recaptcha_response_field"]); if (!$resp->is_valid) { // Incorrect CAPTCHA input echo "Error\n"; die ("The reCAPTCHA wasn't entered correctly. Please go back and try it again.\n(reCAPTCHA said: " . $resp->error . ")"); } else { echo "Valid"; die(); } } add_action( 'wp_ajax_validate_captcha', 'validate_Captha' ); add_action( 'wp_ajax_nopriv_validate_captcha', 'validate_Captha' );
[ 0.030671831220388412, 0.02299482934176922, -0.0026235412806272507, 0.009801085107028484, 0.002946804277598858, -0.00807213969528675, 0.008754989132285118, -0.004412159789353609, -0.01812460646033287, 0.00029672402888536453, -0.019509050995111465, 0.015883734449744225, -0.009697020053863525, ...
[ 0.44977685809135437, -0.15450456738471985, 0.8226168751716614, 0.19176477193832397, -0.2264125496149063, -0.1563900262117386, 0.3156309425830841, 0.059648171067237854, -0.32638031244277954, -0.5011322498321533, 0.013380183838307858, 0.39893728494644165, -0.08526434004306793, -0.36502593755...
Please help me to find out the appropriate English idiom for `Fry the fish using fish's own oil`. This is a Bengali proverb/idiom whose word meaning 'When you are frying some sea fish, initially cook use some oil for frying but within few minute fish emit it's own body oil, and the remaining frying is done by this oil. Literally, this means Somebody has invested some amount of money in some particular purpose but get some extra benefit, as an add-on that save his/her initial investments. Please help me. If I'm not clarify well please ask me. Thanks!
[ 0.007362996693700552, 0.028408724814653397, 0.006264775060117245, 0.024493161588907242, -0.018831733614206314, -0.02122160978615284, 0.008558342233300209, 0.010919326916337013, -0.014822660014033318, -0.00434827757999301, -0.017554940655827522, 0.009930703788995743, -0.01181679405272007, 0...
[ 0.91840660572052, 0.15799082815647125, -0.37535348534584045, 0.32150059938430786, 0.18256361782550812, -0.03800051659345627, -0.15447622537612915, 0.48177003860473633, -0.028835101053118706, -0.2583041489124298, 0.17399844527244568, 0.2644990384578705, 0.14256183803081512, 0.03202004358172...
Let $X$ be a beta distributed variable with parameters $a$, $b$. Let $Y$ be a beta distributed variable with parameters $c$, $d$. Let $Z = \max(X, Y)$. Does anybody know of a fast way to compute the mean and variance of $Z$ given $(a, b, c, d)$? $X$, $Y$ don't necessarily need to have beta distributions, but they should be similar -- in $[0, 1]$. What I really want to find is a way to quickly approximate the distribution of the sample max of a bunch of random variables. Let $Z = \max(\{X_1, X_2,\dots , X_n\})$. If I know parameters (not necessarily beta-distribution) for $X_1, X_2,\dots, X_n$, can I quickly compute a set of parameters (same distribution as $X_i$) to form a distribution that closely approximates $Z$? * * * **EDIT:** To clarify, the purpose of this has to do with programming AI for board games. If we are evaluating a position in a chess game and we determine one move leads to a 60% chance of winning while all others have a 20% chance of winning, then we value the position as a 60% win. This is the minimax algorithm... However, what I'm curious about is whether this can be improved upon. In the simple case, we approximate: $$ \mu_c = \max(\{\mu_1, \mu_2,\dots, \mu_n\}) $$ where $\mu_c$ is the expectation from the current position and $\mu_i$ is the expectation from each of the sub-trees. It seems like we should be able to do better than this. Can we find a function f that replaces max and closely approximates parameters for the distribution of the sample maximum? $$ (\mu_c, \sigma_c) = f\left(\{(\mu_1,\sigma_1), (\mu_2,\sigma_2), \dots, (\mu_n, \sigma_n)\}\right) $$ In this case, we still want $\mu_i$ to be the expectation from each of the nodes. However, $\sigma_i$ does not necessarily need to be variance/standard deviation; just a parameter that represents uncertainty. Just as: $$ \max(\{a, b, c\}) = \max\left(\{\max(\{a, b\}), c\}\right) $$ We should expect: $$ f\left(\{a, b, c\}\right) \approx f\left(\{f(\{a, b\}), c\}\right) $$ So $f$ does not necessarily need to take parameters for more than two nodes. One thing, however, is that $f$ should be relatively fast to compute. If the integrals must be calculated, I think doing double exponential integration might work well, but I'm not sure -- * http://www.johndcook.com/double_exponential_integration.html * http://www.codeproject.com/Articles/31550/Fast-Numerical-Integration
[ -0.012050841003656387, 0.012654402293264866, -0.011223206296563148, 0.0036312269512563944, 0.003497098106890917, -0.011011103168129921, 0.004693306516855955, -0.012412535957992077, -0.012005819007754326, -0.019460363313555717, -0.0107796685770154, 0.007954652421176434, -0.010313620790839195,...
[ 0.3571604788303375, -0.1627926379442215, 0.1636766493320465, 0.022038854658603668, -0.07159743458032608, -0.04718323424458504, -0.1377541720867157, 0.10038086026906967, 0.26912692189216614, -0.2665456533432007, -0.0613771453499794, 0.4360849857330322, -0.6922880411148071, 0.287374436855316...
The title just says it all. My 3.3", Android 2.3.3 emulator (armeabi) boots averagely in 150 seconds on my Ubuntu 12.04, 64 bit system, with 2 GB DDR3 memory. However, sometimes when I tried shutting down using the power button provided in the emulator. I am made to wait indefinitely! This doesn't hurt much, for I usually just close the emulator just like closing a window. However, I somehow can't reconcile the fact and need to know reason(s) why, in spite of booting fairly fast, shutting down takes ages! Does shutting down consume more resources than booting?!
[ 0.008154689334332943, 0.012471728026866913, -0.019048888236284256, 0.01276695728302002, -0.05339012295007706, -0.010684765875339508, 0.006770228501409292, 0.005498708691447973, -0.008087183348834515, -0.004898690618574619, -0.025001827627420425, 0.0085455272346735, 0.007488668896257877, 0....
[ 0.15619559586048126, 0.21814721822738647, 0.9135333299636841, 0.08305485546588898, 0.002473701722919941, 0.12109941244125366, 0.3121856451034546, 0.0849795714020729, -0.16536228358745575, -0.21982280910015106, 0.13609679043293, 0.5105993747711182, 0.017253175377845764, 0.21687264740467072,...
The etymology of _love child_ says it derived as a polite form of "love brat" which was used around the 18th century. My question is when two people are in love and they have a child, could you not call him/her a "love child"? Edit: Why does it have to have the (rather negative) connotation that it is outside of marriage, when love itself is a positive emotion? Funny enough, "love child" makes sense in Indian English, where there is a concept of "love marriage" and "arranged marriage" - as has been earlied posted.
[ -0.006790129467844963, 0.009666969999670982, -0.012402755208313465, 0.01861800067126751, -0.010603438131511211, -0.0006037465063855052, 0.008015093393623829, -0.008160505443811417, -0.01888769492506981, -0.038106903433799744, 0.0044067297130823135, 0.010708943009376526, 0.005157794803380966,...
[ 0.1813967525959015, 0.2382848858833313, -0.003910075407475233, -0.11103525757789612, 0.0754992887377739, 0.022717438638210297, 0.11705219000577927, 0.06845779716968536, -0.07865623384714127, -0.38329851627349854, -0.01141362264752388, 0.05174995958805084, -0.10158193111419678, 0.6437785625...
I have trained linear discriminant analysis (LDA) classifiers for three classes of the IRIS data and struggling with how to make the classification. Here is the procedure: For the Iris data, I have 3 combinations i.e. (0,1), (0,2) and (1,2). So, I trained a simple binary LDA classifier for each combination, and ended up with three classifiers: Classifier(0,1) Classifier(0,2) Classifier(1,2) Now, say I need to classify an input, say k = [1.2, 2.3, 5.0]. What I am doing is passing this input through all the classifiers individually, which are giving me their respective scores, like: Classifier(0,1)[k] = {0: some score, 1: some score} Classifier(0,2)[k] = {0: some score, 2: some score} Classifier(1,2)[k] = {1: some score, 2: some score} In a simple binary case of two classes, what we are taught to do is to take the class with maximum score as the result. My question is, what to do in such a scenario, where I have three results from three different classifiers, and I want to classify the output. Please note that I am not using a multiclass LDA. I am just using a binary LDA for all the possible combinations, a technique which is stated here: http://en.wikipedia.org/wiki/Linear_discriminant_analysis#Multiclass_LDA > quoting the last paragraph of this section: " _Another common method is > pairwise classification, where a new classifier is created for each pair of > classes (giving C(C − 1)/2 classifiers in total), with the individual > classifiers combined to produce a final classification._ " Can somebody please enlighten me about what needs to be done in such a case for classification? Thank you.
[ 0.04014784097671509, 0.014021886512637138, -0.02720983326435089, 0.008225945755839348, 0.0023009739816188812, 0.007967345416545868, 0.011462866328656673, -0.003224940737709403, -0.01746206171810627, 0.023244943469762802, -0.011274667456746101, -0.0021383962593972683, -0.020745377987623215, ...
[ 0.06860510259866714, -0.13807880878448486, 0.29310712218284607, 0.0790836438536644, -0.16127367317676544, 0.39929822087287903, 0.08051031827926636, -0.5659287571907043, 0.2186398208141327, -0.5264466404914856, 0.10544165968894958, 0.29178714752197266, -0.4853654205799103, 0.219381973147392...
I am currently a freelancer (JavaScript) developer, and now I am striving to work more on freelance marketplaces. Some time I want to move from Russia to Canada and find full-time work there. So: How I might provide evidence of expertise in this case? I can not just say in my resume/CV "I was super JavaScript Developer in the MacroHard from 2011 to 2017". Are my accounts (and 4,5/5 star rating) on these marketplaces will be relevant and adequate to show to the potential new employer (urls in the CV, for example)? Or maybe I should just add description of most interesting projects? I have found this article: http://gettingtozen.com/2010/10/how-to-list-your- freelancing-and-self-employment-experience-on-your-resume/ Author gives advice to not include many references to the freelance projects, but his field is not an IT (freelance corporate trainer). In my point of view, it is not very applicable for a Software Developer, since as Software Developer you can explicitly point to the application you worked with. Also, an option here is to refer yourself as "individual entrepreneur" instead of "freelance contractor". My experience in North American employment traditions is zero, so I hope to get advice from colleagues and HRs.
[ 0.009901279583573341, 0.009235154837369919, -0.010291365906596184, -0.0025060579646378756, -0.02054058015346527, 0.01484965905547142, 0.008488791063427925, -0.008759983815252781, -0.01621556282043457, -0.012248094193637371, -0.0063140965066850185, 0.023956121876835823, 0.007423845585435629, ...
[ 0.9610947370529175, 0.5898165702819824, 0.17624452710151672, -0.31207695603370667, 0.2232423573732376, 0.05348159000277519, 0.4969429075717926, 0.6752103567123413, -0.3186493515968323, -0.6006672382354736, -0.04762047156691551, 0.04437852278351784, 0.5050777792930603, 0.1390671730041504, ...
Just installed (netinstall) debian wheezy on a second machine newer machine however the terminal print out is noticeable slower than on the old machine although the new have faster CPU, RAM and HDD (an SSD). When login in on the new machine with ssh it is as fast as I expect it to be. It is obvious that the out put speed is related to computer/display and not just the computer. For example the `dpkg -l` via virtual console takes 10s to print out and is instant via ssh. X is not installed it is just a fresh installation of Debian Wheezy. I assume this related to KMS but I am not sure where to start.
[ -0.007692368701100349, -0.005585175473242998, -0.01565001904964447, 0.014442306011915207, -0.04710789769887924, -0.02572166919708252, 0.008508933708071709, -0.01061788946390152, -0.010910049080848694, -0.035181500017642975, -0.01628810353577137, 0.005670757964253426, -0.012620593421161175, ...
[ 0.8224152326583862, -0.37654489278793335, 0.36816781759262085, 0.054581690579652786, 0.04721718281507492, -0.3555934429168701, 0.16212108731269836, 0.1184546947479248, -0.07999364286661148, -0.8529098033905029, -0.22314733266830444, 0.5718317031860352, -0.4709547460079193, -0.1089287474751...
I am having a hard time trying to come up with an fully expandable keyval parsing macro which deals only with specific keys/values. I wouldn't even attempt this in full generality, that's way beyond me. Here is as far as I've got: \documentclass{article} \usepackage{etoolbox} \usepackage{ifthen} \begin{document} \makeatletter \def\foo@bar{value} \def\thing@i#1{% \thing@ii#1&} \def\thing@ii#1=#2&{% \ifthenelse{\equal{#1}{form}} {#2} {}} \catcode`\==11 \csuse{foo@\thing@i{form=original}} \end{document} Aside from the problem of having to catcode "=" (and then somehow having to catcode it back again later when "=" doesn't mean "=" any more), This doesn't work due to some expansion issue with the `\ifthen` it seems. I am a bit numb looking at this now, I've tried using `\ifx` etc. but no luck. There seems to be no way to get fully expandable keyval processing and so I'm reduced to doing something which hard-codes the possible keys etc. I have to have this because I will have macros with keyval args which need fully expanding to construct csnames.
[ 0.018472693860530853, 0.017932206392288208, 0.0048352633602917194, 0.010523542761802673, 0.005733858793973923, 0.012482081539928913, 0.007106174249202013, 0.014509955421090126, -0.014255823567509651, 0.012451467104256153, -0.006910449825227261, 0.0053089275024831295, -0.022507060319185257, ...
[ -0.09020292013883591, 0.1628628820180893, 0.25723618268966675, -0.13639076054096222, 0.29664096236228943, -0.03782176226377487, 0.13535359501838684, -0.20585797727108002, -0.039148133248090744, -0.5772541761398315, 0.03519453480839729, 0.735910952091217, -0.1512129008769989, -0.21508377790...
Having a HDD Drive on a USB hub with two partitions. Can I mount first partition to Computer A and second partition to Computer B without trouble? Or even better, would it be possible to mount one partition to both computers?
[ 0.0002295000886078924, 0.051951661705970764, -0.015435534529387951, 0.04178876429796219, -0.03266500309109688, -0.014764603227376938, 0.01822996698319912, 0.00804190244525671, -0.040570180863142014, -0.024392638355493546, -0.051647573709487915, 0.029599269852042198, -0.02655860036611557, 0...
[ 0.37173956632614136, -0.17723217606544495, 0.32095128297805786, 0.3902331590652466, 0.016914593055844307, -0.13497640192508698, -0.31082990765571594, -0.10330671817064285, -0.16149157285690308, -0.9903650879859924, 0.04136476293206215, 0.6512137055397034, 0.03486984968185425, 0.06913194060...
When you have this construct: > . . . is a key factor in the **making** and **controlling** of the water. Should you leave only the last verb in the gerund: > . . . is a key factor in the **make** and **controlling** of the water.
[ -0.01262896228581667, 0.018103085458278656, -0.015014585107564926, 0.013107583858072758, -0.008669092319905758, 0.017135348170995712, 0.010357563383877277, -0.02352943643927574, -0.013002501800656319, 0.0419936329126358, -0.003954560030251741, 0.012847686186432838, 0.002301661064848304, 0....
[ -0.5017002820968628, 0.12405800074338913, 0.33999642729759216, -0.388207346200943, -0.5923956632614136, 0.12778568267822266, 0.7266220450401306, -0.3753984868526459, 0.027137069031596184, -0.9078497290611267, -0.4465145468711853, 0.3532656133174896, 0.36684706807136536, -0.0886089950799942...
I'm trying to better understand some of the theory behind fitting models that have a nonlinear link between the response and the predictors. set.seed(1) #Create a random exponential sample n<-1000 y<-rexp(n=n,rate=.01) y<-y[order(y)] x<-seq(n) df1<-data.frame(cbind(x,y)) plot(x,y) Now I will attempt 4 different models fits and explain what I expect as a result versus the actual result. m1<-lm(I(log(y))~x,data=df1) summary(m1) out1<-exp(predict(m1,type='response')) lines(out1) I expect the m1 model to be the worst fit. This is OLS with the log of the response taken before fitting the model. By taking the log of the response I have a model that, when exponentiated, cannot be negative. Therefore, the assumptions of normally distributed residuals with a constant variance cannot hold. The graph appears to under weight the tails of the distribution significantly. Next, I will fit the model using a GLM. m2<-glm(y~x,data=df1,family=gaussian(link='log')) summary(m2) out2<-predict(m2,type='response') lines(out2,col='red') I expect the m2 model to be of a slightly different fit than m1. This is due to the fact we are now modeling log(y+ϵ)=Xβ rather than modeling log(y)=Xβ+ϵ. I cannot justify if this fit should be better or worse from a theoretical standpoint. R gives an R-squared figure for lm() functions calls but not glm() (rather it gives and AIC score). Looking at the plot it appears that m2 is more poorly matched in the tails of the distribution than m1. The issues with normally distributed residuals should still be the same. For m3, I relax the residual distribution assumptions by changing to a Gamma family residual distribution. Since an exponential distribution is a Gamma family distribution, I expect this model to very closely match the sampled data. m3<-glm(y~x,data=df1,family=Gamma(link='log')) summary(m3) out3<-predict(m3,type='response') lines(out3,col='blue') AIC suggests that m3 is a better fit than m2, but it still does not appear to be a very good fit. In fact, it appears to be a worse estimate in the tails of the distribution than m1. The m4 model will use nls just for a different approach. m4<-nls(y~exp(a+b*x),data=df1, start = list(a = 0, b = 0)) summary(m4) out4<-predict(m4,type='response') lines(out4,col='yellow') #for direct comparison t1<-cbind(out1,out2,out3,out4,y) This appears to very closely match m2. I also expected this to very closely fit the data. My questions are - how come does m3 not more closely match my data? Is there a way using glm() to more closely fit this model? I realize that if I am randomly sampling then the tails of my sample data will not fit my model exactly, but these appear to be nowhere close. Perhaps letting n approach infinity would all one (or more) of the models to converge, but rerunning the code with n=100,000 actually appears to be worse fitting models (perhaps that's because with 100,000 random samples more outliers were selected and due to how the graph is present an undue amount of focus is given to these outliers?). I also realize that an exponential distribution is not the same as exponentiating a normal distribution. That is to say, I realize that just because I took the log the of response doesn't mean that I should get a "perfect model" as a result; however, in m3 when I am fitting the models to a Gamma family distribution, I expected to be able to get a very close fit. Thanks in advance.
[ 0.010786769911646843, 0.013293277472257614, 0.0005045398138463497, 0.017137940973043442, 0.004584941081702709, -0.014505267143249512, 0.005331853870302439, 0.00409957580268383, -0.010527674108743668, -0.02468929812312126, -0.004241744987666607, 0.008585125207901001, 0.0003914421540684998, ...
[ 0.45972076058387756, -0.07344299554824829, 0.441285103559494, -0.006646730471402407, -0.00757815083488822, 0.22276578843593597, 0.04869481548666954, -0.6806809902191162, -0.283544659614563, -0.5045070648193359, 0.3253834843635559, 0.40812405943870544, -0.17192934453487396, 0.33519294857978...
I found if I search using grep without specifying a path, like `grep -r 'mytext'` it takes infinitely long. Meanwhile if I search with path specified `grep -r 'mytext' .` it instantly finds what I need. So, I'm curious, in first form, in which directory does grep search? UDATE: grep version: grep (GNU grep) 2.10
[ 0.010057306848466396, 0.01574772596359253, -0.0198114812374115, 0.010136468335986137, -0.009733300656080246, 0.015275373123586178, 0.0073706419207155704, 0.016443470492959023, -0.024961894378066063, -0.02038596384227276, -0.00314171239733696, 0.00684773875400424, 0.008039085194468498, 0.00...
[ -0.17343781888484955, 0.13931842148303986, 0.6184719204902649, -0.10961591452360153, -0.23929151892662048, -0.06622171401977539, 0.1510450541973114, 0.05289541557431221, -0.4859309494495392, -0.5267130732536316, -0.45572078227996826, 0.4162857234477997, -0.6889858245849609, -0.127824962139...
I'm trying to replace the letters in the math font so that it matches the main font. Curiously, this works for the italic version, yet not for bold and roman. Why is that? \documentclass{article} \usepackage{unicode-math} \defaultfontfeatures{Extension = .otf} \setmainfont[ItalicFont = *-Italic, BoldFont = *-Bold]{BemboStd} \setmathfont{latinmodern-math} \setmathfont[range=\mathrm]{BemboStd} \setmathfont[range=\mathit]{BemboStd-Italic} \setmathfont[range=\mathbf]{BemboStd-Bold} \begin{document} \noindent abc \textit{abc} \textbf{abc} \\ $ \mathrm{abc} \ abc \ \mathbf{abc} $ \end{document} ![enter image description here](http://i.stack.imgur.com/Bha7w.jpg) **Edit:** Just in case it might be helpful to the next person, I'm now using \setmathfont[range=\mathup/{latin,Latin}]{BemboStd} \setmathfont[range=\mathit/{latin,Latin}]{BemboStd-Italic} \setmathfont[range=\mathbfup/{latin,Latin}]{BemboStd-Bold} as to only substitute the normal letters, and leave all the greek and `\partial` glyphs alone.
[ 0.023922231048345566, 0.0030138378497213125, -0.012502511963248253, 0.009308749809861183, 0.007132484577596188, -0.018629051744937897, 0.008246750570833683, 0.012353307567536831, -0.014564123004674911, -0.027172520756721497, -0.010854624211788177, 0.0033888225443661213, -0.000891648582182824...
[ 0.30722007155418396, 0.2909437119960785, 0.3663667142391205, -0.07184498012065887, 0.00995546206831932, 0.20079964399337769, 0.17643816769123077, -0.021992437541484833, -0.05299091711640358, -0.6150906682014465, -0.06005552038550377, 0.5807364583015442, -0.10399116575717926, 0.018585406243...
I'm dealing with a homework. We were given a software, with inputs, that we should measure and then verify that the time complexity is O(n^3). I measured the software, picked best times from each batch (since that's the least affected one), and ended up with this: 304 70000 574 480000 775 1190000 850 1580000 1070 3130000 1557 9740000 1965 19540000 Now I'm really lost. I know how to calculate a model using least squares, but I have no idea how to do that for O(n^3). Plus I don't know how I would verify that the model is correct.
[ 0.008157025091350079, 0.005808038637042046, -0.02497701905667782, 0.003872533096000552, -0.015565788373351097, 0.005301673896610737, 0.00578298419713974, -0.020085997879505157, -0.01122964359819889, 0.013249331153929234, -0.0013684378936886787, 0.003280526027083397, 0.015286078676581383, 0...
[ 0.30254629254341125, 0.4257838726043701, 0.11877688020467758, -0.1525392085313797, -0.10756023973226547, 0.42749714851379395, 0.21912620961666107, -0.272015780210495, -0.6042451858520508, -0.4911452531814575, 0.04132438078522682, 0.15458855032920837, -0.05893581360578537, 0.388226896524429...
Consider a horizontal pipe that has water being pumped continuously through it, with a leak at a given point. What will the leak's effect be on the pressure in the pipe, both above and below it in the flow?
[ -0.022463064640760422, 0.028047919273376465, 0.0058353799395263195, 0.036092158406972885, 0.01868181861937046, -0.014599915593862534, 0.011475996114313602, 0.011997119523584843, -0.02363402582705021, 0.023005418479442596, 0.0016288054175674915, 0.026090551167726517, -0.009512053802609444, ...
[ 0.5797576308250427, -0.18669261038303375, 0.4154234230518341, 0.24906133115291595, 0.357504665851593, 0.4519062340259552, -0.1110665574669838, -0.43986716866493225, -0.3330022394657135, -0.28779271245002747, -0.06107531487941742, 0.5223652124404907, -0.1703377217054367, 0.1729859858751297,...
I'm trying to align the baseline of the first line of a list environment with the baseline of a box. After working on this in various forms, I thought I was ready to tackle this issue. I also thought I was making good progress. But testing my approach in various situations has revealed that I'm still missing something. I'm feeling a bit sheepish about asking yet another question about lengths and vertical alignment. But after working on this over the past 24 hours, I feel I'm out of ideas of what to do next. My MWE consists of three parts: 1. A file, `newlist.tex`, which defines a new environment to facilitate aligning a listing environment with a box's baseline, 2. A file, `showandtell.tex`, which contains several macros designed to help show the heights and values involved, 3. And the file to wrap around the MWE. The file `newlist.tex` \makeatletter %% A box to line up against \newsavebox{\aeQuickBox} \def\aeQuick{\savebox{\aeQuickBox}{Tq \rule[-2ex]{0.1pt}{3ex}jay}%%' \usebox{\aeQuickBox}%%' \setlength{\dp\aeQuickBox}{\dp\aeQuickBox}} %% THE STRUT \newlength{\ae@strut@ht} \setlength{\ae@strut@ht}{0.7\baselineskip} \newcommand{\aestrut}{\makebox[0pt][r]{\rule{2pt}{\ae@strut@ht}}} %% MY LIST %% new lengths created for access outside of environment \newlength{\ae@enum@topsep} \newlength{\ae@enum@partopsep} \newlength{\ae@enum@parsep} \newlength{\ae@enum@parskip} \newenvironment{alignedenum} {\begingroup \begin{list} {\bfseries\textbullet} {%% horizontal dimensions \setlength{\labelwidth}{1.5em}%%' \setlength{\labelsep}{0.5em}%%' \setlength{\itemindent}{0em}%%' \setlength{\leftmargin}{\dimexpr \wd\aeQuickBox + 0.5em +\labelwidth +\labelsep -\itemindent \relax}%%' %% vertical dimensions %% test whether `\partopsep` has been added \ifvmode \global\setlength{\ae@enum@partopsep}{\partopsep}%%' \else \global\setlength{\ae@enum@partopsep}{0pt}%%' \fi %% next is true only if this environment is within another list \ifnum\@listdepth>1\relax \global\setlength{\ae@enum@topsep}{\topsep}%%' \else \global\setlength{\ae@enum@topsep}{0pt} \fi \global\setlength{\ae@enum@parsep}{\parsep}%%' }%%' %% set \ae@enum@parskip here because wrong value of `\parskip` is %% called within arguments to `list` environment. \global\setlength{\ae@enum@parskip}{\parskip}%%" %% avoid `\lineskip` if being aligned with a very deep box \nointerlineskip \myremovevspace } {\end{list}\endgroup} %% VERTICAL SPACE REMOVAL \newcommand{\myremovevspace} {%%\par\nointerlineskip \vspace{%%' -\dimexpr \ae@strut@ht +\ae@enum@topsep +\ae@enum@partopsep +\ae@enum@parsep +\ae@enum@parskip +\dp\aeQuickBox \relax}} \makeatother The file `showandtell.tex` \usepackage{xcolor} \makeatletter %% cumulative lengths for stacking and making lengths visible \newlength{\ae@tmp@dim@a} \newlength{\ae@tmp@dim@b} \newlength{\ae@tmp@dim@c} \newlength{\ae@tmp@dim@d} %% COLORS %% orange = strut height %% blue = topsep %% gray = parskip + parsep %% red = aeQucikBox depth \newcommand{\aeStackLengths} {\bgroup %%'---------------------------------------------------------------------- \setlength{\ae@tmp@dim@a}{0pt}%%%' \setlength{\ae@tmp@dim@b}{\dimexpr\ae@tmp@dim@a+\ae@strut@ht\relax}%%' \setlength{\ae@tmp@dim@c}{\dimexpr\ae@tmp@dim@b+\ae@enum@topsep}%%' \setlength{\ae@tmp@dim@d}{\dimexpr\ae@tmp@dim@c+\dimexpr\ae@enum@parskip+\ae@enum@parsep+\ae@enum@partopsep}%%' %%'---------------------------------------------------------------------- \color{orange!80}%%' \raisebox{\ae@tmp@dim@a}[0pt][0pt]{\makebox[0pt][l]{\rule{3pt}{\ae@strut@ht}}}%%' \color{blue!70}%%' \raisebox{\ae@tmp@dim@b}[0pt][0pt]{\makebox[0pt][l]{\rule{3pt}{\ae@enum@topsep}}}%%' \color{gray!50}%%' \raisebox{\ae@tmp@dim@c}[0pt][0pt]{\makebox[0pt][l]{\rule{3pt}{\dimexpr\ae@enum@parskip+\ae@enum@parsep+\ae@enum@partopsep}}}%%' \color{red}%%' \raisebox{\ae@tmp@dim@d}[0pt][0pt]{\makebox[0pt][l]{\rule{3pt}{\dp\aeQuickBox}}}%%' \egroup } %%-------------------------------------------------------------------------- %% VISUALLY REP FOR DIMS %%-------------------------------------------------------------------------- %% show baseline \newcommand{\aeshowbaseline}{\makebox[0pt][l]{\color{blue}\rule{0.1pt}{1ex}\rule{2in}{0.1pt}}} %% show depth \newcommand{\aeshowdepth}{\makebox[0pt][r]{%%' \color{red}%%' \rule[-\dp\aeQuickBox]{2.25in}{0.1pt}%%' \rule[-\dp\aeQuickBox]{0.1pt}{\dp\aeQuickBox}%%' \makebox[0pt][l]{\rule[-\dp\aeQuickBox]{1.5in}{0.1pt}}%%' }} %%-------------------------------------------------------------------------- %% SHOW VALUES FOR DIMS %%-------------------------------------------------------------------------- \usepackage{pgffor} \newcommand\aeshowbox[1] {\begin{minipage}[t]{2in}\tiny\ttfamily \foreach \x in {#1}{ \makebox[1cm][r]{\x} = \the\csname \x\endcsname\\ } \end{minipage}} \def\aeshowLengths{\aeshowbox{topsep,partopsep,parsep,parskip,%%' ae@enum@topsep,ae@enum@partopsep,ae@enum@parsep,ae@enum@parskip}} \makeatother The wrapper file: \documentclass{article} \input{newlist} \input{showandtell} \usepackage{lipsum} \begin{document} \noindent \aeshowbaseline\aeQuick \par \begin{alignedenum} \item \aestrut\aeStackLengths\aeshowbaseline\, My first line: \aeshowLengths \item second line \end{alignedenum} \begin{enumerate} \item \aeshowbaseline\aeQuick \begin{alignedenum} \item \aestrut\aeStackLengths\aeshowbaseline\, My first line: \aeshowLengths \item second line \end{alignedenum} \end{enumerate} \begin{enumerate} \item \verb=\par= inserted \begin{enumerate} \item \aeshowbaseline\aeQuick \par \begin{alignedenum} \item \aestrut\aeStackLengths\aeshowbaseline\, My first line: \aeshowLengths \item second line \end{alignedenum} \end{enumerate} \end{enumerate} \begin{enumerate} \item No \verb=\par= inserted \begin{enumerate} \item \aeshowbaseline\aeQuick \begin{alignedenum} \item \aestrut\aeStackLengths\aeshowbaseline\, My first line: \aeshowLengths \item second line \end{alignedenum} \end{enumerate} \end{enumerate} \end{document} These produce: ![enter image description here](http://i.stack.imgur.com/oHDF9.png) As seen from the image, everything seems to be working fine until I've buried my new environment within two other lists. I don't understand how I've not taken into account all the space. I'm also finding it difficult to locate the documentation for how the various parameters are set at each list level. I can't find much of anything useful in `source2.pdf`.
[ 0.0002346578985452652, 0.009601708501577377, -0.007590140216052532, 0.006983937695622444, 0.005860987119376659, 0.015292515978217125, 0.0037091851700097322, 0.01674535870552063, -0.011616314761340618, 0.005719403270632029, -0.0029189782217144966, 0.008168196305632591, 0.004131676629185677, ...
[ 0.7655162811279297, 0.08204780519008636, 0.3781522810459137, -0.1413852423429489, -0.07380344718694687, 0.09124608337879181, 0.03343161195516586, -0.5493816137313843, -0.2213742733001709, -0.8432143926620483, 0.14554981887340546, 0.43407031893730164, -0.09350139647722244, 0.516694843769073...
How do I use skill points in Mafia Wars? Is there a recommended ratio for energy/ stamina/ attack/ defense/ health?
[ -0.009993660263717175, 0.03643236309289932, 0.025160379707813263, -0.022848911583423615, 0.020717140287160873, 0.0257467832416296, 0.02082030475139618, -0.0425146110355854, -0.02133096754550934, 0.03481481969356537, -0.031894706189632416, 0.04917212203145027, -0.021374400705099106, 0.02610...
[ 0.9046189785003662, -0.1906113624572754, 0.25873667001724243, 0.1686544418334961, 0.23409351706504822, 0.052351225167512894, -0.0798315480351448, -0.6502600312232971, -0.05681482329964638, -0.31837528944015503, 0.38221630454063416, 0.7824157476425171, -0.08039838820695877, -0.4153604805469...
Quotation from The Pastyme of Plasure Especially,I don't know "sprynge, ryall, chefe and orygynal". > "O Mayster Lydgate! the most dulcet sprynge > Of famous rethoryke, with balade ryall > The chefe orygynal." > _—"The Pastyme of Plasure," by Stephen Hawes, 1509._ ![enter image description here](http://i.stack.imgur.com/Sny0O.png)
[ 0.014526420272886753, 0.01656358130276203, -0.018638193607330322, 0.008437956683337688, -0.0030264928936958313, 0.00334219541400671, 0.011488882824778557, -0.024154189974069595, -0.013896162621676922, 0.030127305537462234, -0.005565481260418892, 0.008811616338789463, -0.010901892557740211, ...
[ -0.44660571217536926, 0.2909615933895111, 0.07652273029088974, -0.45046547055244446, -0.10849129408597946, 0.8437833786010742, 0.14182858169078827, 0.04203237593173981, -0.3915991485118866, -0.2686806917190552, -0.17825467884540558, 0.04061432182788849, -0.28001052141189575, -0.00795860774...
I have a small 'brochure' type site with 6 pages, i can see them all the pages showing up in google when i search for my site. But in webmaster tools under the sitemaps section it says 6 submitted, (the blue bar of the graph), but the indexed pages - the red bar is showing 0 indexed pages, even though they seem to be indexed ? any idea why this is ? I dont really think its that important as the pages are still indexed, but it just seems odd. * * * UPDATE 9/3/12 having just looked in google webmaster its showing that there are 11 pages indexed, under the health > index status tab.. but under the optimization sitemap tab it shows 6 urls submitted but only 1 indexed ? please see images bellow index status: ![enter image description here](http://i.stack.imgur.com/n9CSd.png) Sitemap status: ![enter image description here](http://i.stack.imgur.com/TpnsF.png)
[ -0.013050851412117481, -0.005337405018508434, -0.0050953952595591545, 0.0170801542699337, -0.01127618458122015, 0.00009872112423181534, 0.0075445096008479595, -0.01737631857395172, -0.015896689146757126, -0.012804991565644741, -0.003254127223044634, 0.0044560679234564304, -0.0182403512299060...
[ 0.6262438893318176, 0.2690483629703522, 0.6768009662628174, 0.24944090843200684, -0.2120426446199417, -0.19303876161575317, 0.19947734475135803, -0.1085193008184433, -0.19129431247711182, -0.5221636295318604, 0.2100413292646408, 0.3789091408252716, 0.028303245082497597, -0.0591294877231121...
How I can define a new command to use the old-style numbers of the package `gfsartemisia` within the default font in pdflatex ? I tried something like: \newcommand{\oldnum}[1]{{\fontfamily{artemisia}\oldstylenums{#1} \selectfont}} in the preamble to use Some text with default font \oldnum{123} more text with default font but doesn't work.
[ 0.03246762976050377, 0.016629090532660484, -0.009323709644377232, 0.018525490537285805, 0.020764566957950592, -0.025656726211309433, 0.011833151802420616, -0.035569462925195694, -0.019974511116743088, 0.008000999689102173, -0.011593617498874664, 0.0027287660632282495, -0.0044381930492818356,...
[ 0.12472313642501831, -0.0485040582716465, 0.6949223279953003, 0.14140565693378448, -0.20208661258220673, -0.18728552758693695, 0.24281036853790283, -0.2908805310726166, 0.11518760025501251, -0.563069224357605, 0.12785987555980682, 0.6522104144096375, -0.4454537630081177, -0.065838403999805...
One of the things I enjoy about the Saints Row series is the side activities: Insurance Fraud and that kind of thing. I'd like to play more of them, but right now I only have activities unlocked in one area, the central Morning Star-controlled area. (Nine of sixteen are currently unlocked.) The other parts of town have 0 of 15, 0 of 13, and 0 of 12 unlocked. What do I have to do to unlock activities in other parts of town? I've completed 9 missions overall: the last one was Trafficking. Do I have to complete more missions or more activities? Do I have to reach a certain level of respect? Can someone help a fellow Saint out?
[ 0.0025947722606360912, 0.007950172759592533, -0.004647672642022371, -0.004087792243808508, 0.002648312831297517, 0.014942951500415802, 0.008522173389792442, 0.002694827038794756, -0.017477117478847504, 0.018282778561115265, -0.015792585909366608, 0.008745855651795864, -0.024288661777973175, ...
[ 0.12685205042362213, 0.09811583906412125, 0.33563661575317383, 0.11741899698972702, 0.09369329363107681, 0.055465757846832275, 0.7655021548271179, 0.42833763360977173, -0.5312315821647644, -0.49464303255081177, 0.28596919775009155, 0.5640750527381897, -0.28844738006591797, -0.0512097589671...
<ul class="list"> <?php $paged = (get_query_var('paged')) ? get_query_var('paged') : 1; $args = array( 'category' => '8', 'orderby' => 'post_date', 'order' => 'DESC', 'post_type' => 'post', 'post_status' => 'publish', 'posts_per_page' => 9, 'paged' => $paged, ); $postslist = get_posts( $args ); foreach ($postslist as $post) : setup_postdata($post); ?> <li> <h2><a href="<?php the_permalink() ?>"><?php the_title(); ?></a></h2> </li> <?php endforeach; ?> </ul> <?php wp_paginate(); ?> <?php wp_pagenavi(); ?> <?php next_posts_link() ?> <?php previous_posts_link() ?> The code above successfully list the most recent 9 posts, however none the pagination works. I have installed WP-Paginate and WP-PageNavi plugins for testing.
[ -0.003213044023141265, 0.021469397470355034, 0.005285643972456455, 0.008541608229279518, 0.008277980610728264, -0.0026175195816904306, 0.006628384813666344, 0.011659517884254456, -0.007577816024422646, 0.015331972390413284, -0.01054423302412033, 0.005914382636547089, -0.01264098845422268, ...
[ -0.4213152229785919, -0.1688496321439743, 0.8992218971252441, -0.3268548250198364, -0.38768336176872253, 0.7129770517349243, -0.018293526023626328, -0.3360403776168823, -0.20928023755550385, -0.6028940677642822, -0.26745498180389404, 0.45739755034446716, -0.42207422852516174, 0.26881206035...
Can this template at Wikipedia be true? It seems to suggest that Einstein- Cartan theory, Gauge theory gravity, Teleparalleism and Euclidean Quantum Gravity are fully compatible with observation! It also suggests that Loop Quantum Gravity and BEC Vacuum Theory among others, are experimentally constrained whereas string theory/M theory are disputed! What I understand by "Fully compatible with observation" is that all its predictions are confirmed by experiments and it has been found to be more accurate than General Relativity. Has such evidence really been found? Or am I misinterpreting "Fully compatible with observation"? Maybe it means it has been tested only when it reduces to General Relativity? But if that where the case, shouldn't M theory/String theory also be listed under "Fully Compatible" since their predictions also go down to Classical General Relativity at the low-energy, classical limit, if all other forces (other than gravity?) are gotten rid off? What I understand by "Experimentally constrained" is that it is true given certain modifications. However, as far as I know, Loop Quantum Gravity violates Lorentz symmetry and has thus been experimentally "excluded" while BEC Vacuum theory isn't even mainstream? What I understand by "Developmental/Disputed" is that it is still undergoing development OR it has **almost** been experimentally proven wrong but it is still not settled in mainstream physics. If LQG doesn't go to the excluded section, it should at least come here? Since the violation of Lorentz symmetry has been disproven according to this. So my question is "Is this template really reliable?".
[ -0.018064282834529877, 0.022477280348539352, -0.005591236986219883, 0.011598465964198112, 0.03272590786218643, -0.012231018394231796, 0.00953653734177351, -0.016454841941595078, -0.011280848644673824, -0.02800372987985611, -0.007405529264360666, 0.017224445939064026, -0.013067563064396381, ...
[ 0.5836189985275269, 0.12155649065971375, 0.11353344470262527, 0.4963049590587616, -0.22716568410396576, -0.3861129581928253, 0.16710607707500458, -0.0061068544164299965, -0.2430395483970642, -0.3408103287220001, 0.08040440082550049, 0.3803614675998688, -0.17409387230873108, 0.6256279945373...
Let's say I light a wall with two spotlights: One red and one green one. Where they overlap, I'll see a yellow area at the wall. My question is, whether this is caused by an modification of the frequency/wavelength or simply by my eye combining the two incoming lights. Light is "added", wavelength is modified: ![enter image description here](http://i.stack.imgur.com/6DOCu.png) The eye combines two separate lights: ![enter image description here](http://i.stack.imgur.com/bdPjA.png)
[ -0.02172243595123291, 0.010497244074940681, 0.00080257433000952, 0.0033710391726344824, -0.01389558706432581, -0.023805109784007072, 0.006467107683420181, 0.019846569746732712, -0.0077181593514978886, 0.011445444077253342, -0.013355249539017677, 0.018728282302618027, -0.0020798020996153355, ...
[ 0.758787989616394, -0.02310454100370407, 0.30330711603164673, -0.2996175289154053, -0.13428108394145966, -0.018552768975496292, 0.2178107053041458, -0.15458577871322632, -0.4736518859863281, -0.46252089738845825, 0.04342764616012573, 0.4570766091346741, -0.21664492785930634, -0.03480553254...
**Short Question** Is there a typical way to name 'public' and 'private' members of an OO C project? **Background** I fully understand that public and private members do not _really_ exist in the C language. However, like most C programmers, I still treat members as public or private to maintain the OO design. In addition to the typical OO methods I have found my self following a pattern (see example below) that makes it easier for me to distinguish which methods are _meant_ for the outside world vs the private members that may have fewer checks / are more efficient etc... Does a standard or best practice exist for such a thing or is my example below a good way to approach this? **Example Header** #ifndef _MODULE_X_H_ #define _MODULE_X_H_ bool MOD_X_get_variable_x(void); void MOD_X_set_variable_x(bool); #endif /* _MODULE_X_H_ */ **Example Source** // Module Identifier: MOD_X #include "module_x.h" // Private prototypes static void mod_x_do_something_cool(void); static void mod_x_do_something_else_cool(void); // Private Variables static bool var_x; // Public Functions - Note the upper case module identifier bool MOD_X_get_variable_x(void) {return var_x;} void MOD_X_set_variable_x(bool input){var_x = input;} // Private Functions - Note the lower case module identifier void mod_x_do_something_cool(void){ // Some incredibly cool sub routine } void mod_x_do_something_else_cool(void){ // Another incredibly cool sub routine }
[ -0.0028559300117194653, 0.01257333904504776, -0.013573229312896729, 0.018218159675598145, -0.011244612745940685, 0.0076883588917553425, 0.006727498024702072, 0.013238340616226196, -0.01688835211098194, 0.01217544823884964, 0.003189413109794259, 0.019970756024122238, 0.02679634839296341, 0....
[ 0.7031237483024597, 0.3633716106414795, -0.2144962102174759, -0.009884781204164028, 0.0024617116432636976, -0.36089807748794556, 0.36063694953918457, 0.2729676067829132, -0.38421982526779175, -0.16141322255134583, -0.19611994922161102, 0.31954383850097656, -0.0026292179245501757, -0.063715...
I would like to remove the leading space that appears on the left before the equation on the latex below. \documentclass[preview]{standalone} \usepackage{amsmath} \usepackage{bm} \usepackage{varwidth} \usepackage{amsfonts} \usepackage{yfonts} \usepackage[T1]{fontenc} \usepackage{amssymb} \begin{document} \begin{varwidth}{\linewidth} \large \begin{align*} &{\bf y}_{l} = \sum _{k=1}^{L} {\bf H}_{l,k} {\bf x}_{k}+ {\bf w}_{l} &{{(1)}} \end{align*} \end{varwidth} \end{document}
[ 0.01538706012070179, 0.006016760598868132, -0.007713861297816038, 0.01408969983458519, -0.01101910974830389, 0.008270220831036568, 0.006743657402694225, 0.012106507085263729, -0.0071642412804067135, 0.011513330042362213, -0.012592807412147522, 0.006730967201292515, -0.012864705175161362, 0...
[ -0.4411158561706543, 0.25266602635383606, 0.5012724995613098, -0.19992001354694366, 0.30316224694252014, -0.1517445296049118, -0.014301279559731483, -0.04602827504277229, 0.11981195956468582, -0.6177623867988586, -0.1953754872083664, 0.4163891077041626, -0.2931860685348511, 0.0786903873085...
There are reports that `remove_filter` does not work under some circumstances, and that we should provide alternatives. However, the article does not really make it clear under which circumstances this happens. I am looking for code examples that will break `remove_filter`, with WordPress and PHP version, and eventually other relevant info, provided. I think the following snippet should be a useable template: <?php include('wp-load.php'); function filtertest_function($value) { return 'Filtered'; } var_dump(apply_filters('filtertest', 'Original value')); add_filter('filtertest', 'filtertest_function'); var_dump(apply_filters('filtertest', 'Original value')); remove_filter('filtertest', 'filtertest_function'); var_dump(apply_filters('filtertest', 'Original value')); This will return the expected: string(14) "Original value" string(8) "Filtered" string(14) "Original value" I believe the error comes up in some cases where the filter is part of a class, or multiple filters are used. I understand `_wp_filter_build_unique_id` is involved. Please point out where in that code the source of the error is.
[ 0.018783317878842354, 0.018563294783234596, -0.004424624145030975, 0.02458696998655796, -0.0108247185125947, 0.011667352169752121, 0.006288203410804272, 0.0175446979701519, -0.016691014170646667, 0.008499120362102985, -0.012819314375519753, 0.013385279104113579, -0.014093417674303055, 0.00...
[ 0.6046984195709229, -0.07264413684606552, 0.27197661995887756, -0.10197776556015015, -0.06659578531980515, -0.3664487302303314, 0.48096513748168945, -0.08451031893491745, -0.08167938143014908, -0.5901182293891907, -0.12372244894504547, 0.44577306509017944, -0.5799673199653625, 0.1041255816...
we have ran an experiment where we compared three interaction techniques for a 3d docking task. So we had two factors: the aforementioned technique type and a factor representing the direction translation (i.e.: if participants had to move an object that appeared close to their viewpoint and move it in depth or vice versa). Each trial was repeated 5 times. Some of those trials were skipped because of the difficulty. If I run a regular repeated measures anova then each participant that even a single missing value will be dropped from the analysis. This means that I'd have to remove more than half of the participants. By reading around it seems I can use a linear mixed model instead. My doubt is, can I use a mixed model for this type of situation? I am confused as to whether mixed models are only relevant when you have a between-subjects factor such as the classic treatment/control groups. In my case every participant was subjected to the same conditions. There were no between- subjects factors. I ran the mixed model analysis by using Technique, Direction and Repetition as _Repeated_ , the ID of each participant as the _subject_ and technique and direction as fixed factors. Are my assumptions correct or did I do a terrible mistake? If so, what alternatives do I have when dealing with missing values? Thanks!
[ -0.0019715232774615288, 0.020851098001003265, -0.0054975152015686035, 0.022808922454714775, -0.008347723633050919, -0.005528507754206657, 0.007753528188914061, -0.005795851815491915, -0.010264935903251171, 0.003701299661770463, -0.013141661882400513, 0.015659982338547707, 0.00111479638144373...
[ 0.01128574088215828, -0.0571475513279438, 0.4887724220752716, 0.06560993939638138, -0.12749682366847992, 0.4165930449962616, 0.29994022846221924, -0.39998865127563477, -0.44924843311309814, -0.33421069383621216, 0.2069980502128601, 0.3298249840736389, -0.12238642573356628, -0.0401696339249...
This question is not the same as my last one. How do you find the $n$-th derivative where $n$ is a variable? For example, you can find the nth derivative for a specific $n = 3$ D[Log[1 + x], {x, 3}] but how do you get _Mathematica_ to show the $n$-th derivative for $n$ as a general variable? For example, from Wolfram Alpha
[ 0.011764595285058022, 0.002493432257324457, -0.025372348725795746, -0.002867721486836672, -0.01781981997191906, -0.005465792492032051, 0.006155389826744795, -0.002654122421517968, -0.023961534723639488, -0.011000786907970905, 0.003339821705594659, -0.0004132761969231069, -0.01420971658080816...
[ -0.1074674129486084, 0.28146690130233765, -0.043563615530729294, -0.00502760661765933, -0.08026950061321259, -0.07896154373884201, -0.1295185387134552, -0.283631831407547, 0.023615902289748192, -0.3426534831523895, 0.0953320562839508, 0.5800819396972656, -0.39274248480796814, 0.28155866265...
I'm trying to transfer a file between two servers and i'm getting different errors. **Option 1: Logged at OLDSERVER via SSH** > scp file.tar.gz root@IPADDRESS:/var/www/. The error in this case is > /usr/bin/ssh: no such file or directory **Option 2: Logged at new server via SSH** > scp OLDUSER@OLDURL:/var/htdocs/file.tar.gz /var/www/ The error in this case is > ssh: connect to host OLDURL port 22: Connection refused Do you know what could be the problem?
[ -0.00774266105145216, 0.003448406234383583, 0.0026886598207056522, 0.02039295807480812, 0.01532457023859024, 0.01131410337984562, 0.009548744186758995, -0.03265450894832611, -0.021889880299568176, -0.016440026462078094, -0.015137377195060253, 0.009004151448607445, -0.008735626935958862, 0....
[ 0.0694575384259224, 0.024379640817642212, 0.3751027584075928, -0.03660156577825546, 0.11816712468862534, 0.3851845860481262, 0.5410919785499573, -0.016910158097743988, -0.3911302387714386, -1.0616974830627441, -0.12083996087312698, 0.3322950005531311, -0.38046035170555115, -0.0146745471283...
If A is inside B, it might be said that the size of B constricts the maximum size of A. What is the word to describe A's limiting effect on the minimum size of B? What I'm trying to describe is how an increase in size of A affects B. For added bonus, a better word than 'constricts' is also useful.
[ 0.001530705951154232, 0.010977276600897312, -0.015891099348664284, 0.029453638941049576, 0.019748318940401077, -0.01681796833872795, 0.011263465508818626, -0.0224803127348423, -0.018260527402162552, -0.013864763081073761, -0.011483187787234783, 0.012551813386380672, 0.006183546502143145, 0...
[ 0.1714470088481903, -0.07158993929624557, 0.12247352302074432, -0.29193368554115295, -0.09827779233455658, -0.0008850029553286731, 0.28405508399009705, -0.3260557949542999, -0.4126865863800049, -0.4033467769622803, 0.009611906483769417, 0.19204315543174744, -0.3525274097919464, 0.445393681...
I am looking for papers, preferably in economics, where simple OLS is used. The reason is that to convey the workings of linear regression, I want to have real data from real papers. Most research uses simultaneous models or other bells and whistles. Also few researchers release their data. I could generate my own data, but as I said, I want to have the whole story of the data and research behind it. Wooldridge (Introductory Econometrics) does a lot of simple analyses from past papers, but as far as I know, these are just simplified version of the real models.
[ 0.026544328778982162, 0.01763647235929966, -0.014590926468372345, 0.010313550941646099, -0.012131256982684135, 0.000661578553263098, 0.007498506456613541, -0.00020127743482589722, -0.011268310248851776, -0.01860254444181919, -0.0039829071611166, 0.01368942391127348, -0.005049349740147591, ...
[ 0.3196541965007782, 0.16327917575836182, -0.10093989968299866, 0.11069844663143158, -0.24758322536945343, -0.09444117546081543, 0.18394121527671814, 0.19270767271518707, -0.057979632169008255, -0.109367236495018, -0.015966711565852165, 0.3411030173301697, 0.0767432153224945, 0.187389001250...
I have got a problem with an abstract. I created a PHD Thesis in Latex. I use _thesis_ document style. I want abstract to be doublespaced. I put that code to obtained it: {\doublespacing \newpage\ifpdf\pdfbookmark[1]{Acknowledgement}{label:ack}\fi\input{acknowledge.tex} \newpage\ifpdf\pdfbookmark[1]{Abstract}{label:abst}\fi \input{abstract.tex} } Unfortunately, everything is doublespaced, except the end of the abstract (see picture): ![The interline attribute is changed in the end of the abstract](http://i.stack.imgur.com/fLg4J.jpg) Does anybody has any suggestions? Regards.
[ 0.0015540902968496084, 0.006807927042245865, -0.006399739999324083, 0.018594060093164444, 0.0011404547840356827, 0.01770015060901642, 0.007502540480345488, 0.0014584683813154697, -0.012084666639566422, 0.001512786140665412, -0.000793911749497056, -0.0009384970180690289, -0.013999145478010178...
[ 0.018917441368103027, 0.22583457827568054, 0.539304792881012, 0.061774466186761856, -0.3959611654281616, -0.17757590115070343, -0.06734810769557953, -0.20561078190803528, -0.30235201120376587, -0.2943844199180603, 0.15707293152809143, 0.4186881184577942, -0.452239066362381, -0.063057161867...
> **Possible Duplicate:** > When is "L" doubled? If I am using the word _model_ in the context of financial models and the UK, then which of these words should I use? Also, are there any key difference in the meanings of the two?
[ 0.009081011638045311, 0.014253312721848488, -0.004982940852642059, 0.025054194033145905, -0.031849801540374756, 0.006959743797779083, 0.010618496686220169, 0.023299021646380424, -0.02507965825498104, -0.03901034966111183, -0.007078845519572496, 0.012007893063127995, -0.01200666930526495, -...
[ 0.3162161707878113, -0.0380336195230484, 0.36274781823158264, -0.39402613043785095, -0.014412324875593185, 0.3732149302959442, -0.2810797393321991, 0.15798528492450714, -0.4679573178291321, -0.6766372323036194, 0.0072050197049975395, 0.4895555377006531, -0.11693892627954483, -0.10992203652...
I've got an alias set up to launch my text editor in a way that keeps it local to the specific desktop I'm working on in Gnome Shell: alias geany="geany --socket-file=/tmp/geany-sock-$(xprop -root _NET_CURRENT_DESKTOP | awk '{print $3}') ${1+"$@"}" I'd like to emulate this with the actual Gnome Shell launcher. As it stands, I have the launcher command set to: geany --socket-file=/tmp/geany-sock-$(xprop -root _NET_CURRENT_DESKTOP | awk '{print $3}') %F Which will point to the open Geany session in the current window if one is already open, but it won't launch a new session. I'll get a spinner, and then it just quits rather than launching the app. Is there any way to do this?
[ -0.004862971603870392, 0.006723546888679266, -0.0019458949100226164, 0.012246823869645596, -0.02806021273136139, -0.0085399579256773, 0.00789966993033886, 0.004926631227135658, -0.01383689884096384, 0.01586061716079712, -0.0060408213175833225, -0.00239450391381979, 0.007727089338004589, 0....
[ 0.118752121925354, 0.19212225079536438, 0.2468920648097992, -0.239462748169899, -0.21558871865272522, 0.10154049843549728, 0.4945806562900543, 0.1218595951795578, -0.14262337982654572, -0.7008410096168518, 0.010892477817833424, 0.7476772665977478, -0.2632963955402374, -0.03869597241282463,...
I am currently running a DV server with MediaTemple and would like to have my domains: abc.com and cba.com mapped to the subdomains I set up with the WP network one.mydomain.com and two.mydomain.com. The goal is keep the urls: abc.com and cba.com viewable in the browser but the zone files are just referencing the subdomains I have set up. Sorry if this kinda of confusing, it is a major headache. Any help would great!
[ -0.001644784933887422, 0.015950357541441917, 0.00028954504523426294, 0.022933585569262505, 0.010311913676559925, 0.018147524446249008, 0.011317047290503979, 0.015635639429092407, -0.01974066160619259, -0.012133548967540264, -0.015695197507739067, 0.0016880676848813891, -0.022502591833472252,...
[ 0.8973081111907959, 0.3292696177959442, 0.4932389557361603, -0.13239024579524994, -0.05915812775492668, -0.2555745244026184, 0.18924269080162048, 0.15123341977596283, -0.37291207909584045, -0.8163592219352722, 0.31302332878112793, 0.057100486010313034, -0.1630362868309021, 0.72008514404296...
Functions such as `GraphDistanceMatrix[]` or `AdjacencyMatrix[]` are available without invoking Needs["GraphUtilities`"] . Why isn't the whole package available when _Mathematica_ starts? Are the functions different after having loaded `GraphUtilities`?
[ 0.004913076758384705, 0.008135942742228508, 0.000009246756235370412, 0.018801333382725716, -0.01361128594726324, -0.017811348661780357, 0.010582298040390015, 0.013036617077887058, -0.017506273463368416, -0.005538399796932936, -0.015495809726417065, 0.016738329082727432, -0.009257434867322445...
[ -0.24104264378547668, -0.2616093158721924, 0.4058297872543335, 0.381336510181427, -0.12429766356945038, -0.3129655718803406, 0.019584838300943375, 0.06712029129266739, -0.20530688762664795, -0.5144727826118469, -0.1300250142812729, 0.5851854681968689, -0.5299212336540222, -0.16251035034656...
I went back to Skyrim after a long time off to check out Hearthfire. I read that one can construct a new house, protect it from baddies, etc. However, the only thing I've been able to do is modify one of my old, existing houses to swap out an alchemy lab for a kid's bedroom, and adopt a kid. Is that it? I suspect I may have another quest or two to go, but I don't see it in my quest log (which is still somewhat full...) What am I missing?
[ 0.0184269268065691, 0.02620154805481434, -0.0035455129109323025, -0.007287175860255957, -0.006722689140588045, 0.0018312577158212662, 0.006433435715734959, -0.0032983210403472185, -0.019271813333034515, 0.009906225837767124, -0.00690917344763875, 0.009700242429971695, -0.007566083688288927, ...
[ 0.7638663649559021, 0.24612727761268616, 0.18155375123023987, 0.0023826591204851866, -0.046699944883584976, -0.27132076025009155, 0.19426481425762177, -0.15055444836616516, -0.6052024960517883, -0.42815515398979187, 0.13285088539123535, -0.38234108686447144, 0.2343025952577591, 0.360819995...
I'm new to LaTeX and trying to use `hyperref`, and having trouble with `\usepackage{hyperref}`. As recommended in posts, I included `\usepackage{hyperref}` as the last `\usepackage{}` instruction. No other problem when I comment out `\usepackage{hyperref}`. I've successfully compiled the testams.tex document that came with the `hyperref` download from TUG and uses \usepackage[% verbose, colorlinks=true, naturalnames=true, linkcolor=blue, ]{hyperref} I am receiving this error: !Latex Error: \RequirePackage or \LoadClass in Options Section This is my preamble and document structure: \documentclass[11pt,reqno]{report} \usepackage{geometry} \geometry{letterpaper} % ... or a4paper or a5paper or ... %\geometry{landscape} % Activate for for rotated page geometry \usepackage[parfill]{parskip} \usepackage{graphicx} \usepackage{amssymb} \usepackage{amsmath, amsthm} \usepackage{epstopdf} \usepackage[onehalfspacing]{setspace} \usepackage[comma,authoryear]{natbib} \usepackage{titling} \usepackage{url} \usepackage{caption} % Activate to format captions \captionsetup[table]{font=bf} % Make table captions bold %\usepackage[format=hang,font={small,bf}]{caption} %Makes all captions bold \usepackage{hyperref} \title{Title} \subtitle{Subtitle} \author{New Author} \date{\today} \begin{document} \maketitle \tableofcontents \chapter{Begin} \section{Continue} \begin{table}[htbp] \begin{center} \caption{Caption} \begin{tabular}{ccc} %table \end{tabular} \label{tab:first} \end{center} \end{table} \newpage \appendix \chapter{App 1} \section{App 1.1} Lots of tables. \pagebreak \begin{singlespace} \bibliographystyle{plainnat} \bibliography{bibliography} \end{singlespace} \end{document}
[ 0.007125796750187874, 0.0056932782754302025, -0.007285436149686575, 0.023290345445275307, 0.02297329716384411, 0.022084973752498627, 0.008680786937475204, -0.013486361131072044, -0.015712028369307518, -0.033932916820049286, 0.005376085173338652, -0.00712403142824769, -0.008692045696079731, ...
[ 0.2180107831954956, 0.2331894189119339, 0.40239307284355164, -0.24343013763427734, 0.4923398792743683, 0.0528070330619812, 0.718424379825592, -0.02446315437555313, 0.18583911657333374, -0.9000628590583801, -0.12674511969089508, 0.7828508615493774, -0.07557647675275803, -0.03638250753283501...
I want to meta-analyse some studies which I've already collected and review. However, the studies did not used control groups, due to the types of questions used. How can I do the meta-analysis or which software can I use? Briefly my problem is as follows: I reviewed studies which the goal was to evaluate the perception of certain samples of their own behaviour. For each study, the proportion of the sample which claims to adopt the behaviour in analysis, is equivalent to the effect size. Do you have any idea how can I determine the average effect size? Thank you
[ 0.012768992222845554, 0.0048402114771306515, -0.009999352507293224, 0.010385598056018353, -0.0023236956913024187, 0.00433074776083231, 0.005851873196661472, -0.008667878806591034, -0.012997867539525032, -0.01306711882352829, 0.004328913055360317, 0.006166489329189062, -0.013845962472259998, ...
[ 0.38335439562797546, 0.1522504687309265, 0.05164638161659241, -0.041346244513988495, -0.3157879114151001, 0.26609882712364197, 0.3057893216609955, -0.21589317917823792, -0.30695587396621704, -0.4988707900047302, 0.40788155794143677, 0.33525222539901733, -0.2815536558628082, 0.3038831353187...
As the title states - does a unique order exist for any _Mathematica_ list to be sorted by the `Sort[]` function (or as returned by a function that treats lists as sets)? While I'm pretty sure the answer is yes, the kinds of elements you can throw into a _Mathematica_ list seems so general and varied (numbers, strings, symbols, images, graphs, other lists, etc., etc.) that I thought to make sure. (The reason this question occurred to me is an exercise question asked in the book I just started learning _Mathematica_ from: Write a function `SubsetQ[list1, list2]` that checks whether `list1` is a subset of `list2`. My solution is SubsetQ[lis1_, lis2_] := Intersection[lis1, lis2] == Union[lis1, {}] (* thanks to Simon Woods' correction, and assuming Intersection and Union sort canonically *) but that implicitly assumes that the two lists on either side of the equality check will be sorted in the same order.)
[ 0.02403287962079048, 0.019180648028850555, -0.0019720576237887144, 0.014358500950038433, -0.00048200134187936783, -0.0005970560014247894, 0.005900152958929539, 0.012429743073880672, -0.01993354968726635, -0.0036835726350545883, -0.006349523551762104, 0.01004684530198574, 0.003229442983865738...
[ -0.05699336156249046, 0.3108585774898529, 0.19727902114391327, 0.20073190331459045, -0.13104380667209625, -0.10066557675600052, -0.06536008417606354, -0.1711488962173462, -0.5557631254196167, -0.24530069530010223, -0.1962825059890747, 0.04093154892325401, -0.20526912808418274, 0.1036821901...
I have this shell script to install ionCube loader: if [[ $(id -u) -ne 0 ]] ; then echo "Please run as root" ; exit 1 ; fi echo "Welcome to this script to install the ionCube loader on CentOS!" cd /usr/local/src wget http://downloads2.ioncube.com/loader_downloads/ioncube_loaders_lin_x86-64.tar.gz tar zxvf ioncube_loaders_lin_x86-64.tar.gz cd ioncube mkdir /usr/local/ioncube echo "Please specify which PHP version you are using (e.g.: 5.3)." read version cp ioncube_loader_lin_"$version".so /usr/local/ioncube path=$(php -i|grep php.ini | awk 'NR==2{print $5}') read path echo "zend_extension = /usr/local/ioncube/ioncube_loader_lin_"$version".so" >> "$path" When I try to run it on `CentOS 6.5 Final` with this PHP version (`php -v`): PHP 5.3.28 (cli) (built: Jun 23 2014 16:25:09) Copyright (c) 1997-2013 The PHP Group Zend Engine v2.3.0, Copyright (c) 1998-2013 Zend Technologies I get the following error: ./ioncube.sh: line 17: : File or folder doesn't exist Line 17: echo "zend_extension = /usr/local/ioncube/ioncube_loader_lin_"$version".so" >> "$path" The error says that `$path` doesn't exist. When I try to manually run the command of `$path` (`php -i|grep php.ini | awk 'NR==2{print $5}'`), the output is: `/usr/local/lib/php.ini` When I try to run line 17 but replace `$path` with the actual output of the command (of course I'll replace `$version` too): echo "zend_extension = /usr/local/ioncube/ioncube_loader_lin_"5.3".so" >> "/usr/local/lib/php.ini" The command succeeds. I don't understand what's wrong here.
[ -0.0013992716558277607, -0.013276856392621994, -0.01168795209378004, 0.017700523138046265, -0.024617234244942665, -0.0030870242044329643, 0.008114815689623356, -0.01892612874507904, -0.019535504281520844, 0.013679913245141506, -0.00647928798571229, 0.0009179241023957729, -0.00494895875453949...
[ 0.11729422956705093, 0.1724524199962616, 0.23118877410888672, -0.32058343291282654, -0.07580602914094925, -0.14286692440509796, 0.2629390060901642, -0.055497799068689346, 0.11385487020015717, -0.7020277380943298, -0.07746061682701111, 0.7565190196037292, -0.5310875773429871, 0.161551326513...
Does the Nexus 7 have a LED flash for its front facing camera? Can somebody point me to some documentation online that states whether this is true or not?
[ -0.014458760619163513, -0.027226431295275688, -0.012335835956037045, 0.010481679812073708, -0.03893010690808296, 0.02053019031882286, 0.011425230652093887, 0.05069107189774513, -0.034432221204042435, -0.019181450828909874, 0.014640591107308865, 0.02169180102646351, 0.01957331970334053, -0....
[ 0.335748553276062, -0.02193678729236126, 0.2279723584651947, 0.39392611384391785, 0.3298065960407257, -0.224412202835083, 0.3192955255508423, 0.34671106934547424, -0.44682422280311584, 0.05472641438245773, 0.13860388100147247, 0.46148157119750977, -0.5618126392364502, -0.3879952132701874, ...
I'm in a really bad situation with my ToC in the book that I write. Everything is fine until the 9th chapter. There's the "9", some space and then the name of the chapter. But on the 10th (or more) chapter, the "1" of the "10" is under the "9". So the "0" of the "10" is where there should be a space. As a result, there is no space between the number of the chapter and the name of the chapter. I expected that LaTeX would deal with it as it does with the bibliography entries. It leaves some space empty on the start of one digit numbers so they will be all aligned at the end. (For example, when we have [9] and [10], the 0 is under the 9) Let me show you my preamble as well \documentclass[12pt,a4paper]{book} \usepackage[english, greek]{babel} \usepackage[iso-8859-7]{inputenc} \usepackage{kerkis} \usepackage{amsmath} \usepackage{amssymb} \usepackage{amsfonts} \usepackage{amsthm} \usepackage{units} \usepackage{array} \usepackage{framed} \usepackage{extarrows} \usepackage[makeroom,Smaller]{cancel} \usepackage[pdftex]{graphicx} \usepackage[top=3.2cm, bottom=2.8cm, left=2.9cm, right=2.9cm,headsep=9mm]{geometry} \usepackage[Bjornstrup]{fncychap} \usepackage{fancybox} \usepackage{fancyhdr} \usepackage{colortbl} \usepackage{xcolor} \usepackage{footmisc} \usepackage{subfiles} \usepackage{float} \usepackage{slashbox} \usepackage[official]{eurosym} (Here there are some commands but they've got nothing to do with that) \begin {document} \frontmatter \tableofcontents \newpage \thispagestyle{empty} \mbox{} \newpage \mainmatter . . .
[ -0.0019963711965829134, 0.02103770524263382, -0.011748915538191795, 0.024765953421592712, 0.00572122260928154, 0.001963989809155464, 0.009390790946781635, -0.023893628269433975, -0.011669406667351723, 0.002536480315029621, -0.025757428258657455, 0.004714721813797951, -0.012428071349859238, ...
[ -0.05906374007463455, 0.23855969309806824, 0.8710160851478577, -0.09494195878505707, 0.1860647350549698, 0.1176617443561554, 0.11900751292705536, 0.5769615769386292, -0.5203099846839905, -0.40293335914611816, -0.22416847944259644, 0.5461762547492981, 0.09867293387651443, 0.1307938396930694...
I was reading somewhere that the (110) plane has the highest atomic density and is best for p-channel MOSFET performance. How high atomic density influence the performance of P-channel crystal?
[ -0.01462432648986578, 0.006845949217677116, -0.04574768990278244, 0.008791333064436913, -0.017534321174025536, -0.0913945883512497, 0.019052274525165558, 0.004904742352664471, -0.029010308906435966, -0.006585430819541216, -0.0014847108395770192, 0.025553856045007706, -0.02018178068101406, ...
[ 0.659856915473938, 0.27480143308639526, 0.15123985707759857, 0.24358238279819489, -0.2607777714729309, 0.07601151615381241, -0.08058392256498337, -0.3199949860572815, 0.017978550866246223, -0.5335769653320312, 0.08230116963386536, 0.3772946894168854, -0.04234589636325836, 0.465262264013290...
I was going through a lecture slides on maximum likelihood estimation (MLE) and bumped into something I couldn't understand. The lecturer was using MLE to estimate a function $$ y_i=Q(\alpha , \beta ,x_i). $$ Now during MLE estimation (estimating parameters $\alpha$ and $\beta$), the term $$ y_i-Q(\alpha ,\beta,x_i)\quad (1) $$ is directly plugged into Gaussian distribution equations as a random variable. $$ L=\prod\limits_{i=1}^n {\frac{1}{\sigma_i \sqrt{2 \pi}}} \exp{\left(-\frac{{(y_i-Q(\alpha,\beta,x_i))}^2}{2 \sigma_i^2}\right)} $$ Shouldn't $(1)$ be a result of the distribution with random variable as $x$, assuming that the variable $y$ has some noise present in it ?
[ 0.0012534437701106071, -0.005492980591952801, -0.0056901779025793076, -0.0024434244260191917, -0.014189095236361027, 0.006594888865947723, 0.007803613785654306, -0.016244061291217804, -0.008742631413042545, -0.026646530255675316, -0.001209464855492115, 0.002247244119644165, -0.00053744181059...
[ -0.10401806235313416, 0.05110697075724602, 0.2205089032649994, -0.25840824842453003, 0.09966012090444565, 0.5198525786399841, -0.21936212480068207, 0.141916885972023, 0.06584497541189194, -0.6478620767593384, 0.06137163192033768, 0.16599377989768982, -0.36728614568710327, 0.403239548206329...
Say I have a multivariable (several independent variables) regression that consists of 3 variables. Each of those variables has a given coefficient. If I decide to introduce a 4th variable and rerun the regression, will the coefficients of the 3 original variables change? More broadly: in a multivariable (multiple independent variables) regression, is the coefficient of a given variable influenced by the coefficient of another variable?
[ 0.026587873697280884, 0.024852439761161804, -0.005765016656368971, 0.015142505057156086, -0.004702575970441103, -0.019458962604403496, 0.010028612799942493, 0.003852038411423564, -0.016489846631884575, -0.035174306482076645, -0.005171754863113165, 0.02254186011850834, -0.006101250182837248, ...
[ 0.06430105119943619, 0.018407650291919708, 0.3220621645450592, -0.02338004671037197, -0.1790669858455658, 0.25305330753326416, 0.13577435910701752, -0.5789426565170288, -0.20660099387168884, -0.29237619042396545, 0.03392940014600754, 0.5309527516365051, -0.6272464990615845, 0.2802069783210...
I have a table on PostgreSQL 9.2 ("POSTGIS="2.0.1 r9979" GEOS="3.3.3-CAPI-1.7.4" PROJ="Rel. 4.8.0, 6 March 2012" LIBXML="2.8.0""). This is it: CREATE TABLE nearest_point ( id_nearest_point serial NOT NULL, id_iri_radartec integer, id_tramo integer, geom geometry(GeometryZ,4326), CONSTRAINT pk_nearest_point PRIMARY KEY (id_nearest_point) ); ![SELECT nearest_point](http://imgur.com/ipASL2i.png) I need a geometry that represents the point set union of the Geometries. Easy. SELECT id_tramo ,ST_Multi(ST_Union(np.geom)) AS geom FROM nearest_point AS np GROUP BY id_tramo; But I get this: ![Zoom out - 01](http://imgur.com/oAmyaEQ.png) "ST_Union GROUP BY" points in pink. And the other points? ![Zoom out - 02](http://imgur.com/8lpcrSU.png) The arrow indicates where are the points of the table nearest_point. Where is my error?
[ 0.004210826009511948, 0.007773453835397959, -0.011273454874753952, 0.0006604988593608141, 0.005090733058750629, 0.026673993095755577, 0.00578742939978838, 0.013502768240869045, -0.007645390462130308, -0.011134776286780834, -0.009730932302772999, 0.004130409564822912, -0.006051753181964159, ...
[ -0.4209005832672119, 0.17020928859710693, 1.0901172161102295, -0.09159959852695465, -0.1373516023159027, -0.11873795837163925, -0.11156647652387619, -0.039595723152160645, -0.13380786776542664, -0.6504618525505066, 0.19714964926242828, 0.12275844812393188, -0.010237760841846466, 0.41895234...
how can I add a <--more--> tag directly in the template? I mean, the exact php code to use in a template for that shortcode. I need it for a script that makes use of that shortcode to hide content and having to add the "more" shortcode by hand through all the posts would be such a task any help appreciated!
[ 0.012309378013014793, 0.00001980000888579525, 0.005431739147752523, 0.02507658489048481, -0.013141530565917492, 0.010029012337327003, 0.00834696739912033, 0.004042340908199549, -0.0343119353055954, -0.022067371755838394, -0.010794313624501228, 0.005922140087932348, -0.02314022369682789, 0....
[ 0.8557062149047852, 0.07180396467447281, 0.1452929824590683, 0.5296704173088074, -0.19344882667064667, -0.15921960771083832, 0.097376748919487, 0.037325404584407806, -0.004995299968868494, -0.6538862586021423, 0.4957118332386017, 0.551001787185669, -0.10569336265325546, 0.08523430675268173...
I have a process I would like to kill: computer@ubuntu:~$ ps aux | grep socat root 2092 0.0 0.0 5564 1528 pts/1 T 14:37 0:00 sudo socat TCP:xxx.17.29.152:54321 PTY,link=/dev/ttyGPS0,raw,echo=0,mode=666 computer@ubuntu:~$ kill 2092 -bash: kill: (2092) - Operation not permitted <--------------- How to kill ??
[ -0.0008303903741762042, -0.0025546415708959103, -0.0025536497123539448, 0.0071961707435548306, -0.011524722911417484, -0.0051306746900081635, 0.006865145638585091, -0.0022898977622389793, -0.008896971121430397, 0.0055149998515844345, -0.017700621858239174, -0.0008544214069843292, -0.01306980...
[ 0.34073084592819214, 0.1598912626504898, 0.4703514277935028, -0.16840212047100067, -0.05434310808777809, 0.20474103093147278, 0.29600217938423157, -0.20514501631259918, -0.1543664038181305, -0.5009610652923584, 0.16034825146198273, 0.15586216747760773, -0.25840049982070923, 0.3793578743934...
I would like to limit the resource (CPU, Memory, and network bandwidth) consumption of processes on the same server. And it will be good if I can migrate processes from one server to another. I think I am looking for some light-weight virtualization. I found LXC is a good choice. But our 2.6.18 kernel does not support LXC. It is a shared cluster, so, I am not allowed to upgrade the kernel. And I think the "setrlimit" system call will only send signals to the processes when the budget is reached rather than limit the resource consumption as a virtual machine does (Please correct me if I am wrong). Any recommendation for this task? Thanks!
[ 0.027496162801980972, 0.00023326586233451962, -0.004683660808950663, 0.00352025730535388, 0.00929809920489788, -0.0277167446911335, 0.008147016167640686, 0.019427679479122162, -0.016934586688876152, 0.0016184262931346893, -0.006517349276691675, 0.013117766007781029, -0.004419810604304075, ...
[ 0.261638879776001, -0.026773421093821526, 0.4932961165904999, 0.08472529798746109, 0.12919719517230988, -0.08846087753772736, 0.1156526580452919, 0.2954038381576538, -0.21422091126441956, -0.963979184627533, -0.08885279297828674, 0.6444451808929443, -0.49014008045196533, 0.2878093421459198...
Consider a linear unobserved effects model of the type: $$y_{it} = X_{it}\beta + c_{i} + e_{it}$$ where $c$ is an unobserved but time-invariant characteristic and $e$ is an error, $i$ and $t$ index individual observations and time, respectively. The typical approach in a fixed effects (FE) regression would be to remove $c_{i}$ via individual dummies (LSDV) / de- meaning or by first differencing. **What I have always wondered: when is $c_{i}$ truly "fixed"?** This might appear a trivial question but let me give you two examples for my reason behind it. 1. Suppose we interview a person today and ask for her income, weight, etc. so we get our $X$. For the next 10 days we go to that same person and interview her again every day anew, so we have panel data for her. Should we treat unobserved characteristics as fixed for this period of 10 days when surely they will change at some other point in the future? In 10 days her personal ability might not change but it will when she gets older. Or asked in a more extreme way: if I interview this person every hour for 10 hours in a day, her unobserved characteristics are likely to be fixed in this "sample" but how useful is this? 2. Now suppose we instead interview a person every month from the start to the end of her life for 85 years or so. What will remain fixed in this time? Place of birth, gender and eye color most likely but apart from that I can hardly think of anything else. But even more importantly: what if there is a characteristic which changes at one single point in her life but the change is infinitesimally small? Then it's not a fixed effect anymore because it changed when in practice this characteristic is quasi fixed. From a statistical point it is relatively clear what is a fixed effect but from an intuitive point this is something I find hard to make sense of. Maybe someone else had these thoughts before and came up with an argument about when a fixed effect is really a fixed effect. I would very much appreciate other thoughts on this topic.
[ 0.005684149917215109, 0.019278399646282196, -0.01217508502304554, 0.00851479358971119, 0.021902676671743393, -0.011601155623793602, 0.006975636817514896, -0.005202413536608219, -0.006241180002689362, 0.016690917313098907, -0.011412088759243488, 0.011916863732039928, -0.021112248301506042, ...
[ 0.24516232311725616, -0.23817656934261322, 0.23507007956504822, 0.08776145428419113, -0.06794580817222595, 0.26831763982772827, -0.08186664432287216, -0.4153176546096802, 0.13933660089969635, -0.17265474796295166, 0.032486408948898315, 0.44542065262794495, -0.46985214948654175, 0.518587350...
Assume a database with 10,000 elements of data is queried, through a statistical interface, for the average data value. The response is determined using the random-sample query method, with a subset of 500 elements being used. Assuming you had access to all data elements, how could you determine what the least accurate response could be?
[ 0.0022540446370840073, 0.019007030874490738, -0.023980161175131798, -0.0005503497086465359, -0.005744358990341425, 0.0035503082908689976, 0.011245591565966606, -0.007371044252067804, -0.014288441278040409, -0.02573602832853794, -0.019861772656440735, 0.014195717871189117, -0.0036050379276275...
[ -0.007395515218377113, -0.3450070023536682, -0.20035216212272644, 0.4093703627586365, -0.1381998211145401, 0.6749097108840942, 0.0374176912009716, -0.6305100917816162, 0.2069946825504303, -0.2820757329463959, 0.03195945918560028, 0.30239689350128174, -0.37836626172065735, 0.027626000344753...
I want to extract part from a log file column which is like this: xx.xxx.xx.xx#59796: Edit: This is the actual log line: Jan 10 17:38:11 server named[747]: client 21x.x0.x8x.xxx#40649: view external: query (cache) 'domain.TLD/A/IN' denied The part before the "#" in the line above is an IP address, and I want to extract only the IP. The part after the "#" is random numbers and not always the same. I use the below command to grep a pattern, extract the IP column, and then redirect the output to text file, but then I have to use an editor to leave out the extra characters from the extracted column. grep -E 'view external.*denied' /var/log/messages |awk '{print $7}' > view_external_denied_ip.txt If I can extract only the IP without the extra characters in the column, I would use the sort command to sort them ( sort | uniq -c | sort -rn ).
[ 0.00770015362650156, 0.0018821408739313483, 0.0007304133614525199, 0.0011610754299908876, -0.02549278736114502, 0.038244444876909256, 0.0056493403390049934, 0.02113628201186657, -0.01781683787703514, -0.004050225950777531, -0.00429860595613718, 0.000452990410849452, -0.0015623483341187239, ...
[ -0.17168480157852173, 0.03546241298317909, 0.8071599006652832, 0.05584033206105232, 0.22155052423477173, 0.24049195647239685, 0.136392280459404, -0.054725222289562225, -0.2607972323894501, -0.11485070735216141, -0.037655629217624664, 0.28163036704063416, 0.12380649894475937, 0.334567695856...
after turn off skeleton responsive media query and made the site not responsive but the site not show the whole entire it's only show the left side of the site, I would like to make it preview entire like the desktop preview , below is link for I'm talking about, please help!! http://d.pr/i/rPD9 Best regards, Decneo
[ 0.009123273193836212, -0.003676942316815257, 0.0012863469310104847, 0.009986554272472858, -0.020668277516961098, -0.00979661289602518, 0.007958494126796722, 0.010129435919225216, -0.021862592548131943, 0.006072742864489555, -0.014408787712454796, 0.010011277161538601, 0.012638388201594353, ...
[ 0.24035514891147614, -0.2660084664821625, 0.3014090657234192, 0.6430380940437317, -0.476960152387619, -0.011969294399023056, -0.06321191787719727, 0.34537434577941895, -0.36286938190460205, -0.4726540744304657, 0.07826568186283112, 0.38050776720046997, 0.25464990735054016, -0.0071383486501...
Supposedly, newer versions of LyX have a command called \dotfill where one can obtain a string of dots such as in the following attachment. ![enter image description here](http://i.stack.imgur.com/pdNRg.png) I am specifically looking for (lower) dots which are analogous to a horizontal fill. Further, I would prefer to be in Standard mode when doing this (as opposed to in Table of Contents). When I enter math mode and insert \dotfill, nothing happens. How can I obtain these dots? If I have to add something to my LateX preamble, please share. **EDIT** : When I enter \dotfill in LyX (in either inline or displayed mathmode) here is what I get ![enter image description here](http://i.stack.imgur.com/RQ3o2.png)
[ 0.009936630725860596, -0.002819583285599947, -0.013448238372802734, 0.002720837015658617, -0.0024700001813471317, -0.0006793946959078312, 0.007081117946654558, 0.02293439954519272, -0.01360933668911457, 0.023455355316400528, -0.006597995758056641, -0.004686237312853336, 0.009697731584310532,...
[ 0.42819744348526, -0.3244509696960449, 0.7146744728088379, -0.015644101426005363, 0.09354779869318008, 0.0007127492572180927, 0.015413094311952591, 0.1672387272119522, -0.3806358873844147, -0.6233639121055603, 0.18882815539836884, 0.5018476247787476, -0.37284064292907715, -0.11732637137174...
Some of the authors in my blog must always add their posts to a specific category each time they post. Sometimes they forget this so therefore I'm wondering if there are any ways to automatically check and fix this. Here's an example of my desired outcome: * Posts by Author X should always be assigned to category A * Posts by Author Y should always be assigned to category B * Posts by Author Z should always be assigned to category C How do I achieve this?
[ 0.0017482577823102474, 0.014195042662322521, -0.016781289130449295, 0.04273826628923416, 0.027353886514902115, 0.009491117671132088, 0.010013957507908344, 0.010952744632959366, -0.01843593269586563, 0.035061731934547424, -0.027635328471660614, 0.010178019292652607, -0.011086894199252129, 0...
[ 0.442202627658844, 0.18493331968784332, 0.37786585092544556, 0.1386823207139969, -0.3043060898780823, -0.10388209670782089, 0.4976210594177246, 0.06594041734933853, -0.3779979944229126, -0.452870637178421, 0.08914490044116974, 0.05430565029382706, 0.07896995544433594, 0.5210258960723877, ...
$ prstat -t NPROC USERNAME SWAP RSS MEMORY TIME CPU 164 oracle 5446M 5445M 8.3% 13:47:50 1.6% 52 root 576M 728M 1.1% 42:29:50 1.0% 1 noaccess 300M 273M 0.4% 0:20:54 0.0% 1 smmsp 2192K 9912K 0.0% 0:00:06 0.0% 1 nagios 584K 4256K 0.0% 0:01:54 0.0% 6 daemon 8360K 11M 0.0% 0:12:39 0.0% Total: 225 processes, 839 lwps, load averages: 1.68, 1.58, 1.46 Above is the output of `prstat` command on Solaris Box which has `Oracle 11g` installed. The RAM on the Solaris box is `64GB` however the oracle process seems to be swapping around `5.5 GB`, is that normal? Would appreciate any recommendations on getting more fine-grained information!
[ 0.0049036345444619656, 0.0039113243110477924, 0.0040628956630826, 0.017577556893229485, -0.0035009749699383974, -0.000009245704859495163, 0.004484034609049559, -0.0004691242356784642, -0.0021404451690614223, 0.025207415223121643, -0.01254601962864399, -0.0009609294938854873, 0.00284547754563...
[ 0.007612538989633322, -0.09453991055488586, 0.3577786982059479, 0.23424018919467926, -0.0007793174008838832, 0.28949010372161865, 0.21805766224861145, -0.3882855772972107, 0.2082975208759308, -0.19886170327663422, 0.12271429598331451, 0.5725928544998169, -0.5965368151664734, 0.111953929066...
We know that there are a few rules about the rate you build an ÜberCharge at, such as: * You build ÜberCharge faster if the patient isn't fully overhealed. * You build ÜberCharge faster if the patient isn't healed by somebody else. * You don't build ÜberCharge if you aren't healing anybody. Do these rules also apply while you deploy an ÜberCharge? Is healing an ÜberCharged patient going to reduce the length of the ÜberCharge? If the patient isn't taking damage, does it mean the ÜberCharge will be shorter? If your patient dies, does your ÜberCharge deplete much faster?
[ 0.005372193641960621, 0.02810671180486679, -0.004423958715051413, 0.013926067389547825, -0.004756598733365536, -0.013189012184739113, 0.007811509072780609, -0.022501632571220398, -0.012060074135661125, -0.001543642021715641, 0.001347956946119666, 0.01630142517387867, -0.007697938475757837, ...
[ 0.4648315906524658, -0.10023524612188339, 0.520502507686615, 0.5327836871147156, -0.278290331363678, -0.10480809211730957, 0.5486617684364319, -0.3253996670246124, -0.2073812335729599, -0.5450862646102905, 0.3836395740509033, 0.42970359325408936, 0.2651471793651581, -0.3169195055961609, ...
Well just as the title states, id like to know what makes a Linux Distribution "Enterprise" compared to another Non-Enterprise Distros.
[ 0.006679016165435314, 0.022552508860826492, 0.005348114762455225, 0.019187385216355324, -0.08843964338302612, 0.023347798734903336, 0.018649958074092865, 0.0412667840719223, -0.0394410714507103, -0.08059006929397583, -0.0026887618005275726, 0.05093615502119064, 0.07309238612651825, 0.02167...
[ 0.6329421997070312, 0.13916800916194916, 0.015964563935995102, 0.3318488299846649, -0.10721801966428757, -0.4627284109592438, -0.2640535533428192, 0.4409398138523102, -0.18368493020534515, -0.15550707280635834, 0.36430275440216064, 0.7133891582489014, 0.1752190887928009, 0.2585472762584686...
Is there any way to change your starting armor? For example, I started with a Spiral Sallet with Mecha Wings, and a regular Spiral Armor. Can I change these to something else? Or are they stuck like that permanently?
[ 0.002841499401256442, 0.010824409313499928, -0.006196081172674894, 0.0008896504295989871, -0.03204105794429779, 0.001511827576905489, 0.010827627964317799, 0.005621828604489565, -0.02715066820383072, 0.0010418964084237814, -0.008687103167176247, 0.014283454045653343, -0.021930282935500145, ...
[ 0.31694164872169495, -0.1075492724776268, 0.03941856697201729, 0.3707376718521118, -0.21015022695064545, 0.5134875774383545, -0.23431441187858582, -0.1291680485010147, -0.4131900370121002, -0.4691026508808136, 0.08943059295415878, 0.20891225337982178, 0.13844743371009827, -0.09772866219282...
I need to learn how databases work in order to use them more efficiently, and my way of learning is by doing. I want to create my own database system. I am not referring to creating a pseudo-database that would use query to parse files; this would simply be a filesystem interface with a query language. I am talking about the actual structure of a database engine. And since what I have in mind is neither relational nor document-oriented (it's "node-oriented", if that even exists), I would need any resource to be as abstract and high-level as possible. So how would I go about creating that? What resources/tutorials/books can I read to understand? The language does not matter in the slightest. Ideally, the code would be pseudo-code to illustrate the concept, not tied to a particular language, but anything would do. I was not able to find anything on the matter on google (since I am so illiterate on the subject, maybe I am just not entering the right search). If such resources are not available, then I guess something about how to create a client would at least be a step in the right direction.
[ 0.0026043117977678776, 0.013213848695158958, -0.002915371675044298, -0.002976510673761368, -0.011219414882361889, 0.009406440891325474, 0.005220464896410704, 0.01382489688694477, -0.014208122156560421, 0.0017385073006153107, -0.0013867304660379887, 0.004082867875695229, 0.00911039113998413, ...
[ 0.42169955372810364, 0.4095078706741333, 0.010688873939216137, 0.3802807629108429, 0.26640579104423523, -0.053973738104104996, 0.0755971223115921, 0.15995649993419647, -0.0900622084736824, -0.6497024893760681, 0.11698862165212631, 0.4507436454296112, -0.10232535749673843, 0.291703611612319...
I'm using TeXmaker on a Mac. I created my `.tex` file. I compiled it and the editor tells me that the `.log` file is created and the `PDF` file is written. Then, when I click on _view PDF_ , it gives me the error message `file not found`. How can I fix this problem? The problem wasn't occurring in Ubuntu, same `.tex` file and settings. Any help would be appreciated.
[ -0.0018232446163892746, 0.014776510186493397, 0.004168749786913395, 0.011384948156774044, -0.001447727670893073, 0.006302818190306425, 0.006286874413490295, 0.008073071949183941, -0.01794307306408882, -0.027681203559041023, -0.01511751115322113, 0.00890838261693716, -0.004683160223066807, ...
[ -0.12731221318244934, 0.4029048979282379, 0.5825687050819397, 0.10973018407821655, 0.1615876704454422, -0.2555161714553833, 0.27956438064575195, 0.22753234207630157, -0.1904749572277069, -0.9700959324836731, -0.0307131540030241, 0.9198257327079773, -0.5057920217514038, 0.3202005922794342, ...
My hosting's disk space is increasing every day and I just wanted to know, why? I don't upload new files, but it still gets increased by every day. Anyone knows? Thank you.
[ -0.03217745199799538, 0.022120743989944458, -0.005340058822184801, 0.06778867542743683, 0.003154977923259139, -0.014009406790137291, 0.009327375330030918, 0.0025497779715806246, -0.03556923568248749, -0.05122608318924904, 0.008772426284849644, 0.03857690468430519, 0.016916923224925995, 0.0...
[ 0.6258134245872498, 0.2759189009666443, 0.5883283615112305, 0.2208968698978424, -0.15467338263988495, -0.22869202494621277, 0.2964838743209839, 0.7491438984870911, -0.5929312109947205, -0.31615281105041504, 0.3082963228225708, 0.30254989862442017, 0.21370966732501984, 0.494377464056015, ...
I'd previously been renewing my section command in Memoir in order to style the section numbers by prefacing them with a symbol (§) and following them with a period: \renewcommand{\thesection}{§ \arabic{section}.} Werner has helpfully indicated in another post a way of putting the section numbers in the margin, getting rid of headings, and having the text body begin from the section numbers without any skip: \hangsecnum \counterwithout{section}{chapter} \makeatletter \newcommand{\hangsection}{% \savebox{\@tempboxa}{\normalfont\ }% \section{}\hspace*{-\wd\@tempboxa}% } \makeatother \setaftersecskip{0pt}% Could anyone indicate how to format the section numbers (e.g., with a symbol and a period) with such a new command? Many thanks.
[ 0.008186792954802513, 0.02239123359322548, -0.012497544288635254, 0.030448012053966522, -0.008593527600169182, -0.01463015004992485, 0.011172850616276264, -0.012003117240965366, -0.01450884249061346, -0.015911918133497238, -0.014772353693842888, 0.0008719443576410413, 0.002657890785485506, ...
[ -0.08011399209499359, 0.21752268075942993, 0.6317617297172546, -0.1228158101439476, -0.14539550244808197, -0.05763646215200424, 0.3264298141002655, -0.30812868475914, -0.3348197638988495, -0.18913504481315613, -0.543684720993042, 0.048295896500349045, 0.34653326869010925, 0.209266543388366...
To give a brief background of my situation, I am currently writing a group report in LaTeX for my honours project and I'm attempting to use LaTeX for this. There are several sections in this report, each in different .tex files, each contributed by different people and there is also a document class that has been provided to us. Below is an example of the file structure: Master Tex file: - section 1 --Images --TikzInput --IncludePDFs - section 2 --Images --TikzInput --IncludePDFs etc.. Each document (section) complies perfectly fine by itself, including input PDF's and input tikz files. When I attempt to bring them together in a master document using `\include` or `\input`, the document looks for the tikz input files and the IncludePDFs files in the master tex location. I've currently found a way to include different paths for the image folders, is there a similar way for tikz scripts/pdfs? alternatively, is there a way to separately compile each section by themselves and then compile them together at the end? Currently, the only way i can see to do this is to manual change the paths before compiling them all together which is rather arduous for an 400 page document, even with ctrl-f. Thanks in advance for anyone thoughts/input.
[ 0.006461055018007755, 0.009195658378303051, -0.008941670879721642, 0.023704851046204567, 0.03517313301563263, 0.0065393513068556786, 0.006355724297463894, -0.0036430039908736944, -0.020310478284955025, -0.01204235851764679, -0.0037049080710858107, -0.0039928387850522995, -0.00032983592245727...
[ 0.36410272121429443, 0.25740084052085876, 0.362202525138855, -0.04686314985156059, -0.05933109670877457, 0.22540807723999023, 0.16028013825416565, -0.40867260098457336, -0.15249088406562805, -0.6157038807868958, -0.13664180040359497, 0.29341766238212585, -0.15647925436496735, -0.1156411468...
Suppose that I bought an old house, and changed walls and windows, repainted the house, used new wallpaper, bought new sofa, table, etc... For this kind of thing, which word is more appropriate, "decoration" or "remodel" or "upgrade"?
[ -0.013607628643512726, 0.02096981555223465, 0.0027780826203525066, 0.012950179167091846, -0.043227724730968475, -0.024066133424639702, 0.0133597981184721, -0.007349999621510506, -0.009876715950667858, -0.050868064165115356, -0.015660027042031288, 0.009151301346719265, 0.027221349999308586, ...
[ 0.607920229434967, 0.21903030574321747, 0.1961446851491928, -0.2069961428642273, 0.3987357020378113, 0.4243090748786926, 0.2918424904346466, -0.1493011862039566, -0.04833272844552994, -0.5234978795051575, 0.04872836172580719, 0.2773047983646393, 0.10350387543439865, 0.35678187012672424, ...
I'm using a custom function to create a shortcode that displays the latest blog post on the home page of a template. But I'm trying to NOT have it display any images. I know I can use the Advanced Excerpt plugin to have it remove images, but the issue is that it will also remove images from the index.php feed which I want to keep which is using `the_excerpt()` in the template. Here's my custom function that creates the shortcode: function my_recent_news() { global $post; $html = ""; $my_query = new WP_Query( array( 'post_type' => 'post', 'posts_per_page' => 4 )); if( $my_query->have_posts() ) : while( $my_query->have_posts() ) : $my_query->the_post(); $html .= " <article> <span class=\"date\">" . get_the_date() . "</span> <h2><a href=\"" . get_permalink() . "\">" . get_the_title() . "</a></h2> " . get_the_excerpt() . " </article> "; endwhile; endif; wp_reset_query(); return $html; } add_shortcode( 'news', 'my_recent_news' ); I've posted something about this before: get excerpt without images but the solution was to use the Advanced Excerpt plugin, but for this I am trying to use the excerpt on the home and the blog feed page but I want to preserve the `img` markup on the blog feed and remove the img tag from the custom shortcode. I've tried to used just `the_excerpt()` in that custom shortcode function, but that just seems to break the whole function and display some really odd stuff. I'm not too sure if I need to have a filter somewhere to strip that out or what. I'm also not sure that if I do need a filter, where that would go? Before the loop, after the loop, or does it need it's own constructed argument?
[ 0.021045316010713577, 0.006520153488963842, -0.003753814846277237, 0.008753372356295586, -0.01543337106704712, 0.003969247918576002, 0.009221930988132954, 0.0065554967150092125, -0.016844360157847404, -0.019756225869059563, -0.013079429976642132, 0.015564342960715294, 0.012548177503049374, ...
[ 0.28679537773132324, 0.0612160824239254, 0.7970380783081055, 0.06230916082859039, 0.11684568226337433, -0.30813920497894287, -0.129488006234169, 0.40857866406440735, -0.21862225234508514, -0.61715167760849, 0.03306358680129051, 0.3124261200428009, -0.2273649126291275, 0.4488297402858734, ...
1. Here is a piece of solution to automate a dictionary using the links given in the first following comments. There are two problems to solve : * When one definition starts on one page and finishes on another, the words in the headinga are not the good ones. In my example below, take a look at the page `1` where the last word must be `adhesive` and not `adhesion`, and in the page `2`, the first word must be `adjacent` and not `adhesive`. Is there a way to solve this problem ? * Secundly, I would like the letter A of the section to be centerd using `\textbf{\textsf{...}}` for formatting. Here is the code that I've wanted to improve. % Sources : % 1) http://tex.stackexchange.com/questions/30392/how-could-one-setup-a-layout-for-a-dictionary-if-possible % 2) http://tex.stackexchange.com/questions/26122/indexing-an-interval-of-words-on-top-of-every-page \documentclass[twoside]{article} \usepackage{multicol} \usepackage{ifthen} \usepackage{fancyhdr} \usepackage{enumerate} \usepackage{lipsum} % Empty \sectionmark \renewcommand{\sectionmark}[1]{} \fancyhead[L]{\textsf{\rightmark}} \fancyhead[R]{\textsf{\leftmark}} \fancyfoot[C]{\textbf{\textsf{\thepage}}} % Entry command : \dict{<word>}{<gender>}{<text>} \newcommand{\dict}[3]{% \par\vspace{0.25\baselineskip} % \textbf{\textsf{#1}} \textit{#2} #3 % \markboth{#1}{#1} } \pagestyle{fancy} \begin{document} \section*{A} \begin{multicols}{2} \dict{adequate}{n}{\lipsum[1]} \dict{adhere}{n}{\lipsum[2]} \dict{adherence}{n}{\lipsum[3]} \dict{adhesion}{n}{\lipsum[4]} \dict{adhesive}{n}{\lipsum[5]} \dict{adjacent}{n}{\lipsum[6]} \dict{adjective}{n}{\lipsum[1]} \dict{adjoin}{n}{\lipsum[2]} \dict{adjourn}{n}{\lipsum[3]} \dict{adjournment}{n}{\lipsum[4]} \dict{adjunt}{n}{\lipsum[5]} \dict{adjust}{n}{\lipsum[6]} \end{multicols} \end{document} 2. The solution found is the following one even if there is one remaining disturbing behavior, even if it is logical : in the last page of the output of the following code, `zero` appears in both headers... % Sources : % 1) http://tex.stackexchange.com/questions/30392/how-could-one-setup-a-layout-for-a-dictionary-if-possible % 2) http://tex.stackexchange.com/questions/26122/indexing-an-interval-of-words-on-top-of-every-page % 3) http://tex.stackexchange.com/questions/30947/how-to-automate-a-dictionary-sorting-headers/31017#31017 \documentclass[twoside]{article} \usepackage{multicol} \usepackage{fancyhdr} \usepackage[bf,sf,center]{titlesec} % Headers and footers \fancyhead[L]{\textsf{\rightmark}} \fancyhead[R]{\textsf{\leftmark}} \fancyfoot[C]{\textbf{\textsf{\thepage}}} \renewcommand{\headrulewidth}{1.4pt} \renewcommand{\footrulewidth}{1.4pt} % Entry command : \dict{<word>}{<gender>}{<text>} \newcommand{\dict}[3]{% \markboth{#1}{#1}% \par\vspace{0.25\baselineskip}% \textbf{\textsf{#1}} \textit{- #2 -} #3% } \pagestyle{fancy} % For testing \usepackage{lipsum} \begin{document} \section*{A} \begin{multicols}{2} \dict{adequate}{n}{\lipsum[1]} \dict{adhere}{n}{\lipsum[2]} \dict{adherence}{n}{\lipsum[3]} \dict{adhesion}{n}{\lipsum[4]} \dict{adhesive}{n}{\lipsum[5]} \dict{adjacent}{n}{\lipsum[6]} \dict{adjective}{n}{\lipsum[1]} \dict{adjoin}{n}{\lipsum[2]} \dict{adjourn}{n}{\lipsum[3]} \dict{adjournment}{n}{\lipsum[4]} \dict{adjunt}{n}{\lipsum[5]} \dict{adjust}{n}{\lipsum[6]} \end{multicols} \section*{M} \begin{multicols}{2} \dict{main}{n}{\lipsum[3]} \dict{material}{n}{\lipsum[1]} \dict{mathematic}{n}{\lipsum[2]} \dict{more}{n}{\lipsum[2]} \end{multicols} \section*{Z} \begin{multicols}{2} \dict{zebra}{n}{\lipsum \lipsum \lipsum} \dict{zero}{n}{\lipsum[4]} \end{multicols} \end{document}
[ 0.002398096024990082, 0.01079266332089901, -0.006270068231970072, 0.02025478705763817, -0.018875200301408768, 0.006258751265704632, 0.006296650506556034, 0.0008900663815438747, -0.011341334320604801, 0.02323019877076149, -0.012256583198904991, -0.004248232115060091, -0.0013533178716897964, ...
[ 0.27015936374664307, 0.08477871119976044, 0.07286291569471359, 0.08302009105682373, -0.26674365997314453, -0.3103378117084503, 0.04786618798971176, -0.20084498822689056, -0.18349513411521912, -0.8534642457962036, 0.0413605198264122, 0.3405153155326843, -0.40850991010665894, 0.0489884912967...
I'm about to flash an update to Android 4.4.2 on my Nexus 4, downloaded from Google Developers, and one of the files that it's going to flash is the new bootloader version. I wanted to know if flashing a new bootloader will re-lock my fastboot, which will require me to unlock it wiping my data, or if it will in any other way cause my data to be wiped.
[ -0.00158514769282192, 0.018044166266918182, -0.002736123511567712, 0.01592729054391384, -0.03849433735013008, 0.0008608419448137283, 0.008700530044734478, -0.0042799366638064384, -0.02133576013147831, 0.02861158549785614, -0.004894118756055832, 0.02908470667898655, 0.016844719648361206, 0....
[ 0.31422191858291626, 0.15523461997509003, 0.46325787901878357, -0.06935221701860428, -0.015484653413295746, 0.0935271829366684, 0.5027371048927307, 0.44869717955589294, -0.37052369117736816, -0.7419565916061401, -0.06640421599149704, 0.550309956073761, -0.3628808259963989, 0.04669160768389...
I've recently been researching Xen, and this has come up as a quistion for me. Is it possible to have more than 1 dom0 running on a single Xen host? The reason I ask is that this is very similar technology to IBM's Power VM and in particular the VIOS used to own the hardware and manage it on the other LPARs behalf. In our company using LPARs and VIOS we always have 2 VIOS per host so that if one VIOS fails for some reason the other continues to provide access to the hardware for the other LPARs.
[ 0.009693166241049767, 0.012271886691451073, -0.016887571662664413, 0.011828161776065826, -0.02412012405693531, -0.006100871134549379, 0.008944607339799404, 0.021673236042261124, -0.01727226935327053, -0.009994669817388058, -0.0033667082898318768, 0.01810223050415516, -0.0007598388474434614, ...
[ 0.6960510015487671, 0.18428383767604828, 0.3144145905971527, 0.24634677171707153, -0.023201437667012215, -0.087640181183815, -0.11189666390419006, -0.11109195649623871, -0.14365461468696594, -0.49851053953170776, 0.020228033885359764, 0.46936482191085815, -0.23301561176776886, 0.3687324225...
I want to have an idea as to how to update the data regularly from the source, if its stored locally in your server or local machine.Lets say MySQL database. Source:It will be a webservice data repository. Data sets: An online data repository has a number of data sets. I want to know which data sets have been updated and what changes have been made. I have done queries in MySQL, but not done anything like this. Any suggestions?
[ 0.017072781920433044, 0.008617059327661991, 0.003215503878891468, 0.02251577563583851, 0.018636515364050865, -0.002638976788148284, 0.006524241529405117, 0.023142196238040924, -0.01630176045000553, -0.022566242143511772, -0.0020708178635686636, 0.024047445505857468, -0.003169024595990777, ...
[ 0.6597934365272522, 0.062231630086898804, 0.32548990845680237, 0.2559921443462372, -0.025382496416568756, -0.037758901715278625, -0.08556948602199554, 0.2334553599357605, -0.1546112447977066, -0.6019209027290344, 0.22953809797763824, 0.5406360030174255, -0.046349965035915375, -0.0188587792...