text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I encrypted both my phone and SD card upon purchasing my HTC Evo 4g LTE. Recently the phone died (would get to the login screen and then reboot). After a factory reset the SD card shows up as damaged. I've tried mounting it in various ways, but neither the phone nor a computer will attempt to mount the drive - they just ask to format it. Is there any hope of recovering the contents?
[ -0.0035281279124319553, 0.00610041618347168, -0.008576956577599049, 0.014530603773891926, -0.026854226365685463, -0.007745293900370598, 0.00900712888687849, 0.015471333637833595, -0.018987473100423813, -0.024513186886906624, -0.02105094864964485, 0.017257317900657654, 0.003946601878851652, ...
[ 0.24750371277332306, 0.2572632431983948, 0.46133434772491455, 0.23559202253818512, 0.18636809289455414, -0.07020352780818939, 0.4557247459888458, 0.10983419418334961, -0.08329784125089645, -0.20841823518276215, 0.05617271363735199, 0.6107190847396851, -0.23626276850700378, 0.34329086542129...
I don't recognise the name Herb Sutter. > * I don't know him. > > * I know him not. > > What's the difference?
[ -0.014788923785090446, -0.006652848329395056, 0.010355040431022644, 0.03597071394324303, -0.026902060955762863, 0.011251071467995644, 0.0140061154961586, -0.035460006445646286, -0.028818437829613686, -0.040749263018369675, 0.011848610825836658, 0.008642269298434258, 0.0238315649330616, 0.0...
[ 0.6356645226478577, 0.4855634272098541, 0.14425215125083923, -0.3911280632019043, -0.0017561380518600345, -0.07489927113056183, 0.3568378686904907, 0.05531050264835358, -0.1278819441795349, -0.20699653029441833, 0.23666004836559296, 0.21526268124580383, -0.25963613390922546, 0.079187735915...
I would like to copy my "settings" from my desktop to my laptop. I am running KDE on Arch. I am not sure what to do with ~/.config, ~/.local, and ~/.kde4 since they have subdirectories with names that match my desktop hostname. If I naively copy everything, I get all sorts of errors/warning when logging in and trying to open my email/calendar/akonadi.
[ 0.006449540611356497, 0.0020698397420346737, 0.0017690518870949745, 0.0210125595331192, -0.008007253520190716, -0.0019640070386230946, 0.009257386438548565, 0.004986100364476442, -0.018845731392502785, -0.009441922418773174, -0.0032812466379255056, 0.012394151650369167, -0.013800572603940964...
[ 0.08719738572835922, 0.3880658745765686, 0.08403380215167999, 0.17656254768371582, 0.05203857645392418, 0.17776493728160858, 0.37802743911743164, 0.34940627217292786, -0.13322584331035614, -1.0807987451553345, 0.017523987218737602, 0.4438035190105438, -0.02148590423166752, 0.36648932099342...
I need to cite a book which includes a schwa ('ə') in the title. I've been able to achieve certain diacritics in bibtex entries using commands like \'{e} (for an e with acute accent). However, when the base letter itself is not ascii, I'm not sure what to do. My acute problem is typesetting an ə which occurs in a booktitle in the references section. More generally, I'd like to know how to use bibtex which can contain arbitrary non-ascii symbols.
[ -0.002017455641180277, 0.007964544929564, -0.010196501389145851, 0.030128180980682373, 0.03283136337995529, 0.010099880397319794, 0.009756430983543396, -0.004667174071073532, -0.017187349498271942, -0.01243925467133522, -0.009475084953010082, -0.0024645919911563396, -0.013961001299321651, ...
[ 0.13112226128578186, 0.5008590817451477, -0.07206790149211884, -0.116005077958107, -0.1588936299085617, 0.05018188804388046, 0.21276000142097473, 0.21103453636169434, -0.030849797651171684, -0.6199498176574707, -0.16646233201026917, 0.12270816415548325, -0.025529297068715096, -0.0166090726...
When I send an email from my official exchange server, the person who receives it, whether through web access or through Outlook, gets the email with HTML coding! It is very embarrassing and I have searched for the solution in vain. How can I stop this behavior?
[ -0.00932654831558466, -0.015684837475419044, -0.01291817706078291, 0.035234350711107254, -0.02384733036160469, -0.01951037347316742, 0.011515557765960693, 0.005625866819173098, -0.02364126220345497, -0.05482123792171478, -0.01182533148676157, 0.0190882608294487, 0.002218944486230612, 0.013...
[ 0.39489486813545227, 0.3092905580997467, 0.30467814207077026, -0.010585574433207512, -0.09306993335485458, -0.20963427424430847, 0.5699618458747864, 0.4595511555671692, -0.05427587032318115, -0.42221882939338684, -0.003928449470549822, -0.20074903964996338, -0.3992144763469696, 0.104292720...
I have a bib file generated by Zotero (can become biblatex if needed) and I use natbib bibliographies through the package revtex4 (I could also use aastex if needed). Zotero kindly outputs every single detail (DOI, URL and abstracts etc). I need the bibliography entries to include only the names, year and journal like so: `Abadi M. G., Moore B., Bower R. G., 1999, MNRAS, 308, 947` (see http://arxiv.org/abs/1002.0583) I've also come across true/false conditions of properties like author, doi, url etc from Bibliography with only initials of names. There must be a simple way just to turn off features like DOI and URL! Thanks Here's MWE: \documentclass[preprint2]{aastex} % \usepackage[authoryear,round,comma]{natbib} % \usepackage[% % style=numeric-comp,sorting=none, % sortcites=true,doi=false,url=false, % firstinits=true,hyperref]{biblatex} \setcitestyle{authoryear,round,comma} \begin{document} hello \citep{christlein_can_2004} \bibliography{Zotero} \bibliographystyle{plainnat} \end{document}
[ 0.024065595120191574, 0.006594116799533367, -0.0038563976995646954, 0.022643927484750748, -0.007922636345028877, 0.03265202417969704, 0.008499856106936932, 0.016212277114391327, -0.015629231929779053, -0.02501744218170643, -0.002586754271760583, 0.003435177495703101, 0.0037124725058674812, ...
[ 0.03355522081255913, 0.4982580244541168, -0.05393698066473007, -0.10509106516838074, -0.3908315896987915, -0.05436002463102341, 0.018916359171271324, -0.04168790951371193, -0.3664652109146118, -0.43249884247779846, -0.05876287817955017, -0.010225274600088596, -0.42386454343795776, 0.361056...
It's cool to add a hyperlink in references like this: ![enter image description here](http://i.stack.imgur.com/TTnZe.jpg) I want to know how to add the blue box in bibtex. For example, the following article's link is "ee". @inproceedings{DBLP:conf/osdi/DeanG04, author = {Jeffrey Dean and Sanjay Ghemawat}, title = {MapReduce: Simplified Data Processing on Large Clusters}, booktitle = {OSDI}, year = {2004}, ee = {http://www.usenix.org/events/osdi04/tech/dean.html}, crossref = {DBLP:conf/osdi/2004}, bibsource = {DBLP, http://dblp.uni-trier.de} }
[ 0.0124126598238945, -0.0011775338789448142, -0.0044449809938669205, 0.01652730442583561, 0.014150384813547134, -0.0020744018256664276, 0.005268475040793419, -0.00028256443329155445, -0.014014698565006256, -0.00760318199172616, -0.004443750716745853, 0.003014465095475316, -0.01850142143666744...
[ 0.34662407636642456, 0.14067348837852478, 0.15518353879451752, 0.1987084597349167, 0.2031182050704956, -0.4298596680164337, 0.03120414912700653, 0.11889421939849854, -0.2618231773376465, -0.6016709208488464, -0.2480354905128479, -0.0016540291253477335, -0.10529635846614838, -0.124227225780...
My question essentially revolves around multi-electron atoms and spectroscopic terms. I understand the idea that the total wavefunction for Fermions should be antisymmetric. Consider as an example, the $2p^2$ electrons in a partially filled p shell; that is, the outer shell of Carbon. The two electrons both have $l=1$, and hence total orbital angular momentum takes the values: $L = L1+L2, L1+(L2-1),...,|L1-L2| = 0,1,2$ and $S = 0,1$ I can sort of intuitively see that $L=2$ must refer to a symmetric spatial wavefunction and hence an antisymmetric spin wavefunction. I can handwave and say that for $L=2$, we must have $m_{l1}=m_{l2}=\pm1$ and hence they must have opposing spin to satisfy PEP which gives S=0 - but I'm not sure how to express that in terms of an actual wavefunction and it seems to be a bit of a circular argument. However, I don't see why $L=1$ must have $S=1$ (a triplet) and $L=0$, $S=0$ (another singlet). Can anyone shed some light on this? Thanks!
[ -0.001678996253758669, 0.006010852288454771, -0.010226819664239883, 0.01724894717335701, -0.02106524258852005, -0.00566142750903964, 0.008101323619484901, -0.002288819756358862, -0.010598357766866684, -0.009280952624976635, -0.008474585600197315, 0.009924654848873615, -0.023985082283616066, ...
[ 0.18729539215564728, 0.18916712701320648, 0.4232346713542938, -0.018651658669114113, 0.21874956786632538, 0.5091779828071594, -0.33773303031921387, -0.42998144030570984, 0.08104772120714188, -0.27531224489212036, -0.13014988601207733, 0.36993837356567383, -0.22357603907585144, 0.2573789656...
I started a new ladder character and am beginning to find random runes in Act II (Normal). What rune words can I reasonably find runes for and create while I'm still in Normal? I know of Stealth (TalEth), but the words on the Arreat Summit aren't listed by level, rather alphabetically. Is there a handy reference that lists them all by max rune level?
[ 0.009338220581412315, 0.024540742859244347, -0.007928676903247833, -0.017374610528349876, -0.02588571608066559, 0.012628386728465557, 0.01259641069918871, 0.005077214911580086, -0.02958683855831623, 0.02823460102081299, 0.01045732107013464, 0.012592772953212261, 0.017488356679677963, 0.033...
[ -0.03989676758646965, 0.06064686179161072, 0.27041196823120117, -0.04524233192205429, -0.4518337547779083, 0.10706043988466263, 0.709097146987915, -0.0979008749127388, -0.14621610939502716, -0.639700710773468, 0.07116129249334335, 0.21586120128631592, 0.21822819113731384, -0.23443402349948...
There are some effects in the game which remove a champion's sprite from the map, rendering them untargetable for a moment, such as Elise's `Rappel`. Does the Summoner Spell `Flash` do this as well, even just for a split second? Or is the champion always on the field and targettable _at all times_? In other words, does `Flash` instantly place the champion at the target location such that there is no moment in time where the champion isn't on the map, or is there a moment when he doesn't exist on the map? I ask because I was playing a game where I was getting double ulted by both Nunu and Morgana. My `Flash` was up, but it wasn't enough to escape the range of their ultimates completely. I am wondering whether, if I timed my `Flash` just right, if I could make myself untargettable for a split second and avoid the effects of one or both of the ultimates when they triggered (i.e., Morgana's stun activated or Nunu's ultimate blew up for tons of damage).
[ -0.013981147669255733, 0.017086973413825035, -0.010273855179548264, 0.006740769371390343, 0.00416935421526432, -0.023385779932141304, 0.006677161902189255, 0.018530189990997314, -0.016693640500307083, 0.031112413853406906, -0.019435923546552658, 0.016331376507878304, -0.00990322045981884, ...
[ -0.10712963342666626, -0.16122470796108246, 0.07330647855997086, 0.18295229971408844, -0.22364579141139984, -0.13480526208877563, 0.4099527597427368, -0.2181522399187088, -0.43276840448379517, -0.37350067496299744, -0.3647744953632355, -0.0929194763302803, -0.004309097770601511, 0.17390057...
Today morning after a long taxi ride (I am travelling this week, persobal vacation), I took my phone from my pocket and it seemed to be turned off. I switched it on and the phone started up as it was the first time. All my settings, config, apps, widgets were gone. I think the phone was factory reset. Not sure if it was locally (accidental click in settings?) or may be remotely. (The phone _was_ fine earlier in the morning). I tried to get the data from my Google account backup (which I remember setting up when I got this phone) but I could not (no 'Restore' option was enabled in settings. But that would be a different question). I had to manually (painfully) set up things again. All my photos ate backed up to google and I think I can get them back. Same for apps, most of the should have data on the cloud somewhere. Is there a way to know what happened? I have a HTC one (M7), and use Google for most services including back ups.
[ -0.020035170018672943, 0.011402414180338383, 0.013322004117071629, 0.013338347896933556, -0.011300833895802498, -0.01629883050918579, 0.005616975948214531, 0.013576891273260117, -0.011167783290147781, -0.01003528293222189, -0.006671724375337362, 0.010598268359899521, 0.025506673380732536, ...
[ 0.2177477777004242, -0.06215117126703262, 0.7227505445480347, -0.05407128110527992, 0.04139871522784233, -0.08918913453817368, 0.6642100214958191, 0.025891387835144997, -0.2229309231042862, -0.6871518492698669, 0.25668126344680786, 0.4603595733642578, -0.19942206144332886, 0.26927763223648...
I am trying to create a color box using `tcolorbox` package. source code is like this- \documentclass[11pt]{article} \usepackage[top=.5in, bottom=.5in, left=1in, right=1in]{geometry} \usepackage{tcolorbox} \tcbuselibrary{skins} \begin{document} \tcbset{skin=enhanced,fonttitle=\bfseries, frame style={upper left=blue,upper right=red,lower left=yellow,lower right=green}, interior style={white,opacity=0.5}, segmentation style={black,solid,opacity=0.2,line width=1pt}} \begin{tcolorbox}[title=Nice box in rainbow colors] With the ’enhanced’ skin, it is quite easy to produce fancy looking effects. \tcblower Note that this is still a \texttt{tcolorbox}. \end{tcolorbox} \end{document} The error I am having is: ! Package pgfkeys Error: I do not know the key '/tikz/upper left' and I m going to ignore it. ! Package pgfkeys Error: I do not know the key '/tikz/upper right' and I m going to ignore it. ! Package pgfkeys Error: I do not know the key '/tikz/lower left' and I m going to ignore it. ! Package pgfkeys Error: I do not know the key '/tikz/lower right' and I m going to ignore it.
[ 0.004256703425198793, 0.009714536368846893, -0.0001671945210546255, 0.016630515456199646, 0.008446749299764633, -0.003289179177954793, 0.008300645276904106, -0.006647731643170118, -0.01628812961280346, -0.009776975959539413, -0.010150877758860588, -0.002657881937921047, -0.01700577512383461,...
[ 0.6771881580352783, -0.22263503074645996, 0.44030559062957764, -0.07555871456861496, 0.0885094553232193, 0.30600786209106445, 0.013479809276759624, -0.0984535813331604, 0.07619354873895645, -0.9157619476318359, 0.1463705152273178, 0.3430514633655548, -0.24072736501693726, -0.01546453032642...
I was wondering if the temperature of an object affects the amount of radiation it absorbs. For example, if I have a box that is hotter, will it absorb more energy as compared to the same cooler box?
[ 0.040188759565353394, 0.013155126944184303, -0.0007629165193066001, 0.03574814647436142, 0.00855018850415945, 0.004191569052636623, 0.012452718801796436, -0.03466150909662247, -0.03191684931516647, -0.032469190657138824, 0.0042792935855686665, 0.03247079998254776, -0.015894951298832893, -0...
[ 0.9551864266395569, 0.2678757905960083, -0.27392131090164185, 0.09220786392688751, -0.14463229477405548, -0.28876739740371704, 0.04504551365971565, -0.24619486927986145, -0.15785089135169983, -0.22611252963542938, 0.0942608043551445, -0.17057831585407257, 0.21724466979503632, 0.43809330463...
> **Possible Duplicate:** > Using BibTeX to make a list of references without having citations in the > body of the document? I've read that "All items listed in the bibliography should be cited in the body of the paper." But if I did not cite any item in the `.tex` file, how can I list the items in the bibliography? That's it, I did not want to cite any item, but to appear in the bibliography section
[ 0.015811262652277946, 0.006510228384286165, -0.002315357094630599, 0.01930672489106655, 0.008257259614765644, 0.0030832139309495687, 0.007445822469890118, 0.012242892757058144, -0.02228720672428608, -0.024776743724942207, -0.01099560596048832, 0.0024868587497621775, -0.02453635074198246, -...
[ 0.378493070602417, 0.5092684030532837, 0.3175862431526184, 0.12097985297441483, -0.05825294926762581, -0.16062483191490173, 0.21709206700325012, 0.021655816584825516, -0.30103355646133423, -0.58620285987854, 0.27882635593414307, 0.5183215141296387, -0.3799785375595093, 0.33011317253112793,...
I'm creating a project that I want to be able to distribute across platforms. I'm writing in Java and AWT which already gives me a pretty large range of devices, but I'm mostly interested in Windows and Linux (Debian/Ubuntu). I'm trying to determine where I should put config files. I have application- wide configuration files and user-specific files. Where are common directories to put these files? Here's my current setup: ## Windows: App Config: `%PROGRAMDATA%\MyApp\config\` User Config: `%USERPROFILE%\AppData\Local\MyApp\` ## Other: App Config: `/opt/MyApp/config` User Config: `$HOME/.MyApp/`
[ -0.01189391128718853, 0.0049825445748865604, -0.004492643289268017, 0.0056198760867118835, 0.0011960200499743223, 0.004244837909936905, 0.006397482939064503, 0.021009813994169235, -0.014151559211313725, -0.022324588149785995, -0.0013145533157512546, 0.009024223312735558, 0.004041776061058044...
[ 0.39704373478889465, 0.17871719598770142, 0.30541881918907166, -0.07581621408462524, 0.3088986873626709, 0.1419791281223297, -0.037700820714235306, 0.08916739374399185, -0.25554606318473816, -0.9529946446418762, -0.02317202463746071, 0.5825478434562683, -0.14168241620063782, 0.060998085886...
When i got Minecraft when it was in Alpha i was glad that apart from having to connect to the internet to get the majority of the game and update it i could just copy the files to my gaming pc and play it offline (since i don't have her connect to the internet) so i am wondering, can Cubeworld be played offline in a similar fashion to Minecraft (copy all the files to another computer) and play it just fine, i'd like to know before i commit to buying a copy as the i doubt the laptop i connect to the net would be able to play it
[ -0.007619312033057213, 0.00789737794548273, 0.004790432285517454, 0.007736279629170895, -0.0007169670425355434, 0.0041298591531813145, 0.006188997998833656, -0.0027106176130473614, -0.017056284472346306, -0.007857853546738625, -0.007210330106317997, 0.01719890721142292, 0.021396733820438385,...
[ 0.2453012764453888, -0.04241453483700752, 0.08603577315807343, 0.30498048663139343, -0.3623043894767761, -0.18356721103191376, -0.18492893874645233, 0.2406817376613617, -0.21998107433319092, -0.43491342663764954, 0.6230899095535278, 0.4549282491207123, 0.14333835244178772, 0.31175136566162...
I'm using Geoserver in conjunction with the google maps API to add a raster layer onto a map. It loads fine, however there's some strange problem with opacity at the edge of the layer where you can see a boundary. Have you any idea how to fix this? ![Screen shot](http://i.stack.imgur.com/qg948.png) (The original is at obstest.heliohost.org/map2.html.) I'm totally new to gis and to Geoserver so sorry if this is an obvious question! Thanks.
[ -0.012848561629652977, -0.001325551187619567, -0.00374865485355258, 0.011771203950047493, -0.006308787036687136, -0.011589921079576015, 0.005549055524170399, -0.00942093413323164, -0.015507993288338184, -0.0031411005184054375, 0.0008839394431561232, 0.012492071837186813, -0.00751165393739938...
[ 0.4486366808414459, -0.02499067410826683, 0.4553208649158478, 0.13531148433685303, 0.05186887085437775, -0.20094197988510132, 0.07854590564966202, 0.37429723143577576, -0.17843502759933472, -0.9416754841804504, 0.162414088845253, 0.4542088210582733, -0.15164946019649506, -0.051756281405687...
I'm currently trying to find my way into the geometric description of Quantum Mechanics. I therefor started reading: Geometry of state spaces. In: Entanglement and Decoherence (A. Buchleitner et al., eds.). Lecture Notes in Physics 768, Springer Verlag, Berlin, New York, 2009, 1-60. A document that can also be found as a manuscript via: http://www.physik.uni- leipzig.de/~uhlmann/PDF/UC07.pdf Even though I thought that I have a solid background in abstract algebra I somewhat got lost in Chapter 2 when he's trying to classify all the *-algebras that represent actual physical systems (starting at page 24 in the document). Do you have some recommendations for texts that introduce the *-algebra language in Quantum Mechanics in a more 'detailed' way. Because I kind of have the feeling that at a certain point Uhlmann just keeps skipping steps and I also lack some of the physical intuition concerning partial traces, canonical traces, purification and all that. From time to time I'd also be happy to see a concrete example. I'm looking forward to your responses. Best regards.
[ -0.008495728485286236, 0.00894390232861042, 0.002708014566451311, 0.006776070222258568, 0.01694047823548317, 0.004654208663851023, 0.004807132761925459, -0.0253479965031147, -0.011337023228406906, -0.0103115513920784, -0.0014765674713999033, 0.01501104049384594, -0.01661071926355362, -0.01...
[ 0.05731172487139702, 0.16464723646640778, -0.04485448822379112, -0.10404710471630096, 0.06581642478704453, 0.13087256252765656, 0.20729295909404755, -0.1656363308429718, -0.27189570665359497, -0.5429393649101257, -0.23081530630588531, -0.12491985410451889, 0.43944472074508667, 0.5961310863...
So, as the title says: I made the foolish mistake of 'using' an alchemical tome, which means I now find it basically impossible to find ANYTHING on my transmutation slab. Is there a data file somewhere I can remove to wipe the settings for the transmutation tablet? Somewhere in a config folder? My plan would be to make the few useful items I can actually find, then wipe the settings and learn those items again... I'm playing SMP, but it's a friend's server so I can get him to remove files if needed!
[ 0.015473773702979088, 0.01834130473434925, 0.005284076556563377, 0.009007763117551804, -0.006411784328520298, 0.00000844523310661316, 0.0066108969040215015, -0.007812179625034332, -0.02258758619427681, 0.0034893490374088287, -0.003756379010155797, 0.01684831827878952, -0.00803558249026537, ...
[ 0.20320232212543488, 0.28627052903175354, 0.26951274275779724, 0.2988034188747406, 0.3455468416213989, -0.1872030347585678, 0.1962098628282547, 0.29753175377845764, -0.24163785576820374, -0.32152068614959717, 0.012388711795210838, 0.4817201495170593, -0.09839028865098953, 0.157137915492057...
I have multiple sub-questions but they are related. * What would the object look like if in were passing by? * What would a star look like if we were traveling near $c$? Would the perspective be a a large blue-shifted disk??
[ -0.030673248693346977, 0.01795004867017269, 0.025403685867786407, 0.02839219570159912, 0.011715946719050407, -0.019596656784415245, 0.012014675885438919, 0.015946460887789726, -0.03189776837825775, 0.019183404743671417, -0.00018784833082463592, 0.015012443996965885, 0.02247890830039978, 0....
[ 0.5521721839904785, 0.03064095228910446, 0.13735489547252655, 0.29143282771110535, -0.33236294984817505, 0.19438786804676056, -0.14475876092910767, 0.4891793429851532, -0.3578126132488251, -0.5633732676506042, 0.34633195400238037, 0.4879298508167267, 0.0860019326210022, 0.4321689009666443,...
The "Introductory Statistics with R" book contains a section that deals with correlations (section 6.4 in the second edition). The book shows Pearson, Spearman and Kendall correlation coefficients computed on the `blood.glucose` and `short.velocity` columns of the thuesen data set. The p-values associated with these coefficients are 0.048, 0.139 and 0.119, correspondingly. The book then says the following: > Notice that neither of the two nonparametric correlations is significant at > the 5% level, which the Pearson correlation is, albeit only borderline > significant. I have several problems with this paragraph. First of all, my naive guess would be that since the non-parametric coefficients do not imply linearity, they will tend to be "significant" more frequently than Pearson's r. Am I right? Secondly, and more importantly, is such a comparison between p-values of different tests applied on the same data legit? (I'm talking about real-life comparisons and not about trivial examples in a text book) If it is, how one need to interpret the notion that linear correlation is "significant", while rank or concordance correlation isn't?
[ -0.0035396020393818617, 0.014765840955078602, -0.015871524810791016, 0.014398006722331047, 0.013918254524469376, -0.0006599676562473178, 0.008131546899676323, -0.02254413813352585, -0.009498772211372852, 0.005091073922812939, 0.0037239943630993366, 0.013804620131850243, -0.016072960570454597...
[ 0.196821928024292, 0.1926301121711731, 0.758755087852478, -0.04611212760210037, -0.2688429355621338, 0.08391798287630081, 0.05394235625863075, -0.5275055766105652, 0.1775042712688446, -0.057604189962148666, 0.08874807506799698, 0.5312336683273315, -0.2169708013534546, 0.12187246233224869, ...
I have a random point defined by: int x = rand.nextInt(9)-4;//-4 to +4 int y = rand.nextInt(4)+2;//+2 to +5 int z = rand.nextInt(9)-4;//-4 to +4 Y happens to be the "up" vector, but I suspect that's irrelevant. I want to rotate this point so that it ends up relative to a vector that passes through (i,j,k) rather than relative to the vector that passes through (0,1,0). Thus if (x,y,z) is (0,6,0) and (i,j,k) is (1,1,0) the result should be about (5,5,0). Essentially I'm trying to draw a random deflection vector based on the initial input force vector. In this case, fractures through 3D voxel rock, but bullet deflection lines off armor make for a good visualization.
[ -0.0007637343369424343, 0.011131186038255692, -0.017625879496335983, 0.009960198774933815, -0.0046384590677917, -0.0018977539148181677, 0.00504634715616703, 0.011291122063994408, -0.008919971063733101, -0.0000457612331956625, -0.0062766848132014275, 0.010329367592930794, -0.00328333117067813...
[ -0.21994446218013763, -0.30209940671920776, 0.4053735136985779, -0.0917641818523407, -0.1926635205745697, 0.5025250911712646, -0.16776609420776367, -0.41486498713493347, -0.21477310359477997, -0.5666503310203552, 0.27191880345344543, 0.028691301122307777, -0.13224726915359497, 0.3029045462...
Is it possible places images at the same horizontal level? \begin{figure}[h] \includegraphics[scale=0.54]{/h.jpg} \caption{Architecture} \label{fig:Architecture}\hfill \includegraphics[scale=0.54]{H.jpg} \caption{Architecture} \label{fig:Architecture} \end{figure} When I try doing this, all the images go to the last page.. no clue what's going on. i'm using TexMaker. username ~ % identify H.jpg H.jpg JPEG 668x449 668x449+0+0 8-bit DirectClass 32.3KB 0.000u 0:00.000 username ~ % identify h.jpg h.jpg JPEG 692x433 692x433+0+0 8-bit DirectClass 38.2KB 0.000u 0:00.000
[ -0.0028782810550183058, 0.003951353020966053, -0.006467834115028381, 0.02914547547698021, -0.003192273201420903, -0.01674421690404415, 0.005792084150016308, 0.013469291850924492, -0.01839485950767994, 0.003831162815913558, -0.01814357191324234, 0.003535057418048382, 0.00033615902066230774, ...
[ -0.03982017561793327, 0.02484411559998989, 0.5762667059898376, 0.16831710934638977, 0.3219466507434845, 0.17680513858795166, 0.05636288598179817, -0.2815017104148865, -0.19591261446475983, -0.7664974927902222, 0.06553489714860916, 0.34791064262390137, -0.09432598948478699, -0.0944135263562...
I have an Xperia X10. Suddenly my virtual QWERTY keyboard has vanished and has been replaced by some other layout. When I enter the "Contacts" list an try to edit I now get a standard phone keyboard layout instead of the QWERTY one. What could have happened?
[ 0.004369617439806461, 0.00941089540719986, 0.0037922763731330633, 0.021084848791360855, -0.011237121187150478, -0.028817756101489067, 0.010690663009881973, 0.030018778517842293, -0.018542367964982986, -0.016129586845636368, -0.022528616711497307, 0.017580639570951462, -0.00737640680745244, ...
[ 0.21561044454574585, 0.29467710852622986, 0.47260230779647827, -0.099370576441288, 0.4427035450935364, 0.11080767214298248, 0.1742178350687027, 0.1888384073972702, -0.35650670528411865, -0.6198171973228455, 0.27499863505363464, 0.6506341695785522, -0.2505440413951874, 0.39069509506225586, ...
There are two different category "Life Sciences" and "General Lab" and there are sub-categorical product on "Life Sciences". Now "General Lab" category want to fetch all the sub-category and products which "Life sciences" have. The code is in category.php. Here is the whole code: ` <div id="container"> <div id="content" role="main"> <?php if ( is_category('general-lab') ) : ?> <?php //FETCHING ONLY GENERAL LAB CATEGORY $paged = (get_query_var('paged')) ? get_query_var('paged') : 1; $the_query = new WP_Query("posts_per_page=10&category_name=life-sciences&paged=".$paged); ?> <h1 class="entry-title"><?php echo single_cat_title("", TRUE); ?></h1> <?php $count=0; while ( $the_query->have_posts() ) : $the_query->the_post(); ?> <?php if($count % 2 == 0) echo '<div class="left">'; else echo '<div class="right">'; ?> <div id="post-<?php the_ID(); ?>" <?php post_class(); ?>> <div class="cat-thumb"><?php echo get_post_meta($post->ID, '_mcf_block-one', true); ?></div> <div class="cat-entry"> <h2 class="entry-title"><a href="<?php the_permalink(); ?>" title="<?php printf( esc_attr__( 'Permalink to %s', 'twentyten' ), the_title_attribute( 'echo=0' ) ); ?>" rel="bookmark"><?php the_title(); ?></a></h2> <?php the_excerpt(); ?> </div> </div> </div> <?php if($count % 2 != 0) echo '<div class="clear"></div>';?> <?php $count++; endwhile; ?> <?php if ( $the_query->max_num_pages > 1 ) : ?> <div id="nav-below" class="navigation"> <div class="nav-previous"><?php previous_posts_link( __( '<span class="meta-nav">&larr;</span> Previous', 'twentyten' ) ); ?></div> <div class="nav-next"><?php next_posts_link( __( 'Next <span class="meta-nav">&rarr;</span> ', 'twentyten' ) ); ?></div> </div><!-- #nav-below --> <?php endif; ?> <?php else : ?> <h1 class="entry-title"><?php echo single_cat_title("", TRUE); ?></h1> <?php $count=0; while (have_posts()) : the_post(); ?> <?php if($count % 2 == 0) echo '<div class="left">'; else echo '<div class="right">'; ?> <div id="post-<?php the_ID(); ?>" <?php post_class(); ?>> <div class="cat-thumb"><?php echo get_post_meta($post->ID, '_mcf_block-one', true); ?></div> <div class="cat-entry"> <h2 class="entry-title"><a href="<?php the_permalink(); ?>" title="<?php printf( esc_attr__( 'Permalink to %s', 'twentyten' ), the_title_attribute( 'echo=0' ) ); ?>" rel="bookmark"><?php the_title(); ?></a></h2> <?php the_excerpt(); ?> </div> </div> </div> <?php if($count % 2 != 0) echo '<div class="clear"></div>';?> <?php $count++; endwhile; ?> <?php if ( $wp_query->max_num_pages > 1 ) : ?> <div id="nav-below" class="navigation"> <div class="nav-previous"><?php previous_posts_link( __( '<span class="meta-nav">&larr;</span> Previous', 'twentyten' ) ); ?></div> <div class="nav-next"><?php next_posts_link( __( 'Next <span class="meta-nav">&rarr;</span> ', 'twentyten' ) ); ?></div> </div><!-- #nav-below --> <?php endif; ?> <?php endif; ?> </div><!-- #content --> </div><!-- #container --> <?php $current_category = single_cat_title("", FALSE); $parent_cat = get_the_category(); $back_to_current = get_cat_name($parent_cat[0]->category_parent); if ( is_category( array( 'life-sciences','consumables','histology','forensics','pharmaceutical' ) ) == $current_category) { ?> <?php echo '<div id="cat-menu"><h3>'.$back_to_current.'</h3>'; wp_nav_menu( array('container_id' => 'sub-page', 'menu' => $current_category ) ); echo '</div>'; } elseif ( is_category('general-lab')) { ?> <?php echo '<div id="cat-menu"><h3>'.$current_category.'</h3>'; wp_nav_menu( array('container_id' => 'sub-page', 'menu' => $back_to_current ) ); echo '</div>'; } else { ?> <?php echo '<div id="cat-menu"><h3>'.$back_to_current.'</h3>'; wp_nav_menu( array('container_id' => 'sub-page', 'menu' => $back_to_current ) ); echo '</div>'; } ?>
[ -0.0020175380632281303, 0.007218530401587486, -0.0010254798689857125, 0.020818905904889107, 0.012497439980506897, -0.006689372006803751, 0.010170339606702328, 0.021549835801124573, -0.016397548839449883, 0.004779216833412647, -0.01833268813788891, 0.0019617995712906122, -0.008483629673719406...
[ 0.04491076245903969, -0.18828731775283813, 0.4866034686565399, 0.214874267578125, 0.08770551532506943, 0.0609453059732914, -0.2913089990615845, -0.07395465672016144, -0.14071379601955414, -0.34168848395347595, -0.3726101815700531, 0.38436952233314514, -0.10781154781579971, 0.50440603494644...
Below is the LaTeX code for which I am getting the LaTeX error missing `\begin document{}`. How can I resolve this error? \documentclass{acm_proc_article-sp} \usepackage{algorithmic} \usepackage{soul} \usepackage[english]{babel} \usepackage{setspace} \usepackage{psfrag} \usepackage{epsfig} \usepackage{graphicx} \usepackage{amssymb,amsmath} \usepackage{graphicx} \usepackage{footmisc} \usepackage{cases} \usepackage{verbatim} \usepackage{fancyhdr} \usepackage{color} \usepackage{url} \usepackage[capitalize]{cleveref} \usepackage{placeins} \usepackage{subfigure} \usepackage{multirow} \usepackage{makecell} \usepackage{amsthm} \usepackage{setspace} \usepackage{moreverb} \usepackage{paralist} \let\proof\relax \let\endproof\relax \def\nref#1{(\ref{#1})} \def\figref#1{Fig.~\ref{#1}} \def\Dirfig{./Figures/} \begin{document} \title{} \maketitle \begin{abstract} \end{abstract} \keywords{D, P, E} \bibliographystyle{plain} \bibliography{allcomm} \end{document} Here is what I get by using \filelist in the log file This is pdfTeX, Version 3.1415926-2.4-1.40.13 (MiKTeX 2.9 64-bit) (preloaded format=pdflatex 2013.10.13) 13 OCT 2013 07:50 entering extended mode **e2sc.tex (F:\phases4en\trunk\PowerAwSC13\e2sc.tex LaTeX2e <2011/06/27> Babel <v3.8m> and hyphenation patterns for english, afrikaans, ancientgreek, ar abic, armenian, assamese, basque, bengali, bokmal, bulgarian, catalan, coptic, croatian, czech, danish, dutch, esperanto, estonian, farsi, finnish, french, ga lician, german, german-x-2012-05-30, greek, gujarati, hindi, hungarian, iceland ic, indonesian, interlingua, irish, italian, kannada, kurmanji, latin, latvian, lithuanian, malayalam, marathi, mongolian, mongolianlmc, monogreek, ngerman, n german-x-2012-05-30, nynorsk, oriya, panjabi, pinyin, polish, portuguese, roman ian, russian, sanskrit, serbian, slovak, slovenian, spanish, swedish, swissgerm an, tamil, telugu, turkish, turkmen, ukenglish, ukrainian, uppersorbian, usengl ishmax, welsh, loaded. (F:\phases4en\trunk\PowerAwSC13\acm_proc_article-sp.cls ("C:\Program Files\MiKTeX 2.9\tex\latex\graphics\epsfig.sty" Package: epsfig 1999/02/16 v1.7a (e)psfig emulation (SPQR) ("C:\Program Files\MiKTeX 2.9\tex\latex\graphics\graphicx.sty" Package: graphicx 1999/02/16 v1.0f Enhanced LaTeX Graphics (DPC,SPQR) ("C:\Program Files\MiKTeX 2.9\tex\latex\graphics\keyval.sty" Package: keyval 1999/03/16 v1.13 key=value parser (DPC) \KV@toks@=\toks14 ) ("C:\Program Files\MiKTeX 2.9\tex\latex\graphics\graphics.sty" Package: graphics 2009/02/05 v1.0o Standard LaTeX Graphics (DPC,SPQR) ("C:\Program Files\MiKTeX 2.9\tex\latex\graphics\trig.sty" Package: trig 1999/03/16 v1.09 sin cos tan (DPC) ) ("C:\Program Files\MiKTeX 2.9\tex\latex\00miktex\graphics.cfg" File: graphics.cfg 2007/01/18 v1.5 graphics configuration of teTeX/TeXLive ) Package graphics Info: Driver file: pdftex.def on input line 91. ("C:\Program Files\MiKTeX 2.9\tex\latex\pdftex-def\pdftex.def" File: pdftex.def 2011/05/27 v0.06d Graphics/color for pdfTeX ("C:\Program Files\MiKTeX 2.9\tex\generic\oberdiek\infwarerr.sty" Package: infwarerr 2010/04/08 v1.3 Providing info/warning/error messages (HO) ) ("C:\Program Files\MiKTeX 2.9\tex\generic\oberdiek\ltxcmds.sty" Package: ltxcmds 2011/11/09 v1.22 LaTeX kernel commands for general use (HO) ) \Gread@gobject=\count79 )) \Gin@req@height=\dimen102 \Gin@req@width=\dimen103 ) \epsfxsize=\dimen104 \epsfysize=\dimen105 ) ("C:\Program Files\MiKTeX 2.9\tex\latex\amsfonts\amssymb.sty" Package: amssymb 2013/01/14 v3.01 AMS font symbols ("C:\Program Files\MiKTeX 2.9\tex\latex\amsfonts\amsfonts.sty" Package: amsfonts 2013/01/14 v3.01 Basic AMSFonts support \@emptytoks=\toks15 \symAMSa=\mathgroup4 \symAMSb=\mathgroup5 LaTeX Font Info: Overwriting math alphabet `\mathfrak' in version `bold' (Font) U/euf/m/n --> U/euf/b/n on input line 106. )) ("C:\Program Files\MiKTeX 2.9\tex\latex\amsmath\amsmath.sty" Package: amsmath 2013/01/14 v2.14 AMS math features \@mathmargin=\skip41 For additional information on amsmath, use the `?' option. ("C:\Program Files\MiKTeX 2.9\tex\latex\amsmath\amstext.sty" Package: amstext 2000/06/29 v2.01 ("C:\Program Files\MiKTeX 2.9\tex\latex\amsmath\amsgen.sty" File: amsgen.sty 1999/11/30 v2.0 \@emptytoks=\toks16 \ex@=\dimen106 )) ("C:\Program Files\MiKTeX 2.9\tex\latex\amsmath\amsbsy.sty" Package: amsbsy 1999/11/29 v1.2d \pmbraise@=\dimen107 ) ("C:\Program Files\MiKTeX 2.9\tex\latex\amsmath\amsopn.sty" Package: amsopn 1999/12/14 v2.01 operator names ) \inf@bad=\count80 LaTeX Info: Redefining \frac on input line 210. \uproot@=\count81 \leftroot@=\count82 LaTeX Info: Redefining \overline on input line 306. \classnum@=\count83 \DOTSCASE@=\count84 LaTeX Info: Redefining \ldots on input line 378. LaTeX Info: Redefining \dots on input line 381. LaTeX Info: Redefining \cdots on input line 466. \Mathstrutbox@=\box26 \strutbox@=\box27 \big@size=\dimen108 LaTeX Font Info: Redeclaring font encoding OML on input line 566. LaTeX Font Info: Redeclaring font encoding OMS on input line 567. \macc@depth=\count85 \c@MaxMatrixCols=\count86 \dotsspace@=\muskip1
[ 0.009646749123930931, -0.0017733434215188026, 0.010983222164213657, 0.01773754507303238, 0.0293439831584692, 0.012101818807423115, 0.008880461566150188, 0.007909683510661125, -0.009626450948417187, -0.004811795428395271, -0.011102184653282166, -0.00042570545338094234, 0.003637358546257019, ...
[ -0.18099480867385864, 0.456122487783432, 0.3508475720882416, -0.19559407234191895, 0.23718449473381042, 0.11020127683877945, 0.3116439878940582, -0.24587316811084747, -0.029802773147821426, -0.9699617028236389, -0.2656424641609192, 0.4783329963684082, -0.25149017572402954, -0.3466888070106...
`man su` says: You can use the -- argument to separate su options from the arguments supplied to the shell. `man bash` says: -- A -- signals the end of options and disables further option processing. Any arguments after the -- are treated as filenames and arguments. An argument of - is equivalent to --. Well then, let's see: [root ~] su - yuri -c 'echo "$*"' -- 1 2 3 2 3 [root ~] su - yuri -c 'echo "$*"' -- -- 1 2 3 2 3 [root ~] su - yuri -c 'echo "$*"' -- - 1 2 3 1 2 3 [root ~] su - yuri -c 'echo "$*"' - 1 2 3 1 2 3 What I expected (output of the second command differs): [root ~] su - yuri -c 'echo "$*"' -- 1 2 3 2 3 [root ~] su - yuri -c 'echo "$*"' -- -- 1 2 3 1 2 3 [root ~] su - yuri -c 'echo "$*"' -- - 1 2 3 1 2 3 [root ~] su - yuri -c 'echo "$*"' - 1 2 3 1 2 3 Probably not much of an issue. But what's happening there? The second and the third variants seem like the way to go, but one of them doesn't work. The fourth one seems unreliable, `-` can be treated as `su`'s option.
[ 0.0014591340441256762, 0.018703417852520943, -0.017600785940885544, 0.01639518514275551, -0.017390325665473938, -0.004301509354263544, 0.0073939720168709755, -0.028261292725801468, -0.016727976500988007, -0.006796862930059433, -0.01893635280430317, -0.005406167358160019, -0.01176857389509677...
[ -0.4274342954158783, 0.2712504267692566, 0.32501348853111267, -0.5315336585044861, 0.15575464069843292, 0.07803153246641159, 0.4290946424007416, -0.38125118613243103, -0.10119280964136124, -0.3032521903514862, -0.7709705829620361, 0.5333868861198425, -0.23680996894836426, 0.348822444677352...
I have setup exim mta on freebsd. I have also setup DKIM on this machine and my mails are properly signed. I want to know how to setup **Domainkeys** along with DKIM on exim so that my mails are signedf both by Domainkeys as well DKIM. This is because gmail honors DKIM as well yahoo honors Domainkeys.
[ 0.007787863723933697, -0.008250203914940357, -0.01307896338403225, 0.021652907133102417, -0.0008195283589884639, 0.03158007934689522, 0.013429269194602966, 0.028856871649622917, -0.01937108300626278, -0.021393761038780212, -0.017069688066840172, 0.008804566226899624, -0.014374583028256893, ...
[ 0.4094054102897644, 0.30916157364845276, 0.6270794868469238, 0.007195251528173685, -0.4249078631401062, -0.2753294110298157, 0.4646415710449219, 0.10781612247228622, -0.11023017764091492, -0.7264000177383423, -0.00040433506364934146, 0.6483162045478821, -0.0918198823928833, 0.2007491886615...
I found Wildberry Princess' Diary and delivered the dessert pizza to get the Receipt, so I can prove two of the three people Lemongrab has taken are innocent. But Peppermint Butler won't sign anything proving their innocence until I can prove all three are innocent. How do I prove the baby is innocent?
[ 0.01306577492505312, 0.011468375101685524, -0.007696004584431648, 0.01784098520874977, -0.04702909290790558, 0.0463237501680851, 0.015172768384218216, 0.019331514835357666, -0.023406505584716797, -0.047043103724718094, -0.01523000467568636, 0.004211861174553633, 0.008535574190318584, -0.02...
[ 0.10976564884185791, 0.4388255178928375, 0.3285467326641083, 0.21009595692157745, 0.21100759506225586, 0.3104088306427002, 0.48044002056121826, -0.19973590970039368, 0.08218562602996826, -0.05630449950695038, 0.07908421754837036, 0.319295734167099, 0.016147172078490257, 0.41640985012054443...
On a tablet I have a situation where I have multiple users with multiple accounts, and I am trying to have the phone's state backed up in such a way that if I upgrade the operating system, every user's data is backed up. I would like to do this without imaging the phone, so that this backup can be applied to say, a newer version of the Android OS. I am ok with the backup being finicky, if there is some significant changes to the OS, so long as minor changes don't break it. I have tried Titanium Backup, and while it works perfectly for a single user, it does not work when multiple users are involved. Neither the user's, nor their data, is backed up. What application that can achieve this? Edit: To elaborate, backing up each user individually would work but it would be slow. We may be doing this on many devices, so this is primarily a way to save us time.
[ 0.012885909527540207, 0.005488151218742132, -0.00755839329212904, 0.008152389898896217, 0.009586736559867859, 0.0050881667993962765, 0.007880909368395805, 0.007288072258234024, -0.016113488003611565, -0.016196858137845993, -0.009769365191459656, 0.013482602313160896, -0.00778006948530674, ...
[ 0.07667937129735947, 0.4573833644390106, 0.1447540819644928, 0.1188773587346077, 0.30531200766563416, 0.2901568114757538, 0.20694074034690857, 0.24301382899284363, -0.16509047150611877, -0.4328421354293823, 0.30026692152023315, 0.5769700407981873, -0.15376386046409607, -0.20811589062213898...
What is the recommended acquisition strategy for having multi-platform, service-connected devices for mobile developers? Is it necessary to have separate phone numbers and service plans for each platform? My guess is that having a Droid, iPhone and Windows 7 phones all on the same plan and same phone number is out.
[ -0.020209500566124916, 0.019626330584287643, -0.0050138747319579124, 0.0025475360453128815, 0.033303771167993546, -0.0006360886618494987, 0.015016138553619385, 0.06379105150699615, -0.022050587460398674, -0.07368352264165878, -0.030751356855034828, 0.041069671511650085, 0.025145336985588074,...
[ 0.17610271275043488, 0.18466123938560486, 0.14200206100940704, 0.2968102693557739, 0.23347873985767365, 0.19304712116718292, 0.01040567085146904, 0.10982439666986465, -0.27418383955955505, -0.31610357761383057, -0.11162526160478592, 0.5587856769561768, -0.2421531230211258, -0.1288235634565...
I'm trying to set 'orderby = none' to my loop but it is not working. Here's my code: $query = new WP_Query(array('showposts'=>2, 'post__in' => array(99,4,5,2,8,55), 'orderby'=>'none')); Could anyone help me? Thanks.
[ 0.021447589620947838, 0.030438276007771492, -0.01240704208612442, 0.028764069080352783, 0.0022655732464045286, 0.008546881377696991, 0.009002485312521458, 0.009909006766974926, -0.013428941369056702, 0.008350662887096405, -0.009155621752142906, 0.002788390265777707, -0.01800723373889923, 0...
[ 0.1406075656414032, 0.17746582627296448, 0.16870850324630737, -0.0799957886338234, 0.10128786414861679, 0.18455471098423004, 0.6340383291244507, 0.23778465390205383, -0.2309311181306839, -0.6173563003540039, 0.12759128212928772, 0.180181622505188, -0.18566299974918365, 0.29915738105773926,...
Is it possible to create a target directory, similar in nature to the `mkdir -p` switch, where I can define a non-existent target directory within my tar command, and tar will create the directory for me? I know I can redirct the output to a directory using `tar -C /target/dir`, but this doesn't work if the target directory is non-existent.
[ -0.015241232700645924, 0.01878937892615795, -0.0076072933152318, 0.0023958857636898756, -0.0048035127110779285, -0.014452321454882622, 0.008836568333208561, -0.004869204945862293, -0.019615497440099716, -0.030460290610790253, -0.013811469078063965, 0.00041291938396170735, 0.00294334511272609...
[ 0.14735598862171173, -0.2776291072368622, 0.006555961910635233, 0.10581012815237045, -0.005576469004154205, 0.0027752036694437265, 0.14580947160720825, 0.061484694480895996, 0.21521885693073273, -0.4396197497844696, 0.20505964756011963, 0.9308915138244629, -0.19571268558502197, 0.025881368...
Is there a visible difference between 60 FPS vs 120 FPS? I am going under the assumption that the monitor is your standard 60 Hz monitor. I've heard arguments that you would want a higher FPS for first-person shooter games then 60 FPS. I am looking for a good answer that actually has some technical merit behind if possible.
[ -0.031018586829304695, 0.00813402608036995, -0.0017252969555556774, 0.0017476431094110012, 0.008548758924007416, -0.04224202781915665, 0.010613969527184963, -0.029528046026825905, -0.018599530681967735, -0.020869433879852295, 0.0008627728093415499, 0.011225401423871517, -0.015847181901335716...
[ 0.7826264500617981, -0.18367762863636017, 0.17464900016784668, 0.30153706669807434, -0.09457779675722122, -0.36250415444374084, -0.03272416070103645, 0.14220762252807617, -0.2226947546005249, -0.47169366478919983, 0.2843020558357239, 0.8998109102249146, -0.01719527877867222, -0.14708429574...
# Background I'm currently working on a codebase for what is to become a forthcoming website's content "engine", where it will take in different types of standardized data (implemented with XML), parse it, and then generate content dynamically and accordingly. The data itself also includes a list of context clues or helper objects for searching, comparing, analyzing, etc., which is being developed concurrently into a group as a data interaction layer and it is completely separate from the engine. Each generated content will emit some embedded "signatures" that can be sniffed by the interaction layer for post-processing (possibly with few generated hidden input values). # The Problem Nonetheless, the immediate problem that I'm facing is that although those list of hidden signatures are related to the content that is generated, it is essentially not needed by the engine to let the browser know how content should be presented initially. Logically speaking, is it usually the engine's job to parse all data no matter what type of data it is given so that the interaction layer can simply manipulate it later? Or, since that data type acts as a helper for data manipulation, should the interaction layer step in and act as the parent for that data type specifically, so that the engine can just focus on content? The root of the problem is that I started separating things to its specialized components for better maintainability in the future but I'm not exactly sure where to stop separating chunks to its essentials. Should I treat data as simply data? That can't possibly be right.
[ -0.008476102724671364, 0.0058924974873661995, -0.00846178736537695, 0.0003580772317945957, -0.0033817316871136427, 0.008146955631673336, 0.006414906121790409, 0.01423615776002407, -0.011745097115635872, 0.013822940178215504, -0.004575124941766262, 0.0013721822760999203, 0.01413099654018879, ...
[ 0.7100033164024353, -0.06937674432992935, 0.602410078048706, 0.12205979973077774, -0.05236239731311798, -0.0275186225771904, 0.19618968665599823, 0.16533637046813965, 0.0006751712644472718, -0.49163731932640076, -0.1213865876197815, 0.25919753313064575, -0.40783706307411194, -0.36722227931...
Is there a standard way to report the percent correctly predicted when predicting a binary outcome? Using glm in r, the results are predicted probabilities. However, in order to make a comparison to another model, I want to report a single percent correctly predicted value from my binary model. Do I simply choose a cutpoint, and if so, how? Here is a simple example of the code. model.results <- glm(binary.outcome ~ predictor1 + predictor2, family=quasibinomial) Thanks,
[ 0.028222033753991127, 0.016531487926840782, -0.004891445394605398, 0.012219615280628204, -0.002879596082493663, 0.001166766625829041, 0.008508408442139626, 0.006612156983464956, -0.01962311938405037, -0.025451237335801125, -0.0035268752835690975, 0.01018448080867529, -0.00710875540971756, ...
[ -0.06323365867137909, -0.20293790102005005, -0.06211180239915848, 0.2465689778327942, -0.24560438096523285, -0.12240351736545563, 0.1965392827987671, -0.10863061994314194, -0.16095776855945587, -0.24385938048362732, 0.0022449481766670942, 0.568453311920166, 0.0063579804264009, 0.2126509100...
I think I heard on a previous StackOverflow podcast that COBOL was used as the programming language for traffic lights (or something like that), so this got me interested. I did a quick Google search and found this little article: > Today, Cobol is everywhere, yet largely unheard of by millions of people who > interact with it daily when using the ATM, stopping at traffic lights or > buying a product online. > > The statistics on Cobol attest to its huge influence on the business world: > There are over 220 billion lines of Cobol in existence, a figure which > equates to about 80 per cent of the world’s actively used code. **There over > a million Cobol programmers in the world.** There are 200 times as many > Cobol transactions that take place each day than Google searches. I didn't really trust the source seeing as how it's on some random PHPBB forum. So how accurate are these figures? Are there really **220 billion** lines of COBOL? I assume a few people/companies still use COBOL, _but how many?_
[ 0.0021975617855787277, -0.0035096686333417892, -0.015206472016870975, 0.0037809559144079685, -0.01682722195982933, -0.011680010706186295, 0.006714574992656708, -0.0026403802912682295, -0.011814046651124954, -0.021040311083197594, -0.007845724001526833, 0.006688945926725864, 0.024757612496614...
[ 0.5074076056480408, 0.3051286041736603, 0.3351120054721832, -0.027561333030462265, -0.05651961639523506, -0.12380961328744888, -0.0380915068089962, 0.7856842875480652, -0.3257746994495392, -0.3627678155899048, 0.01020818017423153, 0.5280420184135437, -0.4734911620616913, -0.285229802131652...
Version 0.18 introduced action groups, including a dedicated abort group. This can be activated using the pop out button next to the altimeter, but I'm wondering if this can be done from the keyboard. Instead of searching for the UI button while my rocket is exploding below me, I'd rather just hit a keyboard hotkey. What is the hotkey?
[ -0.009849276393651962, 0.0014976193197071552, -0.009195609018206596, 0.005381639115512371, -0.028567247092723846, -0.007414381019771099, 0.008782812394201756, 0.026010163128376007, -0.022619331255555153, 0.03396429866552353, -0.01109376922249794, 0.016364090144634247, 0.014202166348695755, ...
[ 0.071998231112957, -0.3429568111896515, 0.6360133290290833, 0.08449498564004898, -0.06818195432424545, -0.1746871918439865, 0.4306035041809082, -0.3090111017227173, 0.03914007544517517, -0.22881893813610077, -0.315512090921402, 0.683540940284729, -0.5404881834983826, -0.8877530694007874, ...
Looking at the band of stability, my first intuition is to conclude (erroneously) that there is a stable isotope of every element that lies close to the belt of stability. Why is this false? (For example, Uranium has no stable isotopes.) Or in particular, why are some number of protons inherently unstable in the nucleus? Again, based purely on my poor intuition, I would assume that some number of neutrons could be arranged with that number of protons to create stability but that is not true in all cases. Why are these numbers of protons (e.g. 92 for Uranium) unstable?
[ 0.01175165269523859, 0.015603211708366871, -0.01323087140917778, 0.016439076513051987, 0.0033960214350372553, 0.022427953779697418, 0.008348032832145691, -0.00022465712390840054, -0.018842553719878197, -0.0030017360113561153, -0.0008295925799757242, 0.008502444252371788, -0.02299638092517852...
[ 0.3181953728199005, 0.15430137515068054, 0.010486932471394539, -0.12380962818861008, -0.05289094150066376, -0.6233022212982178, 0.14656677842140198, -0.20363938808441162, 0.05377478152513504, -0.3941938877105713, -0.08817330747842789, 0.29959800839424133, -0.2992393970489502, 0.63703823089...
When I use levitation and run on my psychic ball I feel like I'm moving faster is this actually the case or is it a sort of illusion? I know I sacrifice maneuverability so it'd make sense that I'm also faster.
[ -0.014187447726726532, 0.009448393248021603, -0.015443571843206882, -0.007233984302729368, -0.04994432255625725, -0.014990581199526787, 0.011458752676844597, -0.04126831889152527, -0.018327917903661728, -0.015514623373746872, -0.004332678858190775, 0.026479266583919525, -0.02565106935799122,...
[ 0.6408271193504333, -0.23518937826156616, -0.145735502243042, 0.2970133125782013, -0.44978097081184387, -0.016494911164045334, 0.4523513913154602, -0.17430157959461212, -0.4372112452983856, -0.6216788291931152, 0.458543598651886, 0.2217758148908615, -0.08568239212036133, -0.026460623368620...
In each application there are concepts like users, customers, etc. and some part of the application tries to manage them. For example: 1. keeping the information of people, organizations, firms, etc. 2. managing users (people who work with the system) 3. managing customers (people who purchase goods/services) 4. distinguishing customers, from users, from ordinary people, etc. 5. ... Do we have a universal term for that area of each application (that subsystem, that module, that whatever)? Or if I want to ask more precisely, can we package customers, users, people, organizations, etc. all in one module, and name it like X? Is this a correct taxonomy of subsystems? Update: Is **Entities** a good name?
[ 0.00989858340471983, 0.001583919976837933, -0.00533750094473362, 0.01587926410138607, -0.0028035300783813, -0.004808782134205103, 0.008028983138501644, 0.01569550856947899, -0.012869585305452347, -0.014975100755691528, -0.021053418517112732, 0.01329113356769085, 0.01943589374423027, 0.0222...
[ 0.17646710574626923, 0.2374802827835083, 0.4076949954032898, 0.14421632885932922, 0.33769071102142334, 0.16820695996284485, -0.07077652961015701, -0.022059308364987373, -0.2921620309352875, -0.526807963848114, -0.38786399364471436, 0.5686346888542175, -0.22018945217132568, 0.05246554315090...
I have a little problem with my python script. I wrote it to convert MODIS data from .hdf into .tif. It will convert all files in a folder using a loop. I do not know which command I must enter to assume the files from the folder as input file. I tried *.hdf but then I got errors indicating that the file doesn´t exist. The output-name of the new .tif should be the input name of the .hdf. At this point in my script it doesn´t work for me. I have no idea what I have to write now in the "gdal_translate"-command. I tried it with *.tif but this doesn't work. Thanks for reply! Cheers #!/usr/bin/python import os import glob from osgeo import gdal from osgeo import ogr from osgeo import osr from osgeo import gdal_array from osgeo import gdalconst from osgeo.gdalconst import * path = '' path = raw_input('Directory? (z.B. C:\Daten\Modis\):') # change to path os.chdir( path ) # checks directory vcheck = os.getcwd() file = glob.glob('*.hdf') for file in glob.glob('*.hdf'): os.system('gdal_translate -of GTiff -a_srs EPSG:4326 "HDF4_EOS:EOS_GRID:"*.hdf":MODIS_Grid_8Day_Fire:FireMask" *.tif')
[ -0.006169913336634636, 0.015251659788191319, -0.0061460696160793304, 0.024421848356723785, 0.005725740455091, 0.00030942982994019985, 0.008588409051299095, 0.031159870326519012, -0.022224370390176773, -0.011498482897877693, -0.012929437682032585, -0.001783379353582859, -0.01131596788764, -...
[ 0.5002937912940979, 0.198195219039917, 0.26328468322753906, -0.2184228003025055, -0.29528388381004333, -0.1139954999089241, 0.3750470280647278, 0.04580249637365341, -0.014140202663838863, -0.7357511520385742, 0.32043981552124023, 0.6935191750526428, -0.44643938541412354, 0.3388346135616302...
Is this worded correctly if it was spoken in an interview? > I am like a clean slate. I do not have any preconceived notions about how > the company runs
[ 0.0006035151891410351, 0.020548025146126747, -0.008877488784492016, 0.01971765235066414, -0.0064867655746638775, -0.0015192307764664292, 0.014609813690185547, 0.0036564671900123358, -0.016886891797184944, 0.041557490825653076, -0.01599857024848461, 0.017013203352689743, 0.023768143728375435,...
[ 0.6590453386306763, 0.4991154372692108, 0.15782581269741058, -0.24886947870254517, -0.24907982349395752, -0.17700909078121185, 0.3931180536746979, 0.2076527178287506, 0.12468273937702179, -0.34271669387817383, 0.09693019092082977, 0.5799831748008728, 0.18659912049770355, 0.1057047322392463...
I saw these topics and sure they help a lot, Resize longtable to width of landscape page How to fit landscape multi-page table to textwidth But still it does not fix my problem for some reason. I am having this table \begin{landscape} \setlength\LTcapwidth{\textwidth} % default: 4in (rather less than \textwidth...) \setlength\LTleft{0pt} % default: \parindent \setlength\LTright{0pt} % default: \fill \begin{longtable}{@{\extracolsep{\fill}}|*{19}{c|}} \caption[Multinomial logistic regression results for daily data of the major Eurozone, the US and the UK market indices, January, 1, 2005, to 20, July, 2012]{Multinomial logistic regression results for daily data of the major Eurozone, the US and the UK market indices, January, 1, 2005, to 20, July, 2012.} \label{grid_mlmmh} \\ & & \multicolumn{6}{l}{Bottom tails} & & & \multicolumn{6}{l}{Top tails} & \\ \cmidrule{2-9} \cmidrule{11-18} % 2 orizonties grammes aristera kai deksia \\ & \multicolumn{3}{l}{(1)} & \multicolumn{3}{l}{(2)} & \multicolumn{3}{l}{(3)} & \multicolumn{3}{l}{(4)} & \multicolumn{3}{l}{(5)} & \multicolumn{3}{l}{(6)} \\ \cmidrule{2-3} \cmidrule{5-6} \cmidrule{8-9} \cmidrule{11-12} \cmidrule{14-15} \cmidrule{17-18}% 2 orizonties grammes aristera kai deksia \\ & Coeff & $\Delta$Prob & & Coeff & $\Delta$Prob & & Coeff & $\Delta$Prob & & Coeff & $\Delta$Prob & & Coeff & $\Delta$Prob & & Coeff & $\Delta$Prob \\ \endfirsthead \multicolumn{3}{c}% {{\bfseries \tablename\ \thetable{} -- continued from previous page}} \\ & & \multicolumn{6}{l}{Bottom tails} & & & \multicolumn{6}{l}{Top tails} & \\ \cmidrule{2-9} \cmidrule{11-18} % 2 orizonties grammes aristera kai deksia \\ & \multicolumn{3}{l}{(1)} & \multicolumn{3}{l}{(2)} & \multicolumn{3}{l}{(3)} & \multicolumn{3}{l}{(4)} & \multicolumn{3}{l}{(5)} & \multicolumn{3}{l}{(6)} \\ \cmidrule{2-3} \cmidrule{5-6} \cmidrule{8-9} \cmidrule{11-12} \cmidrule{14-15} \cmidrule{17-18}% 2 orizonties grammes aristera kai deksia \\ & & & Coeff & $\Delta$Prob & Coeff & $\Delta$Prob & Coeff & $\Delta$Prob & Coeff & $\Delta$Prob & Coeff & $\Delta$Prob & Coeff & $\Delta$Prob \\ \endhead \hline \multicolumn{3}{|r|}{{Continued on next page}} \\ \hline \endfoot \hline \hline \endlastfoot \midrule PIIGS \\ $\beta_{01}$(constant) & 100 & 0.8 & & 0.021 & 0.018 & & 0.043 & 0.146 & & 0.074 & 0.427 & & 0.019 & 0.427 & & 0.019 & 0.427 \\ Log-likelihood \\ $Pseudo-R^{2}$ \\ \\ Non-PIIGS \\ $\beta_{01}$(constant) & 100 & 0.8 & 0.021 & 0.018 & 0.043 & 0.146 & 0.074 & 0.427 & 0.019 & 0.427 & & 0.019 & 0.427 \\ \\ $Pseudo-R^{2}$ \\ \\ \end{longtable} \end{landscape} and it gets off the page all the time. Also, the first "\delta Prob" column on the left gets bigger than the rest all the time.. Can you please help me fitting this table to page width and also ensure that all columns have the same size? I am probably doing something very wrong! Thank you!
[ 0.014088035561144352, 0.010282657109200954, -0.010718625970184803, 0.000026239431463181973, 0.004980748053640127, 0.014962296932935715, 0.008522942662239075, 0.01898365467786789, -0.011355789378285408, -0.019398752599954605, -0.006786355748772621, 0.009639615193009377, -0.004574506543576717,...
[ -0.16000042855739594, -0.08946803957223892, 0.749195396900177, -0.06866953521966934, -0.12549921870231628, -0.0945071130990982, 0.20932884514331818, -0.17452025413513184, -0.3521413207054138, -0.688423752784729, 0.050834983587265015, 0.8593398332595825, -0.1405717134475708, -0.150301307439...
I ran a small pilot study and computed both a p-value and Cohen's d. Now how do I compute the number of samples needed for the full study if I want a given power (e.g. 95%). I hoped I could use this table, but I can't figure out how it's calculated. **Edit:** More details on the study: We're manipulating a single variable (condition A and B). Each participant will have N readings from each condition. The question is, how do we compute how many participants we'll need given the results of a pilot study with one participant.
[ 0.012232494540512562, 0.0176901426166296, -0.007313381414860487, 0.011501665227115154, -0.016401564702391624, -0.010352643206715584, 0.00623004836961627, -0.034081898629665375, -0.01856958493590355, -0.011760996654629707, -0.000016419216990470886, 0.00942203588783741, -0.017080267891287804, ...
[ 0.6333271265029907, -0.2690901458263397, 0.07368545979261398, 0.20244374871253967, -0.15873238444328308, 0.5627625584602356, 0.11943560838699341, -0.6662318706512451, -0.12001947313547134, -0.30600711703300476, 0.4290018379688263, 0.43309614062309265, -0.16882236301898956, 0.29292505979537...
I set all nodes and coordinates that I can be set right away in the at the top of the `TikZ` picture. However, a line can obstruct the node. If the node is after the line, I could use `fill = white` which will white out that portion of the line. Can I achieve something like this without moving the nodes below the offending line or lines? I know we could suggest moving the node, but in some cases, this wouldn't be the case. \documentclass{article} \usepackage{tikz} \begin{document} \begin{tikzpicture}[ every label/.append style = {font = \small}, dot/.style = {outer sep = +0pt, inner sep = +0pt, shape = circle, draw = black, label = {#1}}, dot/.default =, small dot/.style = {minimum size = .1cm, dot = {#1}}, small dot/.default =, big dot/.style = {minimum size = .15cm, dot = {#1}}, big dot/.default = ] \node[scale = .75, fill = black, big dot = {below: \(F\)}] (F) at (2.5, 0) {}; \draw (2.5, 1) -- (2.5, -1); \end{tikzpicture} \end{document} ![enter image description here](http://i.stack.imgur.com/Dn2ln.png) ![enter image description here](http://i.stack.imgur.com/YUv11.png)
[ 0.006267060525715351, 0.0005943134892731905, -0.02326926216483116, 0.008941435255110264, -0.024859964847564697, 0.002767438068985939, 0.00825470220297575, 0.021643733605742455, -0.01416327990591526, 0.023858042433857918, -0.007591491565108299, 0.006445612758398056, -0.002993230242282152, 0...
[ 0.28209203481674194, -0.01077487226575613, 0.5237081050872803, -0.1763201653957367, 0.3425556421279907, -0.06052525341510773, 0.21857969462871552, -0.35882875323295593, 0.06424081325531006, -0.9226949214935303, -0.04997855797410011, 0.32390815019607544, -0.3200116753578186, 0.1041211783885...
I am writing a SOAP based ASP.NET Web Service having a number of methods to deal with Client objects. e.g: * int AddClient(Client c) => returns Client ID when successful * List GetClients() * Client GetClientInfo(int clientId) In the above methods, the return value/object for each method corresponds to the "all good" scenario i.e. A client Id will be returned if AddClient was successful or a List<> of Client objects will be returned by GetClients. But what if an error occurs, how do I convey the error message to the caller? I was thinking of having a Response class: Response { StatusCode, StatusMessage, Details } where Details will hold the actual response but in that case the caller will have to cast the response every time. What are your views on the above? Is there a better solution? ---------- UPDATED ----------- Is there something new in WCF for the above? What difference will it make If I change the ASP.NET Web Service to a WCF Service?
[ -0.007801644504070282, 0.011711488477885723, -0.005787947215139866, 0.007063635624945164, -0.0008393791504204273, 0.009502293542027473, 0.0075822025537490845, 0.003995836246758699, -0.011863800697028637, -0.011100741103291512, 0.002871894743293524, 0.024706009775400162, -0.000943260500207543...
[ 0.47966861724853516, 0.1303350031375885, 0.46185269951820374, -0.22606393694877625, -0.225778728723526, -0.013506133109331131, 0.35099872946739197, -0.4149429202079773, 0.07600996643304825, -0.8384020924568176, 0.018846167251467705, 0.4233678877353668, -0.35925599932670593, -0.111965209245...
I am new to PCA and wanted to do a bit of experimentation on my data set just to see what it looked like (using R). I am not able to give access to the data here since it is confidential. However, if there is some other kind of statistic/visualization you would like to see that would help you answer my questions please let me know and I will provide it. I found the following information about the explained variance: Component Prop.Var 1 0.911804348 2 0.033618098 3 0.020827269 4 0.011772988 5 0.006611746 6 0.005372772 7 0.004464788 8 0.003436401 9 0.002091589 This raises the following questions: 1. Am I justified in removing the other 8 principal components? 2. How do I interpret 91% of explained variance on one component? 3. If I only kept one component what would be the best way to visualize the data? Below is how the graph of the first two principal components looks. The spread of the data like this is not surprising given how little of the variance is on the second component. ![Principal Components 1 and 2](http://i.stack.imgur.com/g9wXP.png) As I mentioned, I am new to PCA so I really do not know if there is even any useful information to be found from this kind of dimensional reduction. Any insight would be appreciated.
[ -0.004060009960085154, 0.01056668721139431, -0.0014046416617929935, 0.014026465825736523, 0.005366222932934761, -0.0035048348363488913, 0.003707063151523471, -0.004040681757032871, -0.008975796401500702, 0.0014214259572327137, -0.0007099105860106647, 0.0059204064309597015, -0.006337874568998...
[ 0.4833288788795471, -0.07808759063482285, 0.4018741548061371, 0.12109649926424026, -0.12834987044334412, 0.30394455790519714, 0.017018992453813553, -0.4938944876194, 0.12611375749111176, -0.21881097555160522, 0.5213168859481812, 0.5311660170555115, -0.20958459377288818, 0.26489749550819397...
I am trying to estimate the probability of an event using a low number of observations. The naive estimator $\hat{p} =\frac{\text{number of positive observations}}{\text{total number of observations}}$ works well when the total number of observations is big enough, but if you have only a few observations, there is a decent chance that you will erroneously conclude to a 0 or 100% probability. I suppose you could set a prior distribution on the estimated probability (say, uniform), and look for better estimators. I suppose this problem has already been tackled many times, so where should I look?
[ 0.01445423811674118, 0.02055765688419342, -0.005350474733859301, 0.009222332388162613, -0.008105970919132233, -0.01410410925745964, 0.007037688046693802, -0.01050504855811596, -0.011493910104036331, -0.004887531511485577, -0.00018482940504327416, 0.008093612268567085, -0.009614363312721252, ...
[ 0.16486406326293945, 0.07104358077049255, -0.08867884427309036, 0.09619823843240738, -0.06580184400081635, 0.109658382833004, 0.2307901829481125, -0.05698025971651077, -0.13056819140911102, -0.655022382736206, 0.13181570172309875, 0.5475385785102844, -0.2933398187160492, 0.1247235909104347...
I've been looking around lately, specifically at some of the MS tools that are available, and I'm noticing a big focus on designer tools and wizards. Not just for UI development but for everything. * Entity Framework has the modeller * RIA Services has the DomainService wizard(s) * Workflow has the whole workflow designer thingy... (I don't know, haven't really used it) There's more, but I think you get the idea. There's lots of designer tools. Using some of these I find that: **They complicate matters beyond the prescribed use cases** (ie all of the tech demo videos) I have been evaluating some of these technologies recently, and trying to work with them I end up having to dissect exactly what the designers, modellers and wizards are doing for me... otherwise I'm lost when I actually have to try to do something with whatever was created. This ends up being a case where I have to fight the tool, or fight its output enough such that I could have just done the whole thing myself without it - and had a _much_ stronger understanding of what's going on. I find this particularly infuriating with the silverlight designer and RIA services domain service wizard. I find myself asking "What good is this tool if I have to figure out its inner workings or re-write half of its output in order to use it?" **They're not as fast** The selling point of these tools is to increase productivity and this point may change over time using the tool, and doesn't necessarily apply to UI designers (though, in some cases it still does -> I'm looking at you silverlight designer). I find that I can hack some code much faster than I can drag-and-drop, resize, move, whatever in a designer. **The UI gets in the way of the model** Maybe this is just me, but when I am using anything reminiscent of a UML design tool I end up spending more time laying everything out so that my lines don't cross and so that I can see it all on the screen than modelling what I'm trying to achieve. **They're no fun** Half of the reason that I code for a living is that I enjoy it. Clicking checkboxes and selecting comboboxes and then fixing everything that comes out isn't fun. **I don't appear to be alone** The community seems to _not_ want these tools either. The best example I can think of at the moment is Entity Framework Code-First. So I ask: 1. Do designer tools _actually_ improve productivity? 2. Are they fun-killers? 3. Is 'the community' _actually_ asking for more designer tools, or are vendors just thinking we are?
[ 0.005076859146356583, 0.003965358715504408, -0.005445113405585289, 0.0006606249371543527, -0.007280871272087097, -0.010999330319464207, 0.0051834462210536, 0.021515104919672012, -0.010225786827504635, -0.0175321064889431, -0.0041988384909927845, 0.0085840355604887, 0.0039969016797840595, -...
[ 0.42088738083839417, 0.10389953851699829, 0.3215271532535553, 0.2665691077709198, -0.5289895534515381, -0.20130017399787903, 0.31367820501327515, 0.006396833807229996, -0.4801643490791321, -0.8461822271347046, -0.009527807123959064, 0.3410642743110657, 0.019206151366233826, 0.3130984902381...
I want to remove the below text links in Edit Post screen: All (8) | Published (5) | Draft (1) | Pending (2) | Trash (2) After searching, i found that this can be done by this code here, but sadly it only works with 'post' type. I'm failed to make it works with my custom post type: add_action( 'views_edit-post', 'remove_edit_post_views' ); function remove_edit_post_views( $views ) { if( get_post_type() === 'movie' ) { unset($views['all']); unset($views['publish']); unset($views['trash']); } return $views; } What's wrong with my code ?
[ 0.007842667400836945, 0.013600323349237442, 0.0022041124757379293, 0.018281947821378708, 0.004257898312062025, -0.004326952621340752, 0.00540422648191452, 0.023095183074474335, -0.016424335539340973, 0.027975846081972122, -0.012775365263223648, 0.014402668923139572, -0.01088804379105568, 0...
[ 0.48353397846221924, 0.06972652673721313, 0.8482369184494019, -0.21071474254131317, 0.02054750733077526, 0.1627892106771469, 0.3721162676811218, -0.2233506441116333, -0.2803598940372467, -0.4429263472557068, 0.05093168839812279, 0.3721824288368225, -0.5216447114944458, 0.209605872631073, ...
Is there a way to operate on Dataset across levels to apply a function `f` to a `key -> value` pair in an Association as follows: pivotApply[f_, key_ -> value_] := key -> f[key, value] For example, take the Titanic dataset and pivot up "class": titanicClass = titanic[GroupBy[Key@"class"], KeyDrop[Key@"class"]]; For this problem, assume only titanicClass. `"class"` key has been deleted and so can't referenced. The motivation is tree-structured data where it would be inefficient to insert the keys `{"1st"...}` at the deeper level. Here `foo` represents a "client" function to be applied to deeper level data but depenent on class values. **EDIT // #age >= 18 &** foo["1st", data_] := Dataset[data][All, #age >= 18 &]; foo["2nd", data_] := Dataset[data][All, "sex"]; foo["3rd", data_] := Dataset[data][All, "survived"]; Is there a way to avoid inefficient downcasting to Normal and back to Dataset to use AssociationMap? Also note there are Datasets nested in a larger one, which should also be cast to Lists (since there is an outer Dataset). AssociationMap[pivotApply[foo, #] &, titanicClass // Normal] // Dataset ![enter image description here](http://i.stack.imgur.com/L4Rik.jpg) Is there a way to query Datasets across levels like this? This seems extremely roundabout, but haven't figured an implementation using, say Keys and Values and related Dataset operators.
[ 0.011695250868797302, 0.00879138708114624, -0.005109413992613554, 0.01012461632490158, 0.01689795032143593, -0.0010213754139840603, 0.007709454745054245, 0.012988580390810966, -0.01170855388045311, 0.03005404584109783, -0.010113758966326714, 0.011240215972065926, -0.008467799052596092, 0.0...
[ -0.12230344861745834, -0.2919963598251343, 0.21292762458324432, 0.15249580144882202, -0.03412528336048126, -0.05004352331161499, -0.04935568943619728, -0.2834039628505707, -0.26667800545692444, -0.03535051271319389, -0.02892615832388401, -0.01597445271909237, -0.5382609367370605, -0.093733...
I looked at the documentation and examples on CTAN, but `circuitikz` is hard to get the hang of. I am trying to create ![wanted](http://i.stack.imgur.com/T9PHD.png) Here is what I have so far: \documentclass{convert = false]{standalone} \usepackage{circuitikz} \begin{document} \begin{circuitikz} \draw (-.5, 0) to[L, l = L] (1.4, 0) to[R, l = R] (5, 0); \end{circuitikz} \end{document} but the spacing is weird, and I can't seem to figure out how to set it up.
[ 0.004276353865861893, 0.0054319798946380615, -0.005610204301774502, 0.02620803192257881, 0.0034636433701962233, -0.016622105613350868, 0.006741202436387539, -0.0022606030106544495, -0.017986871302127838, 0.006878988817334175, -0.0058419350534677505, -0.007572378031909466, -0.0135114816948771...
[ 0.3947880268096924, -0.019492123275995255, 0.42502719163894653, -0.2607954740524292, 0.28977686166763306, -0.2800426185131073, 0.23550887405872345, -0.41069188714027405, 0.0847930833697319, -0.7091464996337891, 0.02561517059803009, 0.6623597145080566, -0.270675927400589, 0.0067454474046826...
I have several `blocks` in a slide: \begin{block}{block 1} block 1 \end{block} \begin{block}{block 2} block 2 \end{block} \begin{block}{block 2} block 2 \end{block} Does anyone know how to overlay them so as to show them block by block by left-clicking mouse?
[ 0.0026962084230035543, 0.014957956038415432, -0.02096717804670334, 0.014000361785292625, -0.03126264736056328, 0.0035021009389311075, 0.006703992839902639, 0.021546218544244766, -0.016606805846095085, 0.020814310759305954, -0.004930619616061449, -0.009810205549001694, -0.0022474327124655247,...
[ 0.23712819814682007, 0.5262476205825806, 0.4790721535682678, -0.06079500541090965, -0.14511580765247345, -0.021438952535390854, 0.01062676403671503, -0.3485298752784729, -0.07548496872186661, -0.5573021173477173, 0.05766349285840988, -0.24162861704826355, -0.0878545269370079, 0.21215239167...
In computer science and programming we talk a lot about "abstraction" by which we mean to create ever "higher" level code to that accomplishes increasingly complex task with less programmer decision making. See Opposite of verb “abstract” and noun “abstraction” for some context. However, abstract means, "to drag away, detach, pull away, divert" which is not what programmers do when we create abstractions in software. For example, back-in-the-day, text was represented by sequences of of byte codes e.g. 078, 111, 119, 032, 105, 115, 032, 116, 104, 101, 032, 116, 105, 109, 101 ... which we "abstracted" to arrays of characters: [N,o,w, ,i,s, ,t,h,e, ,t,i,m,e] ... which we "abstracted" to become a string: "Now is the time" ... which in modern languages is actually a class or similar structure: "Now is the time".methodThatManipulatesString At each stage of "abstraction" the minutia of working with textual information didn't disappear, we didn't remove it, and we really didn't (arguably) turn the textual information into some higher platonic form. All we did was hide all the grubby details and then bolt on yet more functionality with it's own grubby details hidden away. If you dig down into any language or platform, the details are all still there. It is somewhat analogous to a mechanical watch. The watchmakers can tell the time by looking at the gears and springs of a watch but for the rest of us, they hid the watches complexity behind the hands and face of the clock. So, I'm looking for a term, preferably latin, conveys this idea. Programmers already use "encapsulation" but that just means, "to put in." I want a term meaning "to hide, but not remove complexity" preferably something linguistically and conceptually related to "abstraction" itself. I think it's important in training to make sure new programmers don't think that abstracted code contain less complexity under the hood than the non- abstracted code despite it's more simple superficial appearance.
[ -0.00804422702640295, 0.0129552586004138, -0.008303244598209858, 0.008107341825962067, -0.024474836885929108, -0.011156819760799408, 0.007466099224984646, 0.002785210032016039, -0.01052788645029068, -0.0011402848176658154, -0.0038078061770647764, 0.0027938850689679384, 0.010978344827890396, ...
[ 0.18498003482818604, 0.09786774963140488, -0.157216876745224, 0.2712770104408264, -0.4108670651912689, 0.23340755701065063, -0.17586828768253326, 0.044297587126493454, -0.28457438945770264, -0.34921538829803467, -0.1771063655614853, 0.4798128306865692, -0.1826854795217514, 0.12120144814252...
When my post contains approximately more than 30,000 words then the page only shows the post title and no text at all. The same happens with shared host and with WP running on my PC. Any ideas? Any text filter I might disable? Thanks.
[ -0.020072711631655693, -0.0061256783083081245, 0.012806637212634087, 0.036744192242622375, 0.00806424580514431, -0.014245456084609032, 0.010711649432778358, 0.00877724401652813, -0.02024034596979618, -0.010684968903660774, -0.020022569224238396, -0.0015676870243623853, 0.03496353700757027, ...
[ 0.5069217681884766, 0.2054409682750702, 0.44407758116722107, 0.15172149240970612, -0.33862951397895813, 0.1736726611852646, 0.435906320810318, 0.39379456639289856, -0.05814402922987938, -0.7628350257873535, 0.0742473229765892, 0.02251812256872654, -0.09748376905918121, 0.2662893831729889, ...
So I bought CS:GO yesterday when it was on sale and left it to download over night. Today, I opened it up to play, did the usual first-time setup stuff and then got into the game. I tried the weapons course to see what had changed from CS:S and then went on to play a game with others. I tried the "Find Game" option for a bomb mission and after a couple of minutes of searching the pop- up said Creating server.... then Joining server.... but it never did. I opened the console while it was Joining and there were lines continuously generated saying that it tried to reserve a certain IP adress but it failed after three tries and that went on for many addresses. So I thought it was a matchmaking problem and went to the community servers and joined a random one. It never did and it kicked me out after 10 retries. I tried to join multiple servers with no luck. If could guide me into some troubleshooting steps I would be grateful. Thanks in advance,George. P.S: Things I have already tried : Verify local cache,reinstall,restart client,game and pc,all of my ports are open because of a DMZ host, tried increasing the maximum ping allowed to 350,checked to see if I was opted on some beta program but I wasn't. By the way, while joining games the errors in the console included some missing vgui content and some things about STEAM Sockets.
[ -0.024912765249609947, -0.002770360093563795, 0.004143329802900553, -0.004404111765325069, 0.014283588156104088, -0.010808237828314304, 0.00606317026540637, -0.0027330631855875254, -0.014857398346066475, 0.00723971426486969, 0.006590669974684715, 0.022034339606761932, 0.006531764753162861, ...
[ 0.24882198870182037, -0.13881127536296844, 0.24561050534248352, 0.06963537633419037, -0.14633725583553314, 0.00956183485686779, 0.018257776275277138, 0.2511085867881775, -0.13951164484024048, -0.7549214959144592, 0.21685531735420227, 0.4231765568256378, -0.26649823784828186, 0.276973128318...
In _Uncharted 3: Drake's Deception_ , I missed a few cutscenes for various reasons and feel like I've missed a few key plot points. Since there's no shortage of cutscenes, I'm hoping there is some method of watching them again. Is there any method to replay a cutscene without reloading from a save and playing through again?
[ 0.01993093453347683, 0.022403419017791748, 0.011929611675441265, 0.025004664435982704, 0.0015420933486893773, -0.024109741672873497, 0.007782903499901295, -0.040899164974689484, -0.02757132239639759, 0.005824180319905281, -0.014535416848957539, 0.015430513769388199, 0.004170448053628206, -...
[ 0.1691611260175705, -0.18652108311653137, 0.34855562448501587, 0.33424702286720276, -0.15414288640022278, -0.33312925696372986, 0.5372149348258972, 0.18779172003269196, -0.27293092012405396, -0.4905705451965332, -0.22969815135002136, 0.38236168026924133, -0.2494988739490509, 0.209358483552...
I'm not trying to start a flame war. With the same 1080p HD TV, will we be able to tell the difference between the two consoles running the same game? I'm also wondering if developers can actually squeeze more raw power from Wii U than PS3. Is that the case?
[ -0.005840220022946596, 0.003940606024116278, 0.010568363592028618, 0.008594190701842308, -0.06013558804988861, -0.019721051678061485, 0.00876754242926836, -0.021019447594881058, -0.01996735669672489, -0.02280185930430889, -0.005061144009232521, 0.018316276371479034, 0.02186639793217182, -0...
[ 0.8356913924217224, 0.07168493419885635, 0.07531436532735825, 0.454959899187088, -0.39139115810394287, -0.30119577050209045, 0.029890555888414383, -0.13153435289859772, -0.114036425948143, -0.49229469895362854, 0.19967585802078247, 0.36000505089759827, 0.19999359548091888, -0.0656558424234...
I've the problem that, using something related to Fisher-scoring, the gradient, which is usually the sum over a variable times a value which depends upon the parameter we are looking for, the updates will mainly focus on the elements whos scale is large. **Example:** If I would use some spline to estimate the impact on an dependend variable in an smooth fashion, the scale of the spline-columns will be very small compared to a parametric impact of a variable which is time (measured in days) or price (measured in hundreds). **My question is** : Should I use some sort of standardization of variables before optimization to adjust for the difference in scale or should I rather, using something like the `optim()` function in R, provide a scaling-factor within the optimization procedure?
[ -0.0024115622509270906, 0.01433231495320797, -0.0092727430164814, 0.018056990578770638, 0.010134637355804443, -0.0214826762676239, 0.006615432910621166, -0.02430751360952854, -0.0077683450654149055, -0.03564650937914848, -0.009216486476361752, 0.010313456878066063, -0.011006618849933147, 0...
[ 0.11343277990818024, -0.1214957982301712, 0.4442376494407654, 0.007208069320768118, -0.22514314949512482, 0.22287820279598236, -0.2154332399368286, -0.15478499233722687, -0.36155059933662415, -0.5206325650215149, 0.4563887119293213, 0.5159019231796265, -0.016911419108510017, 0.232540428638...
Does the phrase _then before, now once more_ have any meaning in English? Or does it exist just because it rhymes so nicely? Or does it exist at all? Likewise, what about _that time then, once again_?
[ 0.009767157025635242, 0.042500827461481094, -0.022428130730986595, 0.030707476660609245, -0.0020024245604872704, 0.007536604069173336, 0.011422685347497463, -0.002710269531235099, -0.029542790725827217, -0.005848543718457222, 0.003900605021044612, 0.014742732048034668, 0.03843386098742485, ...
[ 0.11762415617704391, 0.27235427498817444, 0.11389756947755814, -0.12004406750202179, 0.1477053314447403, 0.16064639389514923, 0.28852468729019165, 0.42434990406036377, -0.48478806018829346, -0.5072757005691528, -0.15372955799102783, 0.15424898266792297, -0.15170058608055115, 0.319412559270...
I created a non-hierarchical custom post type and, by default, it's sorted by date published. I know I can reorder them by Title with `query_posts()` in the archive template: global $query_string; query_posts( $query_string . '&orderby=title&order=ASC' ); but it takes another SQL query on each archive page. Is there a way to register the sort order natively, eventually to get posts sorted by title even in the admin?
[ 0.018491549417376518, 0.023252418264746666, -0.004224289208650589, 0.01803203858435154, -0.021754756569862366, 0.003030287567526102, 0.010322132147848606, 0.015069140121340752, -0.018383417278528214, -0.004423200152814388, -0.010617660358548164, 0.015343161299824715, -0.010902048088610172, ...
[ 0.24304425716400146, 0.3817084729671478, 0.5545455813407898, -0.07316169887781143, -0.08751634508371353, 0.10968004167079926, -0.01798977330327034, -0.0255991593003273, -0.3647669553756714, -0.5326577425003052, -0.08356674015522003, 0.05085272341966629, -0.10548663884401321, 0.392112731933...
I hope this practical question is not OT and not too trivial for this forum. I am renting an apartment in a duplex with a shared water heater and dryer. Turns out, both water heater and the dryer are connected to my meter. The neighbors seem willing to pay their share for the past 6 months, only how to estimate their fair share? I tried to leave the apartment for a couple of days and take a reading from the electric meter before and after. It appears that in these 48 hours, my meter changed by 10KWH, i.e. $5/day, with everything but the fridge off. Realistic? There are 3 to 4 of them, and they seem to shower a lot! I know, because I don't get much hot water. The neighbors are moving out in a month, so I don't have much time to experiment. There is no way to interrupt the electric circuit to insert a meter; so I am thinking, maybe an induction meter would be feasible? How reliable would it be? EDIT: actually, turns out the dryer plugs into a standard 220V outlet. But heater probably takes much more electricity anyway, right?
[ -0.008276382461190224, 0.00972972996532917, -0.011918289586901665, 0.014755294658243656, -0.0009990418329834938, -0.014236433431506157, 0.007711326237767935, 0.004814269952476025, -0.009281466715037823, -0.016223907470703125, 0.007607477251440287, 0.021331410855054855, 0.006909476593136787, ...
[ 0.8592082262039185, 0.1512860804796219, 0.5929076075553894, 0.10486184060573578, 0.288765013217926, 0.4620644450187683, 0.3599240779876709, -0.20988605916500092, -0.3480828106403351, -0.6614086627960205, 0.47370192408561707, 0.22496774792671204, 0.18210706114768982, 0.20246747136116028, ...
I want to compose a string that contains newlines, so I tried the following: ToString@StringForm["SomeText='`1`'\[NewLine]as well as OtherText='`2`'.", "textA", "textB"] I would expect this to produce a string that looked like: SomeText='textA' as well as OtherText='textB' Instead, I get a string that looks like SomeText='textA' textB'. as well as OtherText=' Directly entering the string `"SomeText='textA'\[NewLine]as well as OtherText='textB'"` gives me the correct result. What is going on here?
[ -0.006767368409782648, 0.02091258205473423, -0.0042788321152329445, 0.008649546653032303, 0.003788175992667675, 0.001193065196275711, 0.0051970393396914005, 0.01055930182337761, -0.010970575734972954, 0.002546069212257862, -0.010036344639956951, 0.0032338653691112995, 0.009452017955482006, ...
[ 0.3170377314090729, -0.23198211193084717, 0.4359252154827118, -0.34614598751068115, -0.1356034278869629, 0.03024641051888466, 0.290149450302124, -0.1914711892604828, 0.09111685305833817, -0.6004700064659119, -0.1753971427679062, 0.29626864194869995, -0.3962567150592804, 0.04195164516568184...
It would be very handy to be able to open a linked graphic file (e.g: pdf) in a graphics editing program (e.g: Illustrator, Inkscape) simply by clicking on the link in the LaTeX editor or by clicking on the graphic in the output pdf. Are there any editors or pdf viewers that support this?
[ -0.011757518164813519, -0.009525713510811329, -0.004662145860493183, 0.02313615381717682, 0.006027875933796167, -0.016800349578261375, 0.010243289172649384, 0.010713758878409863, -0.03042212873697281, 0.007771937642246485, -0.015920979902148247, 0.0010001929476857185, 0.017782075330615044, ...
[ 0.39433324337005615, 0.035999931395053864, 0.3669346868991852, 0.4790189266204834, -0.1029859408736229, -0.4314575493335724, -0.13101612031459808, 0.16007967293262482, -0.38890141248703003, -0.6260883212089539, -0.27776846289634705, 0.55804044008255, 0.04765467718243599, 0.0459899418056011...
I am using LGE Nexus 4 which is running on Android Version 4.4.2. Somehow accidently i do a factory reset on my phone and in return i lost all my emails, messages, pictures i.e, almost everything. Now i am desperately want to recover them but don't know how to do it. I tried to find a solution on google but i got more confused by the given solutions on google. I tried a data recovery software(Dr Drone for android) but the software is not able to recover data from those devices which are running on >Android version 4.2. Technical team of Dr Drone for android want me to root the device but because of complications in the rooting process i don't want to root my device. Is there any way to recover them?
[ 0.002520995680242777, -0.003605867736041546, -0.006170067936182022, 0.006974535062909126, -0.005192249082028866, -0.01478495728224516, 0.0062028635293245316, 0.0338745079934597, -0.017247524112462997, -0.024843379855155945, -0.003368509467691183, 0.020586024969816208, 0.014560888521373272, ...
[ 0.3539273738861084, -0.06435049325227737, 0.7012501358985901, -0.012781518511474133, 0.22342193126678467, 0.05742645263671875, 0.27996209263801575, 0.052511055022478104, -0.2583792209625244, -0.7464895248413086, 0.0606703907251358, 0.6155208349227905, -0.5598762035369873, 0.024855926632881...
I have auto translated articles in portugues and I need to hire people to check the grammars etc. Considering I need portugues (that can speak english a little to make agreements) where should I look for? Thanks
[ 0.028735345229506493, 0.022356070578098297, -0.006211757194250822, 0.016968121752142906, -0.004132741596549749, 0.03477618098258972, 0.013781455345451832, -0.000590475567150861, -0.046908535063266754, 0.038526203483343124, 0.005879650358110666, 0.022321319207549095, 0.010845891200006008, 0...
[ 0.2685720920562744, 0.5650531649589539, 0.039657704532146454, 0.1329362690448761, 0.17399141192436218, -0.05205928906798363, 0.5670506358146667, 0.765573263168335, -0.003521149745211005, -0.7656282186508179, 0.25412988662719727, 0.6715789437294006, 0.40651825070381165, -0.31493380665779114...
I am using Post contents as a slider contents, i need to make the post infinite number of loop, ie, If i view the last post after that first post should display and before first post the last post. Please suggest
[ -0.0003525506763253361, 0.030954653397202492, -0.007876704446971416, 0.014917813241481781, -0.013148396275937557, -0.007056465372443199, 0.014149579219520092, 0.027036143466830254, -0.018832754343748093, -0.010082311928272247, -0.0196700356900692, 0.0020074606873095036, -0.022497618570923805...
[ 0.3954758942127228, 0.06903091073036194, 0.428327351808548, 0.40813055634498596, -0.1898147612810135, -0.07736654579639435, 0.2720867991447449, 0.24388235807418823, -0.1381949782371521, -0.4671200215816498, 0.10790224373340607, 0.0933038592338562, -0.27308377623558044, 0.4325208067893982, ...
I'm not very English literate, but I am annoyed by the use of "that" during the CrossFit games. Announcers, coaches, and athletes all said "that" more times than I can count. * Move _that_ bar. * Use _that_ hip. * Pick up _that_ weight. * Tell me about _that_ workout... "That" is consistently used in place of what I feel should be "the". I do not think I know enough to even ask a good question here... What role does "that" play when describing an object?
[ -0.005776979960501194, 0.0020994055084884167, -0.02087509259581566, 0.006150503642857075, 0.003697830718010664, -0.012330610305070877, 0.006463584490120411, 0.016713544726371765, -0.012726917862892151, -0.008171046152710915, -0.0018026581965386868, 0.008061502128839493, 0.0030659218318760395...
[ 0.27364879846572876, 0.24242693185806274, -0.12435676902532578, -0.20948047935962677, -0.4712035059928894, -0.21123793721199036, 0.27405834197998047, 0.33727121353149414, -0.09519362449645996, -0.5701313614845276, 0.22664989531040192, 0.332485556602478, 0.0032185802701860666, -0.2516949772...
I'm having an issue where my "Course Work" section, which is tabular, is offset by a space or two - whereas ever other section lines up perfectly. I've read up on the tabular environment but could only find information on adjusting spacing between columns and not before them. This is my code and the output, with the red highlighting the undesired space: ![enter image description here](http://i.stack.imgur.com/goFU5.png) \section{\textsc{Course Work}} \begin{tabular}{lllll} Molecular Biomechanics & \ \ & Stem Cell Biology & \ \ & Bioinstrumentation \\ Eukaryotic Molecular Biology & \ \ & Infectious Diseases & \ \ & Biochemistry \\ Graduate Bioinformatics Seminar & \ \ & Differential Equations & \ \ & Genetics \\ Undergraduate Research in MCD Biology & \ \ & Computational Biology & \ \ & Statistics \\ \end{tabular} Any ideas on how I can fix this? _I also apologize beforehand for using "blank columns" to adjust column spacing. I know it isn't the proper way to do that, but when my friend initially taught me some basic LaTeX that was how he did it and I just haven't felt the need to change it yet since learning more about the tabular environment._
[ -0.006517673376947641, 0.0006431146175600588, -0.01177450455725193, 0.024592682719230652, -0.016861867159605026, 0.01068104337900877, 0.006437621079385281, -0.006336156744509935, -0.01427226047962904, 0.00029997341334819794, -0.02125554345548153, 0.0065224431455135345, 0.01067777257412672, ...
[ 0.33957090973854065, 0.06740790605545044, 0.7096433639526367, -0.21147416532039642, 0.2825847268104553, 0.17062793672084808, -0.34064432978630066, -0.15146587789058685, -0.312776118516922, -0.6006171703338623, -0.20037546753883362, 0.4731815755367279, 0.03947319835424423, -0.19052541255950...
When you create a new document in Microsoft Word, you get a standard blank document with the right margins. Moreover when that document is exported to PDF, you are satisfied with the layout. Over and over I have tried to get the best out of LaTeX, to make it standard looking but I have failed. Using `fullpage` or `geometry` packages with custom sizes doesn't give best results. What settings would you use if you were to mimic a Word portrait document in LaTeX ? I'm not concerned with fonts, CM fonts are my favourite.
[ 0.0009597050375305116, 0.0033458119723945856, -0.002795061096549034, 0.009060336276888847, -0.006179601885378361, -0.003325756173580885, 0.007826756685972214, 0.007214952725917101, -0.010483397170901299, -0.0020107473246753216, -0.007971927523612976, 0.005012147594243288, 0.02624142915010452...
[ 0.34553518891334534, 0.02741040475666523, 0.243580624461174, 0.0655662938952446, -0.16779863834381104, 0.0823780819773674, 0.058189284056425095, 0.20321795344352722, -0.3516174554824829, -1.0520251989364624, 0.11014730483293533, 0.5252265334129333, 0.004895782098174095, 0.09601452946662903...
I'm trying to analyse the following sentence. > To understand the importance of this event you should know all the facts. It seems to me that this sentence is complex, and «To understand importance of this event» is dependent adverbial clause. But then, as my grammar book says, it would require a comma before «you», since the dependent adverbial clause comes before the independent clause. So, please help to do the syntax analysis of this sentence and to understand whether the comma is needed.
[ -0.007560106925666332, 0.021727951243519783, -0.011374427005648613, 0.006063608452677727, -0.016027309000492096, 0.005820997059345245, 0.010742620564997196, -0.014268060214817524, -0.013086200691759586, 0.018782606348395348, -0.020286450162529945, 0.0019883059430867434, 0.008514434099197388,...
[ -0.21226388216018677, 0.07147549837827682, 0.2710522711277008, -0.23519957065582275, -0.26394176483154297, 0.18835653364658356, 0.3255489468574524, 0.21123744547367096, -0.30439573526382446, -0.5932456851005554, -0.11547266691923141, 0.4401126801967621, 0.26479393243789673, 0.1756979525089...
I want to edit markdown source in emacs using the It's All Text plugin for firefox. In markdown you have to add two spaces to the line-end to get a new line. Suppose there is such a newline in a markdown source. If I open this in emacs via It's All Text, change something else and save it then the two spaces at the end of the line dissapear which is very annoying. So is there a way to configure emacs not to touch such spaces at line ends in this case?
[ 0.007028705906122923, 0.006830017082393169, -0.015410477295517921, 0.010976523160934448, 0.01326797530055046, 0.003162022680044174, 0.010182294063270092, 0.007436005398631096, -0.016719136387109756, 0.003302075434476137, -0.014839272014796734, 0.0076791029423475266, 0.00024669431149959564, ...
[ 0.42769357562065125, 0.13395573198795319, 0.4715332090854645, -0.021845625713467598, -0.22483742237091064, -0.06128447875380516, 0.24690839648246765, 0.15779279172420502, -0.04569697380065918, -0.5990167856216431, -0.028769047930836678, 0.5896701812744141, -0.5227268934249878, 0.2552377581...
What is the antonym for the verb force? (i.e. antonym of 'urge or force (a person) to an action')
[ -0.025889916345477104, 0.01933085359632969, -0.011808948591351509, 0.03364340960979462, -0.04892235994338989, -0.017443479970097542, 0.01380171999335289, -0.026088213548064232, -0.02082883007824421, 0.05274590104818344, -0.003431985853239894, 0.01752283424139023, 0.05171747878193855, 0.028...
[ -0.036481596529483795, -0.39359474182128906, 0.1007818803191185, -0.3589514493942261, -0.3981373906135559, 0.8753579258918762, 0.4128561019897461, -0.31053823232650757, -0.6750198602676392, -0.14205335080623627, -0.5182551145553589, 0.5674159526824951, -0.23535118997097015, -0.077231064438...
I have a bunch of magnets (one of those game-board thingies) given to me when I was a school-going lad over 20 years ago, and the magnets feel just as strong as it was the day it was given. As a corollary to this question Do magnets lose their magnetism?, is there a way to determine how _long_ a permanent magnet will remain a magnet? Addendum: Would two magnets remain a magnet for a shorter duration if they were glued in close proximity with like poles facing each other?
[ 0.018348628655076027, 0.006383588537573814, -0.01020688097923994, 0.010974965058267117, 0.006568583659827709, -0.005940067581832409, 0.006311881355941296, -0.0012213222216814756, -0.025473453104496002, 0.00538286380469799, 0.007993999868631363, 0.01906907930970192, 0.011025658808648586, -0...
[ 0.5656422972679138, 0.01488091703504324, 0.3926096558570862, 0.11062991619110107, -0.04297354817390442, -0.2105090618133545, -0.03219573199748993, 0.15324845910072327, -0.49491453170776367, -0.053035564720630646, 0.31277996301651, 0.2394234538078308, -0.052889131009578705, 0.57603633403778...
I am trying to install WordPress in a subdirectory. This is a new fresh install. Here is what I did: Assuming my site is `example.com`, and I wanted to install wordpress in `example.com/wordpress` A) I created a database on the server B) I copied all wordpress directories to `example.com/wordpress` C) I moved index.php from `example.com/wordpress` to `example.com/` D) I modified the last line of `index.php` from: `require( dirname( __FILE__ ) . '/wp-blog-header.php' );` and it got modified into: `require( dirname( __FILE__ ) . '/wordpress/wp-blog-header.php' );` E) I modified `wp-config.php` and updated it in accordance with the required DB name, user name, password, and server name Now, when I test hat locally, and when I go to `example.com`, I get redirected to the WordPress installation setup page. However, on the live server, I am redirected to `http://www.example.com/wp-admin/install.php` **(notice the missing wordpress folder, the redirect should be to`http://www.alwarkaa.com/wordpress/wp-admin/install.php`)**. I have done this before, but not sure what is causing the problem. Can you please help? I really ran out of ideas. I tested everything I can think off (including typos), but to no avail. Please help. Thanks.
[ -0.02124662697315216, 0.004231528379023075, 0.0016379689332097769, 0.028747960925102234, 0.018049664795398712, 0.02316700853407383, 0.007710212841629982, 0.005050618667155504, -0.019679974764585495, -0.011444810777902603, -0.01079696323722601, 0.009219912812113762, -0.004234963562339544, 0...
[ 0.20833194255828857, 0.2873729169368744, 0.632853090763092, -0.28873348236083984, 0.05908200889825821, -0.08580916374921799, 0.4486941695213318, -0.04401758313179016, -0.0016549010761082172, -0.7299630641937256, 0.23735861480236053, 0.3805197775363922, -0.3963320851325989, 0.25237888097763...
i want to mod rewrite change extension of any page to what i want. for ex. i want change `index.php` to `index.abc` is it possible with mod rewrite?
[ 0.02169998548924923, 0.025404075160622597, -0.014048677869141102, 0.017914237454533577, -0.003992878831923008, -0.004292658530175686, 0.012472853995859623, -0.007216925732791424, -0.03177851811051369, 0.0038376515731215477, -0.013819390907883644, 0.011454958468675613, -0.009747479110956192, ...
[ 0.2217070460319519, 0.026144297793507576, 0.7218260169029236, -0.1740296483039856, -0.17439088225364685, -0.1270504891872406, 0.26423537731170654, -0.28019946813583374, -0.27064865827560425, -0.6102180480957031, 0.029039565473794937, 0.528225302696228, -0.6087156534194946, 0.07637967169284...
I am using the `mathtools` package in order to right-align entries in small matrices. My code is below. The issue is that the delimiters in the first matrix are not the appropriate size; they should be the same size as the delimiters in the second matrix. I'm sure the issue is because of right- aligning. How can I fix the delimiters in the first matrix without resorting to using `\bigl(` and `\bigr)`? \documentclass[11pt]{article} \usepackage{amsmath,amssymb,amsfonts} \usepackage{mathtools} \begin{document} Here is a sentence. \begin{enumerate} % \item $A = \left( \begin{smallmatrix*}[r] 1 & 2 \\ 5 & 7 \end{smallmatrix*} \right)$ % \item $A = \left( \begin{smallmatrix*}[r] 1 & -1\\ 2 & 3 \end{smallmatrix*} \right)$ \end{enumerate} \end{document} ![picture](http://i.stack.imgur.com/NpFS9.png)
[ -0.00014485185965895653, 0.010151755064725876, -0.012645356357097626, 0.011405698023736477, -0.010507463477551937, 0.015500243753194809, 0.007799006067216396, 0.01757979765534401, -0.010257348418235779, -0.023258991539478302, -0.012596779502928257, 0.013734666630625725, -0.015474183484911919...
[ -0.1499139815568924, 0.41559481620788574, 0.3710121810436249, -0.21057343482971191, 0.1956564486026764, -0.021410349756479263, 0.1283988505601883, -0.29736873507499695, -0.15751633048057556, -0.6523658633232117, 0.016687465831637383, 0.6551885604858398, -0.23627717792987823, -0.12535576522...
I'd always thought that the word "bastard" had a negative and impolite meaning, but when I came across this wine (Fat _bastard_ Chardonnay) I started to doubt my belief. Can some native speaker explain if this usage of "bastard" is possible in positive sense nowadays? If so, what is the positive sense in this case? ![enter image description here](http://i.stack.imgur.com/irEsG.jpg)
[ -0.012754925526678562, 0.017814472317695618, -0.0008636779384687543, 0.03248687833547592, 0.003421097993850708, -0.013158002868294716, 0.006023078691214323, -0.0012236714828759432, -0.015003806911408901, -0.01468454860150814, -0.00007626580190844834, 0.008999373763799667, 0.02262825518846512...
[ 0.46509918570518494, 0.35301533341407776, -0.137835294008255, -0.16634628176689148, -0.5669378638267517, 0.029843417927622795, 0.5687071084976196, 0.3739958703517914, -0.41573366522789, -0.5401075482368469, 0.1873982548713684, 0.3118741512298584, 0.2737630009651184, 0.6679372787475586, 0...
When reading Scrum Guide, as the official text for scrum, I find out there is no specific solution to provide software testing in scrum. (the only hint is on page15) I'm a little vague on whether scrum is considered a software development methodology or not? If it is not, then how come some of its practices opposes _Extreme Programming_? (I know that in scrum guide, the author notes that scrum is a framework not a methodology, but still I'm not pretty clear on that) And what's more, I'm not sure if there are any other important textbook that I'm missing so far about scrum. I need them to be official or of great deal of public acceptance.
[ -0.022337544709444046, 0.0010120676597580314, 0.006255071610212326, 0.011242292821407318, 0.008962901309132576, -0.009772537276148796, 0.007306198589503765, 0.022649899125099182, -0.014609604142606258, -0.035417843610048294, -0.010866220109164715, 0.01644187606871128, 0.0017039794474840164, ...
[ 0.8073417544364929, 0.4519813060760498, 0.026621565222740173, -0.07521620392799377, -0.319751501083374, -0.5532492399215698, 0.26820603013038635, -0.12764020264148712, 0.14026592671871185, -0.07299932092428207, 0.18150974810123444, 0.307391494512558, -0.19792282581329346, 0.140841528773307...
I don't know the math to do this, so I am asking here if someone can work this out with all of the details I'm providing. Total mass EST.: 2,400 lbs. Length from front to back: 14 feet 6 inches. Ground clearance (space between floor and bottom of car base): 4.7 inches. Height (from ground to top of roof): 4 feet and 5.7 inches. Wheelbase (distance from center of front and rear wheels): 8 feet and 1 inch. My simple question is, how much force would it take to lift the back and front of the car from the bumpers in the air? If you want to really answer another question, **how much is lifted from one- side of either the back and front wheels**? The front bumper is a bit closer to the ground compared to the back one (almost 2 inches), so lifting depth must account somewhat. Last, the surface is very flat paved cement. Someone told me with a ballpark guess that the back is 600 lbs., one-side of the back is 240 lbs., one-side of the front is 565 lbs., and the whole front is 1,130 lbs.
[ -0.011592129245400429, 0.0063880993984639645, -0.013672919943928719, 0.0009113319683820009, 0.012487063184380531, -0.005669602192938328, 0.007931599393486977, -0.0041291373781859875, -0.009658722206950188, -0.013304011896252632, 0.004076155833899975, 0.012284702621400356, 0.00680072326213121...
[ -0.026787864044308662, 0.05626736953854561, 0.5392501354217529, 0.5126630067825317, -0.14723685383796692, 0.5937312245368958, 0.1315440982580185, -0.24785567820072174, -0.2813034951686859, -0.15616446733474731, 0.7097054719924927, 0.4867108166217804, 0.6578538417816162, 0.00112289923708885...
I was just prompted to update my Swype Beta to version v1.3.1.9274 which I did. During the process, I was reassured that I won't be losing any of my data. However, upon completion of the installation, I see that I've lost my personal dictionary. Another oddity: in older versions of swype, when you typed a non-English word, you were prompted to add it to your swype dictionary. Then, they changed this, such that all new words were automatically added to your dictionary. With this new v1.3.1.9274 version, however, swype seems to have gone back to the "add to dictionary" prompt. Any clue as to what may be going on? I'm using JB on my Galaxy Nexus.
[ -0.004144961480051279, -0.006467554718255997, -0.0031269537284970284, 0.013349445536732674, -0.0036295121535658836, 0.0012281425297260284, 0.005700232461094856, -0.006120139267295599, -0.014890067279338837, 0.007922879420220852, -0.008337602019309998, 0.01195528544485569, 0.00902917794883251...
[ 0.3368854820728302, 0.23091185092926025, 0.40204647183418274, -0.15295889973640442, 0.07860585302114487, 0.022081172093749046, 0.6191307902336121, 0.11535414308309555, 0.013963628560304642, -0.6128945350646973, 0.12201160192489624, 0.5918054580688477, -0.12160694599151611, 0.19838777184486...
If $\text{P}(M|D)$ is posterior, $a$ is the proportionality constant, $\text{P}(M)$ is the prior and $\text{P}(D|M)$ is the likelihood. I have the the prior distribution, and I know the function that can give the likelihood for any input parameter sample. Then the Bayes rule is: $\text{P}(M|D)= a\times\text{P}(M)\times\text{P}(D|M)$. If I want to sample from $\text{P}(M|D)$ by Gibbs, what is my full conditional distribution? is it $\text{P}(M)\times\text{P}(D|M)$?
[ -0.00014154566451907158, 0.008410828188061714, -0.011700422503054142, 0.014965019188821316, 0.02207564190030098, -0.002921138424426317, 0.007341146469116211, -0.025702709332108498, -0.006138503085821867, -0.0027477419935166836, -0.012564221397042274, 0.000435780209954828, 0.00066150771453976...
[ -0.14347289502620697, 0.040980465710163116, 0.2587099075317383, -0.6085195541381836, 0.13480982184410095, 0.7798635959625244, 0.24850767850875854, -0.5828559398651123, -0.11704885214567184, -0.4080863893032074, -0.07717166841030121, 0.6998847126960754, -0.4049926996231079, 0.32499444484710...
We are considering using BitBucket rather than hosting our Git repositories internally. Does anyone know if this breaks any rules of PCI compliance? I haven't been able to find much information on this.
[ -0.01117801759392023, 0.003202911000698805, 0.01208485383540392, 0.04315295070409775, 0.06401682645082474, -0.0022578847128897905, 0.0129337003454566, 0.003865408943966031, -0.022679386660456657, -0.00969531387090683, -0.008965139277279377, 0.022934583947062492, 0.002998051233589649, 0.018...
[ 0.8225963115692139, -0.0075239744037389755, 0.04665938764810562, 0.16217082738876343, 0.2695474326610565, -0.37696874141693115, 0.2318848967552185, 0.448644757270813, -0.20042401552200317, -0.3730463683605194, 0.13059842586517334, 0.3124762177467346, 0.021456368267536163, 0.141693010926246...
I'm trying to insert some `\noalign`-code in every row of a tabular by using `colortbl` and `\everycr` or `\CT@everycr`. In most cases it works fine, but not if there is a `\rowcolors`-command (from `xcolor`) after my command: `\rowcolors` uses this code: \CT@everycr{\@rowc@lors\the\everycr} This doesn't expand `\the\everycr` and so whatever I put in `\everycr` is always ignored. Is this a bug or did I overlook something? \documentclass{article} \usepackage{array} \usepackage[table]{xcolor} \begin{document} \makeatletter \newcommand\test{\noalign{\hrule height 0.1cm width 5cm}} \rowcolors{0}{green}{blue} %Works after/without \rowcolors: \CT@everycr\expandafter{\expandafter\test\the\CT@everycr} \begin{tabular}[t]{ll} foo & foo\\ foo & foo\\ \multicolumn{2}{l}{blub} \end{tabular} \bigskip %this fails: \everycr\expandafter{\expandafter\test\the\everycr} \rowcolors{0}{red}{green} %\CT@everycr{\@rowc@lors\the\everycr}%inserted by @rowcolors %it would work if \rowc@lors would use this: %\CT@everycr\expandafter{\expandafter\@rowc@lors\the\everycr}% \makeatother \begin{tabular}[t]{ll} foo & foo\\ foo & foo\\ \multicolumn{2}{l}{blub} \end{tabular} \end{document}
[ 0.026176925748586655, 0.013325123116374016, -0.025376617908477783, 0.01777590624988079, -0.012995371595025063, 0.0034783468581736088, 0.007100486196577549, 0.0026660445146262646, -0.010368936695158482, -0.01150873489677906, -0.012545481324195862, -0.000254887156188488, -0.0038188162725418806...
[ -0.06081885099411011, 0.027687640860676765, 0.6066294312477112, -0.1532144397497177, 0.022532355040311813, 0.3188599944114685, -0.2036556899547577, 0.21771268546581268, -0.30660656094551086, -0.6731061339378357, 0.035029418766498566, 0.3914925456047058, -0.4490770399570465, 0.0283482726663...
ls /dev command lists the device files. How to know the **associated drivers** /major_numbers/minor_numbers with those device files?
[ 0.029765238985419273, -0.0013050971319898963, -0.017341142520308495, -0.010264880023896694, -0.053355686366558075, 0.05663735792040825, 0.01662948727607727, -0.016994012519717216, -0.03649650514125824, -0.03564893826842308, -0.0005191441741771996, 0.00467214360833168, 0.03381870314478874, ...
[ 0.23702147603034973, -0.002894336823374033, 0.4435597062110901, 0.04940243437886238, 0.16646632552146912, 0.0472010113298893, -0.03703603520989418, 0.07088594883680344, 0.012369947507977486, -0.6841103434562683, -0.044107384979724884, 0.8079218864440918, -0.16254667937755585, -0.1237811818...
I understand that metals have overlapping of valence and conduction bands. But is this because there exists a partial conduction band within the top part of a metal valence band, or because the conduction band exists, but periodically in the valence region? Is this enabled by the formation of 'covalent' molecules in the metal structure matrix?
[ 0.00274486537091434, 0.017859235405921936, -0.012071254663169384, 0.015503546223044395, 0.03169039264321327, -0.014264559373259544, 0.010736430995166302, -0.003269434440881014, -0.021536733955144882, 0.018564991652965546, -0.008301270194351673, 0.02762625180184841, -0.017251869663596153, 0...
[ 0.6170420050621033, 0.4498407244682312, 0.19006983935832977, -0.03555672988295555, 0.10306527465581894, -0.24859140813350677, -0.03690335899591446, -0.6566862463951111, -0.10850856453180313, -0.37252748012542725, 0.028627479448914528, 0.38472455739974976, -0.022450827062129974, 0.298725008...
I found myself with a sentence like this, using "accept" in the infinitive form after "rather than": > They left the club, rather than accept the terms. But I'm unsure of its grammatical soundness. Conjugating the verb "accept" just sounds wrong: > They left the club, rather than accepted the terms. The gerund sounds right, but I'm not sure why: > They left the club, rather than accepting the terms. What is the grammatically correct way of phrasing this sort of sentence, and why? > [SUBJECT] [PAST_ACTION_1], rather than [PAST_ACTION_2]
[ 0.005097514949738979, 0.02406492456793785, -0.014723879285156727, 0.020888596773147583, -0.03907235711812973, -0.029398126527667046, 0.010485203005373478, -0.016367757692933083, -0.013300208374857903, 0.026050563901662827, -0.0025177872739732265, 0.011862469837069511, -0.008738587610423565, ...
[ -0.22962254285812378, 0.27504730224609375, 0.1945609599351883, -0.4211430549621582, -0.6919082403182983, 0.2620399296283722, 0.7469931244850159, -0.37389886379241943, -0.11576284468173981, -0.2267882078886032, -0.09390426427125931, 0.45065054297447205, -0.06946852803230286, -0.066495984792...
You know when using `the_excerpt()` and no "excerpt" is set on a post it automatically uses the first 50 words or so. Is the same also possible when using `get_the_excerpt()` For instance I'm using `get_the_excerpt()` like this … $return .= sprintf(' <li> <div class="title"><a href="%1$s">%2$s</a><span class="goto">a</span></div> <div class="project-description">%3$s</div> </li>', get_permalink( get_the_ID() ), get_the_title(), get_the_excerpt() ); However if no excerpt is set, the `get_the_excerpt()` function doesn't show a fallback (the first 50 or so words). Any idea how to make that work?
[ -0.0030488059855997562, 0.00510540883988142, -0.005307227373123169, 0.00719856284558773, -0.015557864680886269, 0.005365204997360706, 0.00820566900074482, 0.015191241167485714, -0.01392857450991869, -0.012916112318634987, -0.005831759423017502, 0.0015894263051450253, 0.01203373447060585, 0...
[ 0.22282318770885468, -0.36341559886932373, 0.406864732503891, 0.23821492493152618, -0.12574481964111328, -0.08326668292284012, 0.253958135843277, -0.2967032492160797, -0.2994609475135803, -0.7625332474708557, -0.404436856508255, 0.2568165361881256, -0.18842187523841858, -0.0865366458892822...
I'm trying to make a graphic of a random walk in 1D, but i want that the first and the last point have a color (in this case the first one red and the last blue), but i dont know what i'm doing wrong.. This is my code: RandomWalk[n_, d_] := NestList[(# + Table[Random[Real, {-1, 1}], {d}]) &, Table[0, {d}], n]; OneDim = RandomWalk[5000, 1]; firstpoint = OneDim[[1]] lastpoint = OneDim[[5001]] ListPlot[{{Hue[0], PointSize[.02], Point[firstpoint]}, {Hue[.7], PointSize[.02], Point[lastpoint]}, Line[OneDim]}]
[ 0.009636761620640755, 0.014008148573338985, -0.009240980260074139, 0.0023258249275386333, -0.014524398371577263, -0.00416142214089632, 0.006156880408525467, 0.019721422344446182, -0.010707167908549309, 0.014968049712479115, -0.009801387786865234, -0.0003411706711631268, 0.015346329659223557,...
[ 0.28889718651771545, -0.43555527925491333, 0.6036055684089661, 0.015282217413187027, -0.13540343940258026, 0.13595552742481232, -0.1625034511089325, -0.18163055181503296, -0.4192344546318054, -1.0205053091049194, 0.5772163271903992, 0.10413502156734467, -0.30844104290008545, 0.180778428912...
This is a simple question but a design consideration that I often run across in my day to day development work. Lets say that you have a class that represents some kinds of collection. Public Class ModifiedCustomerOrders Public Property Orders as List(Of ModifiedOrders) End Class Within this class you do all kinds of important work, such as combining many different information sources and, eventually, build the Modified Customer Orders. Now, you have different processes that consume this class, each of which needs a slightly different slice of the ModifiedCustomerOrders items. To enable this, you want to add filtering functionality. How do you go about this? Do you: 1. Add Filtering calls to the ModifiedCustomerOrders class so that you can say: MyOrdersClass.RemoveCanceledOrders() 2. Create a Static / Shared "tooling" class that allows you to call: OrdersFilters.RemoveCanceledOrders(MyOrders) 3. Create an extension method to accomplish the same feat as #2 but with less typing: MyOrders.RemoveCanceledOrders() 4. Create a "Service" method that handles the getting of Orders as appropriate to the calling function, while using one of the previous approaches "under the hood". OrdersService.GetOrdersForProcessA() 5. Others? I tend to prefer the tooling / extension method approaches as they make testing a little bit simpler. Although I dependency inject all my sourcing data into the ModifiedCustomerOrders, having it as part of the class makes it a little bit more complicated to test. Typically, I choose to use extension methods where I am doing parameterless transformations / filters. As they get more complex, I will move it into a static class instead. Thoughts on this approach? How would you approach it?
[ 0.004520290996879339, 0.008431043475866318, -0.00825855415314436, 0.016912508755922318, -0.012761618010699749, 0.01510759349912405, 0.006891719996929169, 0.0028313954826444387, -0.010491683147847652, 0.026722649112343788, -0.00022326712496578693, 0.017204418778419495, -0.006288329139351845, ...
[ 0.4301115572452545, -0.2591894567012787, 0.08340243250131607, 0.3474835157394409, -0.19949623942375183, 0.0016412681434303522, 0.01471993513405323, -0.2146359533071518, -0.3913857042789459, -0.707381010055542, 0.048818349838256836, 0.28072312474250793, -0.30738770961761475, 0.5457481145858...
When displaying a category how can I tell what category it is? Say I'm on `http://www.example.com/category/quotes`, how can I get `quotes` aside from parsing the URL? I tried `get_the_category()` but it returns an array of all the categories of the first post shown in the page.
[ -0.002859930507838726, 0.0028309389017522335, -0.010215548798441887, 0.02272423729300499, -0.01880454272031784, 0.015942638739943504, 0.008092152886092663, 0.014356826432049274, -0.025424249470233917, 0.000008035492101043928, -0.010050123557448387, 0.0025895910803228617, -0.00734181795269250...
[ 0.3804914057254791, 0.2879396080970764, 0.39924412965774536, 0.08770589530467987, -0.7276164889335632, 0.1120356097817421, 0.06644842773675919, 0.0946408361196518, -0.1901760846376419, -0.3713858723640442, 0.040947649627923965, 0.2976931631565094, -0.2550868093967438, 0.4085027575492859, ...
how to write Arabic words between English script in latex? I tried this code but it writes Arabic text in new line \documentclass[a4paper,10pt]{article} %In the preamble section include the arabtex and utf8 packages \usepackage{arabtex} \usepackage{utf8} \begin{document} %start encoding to unicode %Note that your layout must support arabic text when compiling \setcode{utf8} %To start typing in Arabic use the command arabtext hello \begin{arabtext} السَلامُ عَليكم ورَحمةُ الله وبَركاته \end{arabtext} arabic \end{document}
[ 0.0037535966839641333, 0.018998712301254272, -0.0037522432394325733, 0.018314629793167114, -0.01736612804234028, 0.011857746168971062, 0.009866689331829548, 0.014239627867937088, -0.022730093449354172, -0.02373930811882019, -0.008662049658596516, 0.0006118403398431838, -0.019310742616653442,...
[ -0.12878111004829407, 0.43885496258735657, 0.30918967723846436, -0.37188389897346497, -0.030027803033590317, 0.15261328220367432, 0.41610515117645264, -0.23833493888378143, 0.1894165426492691, -0.8071941137313843, -0.3612149953842163, 0.5825784802436829, -0.3198084533214569, -0.18172428011...
Hello I just started using wordpress (complete noob) but I need a bit of help. I have installed wordpress and on a linux/centos webserver and this works fine. However when I use another machine, open a web browser, type in the url, the site looks different from what it looks like on my server. (On my web server, there is a choice to use the Graphical user interface as well) The difference is the images don't appear on my machine (Looks like it can't find the media files). Also when I click on any link (the 'Hello world', login etc.), it says Firefox can't establish a connection to the server at localhost. I can see the images on the webserver and all links are fine. So I don't now what the problem Just to let you know my wordpress url is localhost and my site url is also localhost if that may help.
[ 0.0004062574589625001, -0.007321435492485762, -0.0045652249827980995, 0.016795340925455093, -0.027509283274412155, -0.017387930303812027, 0.0056327031925320625, 0.012742601335048676, -0.01293279230594635, -0.013341797515749931, -0.01201929897069931, 0.00818721204996109, 0.012323614209890366,...
[ -0.3617612421512604, 0.34827443957328796, 0.7780140042304993, -0.2769477665424347, -0.4959835410118103, 0.28178030252456665, 0.43894171714782715, 0.6441311836242676, -0.34717342257499695, -0.8401188254356384, 0.024206586182117462, 0.24677853286266327, -0.18653833866119385, 0.46774888038635...
This question is a follow on from this question - Correlation with non-normal distribution. An ANOVA explanation was given here - Correlation between vegetation and erosion \- that could be applicable to my question. Further to this, I was hoping someone could tell me if there was an advantage to using ANOVA over a kruskal wallis test, and if ANOVA can be used in this way to find the correlation between a categorical varaible and a continuous variable with a non-normal distribution.
[ -0.00631021149456501, 0.01254967786371708, -0.011298540979623795, 0.03322507068514824, -0.0215961504727602, -0.0026672182139009237, 0.010571877472102642, -0.007990056648850441, -0.009666306897997856, -0.012914484366774559, -0.010984224267303944, 0.020153488963842392, -0.024673376232385635, ...
[ 0.1888076364994049, -0.46285808086395264, -0.015167372301220894, -0.05973799154162407, -0.10173594951629639, 0.20922589302062988, 0.046319738030433655, -0.09766475856304169, -0.11135586351156235, -0.3836100399494171, 0.14750301837921143, 0.23623807728290558, -0.15604321658611298, 0.6080200...
I want to compose some pictures as the example picture. All small pictures need to be ranked alignedly, and the subcaption of each row should be near to the pictures. ![1](http://i.stack.imgur.com/bVQeu.jpg)
[ -0.00017692148685455322, 0.01088133454322815, -0.000681706122122705, 0.035627350211143494, 0.008027588948607445, -0.004912347998470068, 0.008730964735150337, 0.012075462378561497, -0.022807519882917404, 0.024583589285612106, -0.04176841676235199, 0.00240746489726007, -0.013090481050312519, ...
[ 0.32128965854644775, 0.22691568732261658, 0.34252113103866577, -0.1544235348701477, -0.029582006856799126, 0.1738465279340744, -0.03821692243218422, -0.26605045795440674, -0.3646811246871948, -0.46898674964904785, 0.055925577878952026, 0.04970254749059677, -0.3467186689376831, -0.010689157...
Hi I have created several projects for several different clients. Each project has OS maps, shapefile layers from Government Bodies, Shapefile layers created, AutoCAD dxf loaded in (in various manners). My question is, is there a programme out there that I can load all my layers into (whilst maintaining their properties), keep all the original files on my computer. Send a link to a client that allows them to open the project, turn on and off various layers, interogate the data (as in simple information tool), measure distance, and print maps, but not be able to edit the layers or save data. I have worked of a GIS browser with various different government agencies but never had to create one. So isthere is a free, easy to use, and easy to learn by the clients one out there? PS I would like to be able to load all the files from QGIS 1.8.0 to this service whilst maintaining their properties, however if this is not possible i can do it over again.
[ 0.00022498227190226316, 0.007925348356366158, -0.0004753995453938842, 0.014280373230576515, 0.016209624707698822, -0.010296590626239777, 0.005468249320983887, 0.01911691203713417, -0.01568155735731125, -0.020477790385484695, 0.005176304839551449, 0.021042414009571075, 0.010815059766173363, ...
[ 0.4460737705230713, -0.07664629817008972, 0.5325388312339783, 0.024464193731546402, -0.20653250813484192, 0.17886227369308472, 0.012467020191252232, -0.06468238681554794, -0.21883247792720795, -0.2980197072029114, -0.11572905629873276, -0.03011874482035637, -0.00844678282737732, 0.08548967...
I'm kinda disappointed by the gameplay of "Become a Legend" mode in PES 11. I'll explain why: I have Fifa 10 on my Xbox and I've played with the "same" mode in there but it's really different. What I'm saying is that in Fifa 10 I control only my player but I can call a pass, a cross etc...in PES I can't do any of that and sometimes the AI of PES it's not really quick in completing a passage or doing something trickier like a 1-2 (that's the term we use in Italy, basically I pass the ball to you and you pass to me the ball back). In Fifa 10 I can do this because I can call the action (and if it's not completed I lose points). Is possible to obtain something like that in PES 11?
[ -0.011378161609172821, -0.0020371752325445414, -0.009883973747491837, -0.00030154624255374074, -0.0070984079502522945, -0.0005646813660860062, 0.005795943550765514, -0.0012470135698094964, -0.013297764584422112, 0.008518010377883911, -0.02364378049969673, 0.011916358023881912, 0.008858886547...
[ 0.12823186814785004, 0.0034392797388136387, 0.3415093719959259, 0.05323087424039841, -0.3496658504009247, -0.4772058129310608, 0.6374866962432861, -0.1403902918100357, -0.2142452597618103, -0.38630911707878113, 0.21922852098941803, 0.239863783121109, -0.23490281403064728, 0.028969550505280...
I have an Xbox 360 gamepad, which I want to use as a PS3 gamepad. I want to make it compatible with the game Shin Sangoku Musou 6 Moushouden (Dynasty Warriors 7) on PC. Is there any emulator software can do this? To clarify, I am trying to find a PS3 Controller emulator. I know there is an Xbox 360 Controller emulator, I am trying to find a similar one but for PS3 controller. I am doing this because Shin Sangoku Musou 6 Moushouden (Dynasty Warriors 7) on PC supports the PS3 controller best. When I use my Xbox 360 gamepad it doesn't recognize the buttons correctly and fully, and vibration also doesn't work.
[ 0.0018634945154190063, 0.00563080282881856, -0.00010159291559830308, 0.01552704069763422, 0.0011156494729220867, -0.011911677196621895, 0.007682198658585548, -0.012502215802669525, -0.014821658842265606, -0.0328855887055397, -0.001781614264473319, 0.015556717291474342, -0.010454581119120121,...
[ 0.0009734425693750381, 0.135642871260643, 0.21047042310237885, 0.30146291851997375, -0.17111434042453766, 0.14497962594032288, -0.13471627235412598, -0.0803954228758812, -0.12669602036476135, -0.5468441247940063, 0.1880272477865219, 0.9024244546890259, -0.11788783967494965, 0.0646449699997...
Trying to follow the suggestions on this thread: splitting a lines layer using points However, how to do so in a windows environment using the GRASS command console is beyond me, so I'm attempting with the Python Console (still with no success). used v.out.ascii module to create .txt of XY coordinates (xycoord.txt) file = xycoord.txt for x, y in file: grass.run_command('v.edit', map='reshydPART_Pirae', tool='break', coords='x,y') ValueError: too many values to unpack Does anyone have any suggestions? I wonder if it has to do with the ascii.output file I'm providing (there are newlines in my txt file). I'm feeling a little lost, especially regarding what to put for the coords parameter. I also can't tell what my error is due to. Thanks in advance.
[ -0.02156458981335163, 0.0008674382697790861, -0.014115285128355026, 0.012062592431902885, -0.03298601508140564, 0.007306019775569439, 0.006945955101400614, 0.0037031397223472595, -0.01431204192340374, 0.004242412745952606, -0.004647393710911274, 0.008614998310804367, 0.003324891673400998, ...
[ 0.3993626534938812, -0.3496343195438385, 0.4259844720363617, 0.05369344353675842, -0.04753298684954643, 0.15800054371356964, 0.10205534100532532, -0.47501543164253235, -0.23147237300872803, -0.8933075666427612, 0.08017633855342865, 0.17217867076396942, -0.38669562339782715, 0.1287623047828...