text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
What do I need to do in order to change the unit size (increase the number of soldiers per unit)? I am stuck on normal size.
[ -0.01313683483749628, 0.011329231783747673, -0.030988164246082306, 0.019259849563241005, 0.024286774918437004, -0.04759444296360016, 0.011709187179803848, -0.022851452231407166, -0.03903694823384285, -0.01635456085205078, 0.012681456282734871, 0.028822317719459534, -0.023183774203062057, 0...
[ 0.3867797553539276, 0.3601239025592804, 0.07823903858661652, -0.3395070433616638, -0.24717484414577484, -0.06341786682605743, 0.6214540004730225, -0.19503915309906006, -0.6081161499023438, -0.7011280655860901, 0.048229534178972244, 0.21136437356472015, -0.10076658427715302, 0.1807864010334...
I know that Emacs is a very powerful editor out there. I try to use it occasionally and want to learn it better. But to learn it better i have to use it more frequently than I am using now. But one big obstacle is that I could not find the basic properties like a project auto completion. Maybe I am lazy but I dont want to write again and again the same long method names. Another one is go to definition stuff, I want to see the real declaration of any class or method or even a variable. Also I know that it can handle the make files but an actual representation of a project alongside the editor would be very good. So, what I am asking is, is there a way to provide these options in Emacs? Or is there an extended version of Emacs that supports these features?
[ 0.008842933923006058, 0.004464947618544102, -0.014705203473567963, -0.008176554925739765, -0.001377291977405548, 0.0010672911303117871, 0.005530523136258125, -0.002504047006368637, -0.015207931399345398, -0.0025038360618054867, -0.006751636043190956, 0.006947195157408714, 0.01388336438685655...
[ 0.1545637845993042, 0.15017040073871613, 0.22082357108592987, 0.2169135957956314, -0.02798861637711525, -0.25795313715934753, 0.2716270983219147, 0.07194660604000092, -0.10096994042396545, -0.5502278208732605, 0.3575548827648163, 0.6478617787361145, -0.09947263449430466, 0.0922636985778808...
I believe it is correct to add the abstract to my TOC but I don't know how. This Answer talking about adding the abstract chapter to the TOC. I'm using Lyx and if I output to .tex file this is what I see. \documentclass[english]{report} \usepackage[T1]{fontenc} \usepackage[latin9]{inputenc} \setcounter{secnumdepth}{3} \setcounter{tocdepth}{3} \makeatletter %%%%%%%%%%%%%%%%%%%%%%%%%%%%%% User specified LaTeX commands. \usepackage{babel} \makeatother \usepackage{babel} \begin{document} \title{My Title} \author{My Name} \maketitle \tableofcontents{} \begin{abstract} %% This isn't a chapter! My Abstract isn't a chapter. * * * Abstract is created with `begin{abstract}` unlike in questions I **LINKED** to and have been marked as a duplicate. In that example abstract is a chapter. Does this affect the correct use of getting my Abstract into the TOC? * * * > Assuming that the abstract and acknowledgement contents is typeset using a > \chapter*{...}, Well it isn't, So what do I do in my case?
[ 0.019826695322990417, 0.00332682766020298, -0.004284084774553776, 0.025165559723973274, 0.009430017322301865, 0.00025114836171269417, 0.007835857570171356, -0.005553967319428921, -0.016091372817754745, -0.0049454038962721825, -0.01880800351500511, 0.008389139547944069, -0.02169830910861492, ...
[ -0.013104486279189587, 0.20524221658706665, 0.44303151965141296, -0.3097987174987793, -0.1041269525885582, -0.0951818898320198, 0.044670313596725464, -0.04983100667595863, 0.1589737981557846, -0.6355852484703064, 0.04436356946825981, 0.6420069336891174, -0.2542547285556793, -0.302844643592...
Yesterday, when I came home, I went to the bathroom (lights off) and i took off my T-Shirt (100% cotton) which I wore under a Shirt (50% cotton, 50% polyester). I _believe_ to have seen a small but well visible light arc. **Question:** Is the amount of static discharge in T-Shirt big enough to produce a visible light arc?
[ -0.005957656539976597, -0.002804761053994298, -0.024555088952183723, 0.019451040774583817, 0.007643005345016718, -0.023831510916352272, 0.010761650279164314, 0.0037461526226252317, -0.024240106344223022, -0.030990799888968468, 0.009918561205267906, 0.019509540870785713, -0.005864656064659357...
[ 1.4020041227340698, -0.03865555673837662, -0.13988825678825378, -0.2207985818386078, -0.144527867436409, -0.19062645733356476, 0.5705820322036743, 0.09498799592256546, -0.2090269774198532, -0.45510879158973694, 0.061105482280254364, 0.28786706924438477, -0.14375357329845428, -0.17013700306...
I run a ps command inside a startup script, if I put in: ps -A -o euser,pid... I get this output: EUSER PID CMD %CPU RSS SZ ELAPSED root 644 ./code 0.0 996 2963 00:00 It's there each time. If I run: ps -C ./code -o euser.... it does not display anything at all. Any suggestions?
[ -0.004620198626071215, -0.010004510171711445, -0.008401506580412388, 0.022180309519171715, -0.020068002864718437, 0.002806371310725808, 0.00869484432041645, -0.0029879941139370203, -0.013854471035301685, 0.006753791123628616, -0.01059506181627512, -0.0016898130998015404, 0.007723037153482437...
[ 0.15028654038906097, 0.13063162565231323, 0.6326255798339844, -0.20644401013851166, 0.20008708536624908, 0.5605137348175049, 0.19255757331848145, 0.2108648270368576, -0.21947866678237915, -0.892094075679779, -0.22237518429756165, 0.21888099610805511, -0.4717366397380829, -0.153948411345481...
I am new on blogging webmaster tools. Usually, when I add new post to my blog, it will automatic update my homepage also. But from January 14th, my homepage doesn't update in the Google SERPs. As a result, I am losing my popularity in the SERPs. Previously when I posted new articles, 70-80% will go to the first page result. But after the problem occurs, none of them reach the top 15 pages of Google SERPs. :( Last 1/12/12, Google webmaster sent me a "Notice of DMCA removal from Google Search" massage indicating one of my URLs contained some infringing content which I deleted after receiving their notice. Not only that, I also cheeked all of my posts if there any additional infringing content available. After removing that, I filled out Google's content removed notification form to notify them. Google sent me a feedback that they received it and suggest "In the future, if you have removed the allegedly infringing content from your site (and won’t put it back), please use the correct form" which also I filled in. Now my question is: 1. Did I do all this right? 2. Although my new posts are indexed in Google with "..", why won't update my homepage which previously automatically updated when a new article was published?
[ -0.012146832421422005, 0.006198737770318985, -0.004243958275765181, 0.01724829338490963, 0.005151247140020132, 0.006913921795785427, 0.008112726733088493, -0.009795526042580605, -0.01740190014243126, -0.02834748663008213, -0.010880162008106709, 0.018087172880768776, -0.006543990224599838, ...
[ 0.43881890177726746, 0.21231448650360107, 0.8417066931724548, -0.021772196516394615, -0.34882721304893494, -0.5698512196540833, 0.38014960289001465, 0.10812480747699738, -0.2645311951637268, -0.6460524797439575, 0.047681089490652084, 0.34838777780532837, -0.05652653053402901, 0.60767650604...
I have difficulty understanding exercise 24 in this document: > Two parallel wires I and II that are near each other carry currents i and 3i > both in the same direction. Compare the forces that the two wires exert on > each other. > > (a) Wire I exerts a stronger force on wire II than II exerts on I. > > (b) Wire II exerts a stronger force on wire I than I exerts on II. > > (c) The wires exert equal magnitude attractive forces on each other. > > (d) The wires exert equal magnitude repulsive forces on each other. > > (e) The wires exert no forces on each other. I think - if you use $F_m=IlB\sin \alpha$, which is Force on electric wire in uniform magnetic field - that $F_{II}=IlB\sin \alpha = 3IlB\sin \alpha > IlB\sin \alpha =F_{I}$ So answer would be b), but how is it possible because you have Newton's third law( the forces should be equal, but does it apply here) and there is not any magnetic field here. So do I use this Lorentz's law or which law do I use?
[ -0.004706701263785362, 0.01287574227899313, -0.027326639741659164, 0.015248814597725868, 0.005383125506341457, -0.024415183812379837, 0.01035485789179802, -0.02326490916311741, -0.02079017646610737, 0.011492613703012466, -0.008996529504656792, 0.02023310959339142, -0.025395028293132782, 0....
[ 0.4076765179634094, -0.2934383153915405, 0.29329952597618103, -0.13641797006130219, -0.6434630751609802, -0.3912423551082611, 0.6348154544830322, -0.8381581902503967, -0.21824856102466583, -0.39974427223205566, 0.25734958052635193, 0.551908552646637, -0.15361866354942322, -0.12491136044263...
Starting from an "a priori" set of models based on my knowledge about potential relations between my dependent variable and the independent variables considered, I use AIC to find best models. More specifically I calculate Akaike weights then Evidence Ratio (ER) and consider that models with a ER < 2 are equally likely. But the same problem remain each time I do that. I selected the best models from a set of them, but I don't know if those models are efficient to predict (or at least represent) my data. I can have selected the best element(s) of the list of the worst models. I do not use $R^2$ in model selection because of the fact that including more variables generally increase $R^2$ value. But ! When the selection by AIC is done and I can consider that models with a ER < 2 are equally likely. Do you find it is correct to calculate $R^2$ or pseudo-$R^2$ for the best "set of models" in order to have an idea of the representativeness of those models and use this value to select the more efficient model? I would be glad to hear your opinions about this! Note: Thanks for suggestions about cross-validation, I will try this. Unfortunatly, I do not have an external dataset to test my models.
[ 0.027234794571995735, 0.02219056338071823, -0.019521858543157578, -0.0016642184928059578, -0.01395668275654316, 0.0044005555100739, 0.007894471287727356, -0.026644514873623848, -0.01174943521618843, -0.0034797866828739643, -0.00031632930040359497, 0.011738586239516735, -0.015470502898097038,...
[ 0.09264339506626129, 0.19214493036270142, 0.04403574392199516, 0.05018315464258194, -0.41596800088882446, 0.44985273480415344, 0.292897492647171, -0.9489789009094238, 0.30385175347328186, -0.5691201090812683, 0.13606026768684387, 0.7114941477775574, -0.02408943884074688, 0.0530614107847213...
i know little about the quantum field theory and also that the permanent magnets have there fields because of exchange of virtual photons across or around the ends. so when we take a dynamo and the constantly changing magnetic field and the emf thus produced, is it possible that it is due to the virtual photons of the magnetic field. Please correct me if I am wrong and please explain the QFT also if possible and veeeerrrryyyy simply please.
[ 0.008382032625377178, 0.008300237357616425, -0.001564071630127728, 0.009799004532396793, -0.014217806048691273, -0.030139004811644554, 0.0084829805418849, 0.0170600526034832, -0.012440282851457596, -0.007674862165004015, -0.017531128600239754, 0.020946241915225983, -0.015276973135769367, 0...
[ 0.655691385269165, -0.07848505675792694, 0.24271105229854584, 0.23288360238075256, -0.08988185971975327, -0.17238160967826843, -0.27392977476119995, -0.08336794376373291, -0.25815698504447937, -0.09939581155776978, 0.34365490078926086, 0.49195393919944763, -0.002641951199620962, 0.86964118...
Our team redesigns the legacy software and there are a lot of design meetings going on. Many of them are not very effective and lack good analysis of the proposed design. Is there any formal way to approach this kind of discussions to make them more effective? I heard about SWOT for example, but afaik it is used more for business projects.
[ 0.02055339328944683, -0.006129398010671139, -0.019012071192264557, 0.00870599877089262, -0.021896101534366608, 0.004649349953979254, 0.008899597451090813, 0.016567235812544823, -0.014176850207149982, -0.017273688688874245, -0.015635719522833824, 0.022663477808237076, 0.010163018479943275, ...
[ 0.45681536197662354, 0.17045064270496368, -0.104007288813591, 0.04492689669132233, 0.10203751921653748, -0.35087859630584717, 0.18342074751853943, -0.08294490724802017, -0.09088537842035294, -0.3363325595855713, 0.15051449835300446, 0.414876252412796, 0.35977277159690857, 0.033399365842342...
Can I omit _there_ in the following question: > How much juice is there in the bottle? When is it possible to omit _there_ in a sentence? Any references to grammar sources are welcome and expected.
[ 0.03531564027070999, 0.023263853043317795, 0.0007089751888997853, 0.04096812382340431, -0.032042257487773895, 0.027608100324869156, 0.01400066539645195, -0.009753227233886719, -0.028722597286105156, 0.0006765717407688498, -0.021408407017588615, 0.016427215188741684, 0.0096608716994524, 0.0...
[ 0.3473118841648102, 0.6113657355308533, -0.05493616685271263, -0.4321543574333191, 0.04045894369482994, -0.0921466052532196, 0.4189850986003876, 0.25282883644104004, -0.22819973528385162, -0.2055513560771942, -0.25341373682022095, 0.4358115792274475, -0.2340913563966751, 0.1883981674909591...
Is there a way to convert an embedded graphic in an ArcMap document, to a non- embedded graphic? I have a number of graphics of little trees. They were embedded into an mxd I have and I want to convert them back to simple graphics, so that I can copy and paste them from one mxd to another without the size changing because of the scale. Is there any way to do this? Thanks, Mike
[ 0.010031403042376041, 0.009260624647140503, -0.00589419761672616, 0.02712022140622139, 0.00013201753608882427, -0.01677856221795082, 0.009813345968723297, 0.006113524083048105, -0.032155126333236694, -0.045266423374414444, 0.0062197367660701275, 0.013864812441170216, -0.003114208811894059, ...
[ 0.4125136137008667, 0.36098426580429077, 0.022049615159630775, 0.7462538480758667, -0.015391314402222633, 0.3199990391731262, 0.15057966113090515, 0.3347112834453583, -0.3383074998855591, -0.6670770049095154, 0.20559649169445038, 0.6491233706474304, 0.010629597119987011, -0.259528964757919...
We run a high-traffic site for a large organization and it was compromised through an editor account. The attacker was able to poison the google cache, and the visible result to us is that if you search for us on using Google, the result (and the pages visible through google cache) show us as a drug warehouse. Our site has been completely audited and is 100% clean now, and we have requested that google reindex our site through their webmaster tools. Even though google bots hit our site almost constantly, we have yet to see the google cache page get recycled. Does anyone know anyone know a quicker way to make this happen? The drug-warehouse page is embarassing to us and our organization.
[ 0.005852401256561279, 0.0025557540357112885, 0.006737377028912306, 0.015295560471713543, 0.005302901845425367, -0.015353781171143055, 0.0066270167008042336, 0.00637080380693078, -0.016596591100096703, -0.01211404800415039, 0.0013923244550824165, 0.01671302318572998, 0.008279042318463326, 0...
[ 0.8347948789596558, 0.3295687139034271, 0.017376640811562538, 0.2337251454591751, 0.5696861743927002, -0.3591153621673584, 0.413566529750824, 0.02862667292356491, -0.07445643842220306, -0.12154103815555573, 0.27412545680999756, 0.14000067114830017, -0.23501498997211456, 0.16488425433635712...
I need to add turn restriction to my application. However, my map provider (Navteq) don't have this one for Venezuela and other countries.What other map provider could have this information?
[ 0.015560138039290905, 0.019174912944436073, -0.010404926724731922, 0.0259457528591156, -0.01208318118005991, 0.035590238869190216, 0.013022230006754398, 0.052175235003232956, -0.029126793146133423, 0.013404025696218014, -0.019839094951748848, 0.022752612829208374, -0.03659540414810181, 0.0...
[ 0.10688728094100952, 0.11734478175640106, 0.42182931303977966, 0.22741380333900452, 0.24292880296707153, -0.2696283459663391, 0.10890097916126251, 0.1981637179851532, -0.5307774543762207, -0.36534497141838074, 0.5671530365943909, 0.29392820596694946, 0.021815147250890732, 0.228864446282386...
I changed my main user's ID from 1000 to 999 and now it no longer shows on the login screen. I expected it not to show, but I also expected the login screen (the default that ships with MATE/Mint 16) to be "smart enough" to show a box where you can manually type in a username if there are no users over 1000 to list. It isn't. Is there a way to change to a different login theme that will allow a username to be entered (from the command line, because I can't log in graphically!)? Is there a way to make the login screen show a username prompt? Maybe some hidden hotkey combination? I can use CTRL+ALT+F2 and login and run `startx`, but the DE response is brutally slow, and it doesn't have the login screen chooser on the menu when MATE is started that way.
[ 0.018086910247802734, 0.0015031786169856787, -0.004905514419078827, 0.00871301256120205, -0.03478192910552025, -0.018770752474665642, 0.007003664039075375, -0.013386187143623829, -0.014015663415193558, -0.028146356344223022, -0.01341572031378746, 0.0035949992015957832, 0.00983630120754242, ...
[ 0.15303702652454376, 0.18115738034248352, 0.5629156231880188, -0.2529718577861786, 0.16007167100906372, 0.4112391173839569, 0.35121068358421326, 0.04473258554935455, -0.4687378406524658, -0.4063252806663513, 0.07986675947904587, 0.17027507722377777, -0.11904400587081909, 0.2595238983631134...
I'm trying to use wp_insert_post to create a front-end submission form so visitors can create a post AND update the meta box fields of that post -- **not the custom fields**. For example, when I use the following in the submission form... <?php update_post_meta($post_id, $meta_key, $meta_value); ?> ...it updates the built-in custom field section of the newly created post, but not the meta boxes that I created. Any ideas?
[ 0.002564790192991495, -0.0015716500347480178, 0.003727308241650462, 0.027098234742879868, -0.005871563218533993, 0.022820577025413513, 0.009233017452061176, -0.007694542407989502, -0.018506614491343498, -0.0027702341321855783, -0.008852032013237476, 0.008328295312821865, -0.01293413247913122...
[ 0.6703166961669922, 0.15517427027225494, 0.08951414376497269, 0.07035351544618607, 0.17777501046657562, 0.1669871211051941, -0.09948160499334335, 0.06674626469612122, -0.024194039404392242, -0.6827446222305298, 0.29616227746009827, 0.20200125873088837, -0.34252065420150757, 0.3156512975692...
I've been using the new theme customizer and it really has me thinking how awesome it would be to allow potential theme buyers/clients etc to play with the customizer for a demo theme (without being logged in). The customizer is located at : `..wp-admin/customize.php`. So what it would take would be to somehow allow open access to this URL **but** nothing else in the back end. I suppose there is probably a security issue with this but I thought I'd get a discussion going on it here in case it's possible to do this safely as I'm sure other people might have the same idea. So the question is : is there any safe way to allow open access to a single page of the admin?
[ 0.001491856062784791, 0.0007495132740586996, 0.004703686106950045, 0.00346945459023118, 0.0006272378377616405, 0.007645749486982822, 0.005671009887009859, -0.00524075236171484, -0.013713154941797256, 0.0007283999584615231, -0.00923252385109663, 0.0079087745398283, 0.0018994510173797607, 0....
[ 0.5204342603683472, -0.06454174965620041, 0.7126103043556213, 0.2883354127407074, 0.1464284062385559, -0.4221387505531311, 0.07902916520833969, 0.09680088609457016, -0.515942394733429, -0.3207261860370636, 0.09451513737440109, 0.3164638876914978, 0.1778111457824707, 0.5175233483314514, 0...
I've read some posts here and other places about how a lot of people don't put much value in certifications but I am beginning to think it may be necessary for me at this point to be able to move to a bigger company, etc. I currently work as a Java programmer with a startup and worked with a small company before that. Now that I'm applying with larger companies the hr people / recruiters have been asking a lot about certifications and some have directly suggested that someone in my position should probably get a few (they were trying to be helpful) since I haven't completed a BS degree yet (I bounced around a bit in college and ended up not finishing but have enough units to finish eventually, just its not something I can do nearly as easily as getting certifications). Anyways, just curious about what people think for someone in my situation where I do have an interested in working for large companies and do not currently have a BS degree (but do have experience already in the field). Any advice on which certifications beyond the SCJP would be appreciated as well
[ 0.007130577694624662, 0.006754291243851185, -0.005125964991748333, 0.004311572294682264, 0.011389939114451408, 0.044751062989234924, 0.003708351869136095, 0.0026730294339358807, -0.017235953360795975, -0.016679342836141586, -0.0011092849308624864, 0.018046580255031586, 0.005758761428296566, ...
[ 1.3009085655212402, 0.4832472503185272, -0.1091248095035553, -0.09607763588428497, 0.23841917514801025, -0.30597710609436035, 0.35546571016311646, 0.1969628781080246, -0.5362550616264343, -0.385039359331131, 0.3624873459339142, 0.45857498049736023, 0.4633685350418091, 0.22029076516628265, ...
> **Possible Duplicate:** > How do I get Google to show links to my site hierarchy in search results? On Stack Overflow and many other websites, I've noticed there is some extra information under the result link. For example: ![Google SO result](http://i.stack.imgur.com/1IH9E.png) The _"6 answers - Jul 31, 2011"_ part is what I'm talking about. Does Google automatically do this or does the website have to do something to make Google add this extra information in search results?
[ -0.00014036893844604492, -0.012379473075270653, -0.00370496092364192, 0.014093579724431038, 0.003608283819630742, -0.013262243941426277, 0.003825616557151079, 0.0072195641696453094, -0.018443528562784195, 0.004896578378975391, -0.0014692756813019514, 0.007259857840836048, 0.00634884322062134...
[ 0.1633801907300949, 0.16311021149158478, 0.4658423066139221, 0.20200221240520477, -0.11671894043684006, -0.170784130692482, 0.34426426887512207, -0.19565920531749725, -0.618107795715332, -0.3978707194328308, 0.17854461073875427, 0.14990563690662384, -0.13029736280441284, 0.4666545987129211...
I dont know much about statistics so bear with me. Lets say I have a set of 1000 workers. I want to figure out who the hardest worker is but I can only measure the amount of work getting done in groups of 1-100 over an hours worth of work. Assuming each worker always does around the same amount of work, over a large number of trials and combinations can I rank my workers by who works hardest? note: this is just a metaphor so dont worry about actually running the tests, just assume I already have a large set of data. **Edit:** when I say "Assuming each worker always does around the same amount of work" I mean each individual does around the same amount of work on a day to day basis. So joey will do around 100 units of work each day and greg will do around 50. The problem is I can only observe the number of units of work done by the group. **More edits:** In regards to the number of workers working at once and the frequency of them working. There could be any number of workers working at the same time. Some workers will probably end up working a lot more than others, that is to say, we can assume some workers will be working almost 90% of the time and others almost never. I know that makes it difficult but I will have a very large dataset so hopefully that makes it a little bit easier. **For each hour we know which workers are working and how much work got done. From that information I want to find out who is doing the most work.** if the data were in JSON format it would look something like this... [ { "work_done": 12345, "Workers": [ "andy", "bob", "cameron", "david" ] }, { "work_done": 432, "Workers": [ "steve", "joe", "andy"] }, { "work_done": 59042, "Workers": [ "bob", "aaron", "michelle", "scott", "henry" ] }, ... ]
[ 0.020498700439929962, 0.019111089408397675, -0.015639521181583405, 0.00467194989323616, -0.02166013792157173, 0.011737095192074776, 0.005616978742182255, -0.018158596009016037, -0.014402929693460464, 0.004055732861161232, -0.00322919012978673, 0.01207278948277235, 0.0070515344850718975, 0....
[ 0.6024823784828186, 0.16642774641513824, -0.07832859456539154, 0.28066056966781616, -0.3000510334968567, 0.1822049617767334, 0.2508848309516907, -0.264813095331192, -0.0989595577120781, -0.76992267370224, 0.21936073899269104, 0.42934706807136536, -0.06798103451728821, -0.5105741024017334, ...
In the WP editor there is a full screen button. It has shortcut `ALT+SHIFT+G`. When the button is pressed it shows a different full screen editor then when the shortcut is used. I tried this on multiple 3.4.1 installations, one of them had no plugins to interfere. Why is this? **EDIT** its a bug: http://core.trac.wordpress.org/ticket/21197
[ -0.01453614141792059, -0.00416702963411808, -0.002477326663210988, 0.01601717248558998, -0.05636206641793251, -0.004973472561687231, 0.007597093936055899, 0.0036498946137726307, -0.012764401733875275, 0.0053897094912827015, -0.01093547884374857, 0.01082457136362791, 0.012340771034359932, 0...
[ 0.04827186092734337, -0.4663397967815399, 0.4997114837169647, -0.05748620629310608, -0.10481549799442291, -0.053589172661304474, 0.2155677229166031, -0.0941755548119545, -0.4098995327949524, -0.8014762997627258, -0.26831161975860596, 0.4459030032157898, -0.7420886754989624, -0.023960318416...
I tried this command but I think something is wrong with `'{}' '{}'`. By the latter, I want `InputXYZ.pdf OutputXYZ.pdf` (I do not want any dupe files, I want to replace the old files to make reading a bit more convenient). $ find . -name "*.pdf" -ok qpdf --password=passMe --decrypt '{}' '{}' \; If you have better and simpler ideas to handle this kind of repetitive pass- typing probs with other programs, do not hesitate to comment or answer. Actually, my teachers use many different passwords and it is quite mess to read the documents. Perhaps, some file with their passwords and then just the `find` -cmd testing every possibility?
[ 0.000608608708716929, 0.013784381560981274, -0.003726362017914653, 0.009369945153594017, -0.023976434022188187, -0.0010715456446632743, 0.0064749508164823055, 0.003654713276773691, -0.021076083183288574, -0.020879115909337997, -0.0013774664839729667, 0.006453918293118477, 0.00295295333489775...
[ 0.0662677213549614, 0.10515642911195755, 0.433663547039032, -0.11331863701343536, 0.012866650708019733, -0.1678200513124466, 0.600304901599884, -0.04949776455760002, 0.1170123964548111, -0.571128249168396, -0.3856918215751648, 0.5297704339027405, -0.4890510141849518, 0.13351012766361237, ...
Thank's to SE (and particularly these two questions: Draw a bivariate normal distribution in TikZ and How to fix a contour plot at top of a 3D box), I've been able to draw what I was looking for with pgfplots except for one detail. In this image: ![this image](http://i.imgur.com/wpxGGrY.png) I would like the contour lines drawn on the surface to be hidden by the surface in the foreground but not in the background (like it would be if you were really looking at such a "valley"). Is it possible to do so ? Here is the code (I downgraded the sampling in order to make it run quicker and don't forget the `-shell-escape` option to make it run with LaTeX). \documentclass{standalone} \usepackage{pgfplots} \usepackage{amsmath} \begin{document} \pgfplotsset{ colormap={whitered}{color(0cm)=(white); color(1cm)=(orange!75!red)} } \begin{tikzpicture} \begin{axis}[ colormap name=whitered, 3d box, width=15cm, view={25}{25}, enlargelimits=false, grid=major, domain=-0.5:4.7, y domain=-2:2, samples=21, xlabel=$x$, ylabel=$\dot{x}$, zlabel={$\text{E}_{\text{m}}$}, colorbar, colorbar style={ at={(1,0)}, anchor=south west, height=0.1*\pgfkeysvalueof{/pgfplots/parent axis height}, title={$\text{E}_{\text{m}}(x,\dot{x})$} } ] \addplot3 [surf] {-0.7+4*exp(-0.5*(x+3))*(3*cos(4*x*180/pi)+2.5*cos(2*x*180/pi)) + 0.5*y*y*4}; \addplot3 [contour gnuplot={number=14,labels={false},draw color=black}, samples=21, ] {-0.7+4*exp(-0.5*(x+3))*(3*cos(4*x*180/pi)+2.5*cos(2*x*180/pi)) + 0.5*y*y*4}; \addplot3 [domain=-0.5:4.7,samples=31, samples y=0, thick, smooth] (x,-2,{-0.6+4*exp(-0.5*(x+3))*(3*cos(4*x*180/pi)+2.5*cos(2*x*180/pi))}); \addplot3 [contour gnuplot={number=14,labels={false},draw color=black}, samples=21, z filter/.code={\def\pgfmathresult{20}}, ] {-0.7+4*exp(-0.5*(x+3))*(3*cos(4*x*180/pi)+2.5*cos(2*x*180/pi)) + 0.5*y*y*4}; \end{axis} \end{tikzpicture} \end{document}
[ -0.0070797959342598915, 0.009817766025662422, -0.009280981495976448, 0.03123289905488491, -0.019038915634155273, -0.030026204884052277, 0.007862193509936333, 0.017193859443068504, -0.01988157257437706, -0.00035568419843912125, -0.015460438095033169, 0.005865762941539288, -0.02533805929124355...
[ 0.13364051282405853, 0.04064276069402695, 0.43854162096977234, 0.2626970708370209, -0.37815794348716736, -0.15237292647361755, -0.052729859948158264, 0.2543271780014038, 0.011386513710021973, -0.3448999524116516, 0.1323024481534958, 0.33460238575935364, 0.18832586705684662, -0.005344430916...
My idea is to write a superset of C# (but question is not language-specific), so that it source-to-source compiles (transcompiles) to C# itself (fall- through switch clauses, default method parameters etc., nothing impossible in C#). First idea was to parse it, make syntax trees, abstract trees etc. but it seems as a bit of an overkill to me, mostly because large portions of code will remain the same. **My question:** Is there a simpler way to do this? One of my ideas was to search for tokens that need modifying (e.g. `switch` in case of fall-through) and then rewrite the code (add `goto case NEXT_CASE` where needed) but is there a better and cleaner way to do this?
[ 0.0035981307737529278, 0.010570220649242401, 0.002319183200597763, 0.019725456833839417, 0.002171942964196205, -0.010951546020805836, 0.0061624799855053425, -0.013913126662373543, -0.01553764846175909, -0.0012285544071346521, 0.005952195264399052, 0.005969802848994732, 0.007807346060872078, ...
[ 0.3182504177093506, 0.20921118557453156, -0.06518999487161636, 0.05980236455798149, -0.06433715671300888, -0.01939280703663826, 0.30770009756088257, -0.035747427493333817, -0.19741959869861603, -0.45983564853668213, 0.061670057475566864, 0.5578463673591614, -0.5187453031539917, -0.25577503...
I'm currently using the latest version of Gingerbread 2.3.7 and Google Plus 1.0.8 on a Nexus S. I'm communicating a lot with the Google Plus Messenger to another user who's using Google Plus on the iPod Touch. We send a lot of messages but the notification bar doesn't show any new messages. I have enabled notification within Google Plus Messenger but still nothing. I have to check Google Plus Messenger on a regular basis to view any new messages received. Is the above a known issue? Is it fixable at the moment or should I wait for the next Google Plus update?
[ -0.018832694739103317, -0.016255896538496017, -0.012002360075712204, 0.024747751653194427, 0.021478962153196335, -0.007837390527129173, 0.0074773915112018585, 0.01498308964073658, -0.01869082823395729, 0.0011851564049720764, -0.007313210517168045, 0.01564713567495346, 0.00005474471254274249,...
[ 0.4517360031604767, 0.13347125053405762, 0.49451327323913574, -0.12173240631818771, -0.10505412518978119, 0.1421731859445572, 0.5783116221427917, -0.07877428829669952, -0.3236270844936371, -0.98093581199646, 0.1906440258026123, 0.5800895094871521, -0.2529480457305908, 0.02614060789346695, ...
I need to create a web mapping application that can show raster layer of historical maps with a time slider. Basically, something like this: http://goo.gl/GYHOX I have different maps of different years of the same are, that can be showed on a time series and moving on with a time slider. The interface can be simple, the important thing is the time slider. I have programming experience only in HTML/CSS, little of PHP. Any suggestions of which tools to use? I had a look at TileMill but the problem is the time-slider that I have no clue on how to build it. Thanks in advance for who will help me.
[ -0.0061454251408576965, 0.008758888579905033, -0.008073958568274975, 0.0010677978862076998, 0.007798550650477409, 0.011070461943745613, 0.006400600541383028, 0.008500101044774055, -0.021875137463212013, -0.007752102334052324, 0.00003179837949573994, 0.008081316947937012, 0.007369229104369879...
[ 0.48548194766044617, -0.013516909442842007, 0.3137493431568146, 0.3323725461959839, -0.03878526762127876, 0.025739163160324097, -0.15509773790836334, 0.34687861800193787, -0.1702650934457779, -0.9314063191413879, 0.4992970824241638, 0.39975371956825256, -0.0884597972035408, 0.4689019918441...
If I have 3 variables, $X$, $Y$ and $Z$, what statistical approach might I use to ask the question "is in population X correlated with $Y$ but is uncorrelated with $Z$"? Obviously, I can assess if $X$ is correlated with $Y$ and separately if $X$ is correlated with $Z$ and have two p-values: p1 <- cor.test(X,Y)$p.value p2 <- cor.test(X,Z)$p.value But I'm not sure how I'd go about sort of "combining" these two results **into one p-value** of my combined hypothesis. Or maybe there is a totally different approach for this question?
[ -0.0003644130192697048, 0.021886412054300308, -0.005157472565770149, 0.009842961095273495, 0.007301962003111839, -0.0037396082188934088, 0.00733791571110487, -0.02637236937880516, -0.010660439729690552, -0.02831324003636837, -0.008651604875922203, 0.0050305891782045364, -0.017159437760710716...
[ 0.3949047029018402, 0.004606401547789574, 0.09469027072191238, 0.07488711178302765, -0.3744504749774933, 0.4518307149410248, 0.14460670948028564, -0.4301013648509979, 0.14036577939987183, -0.2728235125541687, 0.3649466633796692, 0.23437364399433136, -0.33531317114830017, 0.1870547235012054...
I want to produce a boxplot like this in `R`: fishdata <- read.csv("http://dl.dropbox.com/s/4w0utkqdhqribl4/fishdata.csv", header=T) attach(fishdata) boxplot(fishdata[,-1], las=3) which looks like this: ![what it should look like](http://i.stack.imgur.com/JFy0P.png) but when I Sweave this, using this coding: <<echo=F>>= fishdata <- read.csv("http://dl.dropbox.com/s/4w0utkqdhqribl4/fishdata.csv", header=T) attach(fishdata) @ \begin{figure}[htbp] \centering <<fig=T, echo=F>>= boxplot(fishdata[,-1], las=3) @ \rule{35em}{0.5pt} \caption{Boxplot of the fish catches in Bangladesh from 1950 to 2011.} \label{histogram} \end{figure} produces this still, ![boxplot output](http://i.stack.imgur.com/nJWeC.png) Is there a way to get the x-axis titles as above in the coding?
[ -0.00042561045847833157, 0.013190308585762978, -0.004331077449023724, 0.01857006549835205, 0.016459178179502487, -0.017351210117340088, 0.005085034761577845, 0.0039307051338255405, -0.0183967724442482, 0.0008990829810500145, -0.005694535095244646, 0.0016403378685936332, 0.0001989107113331556...
[ 0.25612279772758484, -0.12536390125751495, 0.6274714469909668, -0.2937092185020447, 0.19540558755397797, -0.05796248838305473, -0.07100101560354233, -0.30190208554267883, -0.416551411151886, -0.5641565918922424, -0.07573241740465164, 0.305944561958313, -0.5257906913757324, -0.0243482831865...
There's an achievement in LA Noire for stopping a fleeing suspect with a warning shot. How do I do the warning shot? I tried firing "around" the suspect while he's running, but that didn't work.
[ -0.026431594043970108, 0.02294541709125042, -0.014132768847048283, -0.0036825162824243307, -0.039632830768823624, 0.01521431002765894, 0.014275612309575081, 0.015431567095220089, -0.034943196922540665, 0.028790496289730072, -0.010212747380137444, 0.035313114523887634, 0.0042011430487036705, ...
[ 0.31664544343948364, 0.28828516602516174, -0.36797812581062317, 0.022606005892157555, -0.11222375184297562, -0.23040811717510223, 0.7593808770179749, -0.369529664516449, -0.22255797684192657, -0.10764668881893158, 0.2095404714345932, 0.6358974575996399, 0.02338828518986702, -0.334587514400...
I'm trying to analyse data for my psych undergrad dissertation but struggling with which tests to use because my project is unlike anything I've had experience with previously (and also I don't have a mind for statistics). I have 200 cases of patients who've attended therapy, and the dates of every session they attended. I made extra variables for the number of days wait between each session but I want to know whether different diagnoses (e.g. depression, anxiety...) had an effect on how often patients came for therapy. The trouble is that the diagnostic groups have very different n in each (there are 5 main diagnostic groups, ranging from n = 109 to n = 3!). I'm not sure if there's anyway to do this? Any advice much appreciated. Thanks!
[ 0.015840578824281693, 0.013795528560876846, -0.007736966945230961, 0.02077415958046913, 0.005366849713027477, -0.001016648719087243, 0.006411474198102951, 0.010949475690722466, -0.007803848013281822, -0.016300082206726074, -0.0031664930284023285, 0.001835467992350459, 0.0055974796414375305, ...
[ 0.5481339693069458, 0.33271247148513794, 0.23615391552448273, -0.2302899807691574, 0.17012768983840942, 0.18159693479537964, 0.5479305982589722, -0.11843478679656982, -0.22164663672447205, -0.4675586223602295, 0.5184735655784607, 0.3012424111366272, 0.32721272110939026, 0.36629071831703186...
I have this very strange issue which I discovered on my month-old Note 2. I never had this issue with my previous device which was a Nexus S. I hooked up the phone via USB to my PC, and selected `camera PTP` mode. When I browsed the folders via `Computer\GT-N7100\Phone\DCIM\Camera`, I don't see any videos that I shot. However, when I browse the files through the default `my files` app on the phone, I see them. Why is this? Windows 8 detected the device and installed all drivers too. Because of this, I am not able to move the videos shot with the phone to my PC. Any help?
[ -0.01992511749267578, -0.008665254339575768, 0.001142339315265417, 0.005322781857103109, -0.0140347545966506, -0.014972742646932602, 0.004587041679769754, 0.039456695318222046, -0.012333367019891739, -0.003957832232117653, -0.023630216717720032, 0.01610615849494934, 0.005356628447771072, 0...
[ 0.2213010936975479, 0.10824023932218552, 0.6518498063087463, 0.051088057458400726, 0.15362505614757538, -0.26366594433784485, 0.5174133777618408, -0.0011600445723161101, -0.5492888689041138, -0.35857895016670227, 0.21162132918834686, 0.6252925992012024, -0.6882063150405884, 0.0576950982213...
I'm building a predictive model for a medical condition that happens in both men and women. Physicians have reported that in some cases for women, their menstrual history seems to be a risk factor, though epidemiologists haven't been able to verify this claim. My data includes information about female subjects' menstrual history. My goal is to build a logistic regression model that determines the probability of the existence of this condition. Here's my problem: I'd like to build a model that includes gender and includes menstrual history questions for female subjects, but how do I assign values for male subjects for things like "number of months since last pregnancy"? Is there a general method to deal with this kind of problem?
[ 0.014964229427278042, 0.022986020892858505, -0.00961947999894619, 0.0032450205180794, 0.021600499749183655, 0.0015063080936670303, 0.008183826692402363, -0.0003595855087041855, -0.014294354245066643, -0.005960340145975351, 0.0054798913188278675, 0.008259795606136322, -0.013124852441251278, ...
[ 0.27291783690452576, 0.3051459789276123, -0.3355804681777954, 0.08737794309854507, 0.13881352543830872, 0.15431153774261475, 0.21916519105434418, -0.024754377081990242, -0.038159579038619995, -0.3245665729045868, 0.6215869784355164, 0.2955875098705292, -0.1000225841999054, 0.62661159038543...
I hear a lot of people saying "Send me an invite". I always thought that it was an 'invitation'. Is "sending one an invite" accepted usage? Or is it incorrect? If I need to get my wedding invitation printed, should it read "Wedding Invitation" or a "Wedding Invite"? Is this a US/UK usage difference?
[ -0.02886541560292244, 0.0019822383765131235, -0.012825929559767246, 0.014035732485353947, -0.03287091106176376, 0.016348252072930336, 0.013157851994037628, 0.03144286945462227, -0.016869915649294853, -0.024839965626597404, -0.005629480816423893, 0.008936909027397633, 0.005813012830913067, ...
[ 0.43136999011039734, 0.13634517788887024, 0.521003246307373, -0.29696714878082275, -0.14023397862911224, -0.24153326451778412, 0.27154237031936646, 0.02430640161037445, -0.3970094323158264, -0.10065699368715286, -0.04840892553329468, 0.1231667697429657, -0.2166755199432373, -0.038172367960...
I'm doing an OLS regression of donations made by individuals to a not-for- profit organisation. The donation amount is the dependent variable and dummy (treatment) variables are the only independent variables. The non-transformed treatment variable is significant (5% level), but when I take a log of the donation amount (not treatment), the treatment variables become highly insignificant. Does this show an outlier problem and should not be used for interpretation of the data?
[ -0.006515027955174446, 0.009309218265116215, -0.011321799829602242, 0.019009806215763092, -0.02658720687031746, -0.02336173504590988, 0.014667716808617115, 0.013320007361471653, -0.012712132185697556, -0.054718613624572754, -0.017219578847289085, 0.004854321945458651, -0.014645918272435665, ...
[ 0.6074625253677368, -0.16480782628059387, 0.27618011832237244, 0.10949303209781647, -0.08800318837165833, 0.3443531394004822, 0.28905388712882996, -0.2500852048397064, -0.02764747105538845, -0.2263621985912323, 0.5031471848487854, 0.14712078869342804, -0.15485569834709167, 0.11488719284534...
A developer on our team litters his code with debug outputs. He passes many parameters into methods that are only used for debugging. Personally I find this clutters up the code. If I need to debug something I prefer to use breakpoints and the watch window instead of reading through all the debug output. I am making major changes to several classes written by this other developer, should I maintain his debug outputs or can I justify removing them? I have only been in this team for a few months and don't want to upset this developer but also I expect he will rarely be updating this part of the system.
[ -0.003938112407922745, 0.02707621082663536, -0.002921781037002802, -0.0035164414439350367, -0.014431033283472061, 0.007931205444037914, 0.005934876389801502, 0.011909489519894123, -0.012304729782044888, -0.034601788967847824, -0.005889712367206812, 0.023642292246222496, -0.00990491732954979,...
[ 0.5557775497436523, 0.392129510641098, -0.331849068403244, 0.05236940458416939, -0.08663030713796616, -0.35543009638786316, 0.5141775012016296, 0.20352894067764282, -0.15218521654605865, -0.43237391114234924, 0.19672203063964844, 0.3077424466609955, -0.301588237285614, 0.5105196237564087, ...
I want to calculate a transparency parameter for a superposition of several volumes containing media using the Beer-Lambert Law, which states that $$I/I_0=exp(-\tau)$$ where $I$ is the transmitted intensity, $I_0$ is the incident radiation intensity, and $\tau$ is the optical depth of the medium. My question here is: Can I replace $\tau$ with $\tau_1 + \tau_2 + ... + \tau_n$ when dealing with n media of same/different optical depths? Does it matter whether the volumes overlap (assuming that they do not interact with each other)?
[ 0.012024017050862312, 0.00730586051940918, -0.0009055619593709707, 0.0028069864492863417, -0.009057548828423023, -0.00693393312394619, 0.007012789137661457, -0.009618650190532207, -0.007652432657778263, -0.016599006950855255, -0.005241365171968937, 0.007180529646575451, -0.005941485054790974...
[ 0.36287885904312134, 0.26593300700187683, 0.591820478439331, -0.11774958670139313, -0.10410130023956299, 0.14228835701942444, 0.07747123390436172, -0.3615514039993286, -0.034506719559431076, -0.5307441353797913, 0.00768819497898221, 0.6656213998794556, -0.23800960183143616, 0.5510918498039...
I am working with a Road Network. I have digitized the lines with AutoCAD Map 3D. Then Imported it to ArcMap. After entering all kind of database I have created Routing Dataset. But unfortunately its not giving the desired route. The most probably problem may be the junction. I think there are few problems with junction between the road. Its a huge database and it cant be check manually every junction. Is there any way to create a network topology to check the junctions easily whether the roads snapped or not? Thanks
[ 0.004508474841713905, 0.006005831528455019, -0.007857882417738438, 0.019363440573215485, -0.018009895458817482, 0.001073729363270104, 0.007730415090918541, 0.020179808139801025, -0.017595436424016953, 0.000592433731071651, -0.009883660823106766, 0.025016263127326965, -0.011538560502231121, ...
[ 0.4329991936683655, 0.3462294340133667, 0.37069758772850037, 0.31543251872062683, -0.14444541931152344, -0.12635701894760132, 0.2051004022359848, 0.21439126133918762, -0.17199809849262238, -0.9942031502723694, 0.3141157627105713, 0.5523782968521118, 0.01054356899112463, 0.1741502285003662,...
Is there a way to configure SoundEx properties in ArcEngine? This query: "100 Fairvew Rd returns "100 Fairview Rd" But this query: "100 Farview Rd" returns no results Is there a way to make the SoundEx a bit less rigid? Here are some concrete examples I would like to get to work if possible (ie. these are my specs) 100 cen => central ave 100 w laur ave -> 100 w laurel ave I'm using the IAddressCandidates2 interface to get my candidates. IArray resultsArray = addressCandidates.FindAddressCandidates(addressProperties);
[ 0.014163307845592499, 0.006740933749824762, -0.01637471839785576, 0.008411010727286339, -0.01815822720527649, 0.0047765676863491535, 0.009530525654554367, 0.014781028032302856, -0.01897261291742325, -0.026194248348474503, 0.003373035928234458, 0.014758417382836342, -0.008346645161509514, 0...
[ 0.42883631587028503, 0.13907256722450256, 0.6505443453788757, 0.09266971796751022, -0.07287844270467758, 0.20943008363246918, 0.16982504725456238, -0.5183537602424622, 0.5017798542976379, -0.7302663326263428, 0.33728116750717163, 0.9442116022109985, 0.12161524593830109, -0.0525191053748130...
My http trafic analyze product which is running on lippcap and wriiten with C++ has some troubles with videos which are opened from facebook. I can classify websites and my program outputs log files which contains `IP/PORT/URL/REFERERWEBSITE/WEBSITE` data. I am parsing "REFERERWEBSITE" information from request header's , "Referer:" field . But when I open youtube from facebook there is not any "Referer:" field. I am checking the video session I can't see anything about "facebook" in URL nor in any other field. Now I am in a situation I must get this information somehow . Any advice will be extremely appreciated.
[ -0.020157689228653908, -0.006526760291308165, 0.0022732592187821865, 0.021973397582769394, -0.009277468547224998, 0.021251073107123375, 0.009799470193684101, 0.02199232392013073, -0.011898867785930634, 0.018383340910077095, -0.0016704609151929617, 0.013557849451899529, 0.02042839303612709, ...
[ 0.10214045643806458, 0.22292496263980865, 0.6290674805641174, -0.026715440675616264, -0.42980003356933594, -0.06276693195104599, 0.46649879217147827, -0.38078856468200684, 0.03985230624675751, -0.7716565728187561, 0.13460224866867065, 0.7814220786094666, -0.24659109115600586, 0.06022627279...
In LDA topic model algorithm, I saw this assumption. But I don't know why chose Dirichlet distribution? I don't know if we can use Uniform distribution over Multinational as a pair?
[ -0.0045121172443032265, 0.04954062029719353, -0.0018661912763491273, 0.033373408019542694, 0.009872188791632652, -0.006532056722790003, 0.019577061757445335, 0.0028676025103777647, -0.030933421105146408, -0.01712547428905964, -0.014430600218474865, 0.03009135089814663, 0.005702439695596695, ...
[ 0.02003548853099346, -0.47795748710632324, -0.1393418312072754, 0.020744066685438156, -0.5399141311645508, -0.18526653945446014, -0.1293286234140396, 0.025218414142727852, -0.22863498330116272, -0.4915587902069092, 0.05862656235694885, 0.4621679186820984, -0.4575563669204712, 0.31703481078...
I have a friend who has just show me his medical prescription for hyperopia (farsightedness) correction and he needs glasses with 4,25 diopters for that, which seemed to be weird for me because I had learned, from the mirror equation, that the maximum correction possible for hyperopia is 4 diopters: $$ \frac{1}{f} = \frac{1}{p} + \frac{1}{p'} $$ If we have $0.25m$ for the normal eye distant point and more than $0.25m$ for the farsighted eye distant point (negative sign, because it's a virtual image), then we would have: $$ \frac{1}{f} = \frac{1}{0.25} + \frac{1}{p'} = 4 - \frac{1}{|p'|} \in\quad ]0,4[, \quad\text{since}\quad |p'| \geq 0.25m \quad\text{and}\quad p'<0 $$ I did some google search and find out that, indeed, hyperopia can reach values even greater, such as 20 diopters, but I can't find pages where doctors explain that with equations or physics teachers explain how things really work in ophthalmology. Either I am doing some terrible mistake, or doctors are doing some terrible mistake, or this equation just don't apply to hyperopia at all... Which one is true?
[ -0.004469471983611584, -0.0029509353917092085, -0.02575460448861122, 0.013234313577413559, -0.0346500538289547, -0.029882926493883133, 0.00967932865023613, -0.0011443649418652058, -0.011697400361299515, -0.020053666085004807, -0.0037163307424634695, 0.003950648009777069, -0.01257316768169403...
[ 0.25760000944137573, 0.43681567907333374, 0.9257588982582092, 0.07750791311264038, 0.13504792749881744, 0.30150991678237915, 0.30040082335472107, 0.01727154292166233, -0.26232531666755676, -0.39242494106292725, -0.07774977385997772, 0.9072957038879395, -0.06909241527318954, 0.2318961918354...
Suppose I have a n*n raster, and I want to create k square blocks (k can be divided exactly by n*n ) for zonal analysis: for example, when n = 4 and k = 4 a 4*4 raster is create with value 1 1 2 2 1 1 2 2 3 3 4 4 3 3 4 4 How can I do this in R?
[ 0.014310619793832302, 0.015074755996465683, -0.01079809945076704, 0.01846177689731121, -0.022552037611603737, 0.015316685661673546, 0.007971742190420628, -0.012828098610043526, -0.016531364992260933, -0.005753496661782265, -0.005725969094783068, -0.0014018758665770292, -0.0061005037277936935...
[ -0.056436583399772644, 0.12956058979034424, -0.022883063182234764, -0.34879070520401, -0.003230659756809473, 0.28494173288345337, 0.24507248401641846, -0.6241097450256348, -0.14815139770507812, -0.6749352812767029, -0.2680203914642334, 0.3410401940345764, -0.39652469754219055, 0.0889341980...
I'm using this code to get category slug on categories archive page outside of the loop: <?php $cat = get_term_by('name', single_cat_title('',false), 'category'); echo $cat->slug; ?> And I would like to know how can I add the result of that code inside 'base' => '/HERE/' of paginate_links php code that looks like this: <?php echo paginate_links( array( 'current' => $current_paginate_page, 'show_all' => true,'prev_next' => false, 'total' => $wp_query->max_num_pages, 'base' => '/HERE/', 'format' => '%#%' )); ?> I just don't know how to correctly write it inside another php code. Please help, thank you! http://wordpress.org/support/topic/single_cat_slug?replies=5 http://codex.wordpress.org/Function_Reference/paginate_links
[ -0.012370122596621513, 0.01367481704801321, 0.0007046664832159877, 0.016550978645682335, 0.007483024150133133, 0.0009730948368087411, 0.007026177365332842, -0.0043373024091124535, -0.015529319643974304, 0.004101999569684267, -0.016160596162080765, 0.00028350914362818, -0.022537579759955406, ...
[ 0.324084609746933, 0.05714680999517441, 0.48536837100982666, -0.23728875815868378, -0.20542968809604645, -0.13967029750347137, 0.18618465960025787, -0.3973281979560852, -0.38573893904685974, -0.28009068965911865, 0.3325009346008301, 0.1632717400789261, -0.3514566421508789, 0.52702492475509...
When the 3DS XL comes out, will we lose all the 3 stars rating, gold wheel, and VR score if we play MK 7 on the new console? I heard in the past that the SD card won't be able to save those data, because it was worried that game players can just duplicate many SD cards that way. But is there some other way, for example, if 3DS can transfer data through WiFi (the StreetPass mechanism), then maybe there can be a way to either copy or move the game achievements to one and only one 3DS console? (or if each game player can log in, and retrieve those achievements on their console... and players usually won't share login with each other).
[ 0.0009295977652072906, 0.012673539109528065, -0.0010942851658910513, 0.014774583280086517, -0.005002859979867935, -0.0069757732562720776, 0.007930257357656956, 0.002652740105986595, -0.01762884110212326, -0.01572439633309841, -0.016946740448474884, 0.021536685526371002, 0.007675636559724808,...
[ 0.3569064438343048, -0.0018087384523823857, 0.8775584697723389, 0.32595664262771606, -0.15467646718025208, -0.17539440095424652, -0.09672495722770691, 0.11249256879091263, 0.0015174646396189928, -0.24348051846027374, 0.05996132642030716, 0.9519839882850647, -0.14312238991260529, 0.10720177...
Does radiated energy of black body radiation due to the transition of electrons to higher energy state? or due to the increasing vibration of atoms themselves?
[ 0.04729357734322548, 0.012538819573819637, 0.009641033597290516, 0.02041012793779373, -0.01811819337308407, -0.07394229620695114, 0.018411003053188324, -0.0113606546074152, -0.033486928790807724, -0.10574520379304886, -0.0335487425327301, 0.05332822725176811, -0.011887168511748314, 0.02155...
[ 0.887618899345398, 0.2193330079317093, -0.29575538635253906, -0.07063771039247513, 0.05995215103030205, -0.3221285045146942, 0.14806614816188812, -0.33249834179878235, -0.22175316512584686, -0.13864272832870483, 0.03386777639389038, 0.5389611721038818, -0.4704495668411255, 0.37432095408439...
What is the name of this equation: $$\frac {d^2\psi}{dr^2}+k^2\psi=0?$$ (I want a Wikipedia link for this equation, but I don't know what its name is.) * * * Point: In this equation, the wave function is only function of space $\psi (r)$, for example $\psi=Ae^{\mathrm i kr}$.
[ -0.01697998121380806, 0.002698199823498726, 0.0090946638956666, 0.011165841482579708, -0.007603928912431002, -0.02930336631834507, 0.007466188631951809, -0.00499091437086463, -0.01945701241493225, 0.010328729636967182, -0.006896377541124821, 0.011528442613780499, -0.01451914757490158, 0.01...
[ -0.14072741568088531, 0.2975860834121704, 0.9062504768371582, -0.29243773221969604, -0.049541521817445755, 0.2566900849342346, -0.26314857602119446, -0.3222007751464844, -0.1950879544019699, -0.42517316341400146, 0.07192431390285492, 0.31468281149864197, -0.33466973900794983, 0.38022914528...
## situation I have a modx site on a VPS with multiple domain and subdomain names. The modx site should use what I call the 'primary' domain name's 'primary' subdomain, ie _`www`.`intendedname`.tld_ . The problem is that as time pass, the site mysteriously starts using another subdomain for links to content like videos, images, and even pages and (internal) links. The other subdomains doesn't serve this content of course. If I clear the modx cache, the original state is restored. However, the problem comes back again later. The VPS has a domain registered and multiple A records pointing to the VPS's IP, as subdomains. There is the 'primary' whan which is intended to be used as the public content server, the other ones are like `docs.` and `test.`, etc. On top of that, I have dynamic-dns service _client_ installed from no-ip on the machine and a dynamic domain-name bound. It gives a completely different domain name. I originally used it for ssh login and to serve a completely different site. An `nginx` server is put into good use to do rewrite the different subdomains to the right places. **edit** The modx templates use Templates use `<base href="[[++site_url]]" />`. ## current attempt to fix The current 'solution' to the problem is to also use the rewrite to rewrite _everything_ to the 'primary' domain and subdomain. In the nginx config file for the site, it utilizes (unsurprisingly) the `rewrite` directive to rewrite the unexpected `server_name` entries (ie. the other subdomains) in a `server` block dedicated to this task. So with this, the main site basically works (sort of) but this renders all the other functions (docs) useless. Before this rewrite was set, the 'solution' was to clear the modx cache on a regular basis. The original modx content is not getting corrupted, only the files in cache are. **What can I do to find out what actual the problem is and fix it?**
[ 0.0043510133400559425, 0.0010430868715047836, 0.0028971415013074875, 0.014383192174136639, -0.011737514287233353, 0.00779961422085762, 0.010055034421384335, 0.019102083519101143, -0.014611467719078064, 0.013369384221732616, -0.022605050355196, 0.004297556821256876, -0.00866363663226366, 0....
[ -0.10977693647146225, 0.15647540986537933, 0.7918758392333984, 0.13894443213939667, 0.05263793095946312, -0.06341366469860077, 0.4247395396232605, 0.10343275964260101, -0.4460297226905823, -0.6619042754173279, -0.4320995807647705, 0.586330235004425, -0.3369876742362976, 0.32878148555755615...
With 100 HP and a level 5 sword, how can I beat the hell level in hard mode? I'm using the Sword of Summoning; would it be better to use the Sword of Fire or the Sword of Life instead?
[ -0.01602381095290184, 0.024689553305506706, -0.014436674304306507, -0.005727414041757584, 0.005737421568483114, 0.009249280206859112, 0.012482433579862118, -0.011247792281210423, -0.02028683014214039, -0.057066045701503754, -0.018277429044246674, 0.017700687050819397, -0.01549238059669733, ...
[ 0.017547478899359703, 0.30432236194610596, 0.3060392141342163, 0.3529258966445923, -0.34650418162345886, 0.20648124814033508, 0.42736947536468506, -0.4337155222892761, 0.07343726605176926, -0.3579566776752472, 0.33321908116340637, 0.8301474452018738, 0.5696796178817749, -0.1915551573038101...
I've performed a random search in hypothesis space $$\{(c,h)| c \in U[1,256]; h\in U[1,100];c \in \mathrm{Z} \text{ and } h \in \mathrm{Z}\}$$ that defines the parameters of a standard multilayer perceptron (MLP) neural network. In each step of the random search, I draw two parameters $c$ and $h$. $c$ defines the number of input features and $h$ defines the number of hidden layer nodes. $c$ and $h$ are integers drawn from a uniform distribution defined above. I train a neural network defined by $(c,h)$ and calculate a misclassification rate and average squared error rate for each model. This is done with $10$-fold cross-validation to estimate the true error for each $(c,h)$. I therefore have an average misclassification rate and an average square error rate over the train sets and the left-out sets for each parameter pair. The question is, how do I chose the best pair of $(c,h)$ and is the method I use here sufficient? There is no reasonably clear point in the results as I'd have hoped. The results over the hypothesis space in the training data is: ![Training sample estimated errors](http://i.stack.imgur.com/ITpEf.png) The results over the hypothesis space in the hold-out data is, ![Hold out sample estimated errors](http://i.stack.imgur.com/kXlYE.png) This question relates to work I've done as part of my masters dissertation, and is related to the question here this
[ -0.003677903674542904, 0.008428352884948254, 0.001498428639024496, 0.007591949310153723, -0.00668658223003149, -0.019012823700904846, 0.005748363211750984, -0.014854716137051582, -0.0076513709500432014, -0.005260204430669546, -0.0059127481654286385, 0.0015242802910506725, 0.01350187882781028...
[ 0.35862505435943604, -0.13776563107967377, 0.3780573606491089, 0.07310662418603897, 0.2819674611091614, 0.19983266294002533, 0.03334517031908035, -0.23238998651504517, 0.009758012369275093, -0.5554614067077637, -0.03599162399768829, 0.34305545687675476, 0.1490553915500641, 0.51149761676788...
> **Possible Duplicate:** > Temporarily increase line spacing I tried to follow these examples here Temporarily increase line spacing but in my case it didn't change anything. I have to use a latex template, which is provided by my university. My title looks currently like this: ![enter image description here](http://i.stack.imgur.com/32Tpk.jpg) as you can see, the third line has not enough spacing. Can anyone help me to fix this issue? You find the entire template (already with my title in it) for downloading on my webspace: http://www.easteregg.ch/DDIS-diploma-thesis-template-en-long- title.zip just run the "ddis-thesis-EN.tex" in the root folder and you will get probably the exact same issue.
[ 0.011188489384949207, 0.006159312091767788, -0.0063622817397117615, 0.026951387524604797, 0.01609121635556221, -0.005083391908556223, 0.005507313646376133, -0.013215652666985989, -0.019736602902412415, -0.016625383868813515, -0.014011614955961704, 0.005883827805519104, -0.011072918772697449,...
[ 0.3312092423439026, 0.11140405386686325, 0.47422948479652405, 0.2269662469625473, -0.17976464331150055, -0.001784206135198474, 0.44170424342155457, 0.16806207597255707, -0.34708619117736816, -0.8317154049873352, -0.029373619705438614, 0.2962811589241028, 0.013133012689650059, 0.13556829094...
I am trying to perform a merged query to get _at most_ 6 posts from either the `post` post type and my `events` custom post type. The `post` query is based off of whether or not the Featured Image is set, and the `events` query is based off of whether or not the post has the meta value _Featured_ selected. Everything works except: **The query only returns 5 posts, and they are not balanced in the sense that [depending on which post I set as Featured/Add Featured Image too] they take precedence over one another.** For Example: If I add a Featured Image to a `post` post the results will be _3 event, 2 post_. If I then check the Featured meta value in an `events` post type the results will be _4 event, 1 post_. Here is the code I am using to query and return posts: <?php $post_types = get_post_types('','names'); $featuredEvents = get_posts(array( 'post_type' => 'events', 'post_status' => 'publish', 'showposts' => 6, 'meta_key' => '_ecmb_featured_event', 'meta_value' => 'on' )); $featuredNews = get_posts(array( 'post_type' => 'post', 'post_status' => 'publish', 'showposts' => 6, 'meta_key' => '_thumbnail_id' )); $mergedPosts = array_merge( $featuredEvents, $featuredNews ); $postIDs = array(); foreach( $mergedPosts as $item ) { $postIDs[] = $item->ID; } $uniquePosts = array_unique($postIDs); $posts = get_posts(array( 'post__in' => $uniquePosts, 'post_type' => $post_types, 'post_status' => 'publish', )); foreach( $posts as $post ) : setup_postdata($post); if( get_post_type() == 'events' ) { ?> This is a Featured Event <?php } elseif (get_post_type() == 'post') { ?> This a post with the Featured Image set <?php } endforeach; wp_reset_query(); ?> Can someone help me with this? **Edit:** Thanks to the answer below, I figured it out and used the following code: $sliderQueryArgs = array( 'posts_per_page' => 6, 'post_type' => array( 'post', 'events' ), 'orderby' => 'post_date', 'meta_query' => array( 'relation' => 'OR', array( 'key' => '_ecmb_featured_event', 'value' => 'on'), array( 'key' => '_thumbnail_id', 'compare' => 'EXISTS') ) ); $sliderPosts = get_posts( $sliderQueryArgs );
[ 0.004030013456940651, 0.010287858545780182, -0.001416089478880167, 0.015632405877113342, 0.030083253979682922, -0.011726953089237213, 0.009545503184199333, 0.01123465783894062, -0.011887448839843273, 0.000023762229830026627, -0.009788711555302143, 0.018117813393473625, -0.012037184089422226,...
[ 0.39034634828567505, -0.007522324565798044, 0.32888248562812805, -0.1705905944108963, -0.27309441566467285, 0.18467271327972412, 0.2206432819366455, -0.30087846517562866, -0.2092418074607849, -0.7798778414726257, 0.1567745953798294, 0.23917630314826965, -0.45067542791366577, 0.160783991217...
Prestige badges were added to Spiral Knights today, given out at various prestige milestones. The first badge is awarded at 5,000 prestige. However, I'm already over 5,000 prestige and mousing over any mention of prestige says my next badge is at 5,000... How can I tell if I've unlocked this badge?
[ 0.015037675388157368, 0.007308165077120066, -0.012402961030602455, 0.023509718477725983, -0.038824860006570816, -0.019850581884384155, 0.011061659082770348, -0.010911252349615097, -0.02049505151808262, -0.0015706110280007124, -0.0035188854672014713, 0.0012627835385501385, -0.0157589465379714...
[ -0.3830147087574005, 0.0793108195066452, 0.8705084323883057, 0.15773867070674896, -0.20327730476856232, -0.19228483736515045, 0.3704107403755188, -0.10522990673780441, -0.6326867938041687, -0.3652079105377197, -0.08890257030725479, 0.5015363097190857, 0.34224897623062134, 0.345373004674911...
There is a "meanings" tag on English.SE, but I would like the opposite way around — I have a meaning already, I want the word (if it exists) for doing just this.
[ 0.00430893013253808, 0.027497675269842148, -0.004382025450468063, 0.04522499814629555, -0.01645587384700775, 0.044060785323381424, 0.009510543197393417, 0.021799704059958458, -0.02464887872338295, 0.025315243750810623, 0.015528588555753231, 0.010343431495130062, 0.006754823494702578, -0.00...
[ 0.9603771567344666, 0.15754766762256622, 0.17849533259868622, 0.07634447515010834, -0.24846303462982178, -0.34608569741249084, 0.345514178276062, 0.12866725027561188, -0.10401272028684616, -0.2456933557987213, -0.05830084905028343, 0.28372037410736084, 0.11316007375717163, -0.0457438752055...
I use Fedora 20 on lxde. `lxrandr` allows me to only choose resolution. When I run xrandr -o right it makes my screen black with just cursor visible. `/root/xorg.conf.new` contains: Section "ServerLayout" Identifier "X.org Configured" Screen 0 "Screen0" 0 0 Screen 1 "Screen1" RightOf "Screen0" Screen 2 "Screen2" RightOf "Screen1" Screen 3 "Screen3" RightOf "Screen2" InputDevice "Mouse0" "CorePointer" InputDevice "Keyboard0" "CoreKeyboard" EndSection Section "Files" ModulePath "/usr/lib/xorg/modules" FontPath "catalogue:/etc/X11/fontpath.d" FontPath "built-ins" EndSection Section "Module" Load "glx" EndSection Section "InputDevice" Identifier "Keyboard0" Driver "kbd" EndSection Section "InputDevice" Identifier "Mouse0" Driver "mouse" Option "Protocol" "auto" Option "Device" "/dev/input/mice" Option "ZAxisMapping" "4 5 6 7" EndSection Section "Monitor" Identifier "Monitor0" VendorName "Monitor Vendor" ModelName "Monitor Model" EndSection Section "Monitor" Identifier "Monitor1" VendorName "Monitor Vendor" ModelName "Monitor Model" EndSection Section "Monitor" Identifier "Monitor2" VendorName "Monitor Vendor" ModelName "Monitor Model" EndSection Section "Monitor" Identifier "Monitor3" VendorName "Monitor Vendor" ModelName "Monitor Model" EndSection Section "Device" ### Available Driver options are:- ### Values: <i>: integer, <f>: float, <bool>: "True"/"False", ### <string>: "String", <freq>: "<f> Hz/kHz/MHz", ### <percent>: "<f>%" ### [arg]: arg optional #Option "NoAccel" # [<bool>] #Option "AccelMethod" # <str> #Option "Backlight" # <str> #Option "DRI" # <str> #Option "ColorKey" # <i> #Option "VideoKey" # <i> #Option "Tiling" # [<bool>] #Option "LinearFramebuffer" # [<bool>] #Option "SwapbuffersWait" # [<bool>] #Option "TripleBuffer" # [<bool>] #Option "XvPreferOverlay" # [<bool>] #Option "HotPlug" # [<bool>] #Option "ReprobeOutputs" # [<bool>] #Option "XvMC" # [<bool>] #Option "ZaphodHeads" # <str> #Option "TearFree" # [<bool>] #Option "PerCrtcPixmaps" # [<bool>] #Option "FallbackDebug" # [<bool>] #Option "DebugFlushBatches" # [<bool>] #Option "DebugFlushCaches" # [<bool>] #Option "DebugWait" # [<bool>] #Option "BufferCache" # [<bool>] Identifier "Card0" Driver "intel" BusID "PCI:0:2:0" EndSection Section "Device" ### Available Driver options are:- ### Values: <i>: integer, <f>: float, <bool>: "True"/"False", ### <string>: "String", <freq>: "<f> Hz/kHz/MHz", ### <percent>: "<f>%" ### [arg]: arg optional #Option "SWcursor" # [<bool>] #Option "kmsdev" # <str> #Option "ShadowFB" # [<bool>] Identifier "Card1" Driver "modesetting" BusID "PCI:0:2:0" EndSection Section "Device" ### Available Driver options are:- ### Values: <i>: integer, <f>: float, <bool>: "True"/"False", ### <string>: "String", <freq>: "<f> Hz/kHz/MHz", ### <percent>: "<f>%" ### [arg]: arg optional #Option "ShadowFB" # [<bool>] #Option "Rotate" # <str> #Option "fbdev" # <str> #Option "debug" # [<bool>] Identifier "Card2" Driver "fbdev" BusID "PCI:0:2:0" EndSection Section "Device" ### Available Driver options are:- ### Values: <i>: integer, <f>: float, <bool>: "True"/"False", ### <string>: "String", <freq>: "<f> Hz/kHz/MHz", ### <percent>: "<f>%" ### [arg]: arg optional #Option "ShadowFB" # [<bool>] #Option "DefaultRefresh" # [<bool>] #Option "ModeSetClearScreen" # [<bool>] Identifier "Card3" Driver "vesa" BusID "PCI:0:2:0" EndSection Section "Screen" Identifier "Screen0" Device "Card0" Monitor "Monitor0" SubSection "Display" Viewport 0 0 Depth 1 EndSubSection SubSection "Display" Viewport 0 0 Depth 4 EndSubSection SubSection "Display" Viewport 0 0 Depth 8 EndSubSection SubSection "Display" Viewport 0 0 Depth 15 EndSubSection SubSection "Display" Viewport 0 0 Depth 16 EndSubSection SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection Section "Screen" Identifier "Screen1" Device "Card1" Monitor "Monitor1" SubSection "Display" Viewport 0 0 Depth 1 EndSubSection SubSection "Display" Viewport 0 0 Depth 4 EndSubSection SubSection "Display" Viewport 0 0 Depth 8 EndSubSection SubSection "Display" Viewport 0 0 Depth 15 EndSubSection SubSection "Display" Viewport 0 0 Depth 16 EndSubSection SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection Section "Screen" Identifier "Screen2" Device "Card2" Monitor "Monitor2" SubSection "Display" Viewport 0 0 Depth 1 EndSubSection SubSection "Display" Viewport 0 0 Depth 4 EndSubSection SubSection "Display" Viewport 0 0 Depth 8 EndSubSection SubSection "Display" Viewport 0 0 Depth 15 EndSubSection SubSection "Display" Viewport 0 0 Depth 16 EndSubSection SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection Section "Screen" Identifier "Screen3" Device "Card3" Monitor "Monitor3" SubSection "Display" Viewport 0 0 Depth 1 EndSubSection SubSection "Display" Viewport 0 0 Depth 4 EndSubSection SubSection "Display" Viewport 0 0 Depth 8 EndSubSection SubSection "Display" Viewport 0 0 Depth 15 EndSubSection SubSection "Display" Viewport 0 0 Depth 16 EndSubSection SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection
[ 0.006734007969498634, -0.000929297530092299, -0.009601820260286331, -0.004306667018681765, -0.03865532949566841, -0.020521610975265503, 0.007922433316707611, 0.004935848526656628, -0.011113995686173439, -0.00017437292262911797, -0.014033547602593899, 0.008166294544935226, -0.0093912798911333...
[ 0.3018081784248352, 0.011627131141722202, 0.9169628620147705, -0.3817972242832184, 0.11932605504989624, 0.08412186801433563, -0.009696812368929386, -0.020994875580072403, -0.130747452378273, -0.8445308208465576, 0.22066077589988708, 0.5172367095947266, -0.44959136843681335, 0.0424851067364...
I'm using `beamer` to create frames. I don't want frame numbers and I want to allow for frame breaks: \begin{frame}[allowframebreaks]{My title} does work, and \begin{frame}[noframenumbering]{My title} How do I combine both? \begin{frame}[noframenumbering allowframebreaks]{My title} nor \begin{frame}[noframenumbering][allowframebreaks]{My title} work. \begin{frame}[noframenumbering, allowframebreaks]{My title} only enables the first option.
[ 0.0014204778708517551, 0.02011772245168686, 0.006172383204102516, 0.0038887138944119215, 0.0024989177472889423, 0.008503277786076069, 0.00636429525911808, 0.003110136603936553, -0.011609699577093124, -0.005379224196076393, -0.004521115217357874, 0.0053308503702282906, -0.002032465999945998, ...
[ 0.5924956798553467, 0.33234497904777527, 0.5796091556549072, -0.2870403528213501, -0.09051040560007095, -0.09471889585256577, 0.41022589802742004, -0.7342125773429871, 0.07581330835819244, -0.725132405757904, 0.02354210987687111, 0.4104641377925873, -0.38507047295570374, 0.1660985499620437...
Is there a way to create an environment for reactions defined with mhchem that behaves similarly to equation in terms of numbering the reactions (including a \nonumber command)? Can I get a list of those reactions? \documentclass{article} \usepackage[version=3]{mhchem} \newcounter{rxnum} \newenvironment{reaction}{ \refstepcounter{rxnum} \noindent %??? }{ %??? \par\noindent } \begin{document} \begin{reaction} \ce{X + Y -> XY} \end{reaction} \begin{reaction} \ce{XX + YY -> X2Y2} \\%<- newline just like in equation \ce{A + B -> C} \end{reaction} \end{document} This is a two part question. The first part is about creating an environment that behaves similarly to equation, but for reactions defined in mhchem. I had a look in source2e in the definition of equation (around l. 293 in source2e) but couldn't quite figure out how to transfer that onto my new environment. The second part is about how to create a list of those reactions containing the reaction number, the page they occur on and optional a description that is only printed in the \listofreactions. For example: In the code I write \begin{reaction} \ce{X + Y ->XY}\descriptionforlist{XY formation reaction}\\ \ce{A +B -> C} \ce{G + H -> J}\nonumber \end{reaction} Then in the list it should appear as **List of reaction** R 1: XY formation reaction 6 R 2: 6
[ 0.018607497215270996, 0.024833595380187035, -0.002167177153751254, 0.024256473407149315, -0.015363728627562523, -0.02420433796942234, 0.008277224376797676, -0.006909566931426525, -0.019729990512132645, -0.029846813529729843, 0.0023861615918576717, 0.002370232716202736, 0.0009433678351342678,...
[ 0.010438378900289536, -0.004743247292935848, 0.27002134919166565, -0.00488697923719883, 0.3584073781967163, 0.01941446028649807, -0.03869917616248131, -0.6175009608268738, -0.2087712585926056, -0.27905958890914917, 0.012734949588775635, 0.43464866280555725, -0.19435454905033112, 0.09475495...
I am attempting to publish an Adobe Captivate project to a WordPress blog. The .swf file is easy enough to handle, but the HTML5 version is giving me problems. The way that Captivate outputs HTML5 is as a folder that includes an index.html file and other folders with various resources that are references in the index.html file (e.g., as ./images/picture.jpg). I have several goals, in decreasing order of importance: 1) Be able to insert the HTML5 content into a WordPress post (including uploading the content). 2) Have the content "window" automatically fill the area WordPress has available, as if it were a simple video. (I ask this question because I tried using iFrames and they have a fixed size, unless I'm missing an option). 3) Be able to restrict access so that someone can only access the file through the WordPress post (rather than, say, directly following the URL for the index.html file). Is this possible?
[ -0.023399632424116135, -0.0059332940727472305, 0.002092928159981966, 0.02634035423398018, 0.031719036400318146, -0.005997293628752232, 0.010388472117483616, 0.028293438255786896, -0.01794050633907318, -0.0065930429846048355, -0.015348388813436031, 0.007494665682315826, -0.010094331577420235,...
[ 0.3875519335269928, 0.23640207946300507, 0.4224538803100586, -0.07483553886413574, -0.4002220630645752, -0.3005411624908447, 0.24527277052402496, -0.025167174637317657, 0.04803238436579704, -0.724236249923706, -0.07771611958742142, 0.28195470571517944, -0.2872141897678375, 0.16219629347324...
I am trying to use `\textsc` for name and section titles in moderncv template. However, the font seems weird. For example, looking at the following two examples, one for normal section title, one with `\textsc`: \section{Education} \section{\textsc{Education}} Looks like `\textsc` also changes the default font. How can I maintain the default font at the same time have the ability to use `\textsc`? Similar problems happens in `\name` too. Any help is much appreciated. ![enter image description here](http://i.stack.imgur.com/x8J4x.png) **Updated:** A MWE can be seen as follows: \documentclass[11pt,a4paper,sans]{moderncv} \moderncvstyle{classic} \moderncvcolor{blue} \name{aaa}{bbb} \begin{document} \makecvtitle \section{{Education}} \section{\textsc{Education}} \end{document}
[ 0.011360068805515766, 0.007504514884203672, -0.012088393792510033, 0.024716584011912346, -0.016506552696228027, -0.009572974406182766, 0.008921287953853607, 0.010753433220088482, -0.013222850859165192, -0.00975031964480877, -0.016004960983991623, 0.005546315107494593, 0.012392925098538399, ...
[ 0.4428131878376007, 0.2681547999382019, 0.4383412003517151, 0.12491526454687119, 0.06651969999074936, -0.2877748906612396, 0.13653099536895752, 0.06177648529410362, -0.029704062268137932, -0.7681350111961365, -0.07281620055437088, 0.4247921407222748, -0.11657605320215225, 0.330870568752288...
I installed Fifa 13 and I got the `E001` error. Luckily, I managed to solve the problem using the Swiftshader 3D rendering program. Now while playing FIFA 13 there is a large icon of Swiftshader on the bottom left corner, and it disturbs me while playing. Is there any way to remove it?
[ 0.017697276547551155, -0.0077940900810062885, -0.009136582724750042, 0.023663662374019623, -0.008818257600069046, -0.01139035914093256, 0.010085346177220345, -0.0014390145661309361, -0.018488913774490356, -0.014217911288142204, -0.010495600290596485, 0.01663065329194069, -0.01865381374955177...
[ -0.07907474040985107, 0.28000545501708984, 0.0637027770280838, 0.1298224925994873, -0.17223107814788818, -0.41888830065727234, 0.34763142466545105, 0.33232438564300537, -0.279672771692276, -0.5394633412361145, 0.33134543895721436, 0.4388238787651062, -0.6566482782363892, -0.297837257385253...
I have set up a basic PostGIS / Tilestache / gunicorn / nginx / Polymaps test environment, rendering my dataset from PostGIS to GeoJSON and shipping it out to be rendered in Polymaps. That works nicely, but for one small detail: Polymaps seems not to be requesting all tiles. If my dataset encompasses a world map, at zoom level 0, it requests 0/0/0.geojson, but not 0/1/0.geojson (the Eastern hemisphere). If I zoom further into the visible dataset, the right halves of it seem to disappear. With a more local dataset (Norway, display centered around lat/lon 60/10) nothing at all is displayed. One correct tile to load and display would for example be 2/4/0.geojson, which spans the whole dataset. Polymaps never loads any tile with an index above 3. If I load the tile directly, ignoring the tiling scheme, it is loaded and displayed correctly. Any clues to what I am missing? Polymaps code below: var po = org.polymaps; var map = po.map() .container(document.getElementById("map").appendChild(po.svg("svg"))) .center({lat: 60, lon: 10}) .zoomRange([0, 20]) .zoom(10) .add(po.interact()) .add(po.hash()); loading the tiled layer: map.add(po.geoJson() .url("/boundary/{Z}/{X}/{Y}.geojson") .id("boundary") .on("load", loadFeature)) loading the one tile directly (works): map.add(po.geoJson() .url("/test.geojson") .id("boundary2") .on("load", loadFeature))
[ -0.01154923252761364, 0.013131538406014442, 0.00202507758513093, 0.023499149829149246, 0.015197225846350193, 0.00032238534186035395, 0.008691288530826569, 0.016866186633706093, -0.015276039950549603, -0.014450770802795887, 0.0060398029163479805, 0.011665388941764832, -0.019092798233032227, ...
[ 0.004952990449965, -0.13135074079036713, 0.6797301769256592, -0.0031959889456629753, -0.146560400724411, 0.34712281823158264, 0.18026334047317505, -0.11055506765842438, -0.1550457775592804, -0.9225090146064758, -0.15098406374454498, 0.39772358536720276, -0.1160392090678215, -0.154140710830...
I was just wondering if there is some sweet and apt word for the phrase " _any time after this moment_ ".
[ 0.005371099803596735, 0.04102011397480965, -0.02636544406414032, 0.036054302006959915, 0.05759337544441223, 0.023320596665143967, 0.01454397477209568, -0.008938482962548733, -0.033224426209926605, 0.0070486548356711864, -0.003418464446440339, 0.014513698406517506, 0.05667271465063095, -0.0...
[ 0.3556729555130005, 0.332734078168869, 0.2443036139011383, -0.09764380753040314, 0.21496449410915375, -0.061744123697280884, -0.05926711484789848, 0.7498286366462708, -0.043612442910671234, -0.33282244205474854, 0.0366034172475338, -0.09217167645692825, 0.30515217781066895, 0.5591418147087...
I am a bit confused about the rule of setting the definite article in a sentence when it is associated to two nouns. 1. > * The beginning and end of the channel. > * The beginning and **the** end of the channel. 2. > * The arguments belong to the request and to data function. > * The arguments belong to the request and to **the** data function.
[ -0.016592655330896378, 0.0160242710262537, -0.0047560944221913815, 0.009433189406991005, -0.029652396216988564, -0.01899106428027153, 0.01190165989100933, -0.00792758073657751, -0.019331714138388634, -0.0029867433477193117, -0.022934790700674057, 0.007391803897917271, -0.015724051743745804, ...
[ 0.14473600685596466, 0.028373869135975838, 0.4788740873336792, -0.10352879762649536, -0.6235716938972473, 0.18595273792743683, 0.1144927442073822, -0.1352434903383255, 0.20907555520534515, -0.5473770499229431, -0.16701295971870422, 0.4453967213630676, -0.20079344511032104, 0.11255811154842...
I am trying to model data on the number of online sales are made within a fixed sale period of 3 days. Data are generated only when the sale is made. I think for this kind of data I will be using a type of time series model for count data with inflated zero samples. But this dataset adds one more layer of complexity of the time limit. I don’t know any time series statistical model that limits the length of time. I may consider the stochastic process as a (Poisson like) draw of the limited number of trials: if it’s hourly then 72, if it’s by minutes then 4320 trials. By then, if it’s by minutes, the data ends up with too many zero observations, which ruins the effectiveness of the model. Do you have any better idea? * * * The data is structured like this: 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-02 19:15:39 45 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-02 22:32:06 46 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-02 23:39:05 47 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-02 23:20:09 48 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 01:32:09 49 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 03:33:11 50 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 02:47:07 51 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 04:05:05 52 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 04:01:08 53 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 06:53:09 55 2011-10-31 07:01:55 2011-11-03 06:59:59 2011-11-03 07:00:54 56 The first and second times (1st and 2nd column) are the start and end times of the deal, which is about 3-day apart as the deal lasts for 3 days, and of course fixed for each deal. The third time (in the 3rd column) is increasing as the deal goes by with the cummulative number of items sold on the last columne as a _positive integer_ in the 4th columne. The data is generated only when the sale is made, so the time between succesive sales is _irregular_ , which could be a key term to search for the appropirate model. After further search for the appropriate model, I found integer autoregressive model (INAR) may be used, but I am not sure... Since the sale is recorded by the second, I may think the data series with T=259200 observations having many (too many!) zero observations. I may transform the data by minutes by summing over the sales numbers withine one minute to reduce to T=4320. This will reduce the ratio of zero-total observations smaller, and apply INAR or some kind of zero- inflated stochastic model (truncated poisson or something like that). Could you give me any good ideas?
[ 0.007194012403488159, 0.012158667668700218, -0.015011684969067574, 0.007869279012084007, 0.02151806466281414, 0.0031514111906290054, 0.009681694209575653, -0.002820833818987012, -0.010342900641262531, -0.0009067677892744541, -0.014703046530485153, 0.007653134874999523, 0.0019545310642570257,...
[ 0.6082017421722412, -0.06589635461568832, 0.4719350039958954, 0.15531209111213684, -0.16248078644275665, 0.17475546896457672, 0.16670215129852295, -0.18437881767749786, -0.4651015102863312, -0.46421539783477783, 0.3625791370868683, 0.5008525848388672, 0.12624113261699677, 0.230600640177726...
What's some alternatives of saying "You heard me" while snapping back at someone (in informal and formal situations both) ? Can I say "You heard it"? Also in a formal setting when dealing with jerks how to say it properly so the expression has some meaning and weight.
[ 0.028484521433711052, 0.0018004816956818104, -0.02129373513162136, 0.017207061871886253, 0.0014647557400166988, 0.00849117897450924, 0.009997784160077572, -0.02234802208840847, -0.02882244810461998, -0.023535968735814095, -0.0003161939384881407, 0.01503984909504652, 0.011189674958586693, 0...
[ 0.39661967754364014, -0.17564772069454193, -0.030491134151816368, -0.14210332930088043, -0.7951579689979553, -0.24880197644233704, 0.8038123846054077, 0.2939053475856781, -0.5584924817085266, 0.2945641577243805, -0.06863589584827423, 0.6083785891532898, 0.307661235332489, -0.03722929954528...
I have several figures and subfigures in a LaTeX file, but only the first few are numbered correctly. Here is the part of the code that is giving me problems: \documentclass[11pt]{article} \usepackage{graphicx} \usepackage{subcaption} \usepackage{amsmath} \begin{document} \begin{figure}[t] \centering \includegraphics[width=5in]{figure1.png} \caption{Caption1} \label{fig:fig1} \end{figure} \begin{figure}[htp!] \caption{caption2} \label{fig:fig2} \centering \begin{subfigure}{\textwidth} \includegraphics[width=\textwidth]{subfig1.png} \caption{subfig1} \label{fig:subfig1} \end{subfigure} \begin{subfigure}{\textwidth} \includegraphics[width=\textwidth]{subfig2.png} \caption{subfig2} \label{fig:subfig2} \end{subfigure} \end{figure} \newpage \begin{figure}[htp!] \centering \ContinuedFloat \begin{subfigure}{\textwidth} \includegraphics[width=\textwidth]{subfig3.png} \caption{subfig3} \label{fig:subfig3} \end{subfigure} \begin{subfigure}{\textwidth} \includegraphics[width=\textwidth]{subfig4.png} \caption{subfig4} \label{fig:subfig4} \end{subfigure} \end{figure} \newpage \begin{figure}[htp!] \centering \includegraphics[width=5in]{figure3.png} \caption{Caption3} \label{fig:fig3} \end{figure} \end{document} Fig.1 and Fig.2 are correctly numbered, but Fig.3 is numbered Fig.1. The subplots in Fig.2 are also numbered incorrectly: the first four are labeled (a)-(d), but the fifth on the third page starts from (a) again. Does `\ContinuedFloat` not work over more than two pages? Here is the output of `\listfiles`: article.cls 2007/10/19 v1.4h Standard LaTeX document class size11.clo 2007/10/19 v1.4h Standard LaTeX file (size option) graphicx.sty 1999/02/16 v1.0f Enhanced LaTeX Graphics (DPC,SPQR) keyval.sty 1999/03/16 v1.13 key=value parser (DPC) graphics.sty 2009/02/05 v1.0o Standard LaTeX Graphics (DPC,SPQR) trig.sty 1999/03/16 v1.09 sin cos tan (DPC) graphics.cfg 2007/01/18 v1.5 graphics configuration of teTeX/TeXLive pdftex.def 2011/05/27 v0.06d Graphics/color for pdfTeX infwarerr.sty 2010/04/08 v1.3 Providing info/warning/error messages (HO) ltxcmds.sty 2011/11/09 v1.22 LaTeX kernel commands for general use (HO) subcaption.sty 2012/01/12 v1.1d Sub-captions (AR) caption.sty 2012/02/19 v3.2f Customizing captions (AR) caption3.sty 2012/01/12 v1.4b caption3 kernel (AR) amsmath.sty 2000/07/18 v2.13 AMS math features amstext.sty 2000/06/29 v2.01 amsgen.sty 1999/11/30 v2.0 amsbsy.sty 1999/11/29 v1.2d amsopn.sty 1999/12/14 v2.01 operator names supp-pdf.mkii omscmr.fd 1999/05/25 v2.5h Standard LaTeX font definitions
[ 0.0056266565807163715, 0.0008486980805173516, 0.0044931573793292046, 0.025343384593725204, 0.03312099725008011, 0.01277111005038023, 0.008776376955211163, 0.013868981972336769, -0.01600193977355957, 0.0078930938616395, -0.019295427948236465, -0.0021884706802666187, -0.0017916469369083643, ...
[ -0.055484380573034286, 0.14028339087963104, 0.5281092524528503, -0.10362434387207031, -0.06812244653701782, 0.4027114510536194, 0.32963600754737854, -0.5089440941810608, -0.4465002715587616, -0.5580408573150635, 0.198158860206604, 0.16841961443424225, -0.03908132389187813, -0.1192286014556...
Suppose three of my party members are already level 99, but one isn't yet. When I get XP, does it all go to the one who isn't level 99, or is it split four ways and wasted on the other three?
[ -0.0020526002626866102, 0.025808408856391907, -0.00044368539238348603, 0.008809858001768589, 0.009886533953249454, -0.0031261034309864044, 0.010823524557054043, -0.009195258840918541, -0.026631753891706467, -0.012387028895318508, -0.015566812828183174, 0.019017262384295464, -0.00496959267184...
[ 0.31499922275543213, 0.30242690443992615, 0.24454081058502197, 0.16450273990631104, -0.15203846991062164, 0.24497216939926147, 0.23986870050430298, -0.17414098978042603, -0.4471443295478821, -0.1918582171201706, 0.060702789574861526, 0.18225638568401337, 0.10641219466924667, 0.098125621676...
I'm looking to get a job as a Python programmer. I know the basics of the language and have created a few games with it using pygame. I've also started to experiment with Django. However, looking at the job market, it doesn't seem very many Python jobs are web-related. On the desktop side of things, it doesn't seem like very many companies use the popular GUI libraries like pyQt or wxPython. How are companies actually using Python? What areas should one focus on to land a job as a Python programmer?
[ -0.0019479464972391725, -0.008270775899291039, -0.010977874509990215, 0.0020850286819040775, -0.006870722398161888, 0.0044952817261219025, 0.006908105686306953, 0.018907075747847557, -0.015219567343592644, -0.017429277300834656, -0.007353540509939194, 0.010152328759431839, 0.0129976049065589...
[ 0.770004391670227, 0.3633091151714325, -0.23838944733142853, 0.04328751191496849, 0.06630226969718933, -0.10568279772996902, 0.09408560395240784, 0.5783194303512573, -0.32435503602027893, -0.4592583179473877, 0.2134300023317337, 0.8096885085105896, -0.11780021339654922, -0.0318839773535728...
I am fairly new to *nix, and have come across the need to drop multiple processes, that should be run 100% of the time. to background using `&`. I use the following line in an init.d script to do this (running as the user `user`: `su -c 'process arg1 arg2 -w - | process2 arg1 -r - &' user` (where -w writes to and -r reads from STDOUT, STDIN) Specifically, I know this is not generally acceptable, as the processes aren't well shielded from outside influence. Is it acceptable to create background jobs for "services?" Should I instead use a FIFO/named pipe to handle the interprocess communication? If so, should I still create both processes as background jobs? Is this stable? For specifics, please refer to this mailing list thread. Thanks, Matt
[ -0.001775519922375679, 0.00956043228507042, -0.013702329248189926, -0.0002179304137825966, -0.031194154173135757, 0.0054788244888186455, 0.006255028769373894, 0.012132828123867512, -0.012165696360170841, 0.017013320699334145, 0.000944891944527626, 0.0021068661008030176, -0.006047806702554226...
[ 0.32443323731422424, 0.3220895230770111, 0.2740989625453949, -0.3716515302658081, 0.059246812015771866, -0.05298404023051262, 0.577919602394104, -0.14433978497982025, -0.10850729793310165, -0.47466546297073364, -0.06630624085664749, 0.3445517122745514, -0.3625450134277344, 0.00763016985729...
For my thought experiment, I will create my very own infinite expanding universe similar to our own assuming current cosmological theories are correct, except for a couple things: $$\text{(Hubble's constant) }H_0 = \mathrm{1 s^{-1}}$$ $$\text{ (speed of light) }c = \mathrm{100 m/s}$$ In my universe, Hubble's constant never changes from $t=0\to\infty$ and the current time is $t= \mathrm{1 s}$. Now there is an observer $O_1$ who fires a photon at an observer $O_2$ in which the observers are separated by a distance $$d=\mathrm{100m}$$ Due to the expansion of the universe, $O_1$ is moving relative to $O_2$ at the following velocity $$v = H_0d=\mathrm{100 m/s}$$ Clearly the photon will never reach $O_2$ since the photons can only travel at the speed of light, and already $O_2$ is receding too fast, so the photon could never hope to catch up. I believe then the radius of the Hubble Sphere for my universe is $\mathrm{100 m}$ So, here's what I want to be able to figure out: > What is the largest value we could assign $d$ for which the photon would > reach $O_2$, where $d$ is the proper distance (as described here) at the > current time ($t=\mathrm{1 s}$) from the $O_1$ in the direction of $O_2$? > > Also, how far apart were $O_1$ and $O_2$ at $t=\mathrm {0}$? I am trying to figure it out with these modified constants so I could wrap my head around the math in the real world. If I'm not mistaken there must be some integral calculus going on to figure it out, but I am having difficulty figuring it out because (for one) it has been so long since I took calculus, and (for two) the equations I know about are all with respect to $d$. For example, $v(d) = H_0d = 1d = d$, but what to do from there - and how to account for the increasing spacial distances in the formula. For example at $t = 2\mathrm{s}$ you might think that $O_2$ would be at $d=200\mathrm m$ since it was traveling at $100 \mathrm {m/s}$ and $\mathrm {1s}$ has ellapsed, so it must have moved $\mathrm {100m}$ from the original $d=\mathrm{100m}$. However, $d$ should be the proper distance, and we know that a point at $d=200\mathrm m$ was moving away from $O_2$ at $100 \mathrm{m/s}$ at $t=1 \mathrm {s}$ and there is no way $O_2$ could reach that far because it is outside of $O_2$'s Hubble Sphere (not to mention that hypothetically $O_2$'s proper distance should not change with respect to time. To complicate it further, along the 100m trip during $t \to 2\mathrm s$, $O_2$ has increased velocity proportional to $H_0$. My thought was that the expansion of space must have affect the speed of light, since theories are that distant galaxies can move away at speed greater than the speed of light in relation to the coordinates of an observer. This would then mean that the effective speed of light would decrease as the value of $H_0$ increases, is that right? I know that in the real world the value for $H_0$ is so small ($2.3×10^{−18}\mathrm{s^{−1}}$) and the value for $c$ is so big ($3.0×10^8 \mathrm{m/s}$), there would need to be an insane 26 significant figures in any sort of measurement before the expansion of the universe affected the speed of light. But still - hypothetically with my thought experiment it seems to have a non- zero effect - making me question what it actually means when someone tells you the speed of light is a particular value (like I did earlier by saying it was 100m/s) - would it mean the effective speed given the expansion of the universe, or the speed with relation to proper distances relative to a particular time (i.e. how far it is from where it was emitted, or how close it is to point it was aimed at)? (also I just like Math and would like to know the math behind trying to figure out the answer to my experiment)
[ -0.01716316118836403, 0.010907059535384178, -0.015756282955408096, -0.00368823716416955, 0.0020476579666137695, -0.0074129169806838036, 0.007888887077569962, -0.0056967055425047874, -0.010077251121401787, -0.021099621430039406, 0.011557262390851974, 0.006138382479548454, -0.00149821199011057...
[ -0.20746944844722748, -0.15673775970935822, 1.057925820350647, 0.0583646260201931, 0.23258265852928162, 0.1946488469839096, 0.054807137697935104, 0.5671826601028442, -0.5471203327178955, -0.9101860523223877, 0.1657181829214096, 0.24273094534873962, -0.5010284781455994, 0.8356510996818542, ...
Some people adopt the affectation in a message, memo or email where the sender is identified explicitly in the header of the message that the sender is shown as being "From the desk of Joe Smith" rather than just from "Joe Smith". In some situations, "From the office of" makes more sense - I take this as meaning that someone on the staff of a politician or senior executive wrote the contents, but that the office holder is happy to be considered as supporting the message. Is this what Joe Smith is trying to tell us with this construction?
[ 0.003660130314528942, -0.0032386351376771927, -0.005804209969937801, 0.02096250280737877, -0.02343074232339859, 0.0006968425586819649, 0.008783962577581406, -0.004216799046844244, -0.010356449522078037, -0.0035894473548978567, -0.016581255942583084, 0.016992883756756783, 0.018361972644925117...
[ 0.5173424482345581, 0.07015369832515717, -0.01915118843317032, -0.0725422278046608, -0.17596937716007233, -0.028852304443717003, 0.034194815903902054, 0.3409441411495209, -0.25123363733291626, -0.3271513283252716, -0.3762776255607605, 0.08220205456018448, -0.4260767102241516, 0.04109280928...
How can I fine tune and reduce the vertical spacing of lines so characters from lines above and below are almost pixel-touching each other? Example: ![Example of lines barely touching each other](http://i.stack.imgur.com/OSHYI.png)
[ -0.0065657212398946285, 0.01780572533607483, -0.006477938033640385, 0.02526594139635563, -0.005712374113500118, -0.020615292713046074, 0.01130171399563551, -0.014675811864435673, -0.024208026006817818, 0.004800229333341122, -0.016733065247535706, 0.010968616232275963, -0.0003052954562008381,...
[ 0.14176520705223083, -0.03260987624526024, 0.4206327199935913, 0.10184933245182037, -0.09376877546310425, 0.22178177535533905, 0.24393564462661743, -0.28977206349372864, -0.3121736943721771, -0.7345504760742188, 0.10764152556657791, 0.21808597445487976, 0.05144600570201874, -0.081275552511...
Until recently, I thought that execution of \futurelet\testtoken<token1><token2>... has the following effect: The control sequence `\testtoken` is `\let` to `<token2>` and `\futurelet\testtoken` is stripped from the input stream, so TeX continues with `<token1><token2>...`. In this answer of mine I observed (with TH's help) that this is not quite true with respect to catcodes, and in this nice answer Philippe explains that things can also go terribly wrong in nested alignments. For the sake of completeness let me give two short examples here. \def\activateA{\catcode`A=\active} {\activateA \gdefA{undocumented behaviour?} } An \futurelet\testtoken\activateA A \bye The expected output is "An undocumented behaviour?", but this one gets only when `\futurelet\testtoken` is removed. With the `\futurelet`, the output is just "An A". The reason seems to be that `\futurelet` fixes the catcode of the last `A`. (One still gets that `\testtoken=the letter A`.) The second example is a bit longer and more involved. \def\begintestalign{\show\testtoken $\vcenter\bgroup\halign\bgroup##&##\cr} \def\endtestalign{\egroup\egroup$} \halign{#&#\cr a & \futurelet\testtoken \begintestalign & c \cr test & de \cr \endtestalign \cr } \bye The expected output is ![](http://i.stack.imgur.com/IvxgY.png) (and one would expect `\testtoken=alignment tab character &`), but instead one gets `\testtoken=\outer endtemplate`, and then ! Emergency stop. <recently read> \endtemplate <template> \endtemplate l.7 & c \cr ! ==> Fatal error occurred, no output PDF file produced! I can't fully understand what's happening here, but my main question is: Where do I find this behaviour of `\futurelet` documented? (And honestly, I'd really like to have a `\futurelet` that behaves as expected; the implemented behaviour is rather mean, and I wonder if that's a bug or a feature.)
[ -0.0015402043936774135, 0.013035835698246956, -0.008119771257042885, 0.01289520226418972, 0.007336548529565334, 0.017562896013259888, 0.009418990463018417, 0.0030091393273323774, -0.010142186656594276, -0.0011069206520915031, -0.020177481696009636, 0.0027985479682683945, -0.00407576421275734...
[ -0.07954498380422592, -0.031815893948078156, 0.04157169535756111, -0.00787609163671732, 0.29974961280822754, -0.04498201236128807, 0.3609616458415985, -0.09534990787506104, 0.08771459013223648, -0.43179798126220703, -0.19504135847091675, 0.42777219414711, -0.35723552107810974, -0.073140732...
I'm pretty deep into my latest game of Galactic Civilizations 2: Twilight of the Arnor, and major races are starting to drop out of the game. Except, in Gal Civ 2 instead of sticking around until they are totally eliminated, the major races will surrender to a friendly civilization, deeding their planets over to that friendly civilization. How does the AI determine who to surrender to, and is there some way to game the system? Daddy wants more free planets.
[ -0.010504732839763165, 0.0035980495158582926, -0.004787744954228401, -0.004844955168664455, -0.01583624817430973, -0.011312014423310757, 0.007923786528408527, -0.006448033265769482, -0.016352394595742226, 0.004371075890958309, -0.008715548552572727, 0.011873813346028328, -0.01523318234831094...
[ 0.17245295643806458, -0.13079597055912018, 0.5288345217704773, 0.05194719880819321, -0.13854585587978363, -0.06373796612024307, -0.37614545226097107, 0.30525046586990356, -0.2620592713356018, -0.5758260488510132, -0.11409955471754074, 0.3328711688518524, -0.2346915602684021, 0.458304256200...
For example, the shadow blade shows 3000 (1000) What is the (1000) supposed to represent?
[ 0.043193209916353226, 0.01002498995512724, -0.038902584463357925, 0.05620821192860603, -0.03928413242101669, 0.014451513066887856, 0.025439893826842308, 0.007412445265799761, -0.022243225947022438, -0.05964305251836777, -0.04267800226807594, 0.016710704192519188, -0.059533677995204926, 0.0...
[ 0.6317485570907593, -0.010539274662733078, 0.2542489171028137, 0.44360992312431335, -0.2923482358455658, 0.28181055188179016, 0.27611246705055237, -0.2416432499885559, -0.3732960522174835, -0.3919115364551544, 0.08113901317119598, -0.02412915788590908, -0.13014712929725647, 0.1050679236650...
I need to show one subsubsection title in the main body of the text and another in the contents page. I have been using: \subsubsection[contents title]{text title} This had worked really well. Now one of the titles that I need in the contents has square brackets in and so gives me loads of error messages (I assume because the brackets are getting confused). This is my minimum working example (`\subsubsection[$[\text{PtL}^5\text{(ppy)Cl}]^+$]{Title in text}` is the line causing the problem): \documentclass[12pt,a4paper,oneside]{memoir} \begin{document} \tableofcontents \chapter{1} \subsubsection{Title in text} \subsubsection[$[\text{PtL}^5\text{(ppy)Cl}]^+$]{Title in text} \end{document} Thanks for your help, sorry, I'm still relatively new to LaTeX!!
[ 0.007826752960681915, -0.0022878521122038364, -0.017658255994319916, 0.026762545108795166, 0.0027672830037772655, 0.017823301255702972, 0.00783485360443592, 0.028500646352767944, -0.0180787593126297, -0.024714462459087372, -0.020338762551546097, -0.004010246135294437, -0.014856003224849701, ...
[ 0.06483771651983261, 0.22077199816703796, 0.5397258400917053, -0.008456423878669739, 0.12807777523994446, -0.24128422141075134, 0.1696518510580063, -0.20535412430763245, -0.17335514724254608, -0.5198546648025513, -0.11719251424074173, 0.35625243186950684, -0.019788866862654686, 0.342734277...
First, this is not a copy of Change \textwidth and \textheight in mid-document as I can't set a manual `\pagebreak` and `\newgeometry` or `\restoregeometry` thereafter. I want to change the footer size for the first page of a document to include some kind of legend. The main content is a longtable with varying(!) content length. The legend should only be included on the first page and the footer should only be extended on this page also. This works after the lontable finishes (see `lipsum` text) but not while it continues. The footer stays big, although it does not contain the legend anymore. I'm in serious doubt at the moment that it is possible with pdflatex to do what I want, but I'd like to see that I'm wrong and there is a way to do it as intended. \documentclass[a4paper,11pt]{article} \usepackage[top=2cm,left=2cm,right=2.5cm,bottom=5.5cm,includeheadfoot]{geometry} \usepackage{booktabs} \usepackage{longtable} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage{fancyhdr} \usepackage{lipsum} % Header, Footer \setlength{\headheight}{26pt} % Pagestyle firstpage \fancypagestyle{firstpagestyle} { \setlength{\textheight}{620pt} \setlength{\footskip}{160pt} \lhead{This is my header}% \chead{}% \rhead{\today}% \lfoot{ \smash {% % hide vertical stretch of the following content \parbox[b]{\textwidth} { \Legende This is my normal footer text. } } }% \cfoot{} \rfoot{\thepage}% \renewcommand\headrulewidth{0.35pt}% \renewcommand\footrulewidth{0.35pt}% } % Pagestyle all other pages \fancypagestyle{mypagestyle} { \setlength{\textheight}{660.2pt} \setlength{\footskip}{30pt} \lhead{This is my header}% \chead{}% \rhead{\today}% \lfoot{This is my normal footer text.} \cfoot{} \rfoot{\thepage}% \renewcommand\headrulewidth{0.35pt}% \renewcommand\footrulewidth{0.35pt}% } \newcommand{\Legende}{ \parbox{0.9\textwidth}{some \\ lines \\with \\ a \\ lot \\of \\text }\\[1.5ex]} %Dummy text for table \newcommand{\bla}{bla \\ blubb \\ bla \\blubb \\bla \\} \newcommand{\blubb}{\bla\bla\bla\bla\bla\bla} \begin{document} \pagestyle{mypagestyle} \thispagestyle{firstpagestyle} \renewcommand{\arraystretch}{1.25} \begin{longtable}{c} \toprule \blubb \blubb \blubb \bottomrule \\[0em] \end{longtable} \lipsum \lipsum[1-4] \end{document}
[ 0.0034732480999082327, 0.011058688163757324, -0.005620710086077452, 0.004716420546174049, 0.012371785938739777, 0.0028612983878701925, 0.007406844291836023, 0.017861101776361465, -0.009789369069039822, -0.023371217772364616, -0.007801400031894445, 0.005649343132972717, -0.01057821698486805, ...
[ 0.1730222851037979, -0.11492333561182022, 0.6924179792404175, 0.08336825668811798, 0.10728659480810165, -0.07528844475746155, -0.14773118495941162, -0.050450123846530914, -0.39928439259529114, -0.2362799197435379, -0.009508340619504452, 0.32727789878845215, 0.09163450449705124, 0.021963974...
Disregarding String Theory, which says that electrons and quarks are 1-dimensional, why are electrons and quarks 0-dimensional? The way I look at it is that they have mass, and if we were the same size as an up quark and looked at another up quark, it would look like a 3-dimensional object. So why do we say particles are 0-dimensional as opposed to 3-dimensional? I hope I'm correcting in saying they are regarded as 0-dimensional, the reason I think it is because of this quote from Wikipedia: > _String theory posits that the electrons and quarks within an atom are not > 0-dimensional objects, but rather 1-dimensional..._
[ -0.0031135284807533026, 0.027332443743944168, -0.001910623162984848, -0.0016591253224760294, 0.010529471561312675, -0.023961344733834267, 0.009425675496459007, 0.0010672963690012693, -0.010245177894830704, -0.027584215626120567, -0.007473006844520569, 0.011571155861020088, -0.004681052640080...
[ 0.49485236406326294, 0.2743804156780243, -0.0007539471844211221, 0.03599657490849495, -0.47218576073646545, -0.42210817337036133, 0.3200588822364807, 0.2965993285179138, -0.3825598955154419, -0.20148871839046478, 0.07204879820346832, 0.1621321588754654, -0.5646364688873291, 0.8715666532516...
Once in a great while I stumble across a strange fact about an English phrase or term that originated directly from film or print. For instance, * "Be afraid. Be very afraid." - Originated from the movie The Fly (1986).1 * The name "Wendy" was a rarely-used name until the _Peter Pan_ character Wendy Darling.2 * William Shakespeare invented a lot of words3a and turned many a phrase3b, including "eyeball," "puking," and "laughing stock." My question is, what are your favorite or most astonishing portions of your vernacular whose origins lie in film, music or print media? Regional or local uses are fine too.
[ -0.010410132817924023, 0.010532991029322147, -0.013632828369736671, 0.01153620332479477, 0.00990577694028616, 0.007373310159891844, 0.007016819901764393, 0.026458848267793655, -0.009027193300426006, -0.005281492136418819, -0.0069756717421114445, 0.015147329308092594, 0.02463085949420929, 0...
[ 0.4399643838405609, -0.14899975061416626, -0.26561102271080017, -0.03805650770664215, -0.2750522196292877, -0.18294036388397217, 0.4802719056606293, 0.8104358315467834, -0.21151615679264069, -0.2276497483253479, 0.1758149415254593, 0.11243899911642075, -0.060362666845321655, 0.426249116659...
Is there some book or paper that formalizes statistical mechanics, like some people have done with relativity, and proves the second law of thermodynamics from more foundational axioms?
[ 0.020981287583708763, 0.031886134296655655, -0.00032251261291094124, 0.02737325057387352, -0.02297237701714039, -0.01010698638856411, 0.013554422184824944, -0.02711867168545723, -0.020696045830845833, -0.023730125278234482, -0.026294784620404243, 0.019823571667075157, -0.020186718553304672, ...
[ 0.3680878281593323, -0.11209077388048172, -0.40416353940963745, 0.665248692035675, 0.17890280485153198, -0.18801645934581757, -0.13630233705043793, 0.021985655650496483, -0.023287205025553703, -0.02326248399913311, 0.10534912347793579, 0.03880179300904274, 0.02318708226084709, 0.1339438408...
I have a few questions regarding functionality testing of PBs in a new wp installation (3.0.4): 1. Are pingbacks sent immediately when a post is published, or are they scheduled as a cron job? If the later is correct, how often does the job run and can I trigger it manually? 2. Are there any other terms for PBs to be sent besides having "Attempt to notify any blogs linked to from the article" set to on, and of course a link to another blog in the post content? (for example, should the sender post be public and not private? should the blog be non private?) 3. Should the outbound link be placed in the post content or can it be placed in a custom field of the post in order for the PB to be sent? 4. In case PBs are not sent nor received in my blog, what is the best way to debug and detect the problem? Thanks in advance!
[ 0.0034605362452566624, 0.015319647267460823, -0.0006910722004249692, 0.025902917608618736, 0.02279488369822502, 0.01916692778468132, 0.007586643565446138, 0.019236378371715546, -0.012541355565190315, -0.0004708748310804367, -0.011759590357542038, 0.011957697570323944, 0.001703899004496634, ...
[ 0.9179884791374207, -0.08342372626066208, 0.5148832201957703, 0.05669895187020302, -0.290175199508667, -0.14258253574371338, 0.16130410134792328, -0.5696852803230286, -0.19399531185626984, -0.5340356826782227, 0.0915258377790451, 0.3721325397491455, -0.26335108280181885, 0.0345467776060104...
My source dataset is DWG and destination is GDB. The source dataset has Date attributes as Char in DD-MM-YYYY. The corresponding destination attributes have to be of Date datatype, that too in M/D/YYYY format. I tried DateFormatter transformer, with output date format string as **FME Date (%D)** , I got the error as `The date 'FME Date (05/14/1981)' is not valid. Valid forms for a date type are YYYYMMDD, YYYYMMDDHHMMSS, or HHMMSS` For format string as **%D** , I got error as: `The date '05/14/1981' is not valid. Valid forms for a date type are YYYYMMDD, YYYYMMDDHHMMSS, or HHMMSS` For format string as **%m"/"%e"/"%Y** and **%m/%e/%Y** , again the above error message. As from the error, I cannot perform the desired formation. Is there a transformer available in FME Workbench to perform desired format conversion? Or should I try something like using **`SubstringExtractor`** and then **`StringConcatenator`**? I need to read the documentation thoroughly to see if I can get it working. Thanks
[ -0.025343645364046097, 0.012711910530924797, -0.012204760685563087, 0.011673002503812313, 0.0013562992680817842, -0.0060812304727733135, 0.009475277736783028, 0.025159139186143875, -0.010280899703502655, -0.011940624564886093, -0.01447482779622078, 0.013119785115122795, 0.006160294637084007,...
[ 0.1939641386270523, 0.18342441320419312, 0.6490033268928528, -0.1953924000263214, -0.050419606268405914, 0.05242043733596802, 0.01474896539002657, 0.16442298889160156, 0.2653436064720154, -0.6810563802719116, -0.080448217689991, 0.4234883785247803, -0.06333284825086594, 0.6251455545425415,...
Recently I changed my password on my computer but for the past few days I keep entering the old one when trying to log in or unlock it. Is there a word or phrase in to describe this? Another example would be writing 2013 for the year out of habit even though you know it's 2014.
[ -0.0008587316842749715, 0.00385176669806242, -0.02661161497235298, 0.029228074476122856, -0.020089266821742058, 0.024634726345539093, 0.011344374157488346, -0.015457120724022388, -0.015745747834444046, 0.005386367905884981, -0.008669925853610039, 0.006245344411581755, 0.015108121559023857, ...
[ 0.421305775642395, 0.17692026495933533, 0.2545652687549591, 0.009404747746884823, -0.005212315823882818, -0.06719538569450378, 0.5834366083145142, 0.5506365299224854, -0.8188196420669556, -0.5811816453933716, 0.11655731499195099, -0.43931928277015686, 0.29068103432655334, 0.543627798557281...
The objective is to get a set of all buildings of an Open Street Map section in form of single points in a shapfile. 1. I downloaded the section from osm. 2. Imported the section into QGIS (and later imported it into a PostGis db using osm2pgsql) 3. To transform the polygons (representing buildings) into points I used the Polygon Centroids function Unfortunately some of the buildings are represented by multiple polygons subsequenty I did get multiple points for one building which I is not what I wanted. Searching the web I found out that these polygons should be multipolygons but neither qgis nor osm2pgsql does import these relations so that I could merge the poygons of a building Any ideas how i could get single point list of the buildings?
[ -0.016221914440393448, 0.01477053016424179, 0.0023529171012341976, 0.025168422609567642, 0.0014781233621761203, -0.0009445520117878914, 0.009638052433729172, 0.03955100476741791, -0.016999607905745506, -0.005354853346943855, -0.010711940005421638, 0.012501927092671394, -0.009218783117830753,...
[ -0.1413370817899704, 0.08052613586187363, 0.5180880427360535, 0.21978715062141418, -0.5260694026947021, -0.0740516260266304, 0.5065465569496155, -0.47476544976234436, 0.08759690821170807, -0.5759069919586182, -0.132806658744812, -0.012749100103974342, -0.2176068127155304, 0.109959855675697...
So, I understand that quantum teleportation is the transfer of a quantum state from one particle or system of particles and its correlations to another receiving system consisting of particle(s). Can't this be used in a way to achieve teleportation as depicted in sci-fi? You start out with a translationally entangled particle pair (position and momentum are correlated) as described by this paper: http://pra.aps.org/abstract/PRA/v61/i5/e052104 Now according to these papers, you can theoretically teleport the position and momentum information of particle. Atomic teleportation of the external degrees of freedom, (Their position and momentum) http://pra.aps.org/abstract/PRA/v49/i2/p1473_1 http://prl.aps.org/abstract/PRL/v86/i14/p3180_1 http://iopscience.iop.org/0295-5075/75/6/847 Firstly, I need someone to help me understand **exactly what is being teleported in regards to the papers I've listed**. I understand completely the discrete case of quantum teleportation with spin or polarization, I'm not asking for that. This is what I know: 1) Start out with a translationally entangled pair. 2) Interact an "input" particle with one of the pairs. 3) Make a measurement of the input particle with one of the pairs' position and momentum. (This is the step I don't really understand) But I know that this is analogous to the Bell state measurement in discrete quantum teleportation that is widely described everywhere. 4) This result is communicated to the other laboratory where appropriate "shifts" of position and momenta are done to the entangled pair just like in the discrete case again. So, if position and momentum information is teleported, then does this mean that if the input particle was propagating in the X-direction like a wave- packet, then after teleportation, the receiving particle will now move in the X-direction relative to its original position? This is quite confusing to me. I want to understand this because I want to ask the trillion dollar question, "What if you are to replace the input particle with a input MOLECULE?" For a diatomic molecule, its simply two atoms of the same type. According to this wikipedia article: http://en.wikipedia.org/wiki/Linear_combination_of_atomic_orbitals I know there are vibrational states in molecules which ARE essentially position and momenta information right? If we are to set up TWO translationally entangled pairs of atoms of the same type... And have them simultaneously interact with the molecule in step 3), _then can't we do the appropriate SHIFTS of the entangled pairs of atoms to make them turn into the input molecule?!_ So, the overall effect is that if you start out with H2 in Lab A, and had a translationally entangled pair for each atom, then you'll end up with H2 in Lab B after teleportation which is exactly what we want if we want to stick close to the sci-fi sense. If what I am saying makes sense, then can't this be in principle be extrapolated to larger molecules and eventually cells, and organs, to an entire human? I know that this is a large jump... However, what I'm trying to say is that according to this kinda scheme I'm asking about, its sticking to real physics, and its essentially accomplishing teleportation in the sense people know about. Do I make sense?
[ 0.0038049877621233463, 0.015521653927862644, 0.013436920940876007, -0.0007995889172889292, 0.0028708067256957293, -0.024705037474632263, 0.006871153600513935, -0.010336952283978462, -0.011941561475396156, -0.018086818978190422, -0.011015554890036583, 0.01672990620136261, -0.01094428915530443...
[ 0.5243349075317383, -0.37046071887016296, 0.5199138522148132, 0.39004915952682495, -0.0630556270480156, -0.21250097453594208, -0.26518514752388, -0.24862080812454224, -0.1687571108341217, -0.49082499742507935, 0.03812272474169731, 0.17943750321865082, -0.3499971032142639, 0.524396896362304...
I am testing several frameworks (currently Vafpress) to help me creating custom meta boxes. I need to save a lot of attachments with my posts but noticed that the attachment is saved with its full URL, not the ID. 1. There seems to be no "easy" way to get full data (title, desc, date) from an attachment, when I only have the URL, right? 2. Is there any framework that saves ID instead of the URL? 3. Any other suggestions to save attachments, maybe without a custom meta box?
[ 0.007917312905192375, -0.004030730575323105, 0.002552037127315998, 0.012107216753065586, 0.025726303458213806, 0.015571868978440762, 0.007935164496302605, 0.041831791400909424, -0.015298639424145222, -0.042249228805303574, -0.014978132210671902, 0.008010734803974628, 0.00302319903858006, 0...
[ 0.5002095103263855, 0.0017312170239165425, 0.11416243016719818, 0.28320133686065674, -0.34279024600982666, -0.20934294164180756, 0.23224599659442902, -0.02714589238166809, -0.10825766623020172, -0.5987192988395691, 0.46671485900878906, 0.26129940152168274, -0.20274408161640167, 0.340109169...
This equation: $(\frac{-a}{x})^2=\sqrt{\frac{1}{x}}$ at $a > 0$ and $x > 0$ has a clear solution $x=a^{4/3}$, doesn't it? However, Reduce[(-(a/x))^2 == (1/x)^(1/2) && x > 0 && a > 0] // ToRadicals yields > > a > 0 && x == (-1)^(2/3) a^(4/3) > where ComplexExpand[(-1)^(2/3)] > > -(1/2) + (I Sqrt[3])/2 > is a complex number, numerically `N @ %` yields `-0.5 + 0.866025 I`. Why?
[ 0.0031523837242275476, 0.008319446817040443, -0.006060441490262747, -0.0008949476759880781, 0.004221438430249691, -0.019742682576179504, 0.0046796719543635845, -0.0048509687185287476, -0.008120675571262836, 0.008314432576298714, -0.005220663733780384, 0.0011867518769577146, -0.01784685440361...
[ -0.5377835035324097, 0.03701110556721687, 0.6528357267379761, 0.00117680081166327, 0.22870928049087524, 0.15820784866809845, 0.1951492875814438, -0.4674969017505646, -0.0603012852370739, -0.5064336657524109, -0.2062681019306183, 0.5728374123573303, -0.1734023541212082, 0.06474775820970535,...
Frequently, playing the Scout in TF2, I hear this phrase: > Thanks pally! What does the pally in this sentence indicate? Is it his imaginary friend? A paladin? Is it a nickname for something?
[ -0.028041230514645576, 0.01733255200088024, -0.012378577142953873, 0.01853790134191513, -0.022705845534801483, 0.03405844792723656, 0.013765696436166763, -0.04600352793931961, -0.02158866450190544, -0.02175431326031685, -0.009111358784139156, 0.0029056924395263195, -0.009830793365836143, 0...
[ -0.17904649674892426, 0.03270089253783226, 0.2420622855424881, -0.06884129345417023, -0.1159171462059021, -0.09504441171884537, 0.2552562952041626, 0.22751808166503906, -0.5776012539863586, -0.8187629580497742, -0.21641018986701965, 0.07915529608726501, 0.38360822200775146, -0.241046205163...
I've got a word processor (WordGrinder) which supports Unicode. It can export LaTeX files, which then get run through XeTeX. Unfortunately, it seems that while XeTeX _understands_ the Unicode correctly, it's not doing any kind of font substitution if the default font doesn't have a particular glyph in it. And since the default font is Computer Modern Roman, which has lousy font coverage, the default behaviour if I don't specify a font is that all my Unicode characters just vanish. This is not optimal. For test purposes I've installed Linux Libertine O and hard-coded a reference to the font. More things now work, because Linux Libertine has much better Unicode coverage. But I still have missing glyphs. And I'd very much prefer not to have to hardcode a font reference in my output files because that makes it much less likely that the end user will be able to use the output files. (Linux Libertine O is not a default part of my distribution, for example.) Is there any way I can tell XeTeX to _automatically_ find an appropriate font which supports the Unicode glyphs? I'm quite happy with rendering in Computer Modern Roman and falling back to something else for unsupported glyphs; I don't care if the appearance is mismatched as long as the glyphs are actually _there_. (If the user wants a specific font they can edit the file.) The distribution I'm currently using is the Debian version of xetex-live, but I'd like a distribution-agnostic fix if there is one. I'm happy with requiring XeTeX.
[ 0.019785966724157333, 0.0006673606112599373, -0.010712134651839733, 0.005703731440007687, -0.008659044280648232, -0.021830499172210693, 0.008285840973258018, -0.0002111236099153757, -0.014139516279101372, -0.01887657679617405, -0.01218637079000473, -0.005026895087212324, 0.020374441519379616...
[ 0.1637975573539734, 0.16078796982765198, 0.24455168843269348, 0.06694846600294113, 0.04996601119637489, 0.027608122676610947, 0.3774116039276123, 0.21147693693637848, 0.08340746909379959, -0.7777062654495239, 0.027454877272248268, 0.6253424286842346, -0.18420325219631195, 0.084477424621582...
I am analyzing the influence of social capital on household income. I have collected data through a questionnaire survey for my study.Different dimensions (variables) of social capital were measured using 5-point multiple likert items.The scores on each item was summed up to obtain the single value for each variable.Now i have 6 such variables and I want to perform factor analysis on these 6 variables to obtain a single social capital index for each household.My question is that can I summed up the factor scores (obtained through factor analysis) to obtain social capital index? if my approach is not appropriate, please suggest me some alternatives. I have to use this social capital index in multiple regression for further analysis.
[ 0.016519643366336823, 0.005998439155519009, -0.007519055157899857, 0.014035462401807308, 0.04662130028009415, -0.00783088244497776, 0.01015385240316391, -0.01944199576973915, -0.011151483282446861, -0.021680593490600586, -0.006857812404632568, 0.007715100422501564, -0.02139287441968918, 0....
[ 0.571650505065918, 0.22742170095443726, 0.3715660870075226, 0.23127293586730957, -0.21673795580863953, 0.780281662940979, 0.12217029929161072, -0.2851952910423279, -0.5423942804336548, -0.6573121547698975, 0.3837236166000366, 0.20549431443214417, 0.03765498101711273, 0.6548689007759094, ...
I seem to have too many colour escapes so it only works in some prompts - can someone please correct for me ? PS1='\[\e[0:35m[\u@\e[0:33m\h\e[m \e[0:35m\w]\e[m \e[0:33m`git branch 2> /dev/null | grep -e ^* | sed -E s/^\\\\\*\ \(.+\)$/\(\\\\\1\)\ /`\e[m\] \n$ '
[ 0.006405914202332497, 0.006833801046013832, -0.0025234490167349577, 0.011518307030200958, -0.01256951130926609, 0.007042961660772562, 0.0036867070011794567, -0.007795567158609629, -0.010458286851644516, 0.03200888633728027, -0.004616871476173401, -0.003982916008681059, 0.0018268225248903036,...
[ -0.17788296937942505, -0.12464406341314316, 0.4639822840690613, -0.08596350252628326, 0.3087022304534912, 0.3143060803413391, 0.8768236637115479, 0.03611335530877113, -0.22817207872867584, -0.5488350987434387, -0.33679327368736267, -0.06592623144388199, -0.5543579459190369, 0.0650534853339...
have the following setup Raspberry pi with samba installed and running Windows 8 machine that can see the samba share, access the proper DIR, see all contents In the smb.conf file, at the base of the file, I created the above dir with this [AirLock] comment = AirLock path = /thepath/ writeable = Yes create mask = 0777 directory mask = 0777 browseable = Yes public = Yes read only = no With these settings I can see the whole dir that I want but I cant create files, edit files, or create DIRs. I need to do all of this.
[ -0.00014626327902078629, -0.0004238135297782719, -0.01564886048436165, 0.013049116358160973, -0.007870987057685852, -0.013310744427144527, 0.009236459620296955, 0.0292787104845047, -0.016400303691625595, -0.012504566460847855, -0.00492442911490798, 0.007123721297830343, -0.002617337740957737...
[ -0.21796470880508423, -0.028782490640878677, 0.46203699707984924, -0.13644863665103912, -0.09536571800708771, -0.191371887922287, 0.44861528277397156, -0.1671680361032486, -0.11349323391914368, -0.66715008020401, 0.11747433990240097, 0.7606080770492554, -0.6792749762535095, -0.164153695106...
I have recently found a very nice example of how to create text boxes with a parchment-style layout: http://www.texample.net/tikz/examples/framed-tikz/ Also, here on tex.stackexchange, there are numerous samples employing the `framed` package to produce this kind of layout. However, in my document, I am already using the `mdframed` package to typeset a different kind of frame/box. Is it possible to produce such a layout using the `mdframed` package as well? If so - how? ## Edit I following the suggestion from Werner, I have played around a little and achieved the following: \makeatletter \usetikzlibrary{decorations.pathmorphing, patterns} \renewrobustcmd*\mdf@tikzbox@tfl[1]{%three or four borders \path(0,0)rectangle(\mdfboundingboxwidth,\mdfboundingboxheight);% replace \clip by \path \begin{scope}[mdfcorners]% \clip[preaction=mdfouterline]% [postaction=mdfbackground]% [postaction=mdfinnerline]#1;% \end{scope}% \path[mdfmiddleline,mdfcorners]#1; }% \newenvironment{parchment}[1][]{% \mdfsetup{% frametitlebackgroundcolor=none, innermargin=5pt, frametitleaboveskip=-\ht\strutbox, frametitlebelowskip=-\ht\strutbox, innertopmargin=10pt, innerbottommargin=10pt, tikzsetting={ line width=2pt,orange!30!black!10, decorate, decoration={random steps, segment length=2.5cm, amplitude=1.5mm} }, everyline=false } \ifstrempty{#1}{% \mdfsetup{% frametitle={\fcolorbox{black}{white}{Handout}}, }% }{% \mdfsetup{% frametitle={\fcolorbox{black}{white}{Handout: #1}}, }% }% \begin{mdframed}[] \relax% }{\end{mdframed}} \makeatother However, when I use `everyline=true` instead of `everyline=false`, the pagebreaks in my document change. Can anybody explain? ## Edit 2 I have finally found some time to revisit this issue. Please have a look at this MWE: \documentclass{article} \usepackage[a4paper]{geometry} \usepackage{tikz} \usepackage{etextools} \usepackage[framemethod=tikz]{mdframed} \usepackage{lipsum} \geometry{verbose,tmargin=2.5cm,bmargin=2.5cm,lmargin=2cm,rmargin=3cm,marginparwidth=2.5cm,headheight=15pt} \usepackage{emerald} \let\fancyfont\ECFAugie \makeatletter \usetikzlibrary{decorations.pathmorphing, patterns} \renewrobustcmd*\mdf@tikzbox@tfl[1]{%three or four borders \path(0,0)rectangle(\mdfboundingboxwidth,\mdfboundingboxheight);% replace \clip by \path \begin{scope}[mdfcorners]% \clip[preaction=mdfouterline]% [postaction=mdfbackground]% [postaction=mdfinnerline]#1;% \end{scope}% \path[mdfmiddleline,mdfcorners]#1; }% \newenvironment{parchment}[1][]{% \mdfsetup{% frametitlebackgroundcolor=none, innermargin=5pt, frametitleaboveskip=-\ht\strutbox, frametitlebelowskip=-\ht\strutbox, innertopmargin=10pt, innerbottommargin=10pt, tikzsetting={ line width=2pt,orange!30!black!10, decorate, decoration={random steps, segment length=2.5cm, amplitude=1.5mm} }, everyline=false } \ifstrempty{#1}{% \mdfsetup{% frametitle={\fcolorbox{black}{white}{Handout}}, }% }{% \mdfsetup{% frametitle={\fcolorbox{black}{white}{Handout: #1}}, }% }% \begin{mdframed}[] \relax% }{\end{mdframed}} \makeatother \begin{document} \begin{parchment}[Lorem ipsum] \fancyfont \lipsum[1-60] \end{parchment} \end{document} Here's the list of issues that I would still like to resolve: * The frame borders are sometimes "chopped off". I think this is related to the bounding box/clipping of TikZ, but I don't know how to resolve it. ![screenshot](http://i.stack.imgur.com/tODvi.png) * I would like to give the whole thing a background image that is cropped precisely to the drawn border. Additionally, it would be great if the baselines could be slightly curved and tilted to look more like handwritten script - but I guess I can figure out how to do this - perhaps like this: Cthulhu document
[ 0.008273455314338207, 0.013483624905347824, 0.005925287958234549, 0.022451933473348618, 0.025901298969984055, -0.00009940078598447144, 0.008640632033348083, 0.0017200747970491648, -0.019527113065123558, -0.025090999901294708, -0.014803201891481876, -0.00045383768156170845, -0.012997187674045...
[ 0.734585165977478, 0.22796016931533813, 0.16564759612083435, -0.18298466503620148, -0.15717245638370514, -0.060117918998003006, -0.1999179869890213, -0.3171007037162781, -0.35656505823135376, -0.7453327178955078, 0.2517116069793701, 0.4461895227432251, -0.2691408395767212, 0.11213316023349...
I am currently trying to generate a list of permutations of length 2 of elements of a list of strings. For example: Permutations[{"s1", "s2", "s3", "s4"}, {2}] {{"s1", "s2"}, {"s1", "s3"}, {"s1", "s4"}, {"s2", "s1"}, {"s2", "s3"}, {"s2", "s4"}, {"s3", "s1"}, {"s3", "s2"}, {"s3", "s4"}, {"s4", "s1"}, {"s4", "s2"}, {"s4", "s3"}} which is quite correct. Now, my problem starts when attempting the same operation for a really big list of strings imported from a tsv file - we are talking of a bit more than 10000 strings. All sort of characters can be expected in strings. This is the code that I am using: t1 = Import["/home/bla/work/parsed_names.tsv"] t2 = Table[t1[[i]][[1]], {i, 1, Length[t1]}] Permutations[t2, {2}] After evaluation of the last line, I get no output. If I evaluate it twice, all the symbols defined previously (t1, t2) become undefined, and Permutations finally give this output: Permutations[t2, {2}] Permutations::normal: Nonatomic expression expected at position 1 in Permutations[t2,{2}]. >> Permutations[t2, {2}] Which is the output for an undefined value of t2. I tried to debug the problem: the import operations seems to go fine, each element of t2 is a proper string: Length[t2] 10525 Tally[Table[Head[t2[[i]]], {i, 1, Length[t2]}]] {{String, 10525}} I tried running Trace to understand what is going on, but Mathematica still kept quiet during the first evaluation. The second evaluation again forces all the symbols to be cleared, and the output of Trace is unhelpful: {Permutations[t2,{2}], {Message[Permutations::normal,1,Permutations[t2,{2}]], {MakeBoxes[Permutations::normal: Nonatomic expression expected at position 1 in Permutations[t2,{2}]. >>,StandardForm], RowBox[{StyleBox[RowBox[{Permutations,::,normal}],MessageName],: , "Nonatomic expression expected at position \!\(1\) in \!\(Permutations[\(\(t2, \(\({2}\)\)\)\)]\). \!\(\*ButtonBox[\">>\", ButtonStyle->\"Link\", ButtonFrame->None, ButtonData:>\"paclet:ref/message/General/normal\", ButtonNote -> \"Permutations::normal\"]\)"}]},Null},Permutations[t2,{2}]} Is there something obviously wrong that I am doing? If not, how can I debug this problem further?
[ -0.000043712323531508446, 0.004359683021903038, -0.005173570476472378, 0.015208994038403034, -0.0034356629475951195, 0.001418914645910263, 0.0040655010379850864, 0.0011098558316007257, -0.015178078785538673, 0.010306822136044502, 0.00766243040561676, -0.0035305593628436327, -0.01073711924254...
[ -0.1465800702571869, 0.05791967734694481, 0.2139168083667755, -0.2132561355829239, -0.1382172703742981, 0.29443174600601196, 0.3211548924446106, -0.48361697793006897, -0.28286314010620117, -0.4437556564807892, -0.23723992705345154, -0.04608287662267685, -0.34167248010635376, 0.153860688209...
Is there a way, preferably PowerShell based, to extract all the configuration data for a particular web site, for installing that web site on site at clients?
[ -0.02849956601858139, 0.023684045299887657, 0.01392325945198536, 0.04011514410376549, -0.03698166087269783, 0.03220624476671219, 0.012036141008138657, -0.05763881653547287, -0.01821501925587654, -0.0007465044036507607, 0.009019455872476101, 0.02243484929203987, 0.004124063532799482, 0.0135...
[ 0.3618871569633484, -0.10659118741750717, -0.03377610445022583, 0.5411682724952698, 0.4783068001270294, -0.21580268442630768, -0.26487672328948975, -0.02152223512530327, -0.23375801742076874, -0.3519348204135895, -0.14338648319244385, 0.4486951529979706, 0.01664740964770317, -0.34354045987...
I am doing the 'Battle for Windhelm' war quest; I am sided on the Imperials. I am killing Stormcloaks, but when I do it gives me +1000 bounty in the city. This does not seem how it should be working. What causes this and how can I fix it? I understand there are commands to set my bounty or skip this quest stage, but I'd rather not do either; would rather not get the bounty in the first place. That's how it normally acts.
[ 0.015342329628765583, 0.017566999420523643, -0.0006989396642893553, -0.002677389420568943, -0.030401943251490593, -0.023203382268548012, 0.0056145028211176395, 0.0019827084615826607, -0.018483707681298256, 0.025725403800606728, -0.014989102259278297, 0.024437809363007545, -0.0360294878482818...
[ 0.23786842823028564, 0.11822725087404251, 0.40571483969688416, -0.10265623033046722, -0.6226397752761841, 0.06654604524374008, 0.18918967247009277, -0.31489160656929016, -0.32742762565612793, -0.5945814251899719, 0.1346534639596939, 0.3475860357284546, 0.1533142328262329, 0.100123278796672...
I've seen a couple of descriptions of the basic statistics of a compound Poisson process, basically just simple statements about how to compute the mean and variance given the mean and variance of the underlying processes. Unfortunately I haven't found anything which I feel comfortable citing in a professional context. **Can anybody recommend a good desk reference that would cover compound Poisson processes/distributions?** Preferably it would be general enough to serve as a good desk reference on statistics for an engineer (with a tendency to do a lot of numerical analysis), rather than focusing purely on compound processes. Bonus points if it also gives a good outline of Bayesian inference in general.
[ 0.014274802058935165, 0.012637022882699966, -0.010587665252387524, 0.0424392893910408, -0.007435528561472893, -0.020461902022361755, 0.007901478558778763, -0.0063473996706306934, -0.021722622215747833, 0.004053902812302113, 0.011444643139839172, 0.012724700383841991, 0.004483055789023638, ...
[ 0.3688577115535736, 0.06905195862054825, -0.04377737268805504, 0.3708244264125824, -0.2129751443862915, 0.0073993210680782795, -0.0773693397641182, 0.11719163507223129, -0.2124897986650467, -0.418769896030426, 0.20899665355682373, 0.39053377509117126, 0.22881534695625305, 0.141847655177116...
I want to make this equation fit on one line. I've tried to use `\small`, but it doesn't work. What should I do? \begin{equation} 1^{st}integral= \int_{0}^{\infty}{\frac{2 \pi \zeta}{\alpha} t^{(\frac{-2}{\alpha}-1)}e^{- \pi t^{\frac{-2}{\alpha}} \sum_{i\in K}{\frac{\lambda_i P_i ^{\frac{2}{\alpha}} \rho(D_i\hat{T}_i,\alpha)}{D_i\hat{T}_i^{\frac{2}{\alpha}}}}} e^{- \pi \zeta t^{\frac{-2}{\alpha}}}dt} \nonumber \end{equation}
[ 0.0029080472886562347, 0.021730780601501465, -0.015989070758223534, 0.0010003875941038132, 0.010358190163969994, -0.015757132321596146, 0.007135431282222271, 0.0021707352716475725, -0.015530559234321117, -0.007224017754197121, -0.006965627893805504, 0.002428411738947034, -0.02281531319022178...
[ -0.1437021642923355, 0.14348509907722473, 0.7534707188606262, -0.24551989138126373, 0.27977797389030457, 0.0934419259428978, 0.16954724490642548, -0.6113465428352356, 0.039803627878427505, -0.6275764107704163, 0.07939772307872772, 0.4996473491191864, -0.3261827528476715, 0.3069171011447906...
I was working through the lab on ridge regression and LASSO in ISLR and I came across a strange behavior in the `cv.glmnet` function. When I followed the lab as written I got the following set.seed(1) train <- sample(1:nrow(x), nrow(x)/2) test <- (-train) y.test <- y[test] set.seed(1) cv.out <- cv.glmnet(x[train,], y[train], lambda=grid, alpha=0) plot(cv.out) bestlam <- cv.out$lambda.min bestlam [1] 231.013 For my own benefit I tried it using a different seed (`8675309`) and got back a different result. Any combination of setting the seeds resulted in different answers. I am assuming this has to do with how the 10-folds are changed with the different seeds, however the different `lambda.min` can vary so much I am concerned the package might not be stable. Am I missing something?
[ 0.0038681915029883385, 0.009257972240447998, -0.017081918194890022, 0.01021899376064539, -0.019121820107102394, 0.011092325672507286, 0.00965774804353714, 0.023241911083459854, -0.010432773269712925, 0.006225379183888435, 0.004915551748126745, 0.01177799329161644, -0.0011581761064007878, 0...
[ 0.07799625396728516, -0.2088608592748642, 0.11312820762395859, -0.12958881258964539, 0.24587328732013702, 0.1772249937057495, 0.0880836471915245, -0.47398313879966736, 0.02884923480451107, -0.25677168369293213, 0.38033315539360046, 0.38086315989494324, -0.16667892038822174, 0.4849047362804...
I would like to use a simple bash script (python would be second best) to monitor regularly (hourly) if my mailserver is online and operating. I know that there are dedicated solutions for this task (Nagios, ...) but I really need something simple that I can use as a cronjob. Only to see the mailserver is alive. I know how to talk with a mailserver with telnet, ie: telnet mail.foo.org 25 EHLO example.com mail from: rcpt to: ... but this is interactive. Is it possible to check with a script that the mailserver is communicating? Obviously, I don't want to go the whole way and actually send an email. I just want to test that the mailserver is responding.
[ 0.006682838313281536, 0.0010900995694100857, -0.012223402038216591, 0.0037394780665636063, 0.017119543626904488, -0.011707528494298458, 0.007352486718446016, -0.0031579763162881136, -0.017929688096046448, -0.009564782492816448, -0.012330520898103714, 0.004907014314085245, -0.0131004843860864...
[ 0.47319042682647705, 0.5479510426521301, 0.4477658271789551, -0.027509693056344986, -0.057951200753450394, 0.07126876711845398, 0.43098101019859314, 0.0957862064242363, -0.530427873134613, -0.5744621753692627, 0.20101670920848846, 0.8749241232872009, -0.18234393000602722, 0.032974064350128...
Please consider the following MWE: \documentclass{book} \usepackage[a4paper,twoside,hmargin=3cm]{geometry} \usepackage{tikz,tikzpagenodes} \usetikzlibrary{calc,positioning} \newif\ifoutertikznest \outertikznesttrue % \outertikznestfalse \begin{document} \begin{tikzpicture}[overlay,remember picture,anchor=north west,inner sep=0pt, outer sep=0pt,line width=2pt] \node[draw=red,minimum size=2cm,anchor=north west,inner sep=0pt,outer sep=0pt] at (0,0) (A1) {% \ifoutertikznest % \begin{tikzpicture}[overlay,remember picture,anchor=north west,inner sep=0pt, outer sep=0pt,line width=2pt] % *** \else }; % *** \fi \node[draw=blue,minimum size=2cm,anchor=center,inner sep=0pt,outer sep=0pt] at (current page.center) (A2) {}; \node[draw=green,minimum size=2cm,anchor=north west,inner sep=0pt,outer sep=0pt] at (current page.north west) (A3) {}; \begin{scope}[shift={(0.5,1.0)}] % \node[draw=magenta,minimum size=2cm,anchor=north west,inner sep=0pt,outer sep=0pt] at (0,0) (A4) {}; \begin{scope}[shift={(0.5,1.0)}] % \node[draw=yellow,minimum size=2cm,anchor=north west,inner sep=0pt,outer sep=0pt] at (0,0) (A5) {}; \end{scope} \end{scope} \ifoutertikznest % \end{tikzpicture} % *** }; % *** \fi \end{tikzpicture} \end{document} If you enable the `\outertikznestfalse` line, then there is no additional wrapping in a tikzpicture, and the nested scope shifts look as expected (image left): ![a.png](http://i.stack.imgur.com/Z40vA.png) However, if the rest of the image is "wrapped" under the first node (which requires an extra `tikzpicture`) - and this is the state, if `outertikznesttrue` line is enabled instead - then something is not quite right with the relative shift of the nested `scope`s (image right). While I did expect the extra `tikzpicture` to influence the absolute placement in some way - I didn't expect that it would disturb the relative shifts between the `scope`s. Is this expected behavior? And how could I have the same relative position of the nested scopes (as on image left), even if I have an extra `tikzpicture` wrapper?
[ -0.0037467689253389835, 0.012197558768093586, 0.0019537010230123997, 0.020809264853596687, 0.006062312982976437, 0.018905216827988625, 0.00856664590537548, -0.004335532896220684, -0.01054801233112812, -0.03118051402270794, -0.008244901895523071, 0.00570330023765564, -0.0014518499374389648, ...
[ -0.2715813219547272, 0.009111043997108936, 0.8598592281341553, -0.01566927507519722, 0.23174993693828583, -0.01908719725906849, 0.24424335360527039, -0.40827134251594543, -0.4558286666870117, -0.7273885607719421, 0.07715147733688354, 0.42085739970207214, -0.17502188682556152, 0.01727393642...