text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
What are **the main purposes** to ask someone 'Where do you work?' apart from to find out the type of place he or she works in? I can answer: > I work in a shop. > I work in a hospital. > I work in an office. But what other things can (or should) I use as a reply to this question? * * * ### Extra question: sometimes I see > I work **at** a shop. I work **at** a hospital. I work **at** an office. What's the difference between 'at' and 'in' in this case? | [
0.019695110619068146,
0.007389248348772526,
-0.008008386939764023,
0.011459295637905598,
-0.017822302877902985,
0.029486695304512978,
0.007092618383467197,
0.006581071764230728,
-0.01346707995980978,
-0.006169172003865242,
-0.004014115314930677,
-0.004409450106322765,
0.015740826725959778,
... | [
0.9824830293655396,
0.22700801491737366,
0.23590940237045288,
0.03179054334759712,
-0.061660997569561005,
0.5542812347412109,
0.16168798506259918,
0.13493581116199493,
-0.543279767036438,
-0.3881101608276367,
-0.06460913270711899,
0.4124487042427063,
-0.019191255792975426,
-0.0841943621635... |
If you're using beamer you can do some nice text boxes using: \begin{block}{block title}content of block\end{block} Which gives a text box something like this (the box with the header Esimerkki):  How can I do the same thing in a regular latex document (I really like the look of the beamer text boxes and would like as close to this as possible). Minimal reproducible example: \documentclass{scrartcl} \begin{document} I wish I were inside of a really super cool text box so all the word would stop and look at me and say wow that text is pretty. But alas I am just ordinary \LaTeX{} text. I am sad. \end{document} | [
0.023840513080358505,
0.005554481875151396,
0.0017627824563533068,
0.011638467200100422,
-0.008232759311795235,
0.008382799103856087,
0.006646397989243269,
0.013318756595253944,
-0.013188390992581844,
0.018882140517234802,
-0.015434997156262398,
-0.0007003573700785637,
0.013390541076660156,
... | [
0.41085970401763916,
-0.05118865892291069,
0.45458635687828064,
0.0362485907971859,
-0.26516613364219666,
-0.21025438606739044,
0.3756116032600403,
-0.13832618296146393,
0.03616424649953842,
-0.6631043553352356,
-0.06374556571245193,
0.5548533201217651,
-0.2907509505748749,
-0.028607763350... |
> In solving a problem of this sort, the grand thing is to be able to reason > backwards. That is a very useful accomplishment, and a very easy one, but > people do not practise it much. In the every-day affairs of life it is more > useful to reason forwards, and so the other comes to be neglected. **There > are fifty who can reason synthetically for one who can reason > analytically.** What does the last sentence mean? 1. "Out of 51 people, only one can reason analytically." 2. "A person who reason analytically is a more valuable asset than 50 people who reason synthetically." | [
0.02360374853014946,
0.020538877695798874,
-0.025966254994273186,
0.014816321432590485,
-0.02191765233874321,
-0.015541737899184227,
0.008446617051959038,
-0.0022569322027266026,
-0.01185116358101368,
-0.008609653450548649,
0.0018050898797810078,
0.006344879046082497,
-0.002271583303809166,
... | [
-0.2042974829673767,
0.17615099251270294,
0.0016098490450531244,
0.09899838268756866,
-0.3821558356285095,
0.5423792600631714,
0.46502378582954407,
-0.2851772606372833,
-0.3308548629283905,
-0.8116989135742188,
-0.16950105130672455,
0.45190104842185974,
-0.21711771190166473,
-0.17951238155... |
I have recently installed DD-WRT on my D-Link DIR-615 router (rev. D3). Everything seems fine but I cannot connect to some (most) FTP servers anymore. Any ideas what can cause this issue? -- Passive/active doesn't make a difference. I've discovered something interesting. I've set up a port range forwarding `1024-65535 | Both | 192.168.1.131` for the testing purposes. After that I have enabled or disabled (it doesn't really matters it seems) the UPnP service and it let me connect to FTP but just for a few seconds. | [
-0.014544827863574028,
0.006335795857012272,
-0.014893014915287495,
0.0007091551087796688,
-0.008661116473376751,
-0.013351961970329285,
0.006818010471761227,
0.003525426145642996,
-0.011256947182118893,
-0.013246007263660431,
-0.012817925773561,
0.012297431007027626,
-0.008479283191263676,
... | [
0.2692694664001465,
0.10620290040969849,
0.6175511479377747,
-0.04046986624598503,
-0.1157195195555687,
-0.12192831933498383,
0.5151650309562683,
0.2383464276790619,
-0.36306098103523254,
-0.8490586280822754,
-0.22408662736415863,
0.769910991191864,
-0.22191031277179718,
0.2924432158470154... |
I'd like to ask how can I call the page `www.domain.com/index.php?page=somepage` and make the browser show the user in the URL bar the address `www.domain.com/somepage/`? How do I do that trick? | [
-0.018856119364500046,
-0.006255453918129206,
0.0021619878243654966,
-0.0006911177770234644,
-0.028520086780190468,
-0.010310919024050236,
0.011273767799139023,
0.0017803203081712127,
-0.03353722766041756,
-0.005515603348612785,
-0.015254293568432331,
0.0015892584342509508,
0.005988384131342... | [
0.6148218512535095,
0.16415096819400787,
0.5260897874832153,
0.21323539316654205,
0.06781356781721115,
-0.1554431915283203,
0.20493625104427338,
0.0988820493221283,
-0.06968057155609131,
-0.2851056158542633,
0.2852676808834076,
0.1865462362766266,
-0.03625556454062462,
0.5112189650535583,
... |
Is there a single word for "to paint and/or draw"? If you talk to artists frequently, you find you want one. _Depict_ , _portray_ and similar words work in a sentence like "The artist painted/drew a horse", but I'm looking for something that works in a sentence like "Did you paint and/or draw today?" | [
0.012340323068201542,
0.013814902864396572,
-0.008817077614367008,
0.015571590512990952,
-0.019983289763331413,
-0.024487914517521858,
0.009643957950174809,
0.010456821881234646,
-0.026789424940943718,
0.008790159597992897,
-0.007144732866436243,
0.014150501228868961,
0.01969556137919426,
... | [
0.880834698677063,
-0.10599442571401596,
-0.4378880560398102,
0.2836821675300598,
-0.3023541271686554,
0.38641276955604553,
-0.07838832587003708,
0.2907554805278778,
-0.5616523623466492,
-0.72104412317276,
0.33966898918151855,
0.3805922269821167,
0.06573686003684998,
-0.1995779275894165,
... |
I have a lot of xfce-notes groups containing each a lot of text. I want to transfert them to a new machine without copying the whole home directory. Where does xfce4-notes store its data and configation files? | [
0.011462009511888027,
0.015644723549485207,
-0.0004107120621483773,
0.011054332368075848,
-0.02781900018453598,
-0.017476432025432587,
0.010989709757268429,
0.01980101875960827,
-0.031187638640403748,
-0.0096346540376544,
-0.0011745551601052284,
0.018081124871969223,
-0.000220522764720954,
... | [
0.23908576369285583,
0.3763023018836975,
0.5386956930160522,
0.34543558955192566,
0.20546698570251465,
-0.1420239955186844,
-0.036663006991147995,
-0.10187430679798126,
-0.10030022263526917,
-0.5447726845741272,
0.07579714804887772,
-0.044426329433918,
-0.18782681226730347,
0.1796261966228... |
_imagine_ given a `project` business object and such (simplified) rules: * it's lifecycle is divided into several `stages of evaluation`; * stages flow lineary and represent an evaluation chain; * each stage provides its own `reward value`/algorithm; * promotion of the project is controlled by other user's decision; * the resulting reward is assigned to the project's initiator. _too naive, probably)_ The first thought that comes to my mind is using the `decorator pattern`; because of its structure it looks somewhat applicable. But what if you need to persist additional details provided with the current 'decorated' `state` of project? _I need an extra behavior on each stage_ I've encountered an article about `jBPM`. I believe it has much in common with `WF`. It is surely has the most of what is needed and at the same time it has (overkill) large infrastructure. _but can it be designed without incorporating this much complexity?_ What would you suggest? | [
-0.00039242731872946024,
0.023699766024947166,
-0.0008357870974577963,
0.02652411349117756,
0.003570009721443057,
0.0015248835552483797,
0.0068595511838793755,
0.023399334400892258,
-0.013328161090612411,
0.0073069422505795956,
-0.020529214292764664,
0.01490308903157711,
-0.00789221376180648... | [
0.32299020886421204,
0.2329586148262024,
0.05372059345245361,
-0.007594532798975706,
0.29236921668052673,
0.1871539205312729,
0.5520921349525452,
-0.013870607130229473,
-0.3897963762283325,
-0.5957146883010864,
0.07611565291881561,
0.3236357271671295,
-0.36019647121429443,
0.38036811351776... |
Are you a member of a trade union? Why? Why not? If you are, and don't mind mentioning it, which one? Do you know of any programmers who were helped by being in a union, or would have been helped by being in a union? Do you know of any programmers who were hindered or would have been hindered by being in a union? | [
0.004942958243191242,
0.01711125299334526,
0.0003362719726283103,
0.016547052189707756,
0.01054364163428545,
0.04035341739654541,
0.011053668335080147,
0.015790145844221115,
-0.021878136321902275,
-0.01670016534626484,
-0.003074030624702573,
0.016028838232159615,
0.025012195110321045,
0.01... | [
0.9194577932357788,
-0.08589630573987961,
-0.4807293117046356,
0.2584094703197479,
-0.0783231183886528,
-0.15202945470809937,
-0.10020117461681366,
0.43804940581321716,
-0.7495724558830261,
-0.04046100005507469,
0.039148103445768356,
0.2906782031059265,
-0.059464745223522186,
-0.0520306229... |
When using `listings` and `breqn`, the minus signs (or dashes, according to the unix.SX site then there's some controversy about which it is) disappear from code listing. Here's an example: \documentclass{article} \pagestyle{empty} \usepackage{breqn} \usepackage{listings} \begin{document} \lstset{language=Perl} \begin{lstlisting} #! /usr/bin/perl -w if ($ARGV[0] =~ /^-/) { print "Option given"; } \end{lstlisting} \end{document} with result:  Changing the order of package loading doesn't help. `breqn` does warn that it might break other packages, but it would be really useful to have both working. If it helps, the code listings are in appendices so happen after the equations which `breqn` is meant to help with so I can happily reset anything that got changed. | [
0.014216138981282711,
0.00010252255015075207,
-0.010608009994029999,
0.02468564733862877,
-0.0017069559544324875,
-0.0034482204355299473,
0.008827132172882557,
0.01957016810774803,
-0.01281372457742691,
0.005809489171952009,
-0.009144197218120098,
0.008416098542511463,
-0.017060033977031708,... | [
0.26409071683883667,
0.17282377183437347,
0.3852311372756958,
-0.4362885653972626,
0.021471574902534485,
-0.15983746945858002,
0.09626160562038422,
-0.09200990945100784,
-0.30447107553482056,
-0.45222097635269165,
-0.2990055978298187,
0.5298501253128052,
-0.19066567718982697,
0.14829869568... |
I'm developing a very media-centered site with WordPress and am using the default categories and tags on uploaded media using, in part this plugin Media Categories. I am displaying a sub-menu item under 'Media' for each 'Categories' and 'Tags' successfully using: add_media_page( 'Tags', 'Tags', 'edit_posts' , 'edit-tags.php?taxonomy=post_tag'); It is linking through properly, but it seems somewhere in WordPress this page is assigned as a child of the default Posts type, as when I click it, it opens the 'Posts' menu and displays the Tags page, rather than keeping the 'Media' menu opened and displaying it from there. Is there any way to keep the parent page open (i.e. Media)? This isn't an essential fix, but it would be good for the UI and continuity to have the Media menu stay open, rather than swapping to the Posts menu. I can provide screenshots if this is confusing. Thanks! **EDIT** I've just discovered a cheeky little definition in edit-tags.php: `$parent_file = 'edit.php';` It seems each file corresponding to a submenu item in the WP Admin Menu has this $parent_file set. Now obviously, I can simply change this to 'upload.php' to refer to the Media parent, but I'm wondering if there's a non-hacky way that I can modify this value from functions.php in my theme file? | [
0.0026835831813514233,
0.008039042353630066,
-0.006561265327036381,
0.024481412023305893,
0.005782475695014,
-0.01171056367456913,
0.008047051727771759,
0.008077362552285194,
-0.010813148692250252,
-0.0053438651375472546,
-0.012521571479737759,
0.0088573656976223,
-0.008795488625764847,
0.... | [
0.3880370259284973,
0.2351136952638626,
0.8224334716796875,
-0.08371899276971817,
-0.04755835607647896,
-0.13445910811424255,
0.042591530829668045,
0.0531327910721302,
-0.15763235092163086,
-1.0075644254684448,
-0.015434389002621174,
0.4551539123058319,
-0.1427907943725586,
0.2176377773284... |
I am developing a system that is designed for multiple forms of interfacing. There is a website, but that is connected through an SDK, as well as an HTTP query interface to access data. But to improve speed and quality, I was thinking of creating a system inside IIS that get any message sent to the server, any response, but still let IIS manage SSL and normal socket connections. Is there a way to host my project in IIS without ASP or any other kind of script with extra behavioral events? | [
0.004599050618708134,
0.00968192983418703,
-0.003304023528471589,
-0.0008297875174321234,
-0.008604506030678749,
0.0022170140873640776,
0.006563121918588877,
0.009287850931286812,
-0.015838073566555977,
-0.010470770299434662,
0.0020597330294549465,
0.017246117815375328,
-0.006120817270129919... | [
0.3510657548904419,
0.08804009109735489,
0.010488880798220634,
0.06498703360557556,
0.003047434613108635,
-0.16193591058254242,
0.15737736225128174,
-0.05982811003923416,
-0.21649572253227234,
-0.6434935927391052,
0.019371429458260536,
0.2259030044078827,
-0.20895621180534363,
0.2414906024... |
I've been reading about the star schema (or dimensional) database structure, which puts all measurements main in 'facts' table, and all context for those measurements in 'dimension' table linked to the facts table (I'm doing a horrible job explaining this), so that you can query any measurement by conditioning on the dimension. When I read about this, it struck me as a perfect way to aggregate different csv files with different columns into one store... for example, stock, commodities, and options data in various files go be shoved into a star-schema db with a facts table containing the open/close/high/low/volume, with foreign- key links to dimensions such as time, company, exchange, and 'file pedigree' so that we know where each entry comes from. That would allow you to query for, say, first quarter chinese mining companies with a single, fairly simple sql statement. However, all of the literature on this topic seems to be for large scale 'enterprise' deployments and not small data analysis or research jobs. Does anyone have any experience organizing their data this way? Is there are simpler way? | [
0.00012770527973771095,
0.00626649335026741,
0.003410770557820797,
0.010573143139481544,
0.0041230833157896996,
0.014548752456903458,
0.007518736645579338,
-0.005410307087004185,
-0.014412272721529007,
-0.005599215626716614,
-0.002338060410693288,
0.008409149944782257,
0.011875161901116371,
... | [
0.4335828423500061,
0.2679246962070465,
-0.1366080343723297,
0.24763347208499908,
0.05239060893654823,
-0.11844007670879364,
-0.4151354432106018,
0.053734906017780304,
-0.2108767330646515,
-0.31470155715942383,
0.23072193562984467,
0.3202320635318756,
-0.2827201783657074,
0.752450764179229... |
Assuming I have two SDEs or other datasources with the same datamodel. One contains features the other one is empty. In ArcMap I have layers from both datasources in my table of contents. Now I would like to copy some features that are already selected in the map view from one sde to the other using C# and ArcObjects. What would be the best way to do this? | [
0.0026109954342246056,
0.028271086513996124,
-0.008112379349768162,
0.016244089230895042,
-0.005209364928305149,
-0.02405400574207306,
0.011444210074841976,
0.017224367707967758,
-0.018381288275122643,
-0.0376521572470665,
-0.0017154272645711899,
0.029308363795280457,
-0.021762408316135406,
... | [
0.34346523880958557,
-0.19479456543922424,
0.12666477262973785,
0.21675333380699158,
-0.16898751258850098,
0.01515075284987688,
-0.15262512862682343,
-0.10157311707735062,
0.0700245201587677,
-1.2815415859222412,
0.196998730301857,
0.735504150390625,
0.015985047444701195,
-0.07865064591169... |
What do I have to write in the preamble to bring the bibliography to make parentheses around the volume number of an article? I'm using the `authoryear` style. Here is an example: \begin{filecontents}{filename.bib} @article{Billio.2012, author = {Billio, Monica and Getmansky, Mila and Lo, Andrew W. and Pelizzon, Loriana}, year = {2012}, title = {Econometric measures of connectedness and systemic risk in the finance and insurance sectors}, pages = {535--559}, volume = {104}, number = {3}, journal = {Journal of Financial Economics} } \end{filecontents} \documentclass{article} \usepackage[hyperref=true, maxcitenames=4, isbn=false, dashed=false, style=authoryear, backend=bibtex, firstinits=true]{biblatex} \bibliography{filename.bib} \begin{document} \cite{Billio.2012} \newpage \renewcommand\refname{List of Literature} \printbibliography[heading=bibintoc] \end{document} Any help is appreciated. | [
0.018414780497550964,
0.006207410246133804,
-0.0014927167212590575,
0.019816480576992035,
0.0014243237674236298,
-0.001276333467103541,
0.008190004155039787,
-0.012053610756993294,
-0.014025606215000153,
-0.011316225863993168,
0.0016404251800850034,
0.005619030445814133,
-0.00930112600326538... | [
-0.0034947823733091354,
0.17741534113883972,
0.1668752282857895,
0.15071238577365875,
0.05985753610730171,
-0.156100794672966,
-0.04264667257666588,
-0.18666577339172363,
-0.45471933484077454,
0.15090499818325043,
0.03239087760448456,
0.380868524312973,
0.4517558217048645,
-0.0497785918414... |
I have the following problem while playing some (Bethesda) games, if i set the game to fullscreen (non-windowed). In the upper right corner of the screen, it looks like there is the regular " windows-window" border quickly blinking. Its's always the upper right corner (the three buttons your normally see in a window plus some extra empty space) and it annoys me to no end. It looks as if the game is trying to switch between fullscreen and windowed mode. I have only encountered the problem with bethesda games so far. I find it very hard to explain what exactly happens, but I hope some of you have encountered the same and know a solution * * * The card I am using is a Sapphire hd 7850 As far as I know, I am running the latest drivers for this card. I do have some Rainmeter widgets on my desktop | [
-0.008422735147178173,
-0.002458228264003992,
-0.004639340564608574,
0.0011614890536293387,
-0.020760636776685715,
-0.02755379118025303,
0.008303298614919186,
0.007488253526389599,
-0.010789105668663979,
0.037181731313467026,
-0.021489188075065613,
0.013104395940899849,
-0.001224042032845318... | [
-0.1242327094078064,
0.13681147992610931,
0.30429232120513916,
-0.05610692501068115,
0.10440357774496078,
-0.197721466422081,
0.21770665049552917,
0.5328221917152405,
-0.5219743251800537,
-0.6841266751289368,
0.06521573662757874,
0.6846131682395935,
-0.06574337184429169,
0.3914388716220855... |
Where does the phrase "egg on my face" come from, and what is its meaning? | [
-0.017703263089060783,
0.04486415907740593,
-0.0037461791653186083,
0.023375043645501137,
-0.05626317858695984,
0.07598337531089783,
0.016864517703652382,
0.018566584214568138,
-0.018543817102909088,
-0.0003509145462885499,
-0.03759206086397171,
0.04647532105445862,
0.0077680847607553005,
... | [
0.5975940227508545,
0.7686212658882141,
0.16230426728725433,
-0.24676291644573212,
0.046619564294815063,
0.5269690752029419,
0.008947530761361122,
0.7225271463394165,
-0.30331459641456604,
-0.08893629908561707,
0.20147791504859924,
0.1250319927930832,
0.47718361020088196,
0.168729335069656... |
When you start a download on Xbox Live from your console it will automatically download to the HDD(Hard drive). If there is no space on the drive, it allows you to choose a different storage device, providing there is space. So, where's the option to download to USB from the beginning? Basically, I have a game on my USB. Inserting the USB brings up the game as a trial. So, I have to re-download the game from the marketplace. Although, instead of ipdating the USB, it will download to the HDD. If I place another download ahead of it, it should give me an option to download to a different device if there's space. Or if there isn't enough space on the HDD it will give me an option. So, there isn't an option to select a device first. Meaning, I have to clear space on the HDD, copy to the HDD from the USB, click download again. It will go from 1%-100% immediatly. Copy back to USB. | [
-0.017173830419778824,
-0.007670542225241661,
0.0013894850853830576,
0.004991226829588413,
0.02022145316004753,
-0.02605361118912697,
0.008634889498353004,
-0.023842722177505493,
-0.01578298956155777,
-0.009443651884794235,
-0.00022852059919387102,
0.005998542532324791,
0.008719722740352154,... | [
0.44142377376556396,
-0.14318659901618958,
0.39150357246398926,
0.36018243432044983,
-0.05197927728295326,
-0.18012136220932007,
-0.2403583526611328,
0.17260095477104187,
-0.07415451854467392,
-0.667189359664917,
0.09619062393903732,
1.0576400756835938,
0.06905889511108398,
0.2007245868444... |
I am authoring a WP plugin, and I have to generate some JS (jQuery) code in "realtime". It is working fine in general, though I've stuck with one problem with the WP 3.5.1 and its default Twenty Twelve theme: there the scripts are inserted in the footer of the page, not in the header, causing my code to throw errors. I've looked through all the theme's files (header, footer, functions), and I can't seem to find the reason for that. Looks like the `wp_footer()` is inserting the block, but I can't find any hook or anything else that tells WordPress to insert scripts there. So, my question is: how to make jQuery and other scripts go in the header by default, not in the footer? What does make the theme work this way (maybe the default WP behavior changed?) | [
-0.005395529791712761,
0.0024520722217857838,
-0.006691369228065014,
0.02062823250889778,
-0.00042794598266482353,
0.010662621818482876,
0.007166681345552206,
0.008849658071994781,
-0.011558995582163334,
-0.024279922246932983,
-0.01748069003224373,
0.005936558358371258,
-0.003279234748333692... | [
0.5243975520133972,
0.0025632281322032213,
0.2135608196258545,
-0.05888359621167183,
-0.016689712181687355,
-0.3188250660896301,
0.23901861906051636,
-0.07445672154426575,
-0.09763072431087494,
-0.5596698522567749,
-0.04738582298159599,
0.522336483001709,
-0.17373375594615936,
-0.116501204... |
totally new to LaTeX, this question occured today: I have a table made of 17 columns. Every column is 0,6cm wide. \begin{tabular}{p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0.6cm}|p{0,6cm}} That works perfectly and fine. But i also have on several rows 2 cells which are combined with \multicolumn like this: \multicolumn{2}{|p{1.2cm}|}{\cellcolor{grey}Inverse dependency ratio} Because this multi column combines two differenct cells i figured out 0,6 + 0,6 = 1,2. Right? So i gave this a width of 1,2cm... However this happens:  The combined \multicolumn doesnt fit to the 2 columns at the bottom of it. What did i do wrong? | [
0.008795476518571377,
0.013947099447250366,
-0.004481861833482981,
0.009441863745450974,
0.0003119629109278321,
0.009909311309456825,
0.0029099714010953903,
-0.007743955589830875,
-0.008913209661841393,
-0.009542266838252544,
-0.0033411425538361073,
-0.007387477904558182,
-0.0063527100719511... | [
-0.18871070444583893,
0.2195272147655487,
0.6702460646629333,
-0.18209011852741241,
-0.05078957974910736,
0.2853047251701355,
0.09341099858283997,
-0.5388984680175781,
-0.333649605512619,
-0.31072941422462463,
-0.20851239562034607,
0.12885309755802155,
-0.2040041834115982,
-0.1616795212030... |
I've searched a lot but didn't find an answer ... My Wordpress has got 4 main areas, each one also has a news (posts) section and an own submenu. If I load a post (with single.php) is it possible to get the corresponding submenu to that area? I'm thinking about sub stringing the slug to get the right area. Help is really appreciated. Thx, Oliver Additional Info: * There is a sidebar which I also show in single.php * The sidebar should contain the menu tree of one area * A post is not connected to the menu | [
0.011875379830598831,
0.020300880074501038,
0.013957059942185879,
0.01871553435921669,
-0.009469980373978615,
0.007262644357979298,
0.007541991770267487,
0.03051256202161312,
-0.020114019513130188,
-0.015053393319249153,
-0.014252156019210815,
0.018489152193069458,
-0.005227166227996349,
0... | [
0.5885277390480042,
0.07546501606702805,
0.3452858030796051,
0.32281437516212463,
-0.4335958659648895,
0.12693670392036438,
0.6897372007369995,
0.39726102352142334,
-0.3996872305870056,
-0.5239654183387756,
-0.0956953689455986,
-0.07183805108070374,
0.0011347088729962707,
0.168435811996459... |
Stephen Hawking says in his latest book _The Grand Design_ that, > Because there is a law such as gravity, the universe can and will create > itself from nothing. Is it not circular logic? I mean, how can gravity exist if there is no universe? And if there is no gravity, how can it be the reason for the creation of universe? Also, if the universe doesn't exist, how can _it_ create _itself_? The very sentence doesn't make sense to me. It seems so absurd and illogical that I've never heard such sentences even in philosophy. On what ground, Stephen Hawkings _claims_ this? | [
-0.009876634925603867,
0.014317534863948822,
0.008152706548571587,
-0.0002561560831964016,
-0.009676429443061352,
-0.0032506342977285385,
0.00891108438372612,
-0.005407444201409817,
-0.009324440732598305,
-0.006450118962675333,
-0.013730975799262524,
0.0025067266542464495,
-0.000936737051233... | [
0.5155821442604065,
0.33222219347953796,
-0.18008562922477722,
0.07253279536962509,
-0.05858701094985008,
0.028405578806996346,
0.05256989225745201,
0.06497681885957718,
-0.18040482699871063,
-0.2782011330127716,
-0.31483766436576843,
0.35560619831085205,
-0.0930568277835846,
0.76896893978... |
Could you recommend an R package for estimating a (frequentist) multilevel Weibull regression model? I need to model random intercepts, random slopes, as well as a cross- classified structure. **UPDATE:** It seems like there is currently no "easy" solution for that. I decided to leave it for now by estimating a multilevel discrete hazard model with `glmer` and a multilevel Cox PH model with `coxme` (proposed by EddieMcGoldrick). With regards to the latter, I have still to figure out if implementing a cross-classified structure is possible. | [
0.010600496083498001,
0.013503550551831722,
-0.015524821355938911,
0.007962736301124096,
-0.008440334349870682,
-0.009366612881422043,
0.008584825322031975,
0.016593337059020996,
-0.015373805537819862,
0.000738183269277215,
0.0025094784796237946,
0.009247638285160065,
0.0007076971232891083,
... | [
-0.11027166247367859,
-0.02145085670053959,
-0.007770768832415342,
-0.06886471062898636,
-0.09480114281177521,
0.06639126688241959,
-0.05584125220775604,
-0.2744329273700714,
-0.2915576696395874,
-0.5716062784194946,
-0.010670267045497894,
0.3481498062610626,
-0.34229135513305664,
0.228960... |
For example I have "HP 1020 LaserJet" local USB printer successfully installed on a CUPS. It uses one connection. If I get another HP 1020 LaserJet printer it wouldn't print, I am having to modify the printer and change it's connection. Why? How can I avoid this? I know that it's illogical to use the same type printer on the same computer, but that's my environment. How to make CUPS to use the same Connection for all printers of the same type, model, manufacturer etc.? Thank you!!! **EDIT:** I have found out that it's not possible through the configs or any other standard ways. The only way is to find a nice workaround. | [
-0.0024468842893838882,
-0.007005690131336451,
-0.004964034538716078,
0.02664305455982685,
0.002162486780434847,
-0.01659286953508854,
0.009799369610846043,
0.005997400730848312,
-0.01690736785531044,
-0.055489324033260345,
0.0010557151399552822,
0.005354725755751133,
0.0052881487645208836,
... | [
0.38256680965423584,
0.29284045100212097,
0.28468239307403564,
0.24226319789886475,
0.1940673142671585,
-0.08708080649375916,
-0.229371577501297,
-0.10502233356237411,
-0.5026988387107849,
-0.9940474033355713,
0.2536446750164032,
0.4084055721759796,
-0.4273246228694916,
0.11271628737449646... |
When a form is run through an approval process, I want to create statuses for different stages of said approval. For now, I have the following: * Requested (Initial request was submitted) * Approved formally (Request has been checked formally, meaning that required data is present and it doesn't contain any formal errors, such as wrong date format, too few characters, etc.) * Approved ...? * Approved financially (The request can be paid for) * Resolved (Request has been processed) * Closed (Request has been shipped) What I am missing is a word for _the requests content has been checked it makes sense, is correct in the domain and can be implemented/resolved_. I was thinking about "Approved contextually" but that doesn't quite sound as what I want. | [
-0.00925388652831316,
0.01399527583271265,
0.004565896466374397,
0.023923855274915695,
-0.003945514559745789,
0.004893068689852953,
0.008805854246020317,
0.013867340981960297,
-0.013289466500282288,
0.018952472135424614,
-0.010094335302710533,
0.003176681697368622,
0.0059156641364097595,
0... | [
0.3690313994884491,
0.10047633200883865,
0.579586386680603,
0.1858402043581009,
0.17518475651741028,
-0.4306807219982147,
-0.015365558676421642,
-0.5089412927627563,
-0.3740275204181671,
-0.3459068536758423,
-0.09318671375513077,
0.06234792247414589,
0.12242511659860611,
0.0153729645535349... |
We do a lot of mass emailing of our contacts to promote events, send out newsletters, etc. Some people read and react, some people unsubscribe, but I fear that some might actually mark the email as spam. Is there any way to figure out whether my domain has been added to email blacklists or spam registries? Also, if I use a service like MailChimp to send the emails, how would this work? If one unscrupulous customer was using MailChimp for evil, wouldn't it affect all of their customers? | [
-0.0010842353804036975,
0.0032128882594406605,
0.012205826118588448,
0.02246681973338127,
0.015067240223288536,
0.002274727215990424,
0.007201827596873045,
0.013044854626059532,
-0.016575342044234276,
-0.006723830942064524,
-0.01155065931379795,
0.017741018906235695,
-0.009698284789919853,
... | [
1.115466594696045,
0.27878472208976746,
0.3240249752998352,
0.030088113620877266,
-0.10462172329425812,
-0.3825359642505646,
0.5861868858337402,
0.21763451397418976,
-0.09627851098775864,
-0.06624172627925873,
0.24403604865074158,
0.3875502645969391,
-0.291974276304245,
0.2746705412864685,... |
I am looking for a closed form for the below integral but since I don't have the necessary backgrounds I am not able to solve it: i know the final solution is in the form of modified Bessel functions and polynomials and also I know it might be related to moments of multivariate inverse gaussian distributions: here is the integral assuming z and y are vectors. and $\lambda$ and $\Gamma$ are matrices. $\Gamma$ can be diagonal though. $U$ is a diagonal matrix containing $\sqrt z_i$ as its diagonal elements. $$ U = \begin{bmatrix} \sqrt{z_1} & \cdots & 0 \\\ \vdots & \ddots & \vdots \\\ 0 & \cdots & \sqrt{z_d} \end{bmatrix}$$ $$ \lambda = \begin{bmatrix} \lambda_{11} & \cdots & \lambda_{1d} \\\ \vdots & \ddots & \vdots \\\ \lambda_{d1} & \cdots & \lambda_{dd} \end{bmatrix}$$ $$ \Gamma= \begin{bmatrix} \Gamma_{1} & \cdots & 0 \\\ \vdots & \ddots & \vdots \\\ 0 & \cdots & \Gamma_{d} \end{bmatrix}$$ $$ \int^{\infty}_{0} \frac{1}{\sqrt{z_1 ... z_d}} \exp^{(-\frac{1}{2} y^T(U \Gamma U^T)^{-1} y)} \exp^{(-\sqrt{z}^T\lambda\sqrt{z})} \ \ \mathrm{d}z_1...\mathrm{d}z_d$$ thank you | [
-0.019341792911291122,
-0.004487785045057535,
-0.0005887430161237717,
0.01654093898832798,
0.005494584795087576,
0.00252919876947999,
0.006988192908465862,
0.0034385905601084232,
-0.016256067901849747,
-0.007133422419428825,
-0.009130227379500866,
0.007537201978266239,
-0.01061210222542286,
... | [
-0.18992012739181519,
-0.3670729100704193,
0.4766049087047577,
-0.24379202723503113,
-0.2550341486930847,
0.07085860520601273,
-0.14027942717075348,
-0.5436710715293884,
0.30201560258865356,
-0.46629565954208374,
-0.05980921909213066,
0.4448702335357666,
-0.40750324726104736,
0.31365701556... |
What's the recommended way of installing python packages on Arch? Searching for them on the AUR and installing them from there (or create a `PKGBUILD` file to make a package yourself) or using `pip`? I started off by installing stuff from pacman and the AUR and dont know if it would be wise to mix with `pip` packages. | [
-0.011148597113788128,
0.013645944185554981,
0.004489504732191563,
0.01440211571753025,
-0.03497318550944328,
0.006386448163539171,
0.01278704684227705,
-0.019312499091029167,
-0.03245016932487488,
-0.015529623255133629,
-0.0030120403971523046,
-0.004327654372900724,
-0.02266431599855423,
... | [
0.2875517010688782,
0.11816354840993881,
-0.1181408241391182,
-0.09449871629476547,
-0.1308916062116623,
-0.008786329999566078,
0.3934919834136963,
0.2431110143661499,
-0.10347960889339447,
-0.39859774708747864,
0.1118786558508873,
0.5971993803977966,
-0.4421265423297882,
-0.19475701451301... |
As a followup to my question "How to find out if an wp-admin action edited a file?" I now could use a list of actions and files that can actually cause an update or change to .php-files in a default Wordpress installation on the file system. Right now I think of: * Adding themes * Editing themes * Adding plugins * Updating plugins * Updating core Did I miss something? | [
-0.014973742887377739,
0.009841915220022202,
0.004528306890279055,
0.023929249495267868,
-0.006903802510350943,
-0.012222809717059135,
0.008251680061221123,
0.0012089133961126208,
-0.016990214586257935,
0.0025286602322012186,
-0.01027579978108406,
0.009760540910065174,
0.01025257259607315,
... | [
0.48243027925491333,
0.08566495031118393,
0.3485499322414398,
0.17795099318027496,
-0.2425268143415451,
-0.23617413640022278,
0.328952819108963,
-0.05286794900894165,
-0.6026334166526794,
-0.4289494752883911,
0.21712066233158112,
0.34620222449302673,
-0.27570122480392456,
0.174905017018318... |
Maybe this is a silly question, the answer is probably no; but I gotta ask. For example, if I have an Ubuntu Live USB running on a machine with Ubuntu already installed, might the Live OS ever look for any kind of useful configuration info on the OS already installed on the hard drive without asking me first? Does any *nix live distro have this kind of behavior? | [
0.01215613167732954,
0.014229056425392628,
0.005760709289461374,
0.03291894122958183,
-0.013636560179293156,
-0.0029927834402769804,
0.010770034044981003,
-0.01432295423001051,
-0.026037227362394333,
-0.02384931966662407,
-0.0027985458727926016,
0.010247060097754002,
0.01980813778936863,
0... | [
0.7374196648597717,
-0.02616759017109871,
0.06873776763677597,
0.2478865683078766,
-0.1868952065706253,
-0.3866572976112366,
0.24885118007659912,
0.546878457069397,
-0.2905796766281128,
-0.3053690791130066,
0.2158021181821823,
0.3219285011291504,
-0.44336897134780884,
0.25054094195365906,
... |
Using TeX/LaTeX is it possible to embed multimedia? If so how? | [
0.008800750598311424,
0.021990511566400528,
-0.036098379641771317,
0.02655918337404728,
0.05886152386665344,
0.02102787047624588,
0.02035572938621044,
0.0023349651601165533,
-0.02358751930296421,
-0.08719177544116974,
-0.029439426958560944,
0.029218116775155067,
0.0004341699823271483,
0.05... | [
0.27005642652511597,
0.11032329499721527,
0.29410937428474426,
0.3072713017463684,
0.1267697662115097,
-0.11518673598766327,
-0.33080369234085083,
0.10063639283180237,
0.10985224694013596,
-0.556576132774353,
-0.16113440692424774,
0.7224515080451965,
-0.3821640908718109,
-0.212355256080627... |
I've two configuration files, the original from the package manager and a customized one modified by myself. I've added some comments to describe behavior. How can I run `diff` on the configuration files, skipping the comments? A commented line is defined by: * optional leading whitespace (tabs and spaces) * hash sign (`#`) * anything other character The (simplest) regular expression skipping the first requirement would be `#.*`. I tried the `\--ignore-matching-lines=RE` (`-I RE`) option of GNU diff 3.0, but I couldn't get it working with that RE. I also tried `.*#.*` and `.*\\#.*` without luck. Literally putting the line (`Port 631`) as `RE` does not match anything, neither does it help to put the RE between slashes. As suggested in “diff” tool's flavor of regex seems lacking?, I tried `grep -G`: grep -G '#.*' file This seems to match the comments, but it does not work for `diff -I '#.*' file1 file2`. So, how should this option be used? How can I make `diff` skip certain lines (in my case, comments)? Please do not suggest `grep`ing the file and comparing the temporary files. | [
-0.006312167271971703,
0.010869849473237991,
-0.007595926057547331,
0.01802450604736805,
0.010745402425527573,
0.005450491327792406,
0.007644869387149811,
-0.016311241313815117,
-0.013810021802783012,
0.002101152203977108,
-0.0007807232905179262,
-0.005475505720824003,
0.01711896061897278,
... | [
0.38142335414886475,
-0.06010181084275246,
0.41379114985466003,
-0.17406205832958221,
-0.13268615305423737,
0.14848622679710388,
0.18866859376430511,
-0.2960646450519562,
-0.0869087502360344,
-0.6160993576049805,
-0.00733582628890872,
0.7016061544418335,
-0.5280033349990845,
-0.16201591491... |
I am using floats throughout the text perhaps this messes with spacing. It seems such an easy thing to adjust, yet I find no answer. | [
-0.024243170395493507,
0.031131140887737274,
-0.01913011074066162,
0.04738704860210419,
0.030100196599960327,
-0.013299220241606236,
0.008813994936645031,
-0.004270952194929123,
-0.0408906489610672,
-0.007874876260757446,
-0.01407044567167759,
0.008577232249081135,
-0.013746760785579681,
0... | [
0.49340879917144775,
0.3295610547065735,
0.3615425229072571,
-0.0005013838526792824,
-0.4174719452857971,
0.02302587777376175,
0.2920326590538025,
-0.07026252895593643,
-0.20990513265132904,
-0.7358278036117554,
-0.0158049575984478,
-0.050093624740839005,
0.013552065007388592,
0.6085917353... |
Here is a question about the canonical momentum that I had asked some days ago, but I still have one point that I am not understand. Considering a particle moves in a magnetic field with charge $q$ and mass $m$, its hamiltonian is $$H=\frac{\vec{P}^2}{2m}=\frac{(\vec{p}+q\vec{A})^2}{2m}$$ where $\vec{p}$ is the momentum of the particle, $\vec{A}$ is the vector potential of the magnetic field and $\vec{P}$ is the canonical momentum of the particle. I think, because of the expression of the hamiltonian, the canonical momentum $\vec{P}$ is a conserved quantity. But by the answer in the previous link, it seems that the canonical momentum is not conserved even in a simple example that a particle moves in a homogeneous magnetic field. I am confused about this question. Is the canonical momentum conserved when a particle moves in magnetic field? | [
0.0015869576018303633,
0.00754651241004467,
-0.003292014356702566,
0.0030111228115856647,
-0.006435002200305462,
-0.02856414020061493,
0.006948414258658886,
-0.017033427953720093,
-0.007583808619529009,
-0.004191051237285137,
-0.00644987728446722,
0.009132476523518562,
-0.014392662793397903,... | [
0.06643461436033249,
-0.123466357588768,
0.9549229741096497,
-0.021306093782186508,
0.020136121660470963,
0.08219859004020691,
-0.31231921911239624,
-0.5790802240371704,
-0.27754971385002136,
-0.19742104411125183,
0.08093488961458206,
0.10792206972837448,
-0.20510685443878174,
0.8181913495... |
I currently work for a large-ish manufacturing company that has a variety of pay grade levels. I've put some serious time and effort into my work over the past 4 years and I've managed to make a significant enough impact that they moved me into a new branch of the company to start an in-house software development branch dedicated to serving our international interests. This has implied a ton of traveling and working at night due to time zone differences. While this is a great opportunity and sounds impressive to my friends/family, I'm still working at the lowest possible pay grade level. I assumed with all the additional job responsibilities that I'd also receive some kind of promotion. Is this common? Do I need to speak up in order to get a promotion? Do I even deserve one? Should I just wait it out? I'm lost... | [
-0.004605971276760101,
0.021051054820418358,
-0.004817902110517025,
-0.007702064700424671,
0.0018175255972892046,
0.011956444010138512,
0.005617499817162752,
-0.006537645123898983,
-0.011081336066126823,
-0.00030040647834539413,
-0.008756333030760288,
0.013361101970076561,
0.0051125842146575... | [
0.8492133617401123,
0.4523187577724457,
0.020223429426550865,
-0.07130065560340881,
0.4312084913253784,
-0.06886304914951324,
0.6043052077293396,
0.5047976970672607,
-0.31972044706344604,
-0.49964094161987305,
0.06446384638547897,
0.3118077218532562,
0.38710084557533264,
0.5765212178230286... |
Ampère's law states that $$ \nabla \times \mathbf{B} = \mu_0 \mathbf{J} \tag{1}$$ Simultaneously, from Ohm's law, we know that $$\mathbf{J}=\sigma\left(\mathbf{E}+\mathbf{u}\times\mathbf{B}\right)\tag{2}$$ When equating both currents $\mathbf{J}$, and applying Faraday's law, and knowing that the magnetic field $\mathbf{B}$ is solenoidal, one arrives at the transport equation of the magnetic field $$ \frac{\partial\mathbf{B}}{\partial t}=\nabla\times(\mathbf{u}\times\mathbf{B)}+\frac{1}{\sigma\mu_0}\nabla^2\mathbf{B}\tag{3}$$ In the low magnetic Reynolds numbers limit $Re_m=\mu_0\sigma u L \ll 1$, the externally imposed magnetic field does not change and equation $(3)$ is not explicitly solved, because the diffusional term becomes negligible. In this so-called one-way coupled magnetohydrodynamics (MHD) (valid for most liquid metals), the electric field $\mathbf{E}$ has a potential $\Psi$, which is calculated from the divergence free ($\nabla\cdot\mathbf{J}=0$) condition, using equation $(2)$, with a Lorentz force $\mathbf{J}\times\mathbf{B}$. My question is how should one look at Ampère's law in the context of one-way coupled, or low $Re_m$ MHD? Specifically considering the following two examples **Example 1** Suppose a situation with an insulating boundary (with $\mathbf{n}=\mathbf{\hat{y}}$). This means that $J_y=0$. Furthermore, the imposed magnetic field is constant in time, but varies in space. From $(1)$ $$J_y = \frac{dB_x}{dz}-\frac{dB_z}{dx},$$ which, for an arbitrary, non-uniform, steady magnetic field, would be non- zero. **Example 2** Suppose the externally imposed magnetic field $\mathbf{B}(x,y,z,t)=\mathbf{B}_0$. According to $(1)$, $\mathbf{J}=\mathbf{0}$, and thus there would be no Lorentz force? What is exactly happening at the interface, and am I applying Ampère's law correctly. What am I overlooking? | [
-0.011576970107853413,
0.02524575963616371,
-0.006132264155894518,
0.006870960351079702,
0.010398026555776596,
0.005493839271366596,
0.008239432238042355,
0.002464220393449068,
-0.013796673156321049,
0.009408365935087204,
-0.012674039229750633,
0.013278159312903881,
-0.027874130755662918,
... | [
-0.10471082478761673,
-0.5667650699615479,
0.6103221774101257,
0.043452899903059006,
-0.04602992534637451,
0.5188512802124023,
0.0692744255065918,
-0.8125840425491333,
-0.4151478707790375,
0.05503469705581665,
0.4265729784965515,
0.38887301087379456,
-0.26945507526397705,
0.759750068187713... |
I'm looking for a simple definition of the concept of “client-server” I'd like something similar to this definition of **_state_**. > ... That "thing/information" that you need to remember is called "state". Edit - This isn't a homework question (nor am I a student). My goal is to come up with a compact way of explaining REST to average developers. I didn't want to prejudice the response though. | [
-0.004234958440065384,
0.002899870974943042,
0.0006373786018230021,
-0.004621312953531742,
-0.0010731264483183622,
0.004136841744184494,
0.006300562527030706,
-0.008606556802988052,
-0.013028242625296116,
0.002998539712280035,
0.005778095219284296,
0.014457975514233112,
0.013783350586891174,... | [
0.3378617465496063,
0.1536402851343155,
0.02354704961180687,
0.030143877491354942,
-0.18695402145385742,
-0.27866649627685547,
0.10061399638652802,
0.29327622056007385,
-0.033765289932489395,
-0.5756548643112183,
0.09091028571128845,
0.25424882769584656,
0.19404466450214386,
0.015672570094... |
I am not sure what I am doing wrong here and would appreciate the help I have an array the looks like this: Array ( [85369] => 3 [85368] => 13 [85378] => 23 ) When I try to update the database with update_post_meta( $id, 'key', $MyArr); in the databse the array is not stored. What is stored is: a:0:{} Any help would be appreciated Full Code: function distance($lat1, $lng1, $lat2, $lng2, $unit) { $theta = $lng1 - $lng2; $dist = sin(deg2rad($lat1)) * sin(deg2rad($lat2)) + cos(deg2rad($lat1)) * cos(deg2rad($lat2)) * cos(deg2rad($theta)); $dist = acos($dist); $dist = rad2deg($dist); $miles = $dist * 60 * 1.1515; $unit = strtoupper($unit); if($unit == "K"){ return round ( ($miles * 1.609344) ); }elseif ($unit == "N"){ return round ( ($miles * 0.8684) ); }else{ return round ( $miles ); } } $MilesArray = array(); $bounding_distance = 1; $nearbys = $wpdb->get_results( " SELECT * FROM rvty_geodata WHERE ( geo_latitude BETWEEN ($lat - $bounding_distance) AND ($lat + $bounding_distance) AND geo_longitude BETWEEN ($lng - $bounding_distance) AND ($lng + $bounding_distance) ) " ); foreach ($nearbys as $nearby) { $alt_id = $nearby->post_id; $alt_lat = $nearby->geo_latitude; $alt_lng = $nearby->geo_longitude; if ($rvparks->ID != $alt_id) { $Miles = distance($lat, $lng, $alt_lat, $alt_lng, "M"); $MilesArray[$alt_id] = $Miles; } } asort($MilesArray); if(count($MilesArray)>10){ $MilesArrayChunk = array_chunk($MilesArray, 10, true); $ALTCGArray = $MilesArrayChunk[0]; }else{ $ALTCGArray = $MilesArray; } update_post_meta( $rvparks_id, 'alt_camps', $ALTCGArray); | [
-0.0037212781608104706,
0.017612507566809654,
-0.00013187757576815784,
0.01373615674674511,
0.009329793974757195,
-0.004913488402962685,
0.0056876991875469685,
0.004564991686493158,
-0.012580406852066517,
-0.02195279486477375,
-0.005987992975860834,
0.011530420742928982,
-0.01483459025621414... | [
0.028226234018802643,
0.15827351808547974,
0.5083338618278503,
-0.28770580887794495,
0.24549953639507294,
0.275539755821228,
0.34575802087783813,
-0.3285980522632599,
-0.00840427353978157,
-0.5335925221443176,
0.11063339561223984,
0.8858706951141357,
-0.29045310616493225,
-0.06271179765462... |
I'm **not** an advocate of the no-www movement. I like the www because it adds as a buffer to distinguish between our public and private/static sites. The problem is that with one of our sites, our traffic is split pretty much 50/50 between those that use our www and those that don't. Should I bother rewriting those who hit our non-www site to our WWW site? Or should I just leave them alone? All our google SEO whatnot is on our www site, so I'm not concerned about any of that, only about user perception. Has anyone here had this problem before? I'm not concerned about the technial aspect (that's easy with a quick rewrite rule), primarily the social side. | [
-0.025059688836336136,
0.004463979043066502,
-0.00918673351407051,
0.014166614040732384,
-0.019628871232271194,
0.0029132578056305647,
0.008118788711726665,
-0.003509669564664364,
-0.014261975884437561,
-0.003087787888944149,
-0.002073370385915041,
0.004942059516906738,
-0.004697443917393684... | [
1.1002881526947021,
0.5173963904380798,
0.3405546247959137,
-0.11644274741411209,
-0.48318618535995483,
-0.08011496067047119,
0.19949865341186523,
0.37359172105789185,
-0.09007503092288971,
-0.5710821747779846,
0.02390524372458458,
0.32372570037841797,
-0.077741339802742,
0.254003643989563... |
My dependent variable is categorical (with 3 levels) and my predictor variables are a mix of continuous (age in months, test 1 score, test 2 score, test 3 score) and categorical (gender). I believe I should run a multinomial regression, but when I do the results are really uninterpretable because each age is basically treated as a category (and the same for the test scores). Does anyone know a different test that would be appropriate to use in my case? I would really appreciate any help. | [
-0.0007202034466899931,
0.017564836889505386,
-0.011396289803087711,
0.010914131999015808,
0.028992293402552605,
-0.00006154230504762381,
0.008880873210728168,
0.014422353357076645,
-0.01410050131380558,
-0.031419239938259125,
-0.0029843070078641176,
0.005433131940662861,
0.00026249338407069... | [
0.22380991280078888,
0.3769294321537018,
0.2535525858402252,
0.03568476438522339,
-0.15148714184761047,
0.3940171003341675,
0.5371299386024475,
-0.006540585774928331,
-0.002041432075202465,
-0.19863906502723694,
0.49882569909095764,
0.10845828056335449,
-0.060834210366010666,
0.50913840532... |
As a thought experiment let us assume that we have isolated a magnetic domain. This domain is of finite size and we know its dimensions. Assuming that we can measure an infinitesimal field, will there be a certain region beyond which the field won't be applicable? The instinctive answer to this question is no, but if you think about it we see the magnet's influence on the space around it as the result of equipotential regions then the contention is that only so many discrete equipotential regions are possible (the fact that something is not countable doesn't automatically mean it's infinite). Hence, that line of thought goes, there should be a limit theoretically and practically until which a field can exist. Can you please clarify this sticking point for me? Am I pushing the analogies we use to understand fields too far? What conceptual mistake am I making over here? | [
0.01323392428457737,
0.012503204867243767,
0.0032474116887897253,
0.012450609356164932,
0.008743515238165855,
0.000021480722352862358,
0.006278377026319504,
0.007707028649747372,
-0.008602599613368511,
0.0131409065797925,
-0.00630310270935297,
0.015652945265173912,
-0.012568244710564613,
0... | [
0.10472624748945236,
-0.45131251215934753,
0.1364727020263672,
0.2914144992828369,
0.041193265467882156,
-0.0033760832156986,
-0.01709757000207901,
-0.07323456555604935,
-0.3843398690223694,
-0.4610370993614197,
0.06448204815387726,
0.14520977437496185,
-0.18788832426071167,
0.716554820537... |
I have an app that does some buffering, only problem is when I click on the map to put in the point/line/polygon vertices to do the buffer, I get popups at every mouseclick. It doesn't interrupt the buffer process but its super annoying. At first thought I wanted to try and disable the popups so that they didn't show when "No information available" was the result. But I figure there will be more times than not when a popup will return some results and thus, need to disable/hide the popup so that it doesn't appear at all while the buffer process is running. I've tried putting map.infoWindow.hide(); within the buffer without success and also saw this post recommending an onClick handler and have been unsuccessful as well. Any tips appreciated. | [
-0.008235879242420197,
0.0028171236626803875,
0.0024650227278470993,
0.009336618706583977,
-0.004012036137282848,
0.0005788425914943218,
0.006548899691551924,
0.013688869774341583,
-0.019747741520404816,
0.018155718222260475,
-0.0031640278175473213,
0.013546541333198547,
-0.00684807961806654... | [
-0.03356561437249184,
0.005892944056540728,
0.37918978929519653,
0.132272407412529,
-0.14690634608268738,
-0.1305588185787201,
0.4723633825778961,
0.37937939167022705,
-0.3609192967414856,
-0.7096887826919556,
0.1423213928937912,
0.21374228596687317,
-0.16790646314620972,
0.101156540215015... |
So, I have Windows 7 and Fedora 16 installed on my old HDD. Everything worked well and fine before I've had my new 3TB drive built in, which I initialized as GPT in Windows. Actually I initialized 1,5TB - the rest remains untouched. After that Fedora won't boot up anymore. Instead it prompts me to maintenance mode, showing something like: [...]/sbin/blkid -o udev -p /dev/sda[number] [...] terminated by signal 15 (Terminated) Whenever I press Ctrl+D it shows one or multiple messages similar to that. Using parted /dev/sdb print shows that the drive as such is recognized as GPT. It also shows up in /etc/fstab. Using older kernels results in the same problem. What should I do ? Edit: I initialized the remaining ~1,5TB in Windows - nothing changed. | [
-0.005220511462539434,
-0.004096038639545441,
-0.003373393788933754,
0.024845413863658905,
0.0012742634862661362,
-0.003432289231568575,
0.007837961427867413,
-0.004675097763538361,
-0.014846556819975376,
-0.002069630427286029,
-0.02355119213461876,
-0.0006077846046537161,
-0.010832260362803... | [
0.4109325408935547,
0.2962511479854584,
0.775099515914917,
-0.15463483333587646,
0.03783269599080086,
0.16785869002342224,
0.11233481019735336,
-0.22574453055858612,
-0.18833474814891815,
-0.7332846522331238,
-0.24850568175315857,
0.9378621578216553,
-0.17500388622283936,
0.285840719938278... |
I am using what I would describe as the most primitive method of doing intersection with OGR. The short script below describes how I do it. Is there a best way of doing this? from osgeo import ogr, osr shp1 = ogr.Open(file1) shp2 = ogr.Open(file2) SpatialRef = osr.SpatialReference() SpatialRef.SetWellKnownGeogCS('WGS84') # Create dst file here dstshp = ogr.CreateDataSource('SomeFilename.shp') dstlayer = dstshp.CreateLayer() # define its attribute fields for dstlayer and create them layer1 = shp1.GetLayer(0) layer2 = shp2.GetLayer(0) for feature1 in layer1: geom1 = feature1.GetGeometryRef() attribute1 = feature1.GetField('FieldName1') for feature2 in layer2: geom2 = feature2.GetGeometryRef() attribute2 = feature2.GetField('FieldName2') intersection = geom2.intersection(geom1) dstfeature = ogr.Feature(dstlayer.GetLayerDefn()) dstfeature.SetGeometry(intersection) dstfeature.setField(attribute1) dstfeature.setField(attribute2) dstfeature.Destroy() # and other features must be destroyed too | [
-0.013414736837148666,
0.007398597896099091,
-0.015609163790941238,
0.009212514385581017,
-0.017460057511925697,
0.02133064903318882,
0.00758935883641243,
0.00573196355253458,
-0.010624159127473831,
0.013361066579818726,
-0.004404442384839058,
0.008899377658963203,
-0.004506238736212254,
0... | [
0.1676001250743866,
0.14372453093528748,
0.8513484597206116,
-0.18487335741519928,
0.06276888400316238,
-0.03864632546901703,
0.3384788930416107,
-0.10479284822940826,
-0.2888716459274292,
-0.9022068977355957,
0.1496959924697876,
0.3665550649166107,
-0.274055540561676,
0.057238657027482986... |
I'm able to create question paper using LaTeX code mentioned below: \documentclass{exam} \usepackage[margin=0.7in,headheight=3.5\baselineskip,headsep=1\baselineskip,includehead]{geometry} \usepackage{multicol} \usepackage{amsmath} \setlength\columnsep{46pt} \begin{document} \begin{multicols}{2} \begin{questions} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question. Let This be the test question.Let This be the test question. Let This be the test question. \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \question The points with position vectors\ $\displaystyle \boldsymbol{61i + 2j,\,10i-9j,\,ai-12j }$ \ are collinear, if\ $\displaystyle \boldsymbol{a = }$ \begin{choices} \choice 1 \choice 2 \choice 3 \choice 4 \end{choices} \end{questions} \end{multicols} \end{document} It generates output as mentioned below:  If you notice Question 9, it's going out of the page margin. This happens very frequently when i add maths symbols, is there any way to avoid this error ? | [
0.01449209451675415,
0.0027702609077095985,
-0.006861985195428133,
0.008026912808418274,
0.006051753647625446,
0.013643398880958557,
0.008680466562509537,
0.0006350540788844228,
-0.009301753714680672,
-0.013544653542339802,
-0.0074889156967401505,
0.005842611659318209,
0.002110159955918789,
... | [
0.23376847803592682,
0.2047973871231079,
0.42231959104537964,
0.19298148155212402,
-0.10572870820760727,
0.15151473879814148,
-0.0250760018825531,
-0.3971969485282898,
0.06850461661815643,
-0.7272886633872986,
0.008239087648689747,
0.04982062056660652,
0.23873062431812286,
0.04086552187800... |
I ask because in the Help for the "Last Man Standing" section, it says: > The town can not have been devastated on a previous attack (a town is > devastated if a night passes and there are no citizens within the town > walls.). That makes it sound like it's possible for people to survive outside, but I've never seen it happen or read anything about someone surviving out there. Anyone know if it's possible? Update: The answer was "no" when this question was asked during Season 1. For Season 2 and beyond, the answer is "Yes". | [
-0.0018444494344294071,
0.022191833704710007,
-0.0024946066550910473,
0.006893325597047806,
-0.022741975262761116,
-0.02727419137954712,
0.00538895046338439,
0.006170144770294428,
-0.014188382774591446,
-0.005377890542149544,
-0.013026850298047066,
0.028263501822948456,
0.018968259915709496,... | [
0.27438798546791077,
-0.21046872437000275,
0.32526713609695435,
-0.002155466703698039,
-0.0667935460805893,
0.13572226464748383,
0.6772254705429077,
-0.10720425099134445,
-0.18199332058429718,
-0.43723365664482117,
0.04591680318117142,
0.06069372221827507,
0.010493705049157143,
0.158109098... |
I have some listings that contain pseudo-code of algorithms. And I use this: `\lstset{caption={Descriptive Caption Text},label=DescriptiveLabel}` to add captions to them. My problem is that the caption at the created file shows `Listing 1: Descriptive Caption Text`. I would like it to write something like `Algorithm 1: Descriptive Caption Text`. How can this be done? | [
0.019016500562429428,
0.00895758904516697,
-0.015612700954079628,
0.014096102677285671,
-0.010704601183533669,
0.018222248181700706,
0.008691168390214443,
0.006123566068708897,
-0.01970045641064644,
0.014186078682541847,
-0.020104462280869484,
0.0015682969242334366,
-0.0018159791361540556,
... | [
0.38329604268074036,
0.4868384301662445,
0.3143419325351715,
0.1565360724925995,
0.1633494347333908,
-0.32743027806282043,
0.021267736330628395,
-0.1598188579082489,
-0.26869305968284607,
-0.5624962449073792,
0.03446120023727417,
0.3324633240699768,
-0.39943644404411316,
0.2660611867904663... |
I know that you can't use an Australian Mastercard on a US PSN Account, i've tried but i'm wondering if a Australian Mastercard will work on a UK PSN Account since DLC for UK Games can be brought from the Australian PlayStation Store and work (I've tried with Tales of Graces F, Atelier Totori, Atelier Meruru, Hyperdimension Neptunia) so i figure maybe the UK Store will allow Australian Mastercard (or Visa Card) i ask this as Agarest Generations of War Zero only has DLC in the UK Playstation Store and i'm wondering if i should waste more time trying to find UK PSN Cards or if i can just use my card | [
0.0029016793705523014,
0.00255199964158237,
-0.00616095494478941,
0.030995944514870644,
0.014815768226981163,
0.02417999505996704,
0.009506886824965477,
-0.027135170996189117,
-0.01707969233393669,
0.0006960890023037791,
-0.015061013400554657,
0.013557633385062218,
0.00321376184001565,
0.0... | [
0.5256426334381104,
-0.07703499495983124,
-0.028979353606700897,
0.434156596660614,
-0.2438146471977234,
0.06692207604646683,
-0.018360959365963936,
0.10026166588068008,
-0.24355414509773254,
-0.36330506205558777,
0.13348351418972015,
0.2927231788635254,
0.18743155896663666,
0.402017265558... |
If a regression model is applied and there exist residuals that is very high or very low (meaning outliers compared to the others), is it good practice to get rid of those observations and then do the regression again, particularly if you have a very big sample of data? | [
0.04268919676542282,
0.0388491116464138,
-0.005421152804046869,
0.011094318702816963,
-0.012909389100968838,
0.0012367211747914553,
0.01122100930660963,
-0.009087725542485714,
-0.008599277585744858,
-0.05234404653310776,
-0.001758214901201427,
0.03240558132529259,
0.017391208559274673,
0.0... | [
0.19888080656528473,
-0.22623737156391144,
-0.16709767282009125,
0.33683326840400696,
0.2047393023967743,
-0.1535857617855072,
0.30663546919822693,
-0.3260520100593567,
-0.2112734317779541,
-0.3588221073150635,
0.13567060232162476,
0.5894555449485779,
-0.03970789909362793,
-0.1122911348938... |
There are a few examples of pairs of words ending with _-ee/-er_ like _employee_ and _employer_ or _advisee_ and _adviser_. What I was curious about is if there was any rule that would describe the relationship of the objects in a pair like this and situations when it's appropriate to create a counterpart for a given word. I'll give you an example. It's relatively common in the computer programming world to see the word _dragee_ , which describes an object that is being dragged with a mouse. I understand that this is a relatively new word and could not be found in any dictionary (I've tried). Is that acceptable to make up words like this one or is it just bad English? | [
0.021203329786658287,
0.007567531429231167,
-0.02019585855305195,
0.020713869482278824,
-0.017671700567007065,
0.030281223356723785,
0.00685918191447854,
0.002806266536936164,
-0.013107864186167717,
-0.007649263832718134,
-0.002304484136402607,
0.00384559971280396,
0.0175415500998497,
0.00... | [
0.47957876324653625,
-0.1724022924900055,
0.2085237354040146,
0.06054820120334625,
-0.2221510261297226,
0.03505559265613556,
0.04704803600907326,
0.24833983182907104,
-0.5728649497032166,
-0.4411962926387787,
-0.002220018533989787,
0.3753587007522583,
-0.3793187439441681,
0.444424688816070... |
The big question I have in my mind: how many developers are brownfield ("enterprise") compared to greenfield (all new code, from the ground up). I'm constantly reading breathless articles about the latest technology, only to find out that It Just Won't Work On Our Enterprise Software codebase. People aren't ready for automated testing (because the logic is in the click- handlers and/or database). People aren't ready for ORM tools because we have horrendous amount of logic in stored procs and triggers. People aren't ready for WPF because our existing stuff is all WinForms. We can't get the latest version of Reactive Extensions because existing code used RX 1.0 and there are breaking changes that will require more testing effort than is justified by the return. Etc., etc., etc. Very few articles seem to be oriented toward the brownfield developer, for whatever reason (can't sell ads for articles that start off with "you probably can't use this, but..."?). So, I'm truly wondering: is the software development industry just chock full of greenfield developers, developing new projects for clients which are then released and enjoy a short existence until complete replacement for whatever reason? Or are there hordes of brownfield programmers silently laboring away in the ADO.NET T/SQL VB.NET software mines, looking wistfully up at the sunshine of Entity Framework 5.0 and Haskell, et cetera? How do we even measure that? Salaries (wages??) paid to software engineers* in the two categories? How do we measure THAT? Maybe... revenue generated from selling said software? (There's an assumption that the crappy old software sold by XYZ Corp. actually has maintainers). **My question: Does anybody have any numbers that speaks to how much of the industry is green field vs. brown field?** | [
0.011142279021441936,
0.0033422233536839485,
-0.013559732586145401,
0.001138296676799655,
-0.009645660407841206,
0.008571888320147991,
0.007622315548360348,
0.019802484661340714,
-0.01017160713672638,
-0.035332709550857544,
-0.018974874168634415,
0.0162032563239336,
0.011744201183319092,
0... | [
0.5211575031280518,
0.21687720715999603,
0.16309760510921478,
0.3626658022403717,
0.21015462279319763,
-0.0014504250138998032,
0.13221308588981628,
-0.04517637565732002,
-0.3218667507171631,
-0.5788393020629883,
0.17090119421482086,
0.41929104924201965,
0.013810932636260986,
0.316393762826... |
Sometimes, people really have a bad day... I think it's interesting to learn why that day was bad and learn from our mistakes made on that day. So... What was your worst software development day like? | [
-0.000332981173414737,
0.04663543775677681,
-0.01794615387916565,
0.011275621131062508,
0.02344180829823017,
0.02122274972498417,
0.010849585756659508,
0.025368083268404007,
-0.04737307131290436,
0.002184577053412795,
-0.013276639394462109,
0.03444086015224457,
0.029089273884892464,
0.0010... | [
0.46946221590042114,
0.2756865918636322,
-0.4175901710987091,
0.23279862105846405,
-0.2148660570383072,
-0.1980503350496292,
0.22687573730945587,
0.5995596051216125,
-0.14834003150463104,
-0.2602745294570923,
0.3686605989933014,
0.4914928674697876,
0.02765415981411934,
0.07611539959907532,... |
I know there are other questions about power recharge and weight, but none of them (that I found) ask or answer this. At first, I assumed the X% modifier to recharge speed actually meant `your recharge speed is (100+X)% the standard value`. That made sense, because the modifier could positive or negative. Then I realized your recharge speed modifier can go all the way down to -200%, which completely invalidates my assumption. So, what the hell do they mean by -200% Recharge Speed? I understand it's a bad thing, but I want to quantify how bad it is. Given a recharge time of 5 seconds, how is this -200% supposed to affect me? | [
0.014822404831647873,
0.024394094944000244,
-0.009241193532943726,
0.011549456045031548,
-0.015494137071073055,
-0.03891180083155632,
0.008453644812107086,
-0.033670585602521896,
-0.013699879869818687,
0.02872428111732006,
0.0017688032239675522,
0.02462402544915676,
0.004902392625808716,
0... | [
0.3502715229988098,
0.021835317835211754,
0.5796833038330078,
0.18585655093193054,
-0.022182874381542206,
-0.08201540261507034,
0.1805516928434372,
-0.34119731187820435,
-0.258621484041214,
-0.027009498327970505,
0.08350946009159088,
0.761787474155426,
0.031713422387838364,
0.0445613898336... |
My main goal is a model of the form `Y ~ D + N` . But lets say N and D correlate strongly so I decide to do a PCA on them I can do.. DN <- structure(list(D = c(0.7, 0.7, 0.8, 0.6, 0.5, 0.6, 0.6, 0.43, 0.43, 0.32, 0.6, 0.6), N = c(84, 62, 83, 67, 72, 69, 68, 59, 81, 52, 63, 83)), .Names = c("D", "N"), row.names = c(NA, -12L ), class = "data.frame") PCA <- princomp(DN, cor = TRUE) summary(PCA) #Importance of components: # Comp.1 Comp.2 #Standard deviation 1.233169 0.6923106 #Proportion of Variance 0.760353 0.2396470 #Cumulative Proportion 0.760353 1.0000000 i feel happy with these results. And I can get my new vectors `Comp.1` and `~Comp.1` from `PCA$scores` which will allow me to do `Y ~ Comp.1 + Comp.2` . However if I want to know which comp explains mostly `D` or `N` I think I use `PCA$loadings`. #Loadings: # Comp.1 Comp.2 #D -0.707 0.707 #N -0.707 -0.707 # Comp.1 Comp.2 #SS loadings 1.0 1.0 #Proportion Var 0.5 0.5 #Cumulative Var 0.5 1.0 I dont understand why the loadings matrix are all equal...Is comp2 mostly D ? Questions: 1. Is this the correct approach for what I am trying to achieve? 2. How can I show which component is mostly explained by D and which by N? | [
-0.012678414583206177,
0.0178898423910141,
-0.014636389911174774,
0.012336213141679764,
0.004111526068300009,
-0.009042841382324696,
0.004263885319232941,
-0.00875412579625845,
-0.0062941899523139,
0.006439622491598129,
-0.004726205952465534,
-0.00019993807654827833,
-0.013929324224591255,
... | [
-0.2190542221069336,
0.2577945291996002,
0.27502191066741943,
-0.14017608761787415,
0.11491304636001587,
0.37219738960266113,
0.14054390788078308,
-0.36816102266311646,
-0.055235788226127625,
-0.34840819239616394,
0.018387379124760628,
0.40709516406059265,
-0.10539963841438293,
0.521545648... |
So I'm still kind of new to linux, but what are the steps towards setting a linux box up in such a way that it can send mail using shell? echo "hello world" | mail johnny@moo.com I mean, I've done the necessary sudo apt-get install mailsystem (or something like that) which sets up the mail command. However, will you also have to set up a .com to point to your linux box as its SMTP server? What else needs to be done? | [
-0.012852849438786507,
-0.002370191738009453,
0.003979576285928488,
0.0052206506952643394,
0.0027053209487348795,
0.0005097917164675891,
0.007790937554091215,
0.0006649162387475371,
-0.021560605615377426,
-0.030526412650942802,
-0.00945942010730505,
0.0034922196064144373,
-0.0167002212256193... | [
0.6294829845428467,
0.2598186731338501,
0.2109827846288681,
-0.24221809208393097,
-0.0684828907251358,
-0.2699165344238281,
0.5877279043197632,
0.21742302179336548,
-0.07873178273439407,
-0.6028184294700623,
-0.011465802788734436,
0.2866135239601135,
-0.1542702466249466,
0.1368311941623687... |
Well im looking for a way to do this. In the index to show each post, with the "thumb" on the left and the content of the right using ul li to do so.(http://ae.tutsplus.com/ for example) So far i figure out that with the usual loop i cant separate text and iamge, so im guessing i need to add some php to the mix? | [
-0.011608131229877472,
-0.00039633733103983104,
-0.01617198809981346,
0.006112890783697367,
0.004513354040682316,
0.010113133117556572,
0.006276783999055624,
0.03404711186885834,
-0.021269256249070168,
0.020356781780719757,
-0.01679326221346855,
0.00894093792885542,
-0.010431572794914246,
... | [
0.6726115345954895,
0.28092047572135925,
0.2896968722343445,
0.06827817857265472,
-0.3767094910144806,
0.03729850426316261,
0.3475323021411896,
0.010949133895337582,
-0.07924192398786545,
-0.36618170142173767,
0.2781819701194763,
0.3390710949897766,
-0.2508890926837921,
-0.1013098582625389... |
With the newest Minecraft Dev Version: 12w32a, a new block was introduced, the beacon block. Apparently, it gives off certain powers in a certain radius, when built upon a 3x3 pyramid, etc., and I've seen videos of the block in work. However, it seems that there are different ores used in making pyramid, the question is, does the type of ore used (iron, gold, diamond, emerald, etc.) change anything? | [
-0.012153769843280315,
0.009983979165554047,
-0.012795329093933105,
0.010761577636003494,
-0.03440209478139877,
0.000752774125430733,
0.007828584872186184,
-0.00325034954585135,
-0.023836448788642883,
-0.025821760296821594,
-0.01977834478020668,
-0.0008851356105878949,
0.007124798838049173,
... | [
0.4142923951148987,
0.16826730966567993,
0.2773280739784241,
0.42953523993492126,
-0.44430941343307495,
-0.29262951016426086,
0.33788737654685974,
-0.24672859907150269,
0.013424653559923172,
-0.35325050354003906,
0.1902940422296524,
0.30583035945892334,
0.00197153864428401,
0.1102741584181... |
Is there a way to see when a page was first published, by checking when it was first cached by Google (obviously its not 100% fool proof because there is a couple of days delay in some cases but it will give you a good idea.) The only other way I could think of checking it published date is if the page and post had a publicly viable time stamp on it, but in the case I'm looking for it, it doesn't have a publicly visible time stamp. | [
-0.0017406299011781812,
0.004890926647931337,
-0.0036517567932605743,
0.006219024769961834,
0.021523581817746162,
0.016465719789266586,
0.00671229837462306,
-0.013160647824406624,
-0.019339879974722862,
0.018218517303466797,
0.008289000950753689,
0.01245076023042202,
0.010023404844105244,
... | [
0.27677643299102783,
-0.06185104325413704,
0.17760275304317474,
0.48759689927101135,
0.2581789493560791,
-0.22964070737361908,
-0.04479267820715904,
0.4645087718963623,
-0.5260398983955383,
-0.47010788321495056,
0.3582475483417511,
-0.16955018043518066,
0.21965035796165466,
0.0681012049317... |
I've got an old PC with debian on it, which works as some kind of router. It is set up as wireless access point using hostapd. I want to setup multiple access points, one on old wifi card with 2.4hz, and one on new one with 5hz. I wonder if there is a way to do that? | [
-0.022845672443509102,
0.013522851280868053,
-0.016221370548009872,
0.004343801643699408,
0.008049153722822666,
-0.015789978206157684,
0.010744079016149044,
-0.01686805486679077,
-0.029045330360531807,
-0.02013050951063633,
0.0026960193645209074,
0.008566351607441902,
-0.005494414363056421,
... | [
0.6272874474525452,
0.21412193775177002,
0.043888747692108154,
0.306314617395401,
0.0766967236995697,
-0.3728533983230591,
0.11653116345405579,
0.24640892446041107,
-0.1509040743112564,
-0.6050177812576294,
0.09266868233680725,
0.3138045370578766,
-0.3128105401992798,
0.1635710895061493,
... |
I've recently started looking at the issue of automated quality control and assurance of geographic data. What software exists that can be used for accomplishing this? Esri has got their product 'Data Reviewer' which seems interesting, are there any comparable alternatives to this product? I would like to tag this question with quality-control and quality-assurance, but can't do so. | [
0.0032322288025170565,
0.005925476551055908,
-0.001750255818478763,
0.01243794709444046,
0.014656668528914452,
0.0034095379523932934,
0.008744146674871445,
0.03517954796552658,
-0.01929514855146408,
-0.013169714249670506,
-0.0083503108471632,
0.02055620588362217,
0.007948514074087143,
0.00... | [
0.47784584760665894,
0.02180282212793827,
0.14564649760723114,
0.2785378694534302,
-0.05973208323121071,
-0.15497241914272308,
0.09668120741844177,
0.31334319710731506,
-0.15020234882831573,
-0.5002099871635437,
0.12160982191562653,
0.28823432326316833,
0.24253857135772705,
0.2930034399032... |
I have problems with the tabu, when I make a table with the code below, the contents of the last column in the first row is not all visible. Why is this?  \begin{table}[h] \caption{Explanation } \label{table:browsing_history} \begin{center} \begin{tabu} to 1.0\textwidth{|c|c|c|c|X[m,c]|} \hline User & Contacts & Timestamp & Restaurant & Ranking of selected restaurant \\ \hline 1 & ~ & 13.6.2013.10.33.43 & Restaurang Upper East & 3 \\ \hline 1 & 2,3 & 13.6.2013.10.34.2 & World Class & 2 \\ \hline 3 & 4 & 14.6.2013.11.10.19 & Secret Recipe & 10 \\ \hline \end{tabu} \end{center} \end{table} | [
0.0037204870022833347,
-0.0007544377585873008,
0.006906122900545597,
0.011611828580498695,
-0.0060261753387749195,
-0.0011985546443611383,
0.0047263819724321365,
0.022154800593852997,
-0.01661721244454384,
0.024208683520555496,
-0.013435854576528072,
0.004728525411337614,
-0.0062708780169487... | [
-0.17572128772735596,
0.22211071848869324,
0.3493141829967499,
-0.08583087474107742,
0.047973427921533585,
-0.07343630492687225,
0.329103022813797,
0.3322611451148987,
-0.4669205844402313,
-0.5915656685829163,
-0.025889363139867783,
0.4399966597557068,
-0.488541841506958,
0.299684166908264... |
The statement was in a poem whose name is " the White Troops Had Their Orders but the Negroes looked like Men". The context:... Who really gave two figs? Neither the earth nor heaven ever trembled... What is the meaning of the question? (figs) | [
-0.0004471172287594527,
0.03712236136198044,
-0.011531279422342777,
0.024913299828767776,
-0.016615817323327065,
-0.0028266599401831627,
0.015221222303807735,
0.020591935142874718,
-0.029567647725343704,
0.0012303560506552458,
-0.017128312960267067,
0.024121712893247604,
0.03739243373274803,... | [
0.3676599860191345,
0.05717797204852104,
-0.7943480610847473,
-0.03575390577316284,
-0.6031251549720764,
0.5292176008224487,
0.5040881037712097,
0.2616250514984131,
-0.4126228988170624,
-0.0764542892575264,
0.3592996895313263,
-0.03640337288379669,
-0.6386878490447998,
-0.04948749393224716... |
I would like to see the number of `todos` I have left. Therefor I tried to redefine the `todo` command as follows: \newcounter{todocounter} \renewcommand{\todo}[2][]{\stepcounter{todocounter}\todo[#1]{#2}} However this seems to cause an infinite loop (the todo inside the definition seems to use the todo in the declaration). Searching only yielded this which does not help in my case. Is there a trick to make this work? (etoolbox/ can I use some expansion time wizardry/would an alias help) I do not want to replace all `todo`s with `numberedTodo` or similar. | [
0.00713000912219286,
0.013727320358157158,
-0.008429445326328278,
0.015573972836136818,
0.005030951928347349,
-0.0016758962301537395,
0.007256209850311279,
-0.005432102829217911,
-0.01388284657150507,
0.009104475378990173,
-0.006477358750998974,
0.0008035280043259263,
-0.0054894983768463135,... | [
0.005848728585988283,
-0.3271489143371582,
0.5177077054977417,
-0.018567264080047607,
0.3284968435764313,
-0.2769405245780945,
0.3559063971042633,
-0.17717866599559784,
0.08309336006641388,
-0.39362940192222595,
0.1760837435722351,
0.5457310080528259,
-0.19202645123004913,
-0.1420935243368... |
I cannot access abd on my computer it is quite annoying. I don't know what the deal is. I have tried to modify the path variable on my system variables. It didn't work. When I try to access it in cmd the command comes back "abd is not recognized as an internal or external command, operable program, or batch file. Any help would be greatly appreciated. | [
-0.016269169747829437,
-0.0035861702635884285,
-0.028322303667664528,
0.02877034619450569,
-0.018450438976287842,
0.008831056766211987,
0.009894870221614838,
-0.005200806073844433,
-0.029148172587156296,
-0.008576941676437855,
-0.0326891727745533,
0.007736916188150644,
-0.008362137712538242,... | [
0.010387754067778587,
0.44918200373649597,
0.3720809817314148,
0.03400535508990288,
-0.0852326974272728,
0.025589637458324432,
0.4136596620082855,
0.17449520528316498,
-0.23216597735881805,
-0.7693153023719788,
0.34938621520996094,
0.6114729642868042,
-0.44199076294898987,
0.74086159467697... |
I'm writing a presentation in beamer. I want to show semi-transparent overlays in the table of contents items. I mean, I want to show the current item text in black color and the next ones in gray. Thanks in advance. | [
0.04570917412638664,
0.0011324705556035042,
-0.016836654394865036,
0.03130612522363663,
-0.023188022896647453,
-0.00812638457864523,
0.015422412194311619,
0.022167516872286797,
-0.027273640036582947,
0.008442698046565056,
-0.0272997859865427,
0.0017829725984483957,
0.00020784471416845918,
... | [
0.36505618691444397,
0.2989984452724457,
0.4353044033050537,
0.278524249792099,
-0.1471192091703415,
-0.13985303044319153,
0.05058737099170685,
0.6136159300804138,
-0.10635367035865784,
-0.6706496477127075,
0.08853407949209213,
0.2301507145166397,
0.1832645833492279,
0.053216297179460526,
... |
This code adds an extra class to all my menu items: add_filter('nav_menu_css_class' , 'special_nav_class' , 10 , 2); function special_nav_class($classes, $item){ $classes[] = 'btn'; return $classes; } How do I confine this filter to my main menu (in theme location 'primary- menu')? Regards, Daniel | [
0.016482340171933174,
0.013876393437385559,
-0.004846386145800352,
0.012953169643878937,
-0.0161127932369709,
-0.02939196303486824,
0.009344146586954594,
0.015434845350682735,
-0.017051486298441887,
0.013314847834408283,
-0.028887080028653145,
0.002592390635982156,
-0.024636484682559967,
0... | [
0.2841712534427643,
0.04640747979283333,
0.48849865794181824,
-0.08404112607240677,
0.1060926616191864,
0.23988090455532074,
0.30480819940567017,
-0.35360562801361084,
-0.2741137742996216,
-0.9156960248947144,
-0.5213378667831421,
0.003183026099577546,
-0.2559601664543152,
0.22218072414398... |
I am using XeLaTeX in conjunction with `amsmath` and `unicode-math` for my reports. After a while, I ran into a weird problem in which compilation would just hang indefinitely. I found out the culprit was me using `\pm` or `\times` symbols (there might of course be others) in a math environment. Does anybody know why this occurs and how to fix it? Edit: Okay, after stripping down my list of packages I observed that compilation breaks as soon as I use the `inputenc` package, for example: \documentclass[11pt,titlepage]{report} \usepackage[T1]{fontenc} \usepackage{fontspec} % XeLaTeX required! \usepackage[latin1]{inputenc} % <------ This package breaks compilation!!! \usepackage{mathtools} \usepackage{unicode-math} \setmainfont[Ligatures=TeX]{Myriad Pro} \setmathfont{Asana Math} \setmonofont{Consolas} \begin{document} $\pm$ \end{document} Not being able to use this package would be a real shame though, because we dutch people sure like our accents at times... | [
0.004847487900406122,
0.00807077344506979,
-0.009790442883968353,
0.013491522520780563,
0.01068983506411314,
-0.004846001975238323,
0.005214050877839327,
-0.0055744946002960205,
-0.01147681288421154,
-0.015399022027850151,
-0.007630591746419668,
-0.004023169167339802,
-0.00688306987285614,
... | [
0.5456992983818054,
0.4911114275455475,
0.114256851375103,
0.13704366981983185,
0.22904139757156372,
-0.1952497512102127,
0.23805133998394012,
0.0720326155424118,
-0.38144123554229736,
-0.4193940758705139,
-0.1929313689470291,
0.5899956226348877,
-0.3117119371891022,
0.09750237315893173,
... |
Say I have 2 calendars. 1 for my personal stuff and 2 for my work related. Is there a way that I could only choose to get alerts from one for time being. Later, add or switch calendars? I am using Google Calendar. | [
0.021128695458173752,
0.016467129811644554,
-0.010972960852086544,
0.02277708426117897,
0.030137434601783752,
0.014352205209434032,
0.011044979095458984,
0.05087116360664368,
-0.029419487342238426,
-0.02251771092414856,
-0.0004125435370951891,
0.002745159203186631,
-0.008287918753921986,
0... | [
0.7965846061706543,
0.151455357670784,
0.3732827603816986,
-0.09875021874904633,
-0.12531019747257233,
0.04533366113901138,
0.27014151215553284,
0.42268308997154236,
-0.32237356901168823,
-0.6774730682373047,
0.36388954520225525,
0.4289122223854065,
-0.13795752823352814,
0.0409082658588886... |
Is there a difference between MPL 2.0 and LGPL 2.1 ( **+static linking exception** )? If yes, what is it? As long as i understand the only difference between them is you can't use trademarks of contributors under MPL[1]. | [
0.009951822459697723,
0.022955261170864105,
0.01521536335349083,
0.014507809653878212,
-0.003299559699371457,
0.07563886791467667,
0.014862635172903538,
-0.006801299750804901,
-0.03679754585027695,
-0.03550038859248161,
-0.02638229727745056,
0.011012648232281208,
0.003996141720563173,
-0.0... | [
0.5918961763381958,
0.26523762941360474,
0.10058949142694473,
-0.03434700518846512,
-0.29364216327667236,
-0.48127317428588867,
-0.01775958016514778,
0.24091112613677979,
-0.44354623556137085,
-0.041931964457035065,
0.07790761440992355,
0.412833034992218,
-0.48239824175834656,
0.1435092687... |
I am looking for a place to play AOE 2 TC online - something like Battle.NET. I used to play in Game Zone, but I think this game is no longer supported. Any ideas? | [
-0.03773488849401474,
0.0024480512365698814,
-0.008346022106707096,
0.005025838501751423,
0.05031323805451393,
0.025960151106119156,
0.010988962836563587,
0.0010933816665783525,
-0.03657408803701401,
-0.00482017919421196,
-0.015613614581525326,
0.024004893377423286,
-0.0015969753731042147,
... | [
0.6292852759361267,
-0.08833060413599014,
0.09707015007734299,
0.26204875111579895,
-0.0015590217662975192,
-0.32060354948043823,
0.1592206507921219,
0.5665194988250732,
-0.047367263585329056,
-0.8364731073379517,
0.3889988958835602,
0.2618180215358734,
0.04155854135751724,
0.2732099592685... |
The Feynman propagator for the free electron field is the Fourier transform w.r.t. $y$ of the time-ordered 2-point VEV $\left<0\right|\mathcal{T}[\hat\psi(x)\hat\psi(x+y)]\left|0\right>$, taking $\hbar=c=1$, $$\mathcal{F}[\left<0\right|\mathcal{T}[\hat\psi(x)\hat\psi(x+y)]\left|0\right>](k)= \frac{k\\!\cdot\\!\gamma+m_e}{k^2-m_e^2+\mathrm{i}\epsilon}.$$ In QED, after renormalization, we obtain, valid in the infrared, when $k^2-m_e^2$ is small (given in this neat form, representing the sum of _many_ perturbative integrals, in, for example, Appelquist & Carrazone, Phys.Rev.D **11** , 2856 (1975)), $$\mathcal{F}[\left<0\right|\mathcal{T}[\hat\psi(x)\hat\psi(x+y)]\left|0\right>](k)= \frac{k\\!\cdot\\!\gamma+m_e}{(k^2-m_e^2+\mathrm{i}\epsilon)^{1-{\alpha_{EM}}/{\pi}}},$$ where $\alpha_{EM}\approx 1/137$. This is often called an infraparticle propagator or a dressed particle propagator (I'm rehearsing this stuff, please bear with me, and feel free to comment on anything that seems to need some change). When we move to renormalization group methods, the coupling constant becomes a function of the renormalization scale. In this case, if we measure the Fourier transform of the 2-point VEV, using the operator $\mathcal{F}[\mathcal{T}[\hat\psi(x)\hat\psi(x+y)]](k)$, the measurement scale is determined by the wave-number $k$, so I take it we can write the 2-point VEV as $$\mathcal{F}[\left<0\right|\mathcal{T}[\hat\psi(x)\hat\psi(x+y)]\left|0\right>](k) {{?\atop =}\atop\ } \frac{k\\!\cdot\\!\gamma+m_e}{(k^2-m_e^2+\mathrm{i}\epsilon)^{1-{\alpha_R}(k)/{\pi}}},$$ where the running coupling constant $\alpha_R(k)$ is $\alpha_R(m_e)\approx 1/137$ when $k=m_e$, at about 0.5 MeV and $\alpha_R(m_Z)\approx 1/127$ when $k=m_Z$, the mass of the $Z$ particle, at about 90 GeV (see here). My understanding is that in QED the function $\alpha_R(k)$ is an increasing function of $k$, even to the extent that there is a Landau pole at finite inverse length $m_L$, $\alpha_R(m_L)=\infty$, but that the high-energy behavior of QED has only been calculated perturbatively, so that an analytic form for $\alpha_R(k)$ is not known. I would like, however, to look at a good article, perhaps a review, that gives as closed a form for $\alpha_R(k)$ as is currently known, in as neat a form as possible (this is an _implicit_ question that may be too much to ask of the literature, given that almost everyone has moved on to supersymmetry, noncommutative geometry, string theory, etc., etc., and QED is obviously not empirically useful at high energy). I would also like better to understand the relationship between the running coupling constant formalism and the Källén–Lehmann formalism. Can we equate the two, at least approximately, $$\frac{k\\!\cdot\\!\gamma+m_e}{(k^2-m_e^2+\mathrm{i}\epsilon)^{1-{\alpha_R}(k)/{\pi}}} {{?\atop \approx}\atop\ } \int\limits_{m_e}^\infty \frac{(k\\!\cdot\\!\gamma+m)f(m^2)\mathrm{d}m^2}{k^2-m^2+\mathrm{i}\epsilon}.$$ Here, $f(m^2)\ge 0$ is undefined for $m>m_L$ if there is a Landau pole in QED, but I suppose $f(m^2)$ would not have any poles and would approach zero fast enough for at least the integral $\int\limits_0^\infty f(m^2)\mathrm{d}m^2$ to exist in a well-defined theory? We could equate the two exactly if we were talking about a scalar Feynman propagator, because we could solve for $\alpha_R(k)$, but the change from $k\\!\cdot\\!\gamma+m_e$ to $k\\!\cdot\\!\gamma+m$ doesn't look good because the two expressions have different effects for different components of the Dirac spinor. Better understanding the 2-point VEVs would of course leave the $n$-point VEVs to think about. Uggh. Finally, there are three explicit Questions here (marked by question marks!), but I would also be interested in any ruminations from this starting point that include references. | [
-0.008170186541974545,
0.011662982404232025,
-0.004172684624791145,
0.0030714550521224737,
-0.0016054825391620398,
-0.013999700546264648,
0.006568955257534981,
0.011740095913410187,
-0.00646722037345171,
0.010411751456558704,
-0.01114571187645197,
0.006441669538617134,
-0.024194395169615746,... | [
-0.2180851846933365,
-0.5903850793838501,
0.4137694537639618,
-0.06592366844415665,
-0.40823158621788025,
0.19706319272518158,
-0.3285697102546692,
-0.7476878762245178,
0.3833109140396118,
-0.30225449800491333,
-0.1194702759385109,
0.8148292303085327,
-0.5633009672164917,
0.565944671630859... |
I want to display each product category's sub products list in the shop page. Actually now in my shop page was displaying the whole list of products. But when ever i click on the list of sub category menu the control was automatically goes to the home page of that site. And i can understand that there is no code error. When i add 2 categories followed by the 4 pre_category in the admin part this problem is occur. So may be its WP-ADMIN settings problem. I done this site by using woocommerce plugin. Can u pls help me whats the problem... | [
0.001036516623571515,
-0.004249130841344595,
-0.0006215265020728111,
0.030351104214787483,
-0.008203119970858097,
0.016023335978388786,
0.008278878405690193,
0.027371663600206375,
-0.018360627815127373,
0.01259414479136467,
-0.027740124613046646,
0.006962230429053307,
-0.013637429103255272,
... | [
0.3725912868976593,
0.055457133799791336,
0.4276498854160309,
-0.0033998729195445776,
-0.0022312512155622244,
0.18893560767173767,
0.03765389323234558,
-0.11025314033031464,
-0.3568846881389618,
-0.34890273213386536,
-0.14847198128700256,
0.6799709796905518,
-0.0932697057723999,
0.34766915... |
I've selected a number of polygons from one layer and want to create a new layer. Because I can't right click using a mac how do I find "Save Selection As". Its not in the list in the "Layer" drop-down from the top toolbar. I'm using QGIS 2.4 with Mac OS X 10.9.4. | [
-0.0011074031936004758,
-0.00823594257235527,
-0.01456585992127657,
0.004617541097104549,
0.016097284853458405,
0.0034657528158277273,
0.010192378424108028,
0.050153594464063644,
-0.024356704205274582,
-0.013442746363580227,
0.000008120723578031175,
0.008765305392444134,
-0.00654376065358519... | [
0.24682901799678802,
0.004432442598044872,
0.14716023206710815,
-0.0272041205316782,
-0.1377703994512558,
0.16018730401992798,
0.22906658053398132,
0.30022621154785156,
-0.18602830171585083,
-1.1422672271728516,
-0.031988322734832764,
0.35621023178100586,
-0.05161390081048012,
0.0357596725... |
I am working with a binary predictive model for data that belongs to A and B. The learning sample that I am using contains 6000 row that belongs to group A and 1000 row that belongs to group B. I would like to make my learning sample equal in number for both variables (i.e. 1000 row that belongs to A and 1000 row that belongs to B). A random sampling technique might be very biased when selecting 1000 out of 6000 rows. What would be the best way to pick these 1000 samples from group A in a way that assures this is not a biased sampling? In other words, I would like to have a sample of 1000 rows that represent as closely as possible the 6000 rows. What technique would do this best? Many thanks, | [
0.023171456530690193,
0.005342214368283749,
-0.02067345194518566,
0.029286280274391174,
-0.015730589628219604,
0.008318254724144936,
0.009089933708310127,
0.013894560746848583,
-0.013214866630733013,
0.013579125516116619,
-0.01697656512260437,
0.012401098385453224,
-0.013751095160841942,
0... | [
0.35710057616233826,
0.08734478801488876,
-0.25213688611984253,
0.3711109459400177,
0.035804204642772675,
0.5237282514572144,
-0.04566891863942146,
-0.0811498686671257,
-0.1932336986064911,
-0.7599606513977051,
0.3367355763912201,
0.3824007511138916,
0.07537990808486938,
0.560686469078064,... |
I have a list of points of a curve in a three dimensional space, like myData = {{1.6251942652167208, 3414.632234882431, 3.248326448352207}, {0.195563463104691, 32.18966154343482, 1.1454060351007385}, {0.17904199044243319, 14.25108939238815, 0.9423477799753123}, {0.170816821901991, 8.653164633472851, 0.7958880823665402}, {0.15676037756308206, 5.631156542779637, 0.67467521725227}, {0.1548948722263085, 4.270874169070245, 0.6275611866853114}, {0.1876950201556172, 4.308999312640106, 0.6090637255823652}, {0.16188251254916763, 2.9764924022024317, 0.5292892925723116}, {0.17262610022897934, 2.739111371641244, 0.52653513586836}, {0.16954431585785373, 2.3070638999549393, 0.49560804724631763}, {0.2097961985882085, 2.6534757543854317, 0.5487657657669014}, {0.19081633352127408, 2.0852454796400357, 0.514435548916157}, {0.19563129613920563, 1.917093872159669, 0.4743406166681895}, {0.21254024573386504, 1.9251337728677473, 0.485236735240245}, {0.21896472970440534, 1.8147535334973488, 0.46531629302089494}, {0.20466136180653818, 1.513876366963674, 0.4249037110237554}, {0.21855394267841366, 1.519108716384235, 0.4380750296933395}, {0.23565318703002316, 1.5488572100526539, 0.44631964709969185}, {0.2380818528018018, 1.4529671243699986, 0.42553379428103266}, {0.2424842011874948, 1.389721482236748, 0.4345346021692877}, {0.2647317878191315, 1.4605691656957687, 0.43267419256049483}, {0.2632502748033041, 1.3557447019366855, 0.4286184053606104}, {0.27205729604073775, 1.334394688581283, 0.44259414855715523}, {0.2793705082711959, 1.302387809445032, 0.4336404208867536}, {0.2837573939406187, 1.2544604741697558, 0.42970441420505734}, {0.2948407139307023, 1.251503834903687, 0.4289108411034688}, {0.29600086351343347, 1.1920618765471473, 0.42889878404407833}, {0.29465197778349606, 1.1243426155800962, 0.4276172983591667}, {0.30183374846249794, 1.1064964396078605, 0.42772557494175767}, {0.30918455260251276, 1.0893160114201885, 0.4275101914583863}, {0.3170183678910846, 1.0761991260380923, 0.4251450946217794}, {0.3170471652361686, 1.028070998453355, 0.4231589533788725}, {0.33509119696542056, 1.062744456154385, 0.4233345521815623}, {0.32971337187673616, 0.9954325606304446, 0.4231326015290622}, {0.33595705730728015, 0.979557033300735, 0.4236769992893109}, {0.3436253098086924, 0.9714372846864234, 0.4241910609852374}, {0.34812463781047914, 0.949934939902976, 0.42012852465803996}, {0.3563732253707738, 0.9443465112229241, 0.4204752685781463}, {0.35992684291158533, 0.9221331266260643, 0.41804460944990646}, {0.36120796960334306, 0.893126929139083, 0.41819781912012344}, {0.3694695389695183, 0.8578713443146384, 0.4150160053145223}, {0.37334083248099226, 0.8414330588993054, 0.4145221485938744}, {0.3766063600562105, 0.8812611721272469, 0.45266007028291844}, {0.3835672074215254, 0.8426364383173903, 0.44224264257818485}, {0.3925576360242645, 0.8145071347303069, 0.441619765516166}, {0.3952083972535039, 0.7698172682169502, 0.4289584794098478}, {0.40767794715674194, 0.7561905560105875, 0.42830082380993906}, {0.4139609495348973, 0.7284690866030674, 0.42802517746189234}, {0.4199909107710205, 0.7017646393366863, 0.4245332938666336}, {0.42517695763043706, 0.6756040929172838, 0.41901542028878747}, {0.438909457948436, 0.6699655678281783, 0.41733783818556264}, {0.44563579799220726, 0.6499029238980055, 0.4129281969548992}, {0.4473498530318954, 0.6211332659501384, 0.4089614451608291}, {0.45464041389498117, 0.6051969313774956, 0.4050849964713832}, {0.46156953220721025, 0.5899702211293141, 0.4029806473833252}, {0.47141055300556484, 0.5805373200701324, 0.39970427922756985}, {0.4792326437523886, 0.5681016242905145, 0.39694835543120627}, {0.4853330654106727, 0.553725319161017, 0.393182829103135}, {0.49135512137375686, 0.5399417830238199, 0.3901371989617769}, {0.5006324150392757, 0.5320562690230357, 0.3875692692973447}, {0.5064951595246885, 0.5194105475320817, 0.3832282043244645}, {0.5115547319598217, 0.5062662671288428, 0.3811347260859499}, {0.51504804958564, 0.4917529304140576, 0.3775795425712839}, {0.5220718511737211, 0.48266019418477496, 0.37470445520090895}, {0.5312779766503062, 0.47705149397591645, 0.37098468508988963}, {0.5324554882538403, 0.4612891859386048, 0.36767364179185347}, {0.5377840374755473, 0.451700594900957, 0.3651018926671713}, {0.5450134070805016, 0.44432218995571343, 0.3621032659850748}, {0.5532192176564933, 0.438742917967833, 0.35988610407498006}, {0.5559200803701375, 0.4272035754877022, 0.3562923772639558}, {0.5632612273324766, 0.42125311233985596, 0.35442749937954365}, {0.5695771900612702, 0.4143325143071145, 0.35207878660073166}} I want to find and plot a find of this curve in the space. I don't have a explicit form of the real function, but I know that it is regular enough to allow a polynomial expansion in the interval of interest. Is this possible? | [
-0.006255925167351961,
0.005464241374284029,
-0.00815645232796669,
0.002134984824806452,
-0.000702492892742157,
-0.001424380810931325,
0.0017915100324898958,
-0.005349708721041679,
-0.0027037523686885834,
-0.004330670461058617,
0.003922965843230486,
0.0024495127145200968,
-0.0067854677326977... | [
-0.48385679721832275,
0.22739127278327942,
0.42798736691474915,
-0.08259637653827667,
-0.04041551426053047,
0.5058163404464722,
-0.011842526495456696,
-0.206038236618042,
0.03234938904643059,
-0.6012596487998962,
0.2205582708120346,
0.3893078863620758,
0.030074343085289,
0.4401733875274658... |
So when choosing what language to use for a project, in an ideal world the language is chosen because it's the right tool for the job. However, I often prefer to use a language that I am _fluent_ in rather than one I would have to learn or that I am only conversational in. Of course language fluency also entails knowledge of the applicable libraries in the language. Just because I really like a fairly general-purpose language like Java doesn't mean I should always use it, but at the same time it doesn't mean I should break out something like Perl every time there's some text processing to be done. How does one find the balance here? | [
-0.0054512713104486465,
0.022205563262104988,
-0.006389806978404522,
-0.012156802229583263,
-0.00989718921482563,
0.0019169265870004892,
0.00567634217441082,
-0.00015640398487448692,
-0.01131104864180088,
-0.0007404959760606289,
-0.001402500318363309,
0.011643188074231148,
0.0078462697565555... | [
0.27274951338768005,
0.23050624132156372,
-0.17223060131072998,
0.08722289651632309,
-0.16572917997837067,
-0.13681045174598694,
0.5516360998153687,
0.4392035901546478,
-0.2713828682899475,
-0.5328859686851501,
-0.08167720586061478,
0.6129873991012573,
0.09639641642570496,
-0.3383348584175... |
As is visible in the screenshot below, my Sim has all their needs met and no negative moodlets, yet their happiness level is pretty much stuck to miserable. Is this a bug, and if so, how do I fix it?  | [
0.020197853446006775,
0.012063511647284031,
0.01833350583910942,
0.015171424485743046,
-0.023673834279179573,
-0.014793378300964832,
0.00555072957649827,
0.008042074739933014,
-0.016788965091109276,
-0.012626465409994125,
-0.028488412499427795,
0.01873289793729782,
-0.0038530693855136633,
... | [
0.0982760563492775,
0.0384589359164238,
0.29866957664489746,
0.17309297621250153,
0.2703849971294403,
-0.0721958801150322,
0.5329653024673462,
0.546592652797699,
-0.4489397704601288,
-0.5240852236747742,
-0.1474330872297287,
0.562443733215332,
-0.16202902793884277,
0.04935559630393982,
0... |
Here is an explanation of how NPS is calculated: http://en.wikipedia.org/wiki/Net_promoter_score I'm interested in testing two net promoter scores to determine if they are statistically different. I read a great answer to calculating margin of error for NPS (see link below), but I'm really interested in testing to see if there is a difference between two scores, because I suspect that our results aren't as "different" from year to year as they appear to be. How can I calculate margin of error in a NPS (Net Promoter Score) result? Is this at all possible? I understand t-tests are typically used to test whether two different record sets are statistically different. But is it possible to test Net Promoter Scores, either with a t-test or some other hypothesis test? Any ideas you have would be a great help. Thank you! | [
0.006861435249447823,
-0.0003811554051935673,
0.002650338923558593,
0.013553658500313759,
-0.023578863590955734,
0.010583354160189629,
0.007173202000558376,
-0.01786874420940876,
-0.023485524579882622,
-0.007590734865516424,
0.011824480257928371,
0.01348847895860672,
-0.00688913045451045,
... | [
0.7353085279464722,
0.0030024335719645023,
-0.024132100865244865,
0.23554188013076782,
-0.28061485290527344,
-0.1964152306318283,
-0.08364052325487137,
-0.1406710147857666,
-0.2466360181570053,
-0.3691859543323517,
0.5564265847206116,
0.224893257021904,
0.03570283576846123,
0.1428741514682... |
I have been designing a model plane for Design Technology for the past month or so, and today I laser cut my final design and assembled, it then tested it. Upon testing the plane does not get any lift, whereas the previous testing model which was virtually the same did. The plane is built using Balsa Wood, and Assembled with hot glue (I used as little glue as possible to reduce weight :) ) Any Ideas? Images: http://imgur.com/a/fdKxZ | [
0.003282431047409773,
0.0056771994568407536,
0.00047142189578153193,
0.006653496064245701,
0.01840820722281933,
-0.043884869664907455,
0.0071857343427836895,
0.0020227706991136074,
-0.011926005594432354,
-0.008124864660203457,
-0.010730545036494732,
0.007889730855822563,
-0.01142838317900896... | [
0.5254084467887878,
0.13195671141147614,
0.3603406548500061,
0.3143922984600067,
0.3754388689994812,
0.42922544479370117,
-0.042861126363277435,
-0.3317984342575073,
-0.13154689967632294,
-0.3232208788394928,
0.23042070865631104,
-0.007663165219128132,
-0.014834444038569927,
-0.06535193324... |
I have a binary classification problem from several features. Do the coefficients of a (regularized) logistic regression have an interpretable meaning? I thought they could indicate the size of influence, given the features are normalized beforehand. However, in my problem the coefficients seem to depend sensitively on the features I select. Even the sign of the coefficients changes with different feature sets chosen as input. Does it make sense to examine the value of the coefficients and what is the correct way to find the most meaningful coefficients and _state their meaning in words_? Are some fitted models and their sign of the coefficients wrong - even if when they sort-of fit the data? (The highest correlation that I have between features is only 0.25, but that certainly plays a role?) | [
0.030750922858715057,
0.019130486994981766,
-0.009181927889585495,
0.014276307076215744,
-0.031728871166706085,
-0.0020588086917996407,
0.00783638283610344,
-0.0020441480446606874,
-0.011543529108166695,
-0.006224276963621378,
-0.009140375070273876,
0.020593779161572456,
0.00445482088252902,... | [
-0.07093803584575653,
0.21372291445732117,
-0.14448273181915283,
0.044722478836774826,
-0.12605659663677216,
0.3813090920448303,
-0.125588521361351,
-0.1765977144241333,
0.0100505156442523,
-0.6339349150657654,
0.09443585574626923,
0.3754029870033264,
-0.20337098836898804,
0.48915803432464... |
According to neural sciences the brain process information and reacts in 1 teen of a seconds the numbers add up to be 10.596674 loest is 30 to 20 years the course of time in reality recorded by our time keeping devices. but whant to know if perhaps i estimated wrong or something or if it really is what it sound like and what i mean by this that like stars from far out galaxies we to are seeing what happen years. Back in the past | [
-0.007945994846522808,
0.012381266802549362,
-0.02504349686205387,
-0.020536968484520912,
-0.001433422788977623,
-0.00040965492371469736,
0.005587525200098753,
0.008379654958844185,
-0.007425189949572086,
0.008886413648724556,
0.004184321500360966,
0.005299901124089956,
0.012918863445520401,... | [
0.44809627532958984,
-0.14026416838169098,
0.5587400794029236,
0.12386930733919144,
0.3261847496032715,
0.2122952789068222,
0.04948508366942406,
0.5497045516967773,
-0.46606898307800293,
-0.7412734031677246,
0.4390810430049896,
0.2433614581823349,
0.3652082085609436,
0.6317325830459595,
... |
I'm using biblatex v2.5 and the biblatex-apa style v5.7. I have a couple citations with the same first author, including one with six co-authors in 2008. This is the _only publication_ with that first author that was published in 2008, so it should just show up as **Ranney, et al. (2008)** \- at least as I understand APA guidelines (section 6.12 / p. 175 in my 6th ed. copy). biblatex is instead printing it as **Ranney, Rinne, et al. (2008)** , which I assume is to avoid ambiguous author lists. But again, I _think_ this should only happen when you have the same first author on two pubs in a given year. This happens with `\nptextcite` and `\textcite`. The extra weird thing is that _other_ citations with this first author work out fine. For example, I have two pubs (with the same author list) correctly typesetting as **Ranney, et al. (2012a)** (and 2012b). But there are other cites with the same problem as well (e.g., **Ranney, Cheng, et al., 2001** , or **Munnich, Ranney, Nelson, et al. (2003)** ). So, I'm happy to do some hunting and hacking, but I'm not really sure where to start. In other words, please provide a partial answer even if you don't figure this one out! (I also actually _like_ the way biblatex is doing it, as I realize it helps me find references more easily. But it's not APA 6e per my reading.) | [
0.005529758520424366,
0.01015525497496128,
-0.022338436916470528,
0.015339775010943413,
-0.017793890088796616,
0.018358416855335236,
0.008234924636781216,
0.015849918127059937,
-0.016545020043849945,
-0.029129037633538246,
-0.005746721755713224,
0.004412192851305008,
-0.02197614498436451,
... | [
-0.09809727966785431,
0.42763563990592957,
0.3020411729812622,
-0.208737313747406,
-0.37604838609695435,
0.10080826282501221,
0.44734278321266174,
-0.29260432720184326,
-0.4734562039375305,
-0.3761832118034363,
0.14770644903182983,
0.24662187695503235,
-0.3509606719017029,
0.04165429249405... |
I guess it's always possible to hide them with js, but is there a code snippet out there for query level child page filter? I only want top level pages to remain. EDIT: my coder came up with it, stay tuned for a solution | [
0.022946568205952644,
0.016650596633553505,
-0.0075701577588915825,
0.015030723996460438,
0.009593067690730095,
0.020560424774885178,
0.00991873536258936,
0.023071708157658577,
-0.03802277892827988,
0.02315094880759716,
0.0067155328579247,
0.025088703259825706,
-0.019560927525162697,
-0.00... | [
0.1854771375656128,
-0.052615247666835785,
0.2651931047439575,
0.45220616459846497,
-0.04858855903148651,
0.023701686412096024,
0.5152332186698914,
0.22744791209697723,
-0.3242606520652771,
-0.39734843373298645,
0.008142259903252125,
0.26879197359085083,
-0.09845185279846191,
-0.0631573423... |
When I had my old laptop, I could run Minecraft quite fine, even with the Amazing Shaders mod. When I got my brand new laptop, I could already imagine the possibilities. 2000+ fps. No lag. I was going to be invincible with my NVIDIA GeForce 630 M. Or so I thought. When I opened Minecraft and made a "Test" world in Creative, I was very happy to see how fast things were running until I saw the fps when I hit F3, 93. 93? It didn't look like 93, but I continued to fly around my world relatively smoothly. One and a half minutes later, the "exponential lag" occurred. Two seconds, one frame. Four seconds, one frame. Eight seconds, one frame. This got so bad, it took a while just to exit the game! Is there any way I can play Minecraft fast ever again? Is it my graphics? CPU? RAM? Please help! **BASIC SYSTEM INFORMATION:** * Windows 7 64 bit OS * Intel Core i7 Quad Core CPU @2.6 GHz * nVidia GeForce GT 630M 1 GB * 8 GB RAM | [
-0.01519903726875782,
0.009158311411738396,
-0.007969912141561508,
-0.014267949387431145,
-0.020518384873867035,
-0.023741589859128,
0.005682654678821564,
-0.030935287475585938,
-0.016489073634147644,
0.000005245208740234375,
-0.008241458795964718,
0.009077806025743484,
0.022068167105317116,... | [
0.39152252674102783,
0.2890201210975647,
0.33834555745124817,
0.20710179209709167,
-0.4386572241783142,
-0.11401715129613876,
0.12432365119457245,
0.14247378706932068,
-0.15997178852558136,
-0.6190387606620789,
0.5194240808486938,
0.6890356540679932,
0.5143160223960876,
0.04171377792954445... |
I found a French army cliché; > “A friend when you’ re a first lieutenant, a companion when you’re captain, > a colleague when you’re major, a rival when you’re colonel, the enemy when > you’re general” introduced in a Japanese translation of “L’ étrange Defaire – Témoignage écrit en 1940,” written by French historian and résistant fighter, Marc Bloch (1886-1944), who was arrested and killed by the Nazis in June 1944, only two months before the liberation of Paris by the Allied Forces. I think it’s a very intriguing axiom to describe the nature of human race - the harder, the higher you climb up, which is common to the races / struggles in every field of politics, business, academy, sports, entertainment and you can name it. Is this an axiom proper to French? We have Japanese saying, “両雄並び立たず- Two heroes can never stand side by side (coexist),” which I think is akin to Chinese cliché, 両虎相闘 - liang hu xiang dou - meaning two tigers in a prairie are distined to fight to death, i.e. Caesar couldn’t stand together Pompeius, and Octavianus couldn’t live and let live Antonius. Both Japanese and Chinese cliché match only the last part of the French cliché. Is there English version to the same effect? | [
-0.014635387808084488,
0.021585267037153244,
-0.007904449477791786,
-0.0070602428168058395,
0.0022762329317629337,
0.009766677394509315,
0.008694948628544807,
-0.025255773216485977,
-0.013464557006955147,
-0.01290181279182434,
-0.011152759194374084,
0.01828247494995594,
0.005568815860897303,... | [
-0.11693532764911652,
0.06266102194786072,
-0.10584599524736404,
0.010740648955106735,
-0.19169940054416656,
0.15453478693962097,
0.5211600065231323,
-0.006288555450737476,
0.1297142505645752,
-0.15134601294994354,
-0.432684987783432,
0.2735663652420044,
0.2587912976741791,
0.0882264897227... |
For example: I have `01.txt` and `02.txt` files and need to append data of these files into new file `new.txt`. It should append data based on `01.txt` followed by `02.txt` files. Before this I have to delete only the first AND last line of the files( `01.txt` `02.txt` ) and then append them into the new file. How Do we accomplish this using UNIX? | [
-0.010648786090314388,
0.013691361993551254,
-0.009241890162229538,
-0.0007740050787106156,
0.0012947862269356847,
0.01009839866310358,
0.007065311539918184,
-0.0021658423356711864,
-0.014545275829732418,
0.00040114024886861444,
-0.015037347562611103,
0.004874452482908964,
-0.006573984865099... | [
0.31871649622917175,
0.040540553629398346,
0.2915898561477661,
0.16914141178131104,
0.134436696767807,
-0.1905776411294937,
0.1704220473766327,
-0.014643982984125614,
-0.1198548749089241,
-0.7348207831382751,
-0.19068969786167145,
0.7778363227844238,
-0.038477130234241486,
0.12626907229423... |
I have arch installed on my hd, but I can't establish a WPA2-encrypted network connection because wpa_supplicant is not installed. On my arch live-usb there is wpa_supplicant and there I can establish the network connection. Is there any way to tell pacman of my live arch to install the wpa_supllicant package to my installed arch on /mnt ? | [
-0.008588899858295918,
0.0008980424609035254,
-0.0069961100816726685,
0.022348150610923767,
-0.020357131958007812,
-0.011814448051154613,
0.01407710649073124,
-0.037049684673547745,
-0.021338123828172684,
-0.02789258025586605,
-0.020286070182919502,
-0.003436754224821925,
-0.0233923792839050... | [
0.13899633288383484,
0.19364061951637268,
0.4236694574356079,
0.19531990587711334,
-0.21758832037448883,
0.25434955954551697,
0.08254779875278473,
-0.22292080521583557,
0.023387445136904716,
-1.0208451747894287,
0.04061164706945419,
0.5673300623893738,
-0.2911437749862671,
0.15478540956974... |
I have been thinking of modeling human timing data using jags where the data comes from an experiment where participants tap in time with a very slow metronome. The data is then a number of measurements of how "off" the tap was compared to the metronome. The data could be thought of as coming from a normal distribution. This would then be mock-up data for 10000 taps (using R): timing_distribution <- rnorm(10000, 0, 300)  The problem is that when participants overshoot the target interval they instead react to the metronome tone. Say that reaction time is also from a normal distribution then mock-up data would be: reaction_time_distribution <- rnorm(10000, 250, 50)  The timing distribution and the reaction time distribution could then be though of as being combined into a joint distribution like this: joint_distribution <- pmin(timing_distribution, reaction_time_distribution)  That is whatever comes first the timing impulse or the reaction to tap after the metronome tone results in a tap. My question is how could one go about modeling this in jags/bugs? What I'm after is someth ing like this model { for( i in 1 : N ) { y[i] ~ min( dnorm( muTiming , tauTiming ), dnorm( muReaction , tauReaction )) } tauTiming ~ dgamma( 0.01 , 0.01 ) muTiming ~ dnorm( 0 , 1.0E-10 ) tauReaction ~ dgamma( 0.01 , 0.01 ) muReaction ~ dnorm( 0 , 1.0E-10 ) } But I guess that `y[i] ~ min( dnorm( muTiming , tauTiming ), dnorm( muReaction , tauReaction ))` is not really possible in jags... | [
0.010387791320681572,
0.004471883177757263,
-0.007834621705114841,
0.007143950089812279,
0.026095539331436157,
-0.028310663998126984,
0.00915128830820322,
-0.006160319782793522,
-0.01557376142591238,
-0.009625854901969433,
-0.006883399561047554,
0.012452887371182442,
0.010843086987733841,
... | [
0.6755618453025818,
-0.4331122934818268,
0.3619004189968109,
-0.01391319278627634,
0.13874216377735138,
0.3577542006969452,
0.04404301568865776,
-0.4749928116798401,
-0.3305665850639343,
-0.25905051827430725,
0.34324216842651367,
0.3492108881473541,
-0.25957241654396057,
0.0381684675812721... |
I'm making a custom Latex class which I am basing on the article class. When compiling a document I get `! LaTeX Error: No \title given.`, even though I use \title in the document. How can I solve this problem? | [
-0.00788371916860342,
0.0007343336474150419,
-0.015399272553622723,
0.01820407621562481,
0.008415691554546356,
0.022787628695368767,
0.011264966800808907,
0.01303461566567421,
-0.023438097909092903,
-0.021267957985401154,
-0.005544483661651611,
0.009929756633937359,
-0.002693192334845662,
... | [
0.1517307609319687,
0.30699166655540466,
0.18363086879253387,
0.12013828754425049,
0.2497977614402771,
-0.2187858372926712,
0.5323403477668762,
0.12210091948509216,
-0.030762406066060066,
-0.46471840143203735,
-0.1310623735189438,
0.6589292883872986,
-0.20586705207824707,
0.241880163550376... |
I'm developing site for art gallery. A have Artist CPT (internal name artsin_artist) which is defined in my functions.php like following: register_post_type('artsin_artist', array( 'labels' => $labels, 'public' => true, 'hierarchical' => false, 'supports' => array('title'), 'has_archive' => true, 'rewrite' => array('with_front' => true, 'slug' => 'artists'), 'publicly_queryable' => true, )); So i have good working archive http://artsindika.ru/artists/ and artist's page http://artsindika.ru/artists/arthur-ter-martirosov/ . Every artist has some number of works, works are not dedicated post type, but it is a simple associative array (i have used AdvancedCustomFields Repeater plugin for this feature). Now I need to have a page with following url structure artists/%artists_cpt_slug%/%work_index%. For example for second work of Arthur Ter-Martirosov i need to have following url structure: /artists/arthur-ter-martirosov/2. I tried to use a lot of recomendations about working with Wordpress rewrites, but i always have archive page as a result of routing. I mean that i need work_index query variable in my template to show approriate work. Now i am using following code for my rewrites: function add_artists_query_vars($aVars) { $aVars[] = "work_index"; return $aVars; } add_filter('query_vars', 'add_artists_query_vars'); function add_artists_rewrite_rule($aRules) { $aNewRules = array('artists/([^/]*)/([0-9]+)?$' => 'index.php?post_type=artsin_artist&post_name=$matches[1]&work_index=$matches[2]'); $aRules = $aNewRules + $aRules; return $aRules; } add_filter('rewrite_rules_array', 'add_artists_rewrite_rule'); Please tell me what I'm missing and what direction I need to move on? | [
0.001426956383511424,
0.010006140917539597,
0.006426098756492138,
0.017225254327058792,
0.01284908689558506,
0.0011911889305338264,
0.00674185948446393,
-0.012536757625639439,
-0.01155693456530571,
0.009397483430802822,
-0.01256383117288351,
-0.0019111617002636194,
0.00899591390043497,
0.0... | [
0.2762625515460968,
0.11196548491716385,
0.7737212181091309,
-0.07025405764579773,
-0.032276175916194916,
0.20366187393665314,
-0.05326814204454422,
-0.45228898525238037,
-0.2730548679828644,
-0.67226642370224,
0.11497975140810013,
0.2897961735725403,
-0.13642843067646027,
0.19840167462825... |
By newton's first law, an object will continue to be in the state of motion of rest until and unless an external force is acting on it. To prove or verify it experimentally today, is there any way out to diminish the problem of friction? | [
-0.016733014956116676,
0.029201095923781395,
-0.002305832924321294,
0.011008108034729958,
-0.023673709481954575,
-0.00018509609799366444,
0.011255244724452496,
-0.0137986084446311,
-0.018236203119158745,
0.006545743439346552,
-0.016799191012978554,
0.04331657290458679,
0.010222846642136574,
... | [
-0.11823694407939911,
-0.20421530306339264,
0.23828396201133728,
0.17510613799095154,
0.2256341278553009,
0.11463990062475204,
-0.028538238257169724,
-0.11539696156978607,
-0.4551936089992523,
-0.16368043422698975,
-0.09756699949502945,
0.38771355152130127,
-0.35419902205467224,
0.60141891... |
I noticed in the Prestige Shop there is an option for 10th level prestige where you can reset all stats. What is the purpose of it? The only use I could think of is using it after you've beat the multiplayer and just feel like beating it again for the hell of it. | [
0.018813855946063995,
0.014591943472623825,
0.0023394059389829636,
0.00559851573780179,
-0.012079590931534767,
-0.0017219387227669358,
0.00942812580615282,
-0.011847729794681072,
-0.021163873374462128,
0.019262906163930893,
-0.01152530312538147,
0.02773776836693287,
-0.01343343686312437,
0... | [
0.22443078458309174,
-0.1759263128042221,
0.051797054708004,
0.3833898901939392,
-0.23310308158397675,
0.006435585208237171,
-0.01210988312959671,
0.1590396761894226,
-0.5358204245567322,
-0.3258930444717407,
0.5130667090415955,
0.6718408465385437,
0.33661723136901855,
-0.17461052536964417... |
When I insert a figure created using GraphicsGrid into my (LaTeX) document, the axis are hardly visible. How can I increase thicknes of the axis in plot produced by Plot and LogLogPlot? | [
0.0008015318308025599,
0.008088765665888786,
-0.009560483507812023,
0.030722232535481453,
0.0013683752622455359,
-0.023315832018852234,
0.011754540726542473,
0.014822294935584068,
-0.018611591309309006,
-0.03438747301697731,
0.001041456707753241,
0.017389141023159027,
-0.010563639923930168,
... | [
0.36219748854637146,
0.006868545897305012,
0.8041964769363403,
0.28878021240234375,
-0.011655794456601143,
0.16199611127376556,
-0.13032490015029907,
0.025690553709864616,
-0.10832767933607101,
-0.482123464345932,
0.141566202044487,
0.2257116436958313,
-0.007491095457226038,
0.268617630004... |
Any help is most appreciated. Thanks! \documentclass{article} \usepackage{amsmath} \usepackage{mathtool} \begin{document} \begin{align} \begin{split} & v \cdot \left\{ Ax + b \\ & Cx - d \right\} \end{split} \end{align} \end{document} | [
0.003870210377499461,
0.004129794426262379,
-0.004182721488177776,
0.015222696587443352,
-0.007371131796389818,
0.006884534377604723,
0.006316436920315027,
0.01440601609647274,
-0.012719343416392803,
0.020579742267727852,
-0.005230572074651718,
0.00015398001414723694,
-0.003952564671635628,
... | [
-0.343687504529953,
0.10290868580341339,
0.42259523272514343,
0.6545714139938354,
0.092340387403965,
-0.34139829874038696,
0.49147555232048035,
0.2590448558330536,
-0.01814911514520645,
-0.6428648829460144,
-0.30113673210144043,
0.3191670775413513,
0.3344613015651703,
-0.06763030588626862,... |
I have installed a multi-site installation of WordPress onto my domain. I then added the necessary code to the `wp-config.php` file and `.htaccess` as instructed by WordPress. I also installed a plugin called Quick Page/Post Redirect Plugin which allowed me to place a 301 redirect onto the main domain as I only want to use the sub domain and not the main domain. Then I also added the following line of code to the `wp-config.php` file to redirect the main domain `define( 'NOBLOGREDIRECT', 'URL Redirect Address' );` The site works fine with a redirect on the main domain and my subdomain runs fine when you type in `subdomain.example.com` or `http://subdomain.example.com`. However when I enter `www.subdomain.example.com` or `http://www.subdomain.example.com` the following error message is returned: > Not Found > > The requested URL / was not found on this server. > > Apache/2.4.9 (Unix) Server at www.subdomain.domain.com Port 80 Any help with this would be much appreciated. | [
-0.0009594755247235298,
0.003951184917241335,
0.003456076141446829,
0.01640968769788742,
0.0051258886232972145,
0.004336453974246979,
0.008044615387916565,
0.010665202513337135,
-0.009544579312205315,
-0.025720465928316116,
-0.007308501750230789,
0.010594856925308704,
-0.0011755144223570824,... | [
0.5973068475723267,
0.11263671517372131,
0.4340946078300476,
-0.2922307252883911,
-0.09516250342130661,
-0.0652821809053421,
0.3307192623615265,
-0.17099982500076294,
0.22552074491977692,
-0.7661265134811401,
0.13563638925552368,
0.5305288434028625,
-0.2889237701892853,
-0.0749449804425239... |
This Post: "Progressive" JPEG: Why do many web sites avoid rendering JPEGs that way? Pros, cons? stat that many browsers do not support Progressive JPEG citing this Wikipedia Article. What I have not found is which browsers do and do not support it. I know FF 7 supports it but I have not tested the other browsers to know. Does anyone have a comparison chart or something similar? Bonus Question: What e-mail clients support Progressive JPEG? (Desktop E-Mail Clients e.g. Outlook & Thunerbird etc) * * * So I decided to run few test to see how browsers handled Progressive JPEGs. Here is what I found: * Firefox 7 - Displayed Correctly * Internet Explorer 7 & 8 - Displayed Photo once finished downloaded (Did not load scan by scan) * Safari 5 - Displayed Photo once finished downloaded (Did not load scan by scan) * Chrome 13 - Display Correctly * Opera 11.5 - Displayed Correctly If anyone has any comments to my test or any thoughts please let me know (Including how older browsers of the ones that did load it correctly work). I am still wondering about email clients if anyone can help with that. | [
-0.0031599269714206457,
-0.007424256298691034,
-0.004902215674519539,
0.02002173662185669,
0.011565133929252625,
-0.01874440535902977,
0.0072700828313827515,
0.024245958775281906,
-0.01691018044948578,
-0.00002399226650595665,
-0.005964125506579876,
0.018292948603630066,
-0.00060701277107000... | [
0.3563900291919708,
0.21325169503688812,
0.1972292810678482,
0.31147465109825134,
-0.815216600894928,
-0.12389089912176132,
0.2910063862800598,
0.03580697998404503,
-0.2830808460712433,
-0.32878100872039795,
-0.10745539516210556,
0.24820829927921295,
-0.2434280514717102,
0.3536176681518554... |
I am making a book with the `fancyhdr` package. I have some headers that put page numbers on the top corners of the pages. On pages that are started by `\chapter`, LaTeX moves the line number down to the middle of the bottom of the page and removes it from the top. (Apparently it does something like changing the `pagestyle` to `plain` for a single page.) That's fine, but I want to do the same thing on a certain single page that is not a chapter beginning. What is the command that LaTeX is using to move clear the header and show the page number in the bottom of the page for a single page? | [
0.012469684705138206,
0.013843998312950134,
-0.011598714627325535,
0.026124900206923485,
0.018203984946012497,
-0.0036000912077724934,
0.007538239471614361,
-0.002701997058466077,
-0.015483168885111809,
0.015057321637868881,
-0.014014600776135921,
-0.00014302623458206654,
-0.0033217263408005... | [
0.5003571510314941,
0.20761638879776,
0.7276841998100281,
0.1741868555545807,
-0.10758654773235321,
-0.07232881337404251,
0.003624429926276207,
0.09303136169910431,
0.007847806438803673,
-0.2622357904911041,
-0.05039013922214508,
-0.004603676497936249,
-0.16881202161312103,
0.3747921288013... |
I have a text classification problem, where there are many different classes, and the text to be classified is very short (about 1 sentence each): this is sentence one, label1 this is sentence two, label2 ... I am using the words in the the text as features with a naive bayes classifier. When computing:  the naive bayes classifier just tabulates counts. However, I wanted to weight the count for each word:  Where the weight for a word/class pair $w_i, c_j$ is determined as follows: 1. Take the subset of the training data where $w_i$ occurs 2. Using this subset, compute the entropy on the distribution over class labels. The higher this entropy value the more "disordered" the class labels are for the given word $w_i$ (meaning this is a bad word with respect to classification), and vice versa. Normalize this by the maximum possible entropy. 3. Take the inverse of the above (now words with a low entropy, good words, have higher value) 4. Multiply the above by the fraction of times $c_j$ occurs in the subset mentioned in (1). The reason for this step, is because the quantity computed in (3), inverse normalized entropy, will be high if the word associates with a certain class, no matter what class. However, we don't want the weight to be high for a word/class pair if the word has a high inverse entropy which is high due to some other class we are not concerned with. # What went wrong? I felt that this weighting mechanism would improve results: the weight for a word/class pair is high when a word is informative of that class. However, after implementing the above, my results were worse than the normal naive bayes classifier with no weights! I am not sure why, any insight would be highly appreciated! (maybe someone with sharper math skills than me can point out a potential flaw?) | [
-0.002072147559374571,
0.013515391387045383,
0.000842341803945601,
0.020857732743024826,
-0.022255368530750275,
-0.00625215470790863,
0.009243296459317207,
-0.01235747616738081,
-0.016706092283129692,
0.034190937876701355,
-0.011568859219551086,
0.011020265519618988,
-0.00154945719987154,
... | [
0.11351758986711502,
0.3366447985172272,
0.2324262112379074,
-0.2778177559375763,
-0.3104185461997986,
0.3499387800693512,
0.2768068015575409,
-0.24037671089172363,
-0.2359543889760971,
-0.7186419367790222,
0.012037497945129871,
0.2875930964946747,
-0.022494113072752953,
0.0151923792436718... |
I have an Oracle 11g database setup with Oracle Locator. I have created a table with the this query: CREATE TABLE geotest (id NUMBER(38) NOT NULL, shape SDO_GEOMETRY); Then I inserted one point with the this query: INSERT INTO geotest VALUES(1,SDO_GEOMETRY(2001,3395,SDO_POINT_TYPE(-111.949439,40.722283,NULL),NULL,NULL)); The result is no data in the query layer in ArcMap. It allows me to create the layer, and validates. But then it makes me go through the advanced properties which I suspect means it can't see any of the rows to automatically pull the SRID. Any guidance would be appreciated! **Update** : So when I insert the metadata into the table using this: insert into user_sdo_geom_metadata values ('geotest','shape',sdo_dim_array(sdo_dim_element('X',-180,180,0.005),sdo_dim_element('Y',-180,180,0.005)),4326); And creating the spatial index, I can see the following point inserted into the database using this query: insert into geotest values(1,'test',sdo_geometry(2001,4326,sdo_point_type(0,0,0),null,null)); So it would appear that a point with all 0s work, but anything beside that doesn't. I haven't found anything in the documentation that relates to this, or even any users complaining about that. Any either tricks that may be useful? **Update 2** : Alright it appears that after I drop the index, and create it again, the points then show up. I am not an Oracle expert, so this confuses me, anyone have an explanation for this behavior? I thought once an index was created, as soon as a point is inserted, the index updates. | [
-0.0012901670997962356,
0.009623860940337181,
-0.006664795335382223,
0.006283339112997055,
-0.00997778307646513,
0.038587912917137146,
0.00851127877831459,
0.02048341929912567,
-0.009220471605658531,
-0.009797562845051289,
0.0011391561711207032,
0.008909327909350395,
-0.011886931024491787,
... | [
-0.2903214395046234,
0.09447185695171356,
0.287055104970932,
0.06403829902410507,
-0.07499197870492935,
0.3423745632171631,
-0.046754390001297,
-0.11796262115240097,
0.06116486340761185,
-0.8168324828147888,
0.1671714186668396,
0.17328916490077972,
-0.24605967104434967,
0.21676187217235565... |
I have a use case in which I need to parse the XML generated by an ArcGIS server, then push that data to another API. The ArcGIS server sits behind the firewall at our customer's site and exposes a REST through which we can query the database. The REST URL looks something like this: http://SERVER_NAME/arcgis/rest/services/APP_NAME/MapServer I plan on stepping through the DOM and extract the records in Java (probably using Xerces). All the fields except the geometry fields are pretty straightforward. The geometry fields are defined as type `esriFieldTypeGeometry` with the data being stored as base64 binary string. My question is: how do I parse the binary data in Java and convert it to a set of coordinates? Here is an extract of the XML: <esri:Workspace xmlns:esri='http://www.esri.com/schemas/ArcGIS/10.1' xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance' xmlns:xs='http://www.w3.org/2001/XMLSchema'> <WorkspaceDefinition xsi:type='esri:WorkspaceDefinition'> ... </WorkspaceDefinition> <WorkspaceData xsi:type='esri:WorkspaceData'> <DatasetData xsi:type='esri:TableData'> ... <DatasetType>esriDTFeatureClass</DatasetType> <Data xsi:type='esri:RecordSet'> <Fields xsi:type='esri:Fields'> <FieldArray xsi:type='esri:ArrayOfField'> ... <Field xsi:type='esri:Field'> <Name>SHAPE</Name> <Type>esriFieldTypeGeometry</Type> <IsNullable>true</IsNullable> <Length>4</Length> <Precision>0</Precision> <Scale>0</Scale> <Required>true</Required> <DomainFixed>true</DomainFixed> <GeometryDef xsi:type='esri:GeometryDef'> <AvgNumPoints>0</AvgNumPoints> <GeometryType>esriGeometryPoint</GeometryType> <HasM>false</HasM> <HasZ>false</HasZ> <SpatialReference xsi:type='esri:ProjectedCoordinateSystem'> <WKT>PROJCS[...]</WKT> <XOrigin>-20037700</XOrigin> <YOrigin>-30241100</YOrigin> <XYScale>10000</XYScale> <ZOrigin>0</ZOrigin> <ZScale>1</ZScale> <MOrigin>0</MOrigin> <MScale>1</MScale> <XYTolerance>0.001</XYTolerance> <ZTolerance>0.001</ZTolerance> <MTolerance>0.001</MTolerance> <HighPrecision>true</HighPrecision> <WKID>102656</WKID> <LatestWKID>2234</LatestWKID> </SpatialReference> <GridSize0>3575.0162431554045</GridSize0> </GeometryDef> <AliasName>SHAPE</AliasName> <ModelName>System-maintained Geometry for a feature</ModelName> </Field> </FieldArray> </Fields> <Records xsi:type='esri:ArrayOfRecord'> <Record xsi:type='esri:Record'> <Values xsi:type='esri:ArrayOfValue'> ... <Value xsi:type='esri:PointB'> <Bytes>AQAAAMAnD4uRGzJBAFYObYdpJ0E= </Bytes> </Value> </Values> </Record> </Records> </Data> </DatasetData> </WorkspaceData> </esri:Workspace> Thanks in advance for any suggestions you guys can give. | [
-0.010884321294724941,
0.005442277528345585,
-0.007626584265381098,
0.008309359662234783,
-0.008356448262929916,
0.01702108234167099,
0.009307997301220894,
0.009020423516631126,
-0.018139371648430824,
-0.02249494381248951,
0.013821111992001534,
0.005645358469337225,
-0.01496849488466978,
0... | [
0.1376710832118988,
0.14162540435791016,
0.32141175866127014,
0.046888768672943115,
-0.06623812764883041,
-0.07533548027276993,
-0.15972621738910675,
-0.09964912384748459,
-0.1306937038898468,
-0.7424513697624207,
0.13358600437641144,
0.2589453458786011,
-0.00902384240180254,
-0.0694322660... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.