text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
Consider the following example: \documentclass{article} \usepackage{pstricks-add} \def\sandsynlighed#1#2{% \psframe[ linecolor=red, fillstyle=solid, fillcolor=blue!50 ](#1,0)(!#1 \bredde add #2 4 div) } \psset{unit=0.75} \begin{document} \def\bredde{1.5 } \begin{pspicture}(-0.9,-0.5)(14.3,11.3) % Akserne. \psaxes[ xlabelOffset=-0.75, Ox=-1, dx=1.5, xLabels={0,1,2,3,4,5,6_{+0},6_{+1},7}, Dy=2, dy=0.5 ]{->}(0,0)(-0.25,-0.25)(14,10.75)[$X$,0][$100 \cdot P(X)$,90] % Sandsynlighederne. \sandsynlighed{0}{18.70} \end{pspicture} \end{document}  As can be seen, the labels on the _x_ -axis is all wrong and I have no idea what I have done wrong. I would like to have the labels as specified by `xLabels`; how do I get this? | [
-0.004796049557626247,
0.01260313205420971,
-0.010968293063342571,
0.02749931812286377,
-0.0015101253520697355,
-0.002927611581981182,
0.007845018990337849,
0.008725711144506931,
-0.009878411889076233,
0.0004896579775959253,
-0.014752320945262909,
-0.0023600012063980103,
-0.02166573889553547... | [
0.04888363182544708,
-0.44388800859451294,
0.684879720211029,
-0.2443169206380844,
0.22945046424865723,
0.0939180999994278,
0.011062774807214737,
-0.4382578134536743,
-0.07955741882324219,
-0.541473388671875,
-0.13412974774837494,
0.23056121170520782,
-0.3726005256175995,
-0.05903631448745... |
I'm currently playing Minecraft beta 1.5, and I started working on a trail system. I'm using gravel for the trails and it looks nice. Unfortunately, I've run out of gravel, and I can't find gravel anywhere but deep underground. I've checked my surrounding area for gravel beaches, but they I have yet to find one. How can I easily find gravel? | [
-0.009143137373030186,
0.011785181239247322,
-0.010709814727306366,
0.006642934866249561,
-0.0010103612439706922,
0.0006171179702505469,
0.006332117598503828,
0.007100570481270552,
-0.02674219384789467,
-0.015384708531200886,
-0.0013360449811443686,
0.013645544648170471,
-0.00777068734169006... | [
0.28538256883621216,
0.30516499280929565,
0.10572893917560577,
0.14738401770591736,
0.5269918441772461,
-0.3831535875797272,
0.4220941662788391,
0.19342486560344696,
-0.3364033102989197,
-0.5663856863975525,
0.337311327457428,
0.3170825242996216,
0.3837414085865021,
0.005752436351031065,
... |
The following MWE is producing different outputs on XeLaTeX and LuaLateX: \documentclass{scrartcl} \usepackage{fontspec} \setmainfont{FoglihtenNo04-070.otf} \begin{document} ffi ew \end{document} XeLateX:  LuaLaTeX:  Is this the intended outcome? The font I chose for testing came from http://www.glukfonts.pl/font.php?l=de&font=FoglihtenNo04. | [
0.0171955693513155,
-0.0010718259727582335,
-0.008996021933853626,
0.004415726289153099,
0.01121637225151062,
0.015403443947434425,
0.009172449819743633,
0.01880348101258278,
-0.008599396795034409,
-0.011737964116036892,
-0.007157356943935156,
-0.00624238234013319,
0.0049444036558270454,
0... | [
-0.0606054924428463,
0.08379393815994263,
0.5387613773345947,
-0.40663379430770874,
-0.19415295124053955,
0.4121594727039337,
0.03332216292619705,
-0.6358758211135864,
-0.34897279739379883,
-0.3195841312408447,
0.013036917895078659,
0.49714377522468567,
-0.28190287947654724,
0.440676063299... |
The Facebook Messenger offers this nice little feature called 'chat heads'. For that, it requests the permission to overlay everything. However, i read somewhere about the fact, that it can also see what is displayed on the screen. Can you tell me if that is true? | [
-0.011717458255589008,
-0.017360452562570572,
-0.00015588129463139921,
0.006400825455784798,
-0.021836623549461365,
0.003601105185225606,
0.008196116425096989,
0.0014396487968042493,
-0.023740123957395554,
0.009423104114830494,
-0.006741679273545742,
0.014278963208198547,
0.03492669016122818... | [
0.7428442239761353,
-0.024062104523181915,
0.2466907948255539,
0.19712309539318085,
-0.0012202977668493986,
-0.38082772493362427,
0.45230644941329956,
0.637094259262085,
-0.4721549153327942,
-0.26634445786476135,
0.23771411180496216,
0.26845604181289673,
-0.18615937232971191,
-0.2926420569... |
I'm trying to insert a custom field (from the Advanced Custom Fields plugin), and I need to place it within a div which I'm wrapping around an img, right before the img-tag. So basically this is the result I want (I use `the_field('description')` here to make it cleaner): <div class="the-content"> <div class="owl-wrapper"> <div class="owl-item"> [ the content from the_field() ] <img src"http://imagesource"> </div> </div> </div> But this is what happens: <div class="the-content"> [ the content from the_field() ] <div class="owl-wrapper"> <div class="owl-item"> <img src"http://imagesource"> </div> </div> </div> I.e. content from the field gets inserted at the top instead of in the div I place it. Isn't it rather strange? Here's my code: function wrapImagesInDiv($content) { if ( in_category( 'projects' ) ) { $pattern = '/(<img[^>]*class=\"([^>]*?)\"[^>]*>)/i'; $replacement = ' <div class="owl-wrapper"> <div class="owl-item $2"> '.the_field('description').' $1 </div> </div>'; $content = preg_replace($pattern, $replacement, $content); } return $content; } add_filter('the_content', 'wrapImagesInDiv'); | [
-0.006423738319426775,
-0.0063724881038069725,
-0.0099564203992486,
0.008145272731781006,
-0.024791933596134186,
0.01301744394004345,
0.007462254725396633,
0.016118116676807404,
-0.014198306947946548,
0.0331464521586895,
-0.008458612486720085,
0.0007856248412281275,
0.003357943147420883,
0... | [
0.347222238779068,
-0.0235599335283041,
0.3794398307800293,
-0.17167961597442627,
-0.0970197319984436,
-0.08316868543624878,
-0.06097682937979698,
-0.464381605386734,
0.014164090156555176,
-0.48192721605300903,
-0.10808190703392029,
0.540756106376648,
-0.4400888681411743,
0.098948910832405... |
My widget has a lot of dependent plugins, and disabling one of the plugins results in erratic behavior throughout my entire WP installation. I've been trying to use the function is_plugin_active('Widget_Class_Name') with little to no luck thus far. Is there an effective way to remove the widget altogether if one of its dependent plugins are inactive? | [
0.009098779410123825,
0.0037394871469587088,
-0.007608277723193169,
0.0241048876196146,
-0.012363998219370842,
0.0017418005736544728,
0.008588358759880066,
-0.0016004135832190514,
-0.018650274723768234,
-0.0015597458695992827,
-0.0111205093562603,
0.011784848757088184,
-0.0016744188033044338... | [
0.42800456285476685,
-0.01957344077527523,
0.21051794290542603,
-0.24083149433135986,
0.04033474251627922,
-0.07839950174093246,
0.6209980845451355,
0.05030416324734688,
-0.1909550279378891,
-0.40533703565597534,
0.057672567665576935,
0.38200920820236206,
-0.5135822892189026,
0.34791332483... |
When I charge my GoPro with a wall-charger I'm able to still use it and take photos, but when I connect it to my Linux computer I can't. I want to be able to programmatically control how my computer is connected to this device. So I can charge it and switch to the storage connection mode to manage files on it. Is it possible to **only** charge the device, so it won't know it's connected to the computer? | [
-0.012912077829241753,
-0.00948923360556364,
-0.0023166572209447622,
0.018118424341082573,
-0.01870642602443695,
-0.040514782071113586,
0.010630323551595211,
0.003102254820987582,
-0.02164645865559578,
-0.0035934599582105875,
-0.001203385996632278,
0.016594184562563896,
0.020531900227069855,... | [
0.4123724699020386,
-0.008668237365782261,
0.4615773856639862,
0.195199653506279,
0.4145093560218811,
-0.18480850756168365,
0.024268552660942078,
0.01211540400981903,
-0.3996896743774414,
-0.3594571650028229,
-0.12440462410449982,
0.788362443447113,
-0.5082889199256897,
-0.0315166451036930... |
The sample "TOCControlContextMenu" demonstrates how to display a context menu from the AxTOCControl. When I run this sample and right-click on the TOC control the context menu pops up and remains displayed until I select a menu item. In my application the right-click displays the context menu but as soon as I release the right-mouse-button the first item in the menu is selected. I can see no substantial differences in the code. Has anyone observered (and possibility overcome) this behavior before? | [
-0.009201908484101295,
0.001443897490389645,
-0.0151730477809906,
0.008804120123386383,
0.017279379069805145,
-0.023270078003406525,
0.009111769497394562,
0.010578124783933163,
-0.02089819498360157,
0.021521564573049545,
-0.030631173402071,
0.01725533790886402,
-0.0049438863061368465,
0.02... | [
-0.020963285118341446,
0.09730972349643707,
0.29599738121032715,
-0.1718658208847046,
-0.0763099193572998,
-0.05853640288114548,
0.11781417578458786,
-0.3136727809906006,
0.17837421596050262,
-0.3720194101333618,
-0.009189583361148834,
0.6250000596046448,
-0.14166347682476044,
-0.069324225... |
I hope I'm not missing anything obvious. Instead of trying to elaborate on my title, let me illustrate with an MWE: \documentclass[twocolumn]{article} \makeatletter \setcounter{secnumdepth}{4} \newcounter{subsubsubsection} \newcommand\subsubsubsection{% \@startsection{subsubsubsection}{4}{\z@}{-3.25ex\@plus -1ex \@minus -.2ex}% {2ex \@plus .2ex}{}} \makeatother \begin{document} \section{Section} \subsection{Sub-} \subsubsection{Subsub-} \subsubsubsection{subsubsub-} \end{document} _**Result:_**  _**Remarks:_** 1. The problem goes away if I use `\subsubsubsection*`, but still exists if I use `\subsubsubsection` without changing `secnumdepth`. 2. Yes, I'm aware of the `\paragraph` command: I don't want to use it because what I'm using `\subsubsubsection` for are not paragraphs. If you would like to suggest some other less repetitive name, I'm open to suggestions. What's going on here? | [
0.003321500029414892,
0.020107027143239975,
-0.0026015439070761204,
0.02141376957297325,
0.008976996876299381,
0.0032845023088157177,
0.007524565793573856,
-0.004244580399245024,
-0.01047675684094429,
0.0062810699455440044,
-0.015758464112877846,
0.0037400114815682173,
0.005749151110649109,
... | [
0.08215949684381485,
0.13643915951251984,
0.3747696280479431,
-0.068464495241642,
0.04796728864312172,
-0.21954959630966187,
-0.022737858816981316,
0.15710891783237457,
-0.27107298374176025,
-0.5695719718933105,
0.05770125612616539,
0.15586471557617188,
-0.04512945935130119,
0.177163615822... |
My current Project is to "map" a turntable in qgis. I know, very unlikely but that's what my teacher wants. My teacher also wants me to measure something in this project (Very specific, isn't it?). And that's where the problem starts. Qgis wants me to define geographical coordinates or something like that. The problem is, that i don't have any coordinates because it's a picture of a turntable and not an actual map. Any ideas? Greeting from Germany. | [
-0.018931252881884575,
-0.0005114342784509063,
0.01027290616184473,
0.01870850659906864,
-0.006352119147777557,
-0.007141231093555689,
0.009698417969048023,
0.004489895887672901,
-0.020568745210766792,
-0.007675121072679758,
-0.005847550462931395,
0.01263515930622816,
-0.00872593093663454,
... | [
0.4596109092235565,
0.1166497990489006,
0.3918841779232025,
0.269512802362442,
0.08904825150966644,
-0.27336132526397705,
0.1835612803697586,
0.7102797627449036,
-0.3124125897884369,
-0.8907658457756042,
-0.021426653489470482,
-0.472871869802475,
0.1282792091369629,
0.1738358587026596,
0... |
I have just managed to get my site running behind a Cloudfront distribution using the custom distribution API, but now i cannot login to the site using the standard old address (since i screwed around with DNS). The way i have it hooked up is as follows: backendblog - real server blog - cloud front distribution told to get data from backendblog visiting blog results in the page being returned (and quite fast too!) but none of the dynamic stuff (search, login, etc) work. comments work grand since they are managed by Disqus. So, i have 2 questions around this: 1: how do i set it so that backendblog will allow me to login without redirecting to blog (if i set the config, it changes all the URLs... want the URLs to be what ever site the page loads on...) 2: how do i configure pages to have higher expire times so they stay in CloudFront longer? Any ever try this? the site, for reference, is blog.lotas-smartman.net (clodfront) or backendblog.lotas-smartman.net (dedicated box in Germany). have seen Pingdom avg response times drop massively in the last week with cloudfront... | [
-0.005086520686745644,
-0.008497112430632114,
-0.012175075709819794,
0.005831239745020866,
-0.010087184607982635,
-0.01090580876916647,
0.010041000321507454,
0.008663838729262352,
-0.015897568315267563,
-0.036005955189466476,
-0.00956219993531704,
0.013744426891207695,
0.005464606918394566,
... | [
0.43349310755729675,
0.13586345314979553,
0.4866502583026886,
0.13631948828697205,
-0.035285353660583496,
-0.18711000680923462,
0.3657383620738983,
0.0988321602344513,
-0.24883031845092773,
-0.6117371916770935,
0.20493508875370026,
0.5454651117324829,
-0.16195671260356903,
0.40294528007507... |
Given a web application running across 10+ servers, what techniques have you put in place for doing things like altering the state of your website so that you can implement certain features. For instance, you might want to: * Restrict Logins/Disable Certain Features * Turn Site to "Read Only" * Turn Site to Single "Maintenance Mode" page. Doing any of the above is pretty trivial. You can throw a particular "flag" in an .ini file, or add a row/value to a `site_options` table in your database and just read that value and do the appropriate thing. But these solutions have their problems. ## Disadvantages/Problems For instance, if you use a file for your application, and you want to switch off a certain feature temporarily, then you need to update this file on all servers. So then you might want to look at running something like ZooKeeper, but you are probably overcomplicating things. So then, you might decide that you want to store these "feature" flags in a database. But then you are obviously adding unncessary queries to each page request. So you think to yourself, that you will throw memcached in to the mix and just cache the query. Then you just retrieve all of your "Features" from memcached and add a 2ms~ latency to your application on every page. So to get around this, you decide to use a two tier-cache system, whereby you use an inmemory cache on each machine, (like Apc/Redis etc). This would work, but then it gets complicated, because you would have to set the key/hash life to perhaps 60 seconds, so that when you purge/invalidate the memcached object storing your "Features" result, your on machine cache is prompt enough to get the the new states. What suggestions might you have? Keeping in mind that optimization/efficiency is the priority here. | [
0.0052974652498960495,
0.014578945934772491,
-0.003588553052395582,
-0.0023488986771553755,
-0.004642089828848839,
0.006548697128891945,
0.006416807882487774,
-0.014619524590671062,
-0.009677071124315262,
0.01827183924615383,
-0.007604843936860561,
0.006917683873325586,
-0.01347776222974062,... | [
0.37212616205215454,
-0.01879844069480896,
0.3777998685836792,
0.15811781585216522,
-0.0494801364839077,
-0.3824259042739868,
0.5960985422134399,
-0.29076892137527466,
-0.42416656017303467,
-0.7569582462310791,
-0.2189992517232895,
0.39124640822410583,
-0.3811739981174469,
0.08902613818645... |
A very common programming data structure is a stack. A stack is simply a one dimensional array. Stack elements are pushed in or popped out of the stack one at a time as LIFO (Last In First Out).  I am trying to build a general routine to represent such stacks using LaTeX2e. The stack has two operations a pop and a push. The popped elements are stored in a macro `\popped@element`, so that they can be pushed into another stack if necessary or typeset material. At this point I provided a`\before@pop@hook` and an `\after@pop@hook`, to enable surrounding material to be added. Not sure if this correct from a design point. To get material in an `fbox` the construction became very weird (see lines 71-72): \def\before@pop@hook{\bgroup\color{blue}\fbox} \def\after@pop@hook{\egroup} **I am looking for improvements/corrections to the code to make it more robust and accept any type of input such as verbatim text, ability to be inserted into tables, i.e, which sections of the code would need to become`outer` etc.** So far I have come up with this MWE: \documentclass{book} \usepackage[latin]{babel} \usepackage{xcolor,lipsum} \parindent0pt \makeatletter \newcommand\lorem{Fusce adipiscing justo nec ante. Nullam in enim. Pellentesque felis orci, sagittis ac, malesuada et, facilisis in, ligula. Nunc non magna sit amet mi aliquam dictum. In mi. Curabitur sollicitudin justo sed quam et quadd.} % Define a new stack by letting it to \@empty \newcommand\newstack[1]{% \let#1\@empty } % #1 stack name % #2 element contents \newcommand{\add@element}[2]{% \def\element{#2}% \push@element{#1} } % #1 stack name \newcommand{\push@element}[1]{% \xdef#1{\element+#1} } %% Add hooks here for typesetting \def\before@pop@hook{\bgroup\color{red}} \def\after@pop@hook{\egroup} \def\before@pop@hook{} \def\after@pop@hook{} \long\def\pop@#1+#2\@nil#3{% \def\popped@element@{\before@pop@hook{#1}\after@pop@hook} \def\popped@element{#1} % \xdef\element{#1} if required? % remaining list \def#3{#2}% } % stack name #1 \newcommand\pop@element[1]{% \ifx #1\@empty Error\else \expandafter\pop@#1\@nil#1 \fi } \begin{document} % Create two stacks \newstack{\stack} \newstack{\tempstack} % add some elements to stack \def\elt@start{Start }\def\elt@stop{Stop} \add@element{\stack}{english} \add@element{\stack}{polish} \add@element{\stack}{australian} \add@element{\stack}{german} \add@element{\stack}{\elt@start greek \elt@stop} % pop some elements from \stack % and put them in boxes Popped from stack \pop@element{\stack} \popped@element@ \pop@element{\stack} \popped@element@ % add last popped element from \stack to \temp % pop it into blue box \def\before@pop@hook{\bgroup\color{blue}\fbox} \def\after@pop@hook{\egroup} \add@element{\tempstack}{\popped@element} \parbox{3cm}{\popped@element@} \def\X{german} \def\Y{\popped@element} \ifx\X\Y \add@element{\stack}{\popped@element}\fi In temp stack \tempstack \def\before@pop@hook{} \def\after@pop@hook{} \add@element\stack{\lorem} \pop@element{\stack} \parbox{3.9cm}{\popped@element@} \end{document} **LaTeX3 and Lua solutions are also welcome, provided they come with long explanations.** _**Update_** It is a pity that I cannot accept all the answers as they are all great and examined the issues from all aspects. I was especially impressed with Frank's and Joseph's answers and their explanations. cjorssen's Lua solution is very promising and can be very useful, if one wanted to parse the contents further. I have accepted Bruno's answer as it was the one with a MWE and clearly demonstrating the use of verbatim text in lists. | [
-0.020336110144853592,
0.012832715176045895,
-0.0031955661252141,
-0.0048580714501440525,
0.04054894670844078,
-0.01577412337064743,
0.006839988753199577,
0.028382407501339912,
-0.014416221529245377,
0.024084296077489853,
0.0013515865430235863,
0.0009623577352613211,
0.015065420418977737,
... | [
-0.02330273948609829,
-0.11211513727903366,
0.2522667646408081,
-0.062369439750909805,
-0.013974540866911411,
0.08365420997142792,
-0.29247528314590454,
-0.10421375185251236,
-0.4805087447166443,
-0.27462515234947205,
-0.3925342559814453,
-0.04651973769068718,
-0.4095767140388489,
0.094852... |
My TeXworks doesn't let me compile my documents anymore due to the error: `The program miktex-pdftex.exe was not found!`:  It is OK in TeXnicCenter but not here. I looked up the net to download the missing part, but couldn't find it. Any thoughts? | [
0.011239411309361458,
-0.00022547975822817534,
-0.0013101527001708746,
0.017141830176115036,
0.04220093414187431,
0.0040635885670781136,
0.006834259256720543,
0.01171739399433136,
-0.019479263573884964,
-0.016251850873231888,
-0.01766553893685341,
0.009048575535416603,
0.010232705622911453,
... | [
0.2046481966972351,
0.4538065493106842,
0.5299754738807678,
0.08972710371017456,
0.4736349284648895,
-0.4678299129009247,
0.4848008453845978,
0.39132076501846313,
-0.20694409310817719,
-0.5980292558670044,
-0.011366989463567734,
0.7257516384124756,
-0.04097090661525726,
0.09526277333498001... |
I have been playing BF3 on my Xbox since it came out, and am now thinking about buying the PC version so I can play it on my laptop when I'm not at home. I am aware that my game stats are kept with EA, so am I right in thinking that if I play BF3 on the PC (or PS3 for that matter), it will retain my rank and stats providing I log in with the same EA Origin account? | [
0.0021518843714147806,
0.01257028616964817,
-0.00005076602246845141,
0.009738036431372166,
0.01961948163807392,
-0.013887207955121994,
0.005963106174021959,
-0.009095770306885242,
-0.020666707307100296,
0.0019693439826369286,
0.013843869790434837,
0.03148075193166733,
0.0006207582191564143,
... | [
0.30396851897239685,
0.1967550814151764,
0.08239556849002838,
-0.1804598718881607,
0.07493198662996292,
0.03199217468500137,
-0.07432959228754044,
-0.011325011029839516,
-0.31681305170059204,
-0.29787302017211914,
0.18859398365020752,
0.4623343050479889,
0.10456274449825287,
-0.14387407898... |
Can I use ProFTPd without using a `chroot` jail (thereby preventing access to anything outside of the FTP root)? I have a requirement to have symlinks in my FTP source that point to locations outside of the directory where I root my FTP service. All of the docs and discussion I've read on ProFTPd talk about how to use the chroot functionality (even within StackExchange), but I'm wondering if I can bypass using that and use a different method to serve my FTP root. Since the symlinks must remain as symlinks, mounting the directories as a way of bypassing the chroot restriction (the clever "solution" to the problem) does not work. | [
0.01412847451865673,
0.015389308333396912,
0.005554782226681709,
0.01901181787252426,
-0.019339678809046745,
-0.029915932565927505,
0.009195788763463497,
0.0014991732314229012,
-0.021620824933052063,
-0.017213353887200356,
-0.023233450949192047,
0.012323256582021713,
-0.017612528055906296,
... | [
0.3213995099067688,
0.013474209234118462,
0.4623458981513977,
-0.1522054225206375,
0.1782696545124054,
-0.41644737124443054,
0.3828428387641907,
-0.10492763668298721,
-0.25911542773246765,
-0.40187567472457886,
-0.12060176581144333,
0.46316155791282654,
-0.4262100160121918,
0.2463528811931... |
I'm using OSM2PO to route for walking directions and I've noticed a lot of the OSM walking paths are not ideally setup. For example there are many paths which get divided up by parking lots, which are obviously walkable, but they are ignored and not converted to roads, so the router believes you can't walk through them. This results in much longer walking distances than necessary. Is there any practical way to deal with this? I guess the same problem occurs when dealing with any "areas" that are walkable (ie. parks without paths through them). Another issue is where sidewalks do not connect to roads. In many situations this again leads to sub par routing results. OSM has these issues all over the place. It sounds like a difficult problem, but a common one, so I'm wondering if anyone has addressed this in a practical way. Thanks for your thoughts! | [
0.0006394940428435802,
0.013815152458846569,
-0.023022515699267387,
0.020369119942188263,
-0.03774525970220566,
-0.0018688086420297623,
0.008546404540538788,
0.019824214279651642,
-0.015666063874959946,
-0.009407326579093933,
-0.021647021174430847,
0.013220416381955147,
0.003944825381040573,... | [
0.4235977530479431,
0.16988690197467804,
0.20242349803447723,
0.21715393662452698,
-0.08433512598276138,
-0.31683358550071716,
0.2882950007915497,
-0.042415037751197815,
-0.4050789177417755,
-0.8914346098899841,
0.2175322324037552,
0.23183509707450867,
-0.3078289330005646,
-0.3413772881031... |
I tried to visualize the graph generated by osm2po and rendered that graph in qgis. Now what I am trying to do is plot the vertices in the graph. So that I can visually see where the nodes/vertices are. I have attached a screenshot of the visualization I have up to now.  I want to actually see the nodes in this I followed the suggestion as given by underdark. However, I could see that the number of nodes and edges is less than that in the database. In the database the number of nodes = 32 and the number of edges = 39. However in my case it is 31 nodes only and 38 edges only. What could be the reason?  | [
0.0036823477130383253,
0.00338973430916667,
0.0037399479188024998,
0.00918403547257185,
-0.028592027723789215,
-0.023516669869422913,
0.005986702628433704,
0.021122176200151443,
-0.020395483821630478,
0.017261872068047523,
-0.0025304043665528297,
0.008560349233448505,
-0.022163450717926025,
... | [
0.16892559826374054,
-0.02015366591513157,
0.3464742600917816,
-0.10997729748487473,
-0.26817774772644043,
-0.05673784390091896,
0.10809242725372314,
-0.15023212134838104,
-0.3192143738269806,
-0.8319784998893738,
0.16660276055335999,
0.34012898802757263,
-0.1822873055934906,
0.22567832469... |
For example if we have two options to use non-linear classifier like SVM with kernel or use linear classifier like linear SVM with data preprocessing like some non-linear dimensionality reduction which one is better? In other words why we should use a complicated classifier if we can do some data preprocessing (or "projection" if I am not mistaken it's called manifold learning) to make data linearly separable? | [
0.02037668228149414,
0.015613673254847527,
0.006616630125790834,
0.018590591847896576,
-0.03606121242046356,
0.006916605401784182,
0.012469356879591942,
-0.009845361113548279,
-0.013129903003573418,
0.0010193631751462817,
-0.01248025894165039,
0.013371188193559647,
0.008780739270150661,
0.... | [
-0.2999977171421051,
-0.5851148366928101,
-0.18298548460006714,
0.6643655896186829,
-0.2274622917175293,
-0.14138579368591309,
-0.4050914943218231,
-0.03641751781105995,
-0.16926611959934235,
-0.7535879015922546,
0.04540275037288666,
0.7347455620765686,
-0.175401508808136,
-0.0048235408030... |
I own two lines (different carriers), and one of them is currently on my Galaxy S4. I am thinking of purchasing another (open) android device to place the other SIM on. Since I have purchased apps from the Play Store, I wish to install them on the new android device as well. Obviously, I have different contacts on each phones and I want to keep it that way. Will it be possible to have different sets of contacts under one google account? | [
-0.006550502963364124,
0.01404389925301075,
-0.005309377331286669,
0.009790302254259586,
0.004130196757614613,
0.010112416930496693,
0.008257281966507435,
0.02141805924475193,
-0.019471311941742897,
-0.05831092596054077,
0.0050843688659369946,
0.015012828633189201,
0.017731664702296257,
0.... | [
0.6774019002914429,
-0.06523352116346359,
0.8331671953201294,
-0.11169010400772095,
0.1413571536540985,
0.6972513794898987,
0.10778409242630005,
0.2932671308517456,
-0.29223284125328064,
-0.6222382187843323,
0.36657461524009705,
0.33735039830207825,
-0.3169674873352051,
0.1631147712469101,... |
Every time I'm using this macro it shifts to the right. Why? \newcommand*{\addLine}[1]{ \begin{tikzpicture}[overlay, remember picture] \coordinate (x) at (#1,0); \draw [open triangle 45-] (0,0) -- (x); \end{tikzpicture} } **Edit** Well, I adapted my real macro (I just provided a smaller example for the sake of demonstration), but I still have a little gap: \newcommand*{\AddNote}[4]{% \begin{tikzpicture}[overlay, remember picture]% \coordinate (x) at (#2,0); \coordinate (a) at ($(x)!(#1.north)!($(x)+(0,1)$)$); \coordinate (b) at ($(a)+(0.5,0)$); \coordinate (c) at ($(b)+(0,#3)$); \draw [open triangle 45-] (a) -- (b) -- (c); \node[right] at (c) {\bf\sffamily\smaller#4}; \end{tikzpicture}% } | [
-0.0013639219105243683,
0.010973826050758362,
-0.010282766073942184,
0.01565234735608101,
0.002244148403406143,
-0.006290746852755547,
0.00653312262147665,
-0.0028148938436061144,
-0.01679142564535141,
-0.0111378263682127,
-0.025126997381448746,
-0.0007249906193464994,
-0.0030889324843883514... | [
0.23213735222816467,
-0.13196216523647308,
0.6122927069664001,
-0.1921396255493164,
0.16442573070526123,
0.136836975812912,
-0.20863083004951477,
-0.07796840369701385,
-0.3847896456718445,
-0.8368790149688721,
0.3751605451107025,
0.5874943137168884,
-0.48858460783958435,
0.0523101165890693... |
Is Skype available for the Motorola Xoom? | [
0.020676566287875175,
0.025528596714138985,
-0.03680254891514778,
0.0032737639266997576,
-0.08216465264558792,
-0.021851859986782074,
0.017990373075008392,
0.05583982542157173,
-0.022625379264354706,
-0.12929093837738037,
-0.053687822073698044,
0.08174408227205276,
0.010648929513990879,
0.... | [
0.4269035756587982,
-0.2773972749710083,
0.1251489371061325,
0.6439062356948853,
0.24125349521636963,
0.1162436455488205,
0.2877305746078491,
0.2509821951389313,
-0.22724317014217377,
-0.327079713344574,
-0.049461401998996735,
0.3331318497657776,
-0.2274494171142578,
-0.6096236705780029,
... |
The title might be a bit vague while I'm not quite sure how to describe it. **What I want:** At every single post (in this case of post type 'event') a logged in user must be able to click a 'Join' button (and when clicked a 'Unjoin' button). The admin must be able to see which users have joined an event and which event. It doesn't matter if this must be a custom field at the User profile info or a new list in the admin. **What I already have:** I have made a custom checkbox list of posts at the User profile info. And I do know how to make a button in at a single post. But it must be connected to this list, so when it's clicked the specific post must be checked at the Users profile page. I'm not sure what the easiest method is to make this, any idea? It's actually a quite simple function. Thanks in advance. | [
0.005941042676568031,
0.006890208460390568,
0.01038098894059658,
0.007142194546759129,
0.0068435585126280785,
-0.0018047629855573177,
0.006147164851427078,
0.022504223510622978,
-0.012243288569152355,
0.03776726871728897,
-0.0034446294885128736,
0.012889199890196323,
0.021963002160191536,
... | [
0.776807427406311,
0.15037323534488678,
0.56329345703125,
0.021279718726873398,
-0.19713523983955383,
-0.4170754551887512,
0.0838848277926445,
0.1710415780544281,
-0.31815922260284424,
-0.3726300895214081,
-0.03653144836425781,
0.12944117188453674,
0.039628565311431885,
0.3560647666454315,... |
I would like to use a full section title for: * the section title itself (of course) * the table of contents * the running heads But, I would like a short name to be used with `\nameref{}`, because otherwise the reference is too long to work into the surrounding text effectively. (The reader finding the section is not a problem because the sections are in alphabetical order and the short version will be the beginning of the section title.) Nameref uses the short title if one is provided, but then the short title ends up in the table of contents as well, which I don't want. Is there a way to get the table of contents to use the long titles, leaving the short title for nameref? Alternatively, if there's a way to specify a different title than the section title itself for nameref to use, that would work as well. I found many other questions and tips about using a short version in page headers and a long version in the table of contents, but nothing that involved `\nameref{}`. The page header is not a problem as I'm using `memoir` and can easily define the page header separately (as the same as the full title, in this case). MWE: \documentclass[twocolumn]{memoir} \usepackage{hyperref} \headnameref % headers \copypagestyle{maudcyclo}{plain} \createmark{section}{both}{shownumber}{\sectionname}{} \nouppercaseheads \makeevenhead{maudcyclo}{\oldstylenums{\thepage}}{\scshape\leftmark}{} \makeoddhead{maudcyclo}{}{\scshape\rightmark}{\oldstylenums{\thepage}} \makeevenfoot{maudcyclo}{}{}{} \makeoddfoot{maudcyclo}{}{}{} \makeevenhead{plain}{\oldstylenums{\thepage}}{\leftmark}{} \makeoddhead{plain}{}{\rightmark}{\oldstylenums{\thepage}} \makeevenfoot{plain}{}{}{} \makeoddfoot{plain}{}{}{} % don't number chapters and below \setcounter{secnumdepth}{-1} \pagestyle{maudcyclo} % use header style \begin{document} \tableofcontents \clearpage \section[Section One][S1]{Section One} \label{sec1} Here is some text for section one. \clearpage \section[Section Two][S2]{Section Two} Here I will reference \nameref{sec1}. \end{document} The above MWE works properly except for the fact that it places `S1` and `S2` in the header, whereas I want `Section One` and `Section Two` there. | [
0.013186430558562279,
0.017771773040294647,
-0.009034322574734688,
0.02566484361886978,
-0.0014098589308559895,
0.01090905349701643,
0.008344515226781368,
0.02346549741923809,
-0.016286902129650116,
-0.023825164884328842,
-0.0098189078271389,
0.0013271502684801817,
-0.0032544489949941635,
... | [
0.1054234728217125,
0.2453337162733078,
0.3964148163795471,
0.017756540328264236,
-0.0007782093016430736,
-0.33001142740249634,
0.13990047574043274,
0.25675052404403687,
-0.25303226709365845,
-0.4809478521347046,
-0.42133522033691406,
0.32762205600738525,
-0.06393060833215714,
0.0156428664... |
This is the code I'm using and the image comes up but with the file location next to it \begin{figure} \begin{center} \includegraphics[scale=0.8]{C:/Users/Lauren/Google Drive/japan_all_6.pdf}} \caption{Plotted data showing a quadractic trend for Japan over $6.0$ $1900-1980$} \end{center} \end{figure} | [
0.0019479201873764396,
0.008411312475800514,
0.0038638643454760313,
0.015202201902866364,
0.027132652699947357,
-0.012320773676037788,
0.0052928500808775425,
0.007329920772463083,
-0.012452269904315472,
0.004876069724559784,
-0.005628679413348436,
0.01081807166337967,
-0.0005100698908790946,... | [
0.15633349120616913,
-0.12841713428497314,
0.5335645079612732,
0.18089231848716736,
0.15887701511383057,
0.09995832294225693,
-0.0636286661028862,
-0.011605164036154747,
-0.3353842794895172,
-0.2959858179092407,
-0.0454905666410923,
0.20710580050945282,
-0.13309870660305023,
0.220501288771... |
how can we deal for example with the product of distributions in physics ?? is there any mean to define with physics $ \delta ^{2}(x) $ or to treat a product of two distributions within the Renormalization framework ? another question let be a metric with a dirac delta derivative $ ds^{2}= g_{a,b}dx^{a}dx^{b} $ and for example $ g_{x,x}= \delta (x-y) $ and $ g_{x,y}= H(x-y) $ (heaviside function) how could we compute and define the Einstein Equations ?? $ R_ {a,b}=0 $ | [
0.0026936400681734085,
0.01900039240717888,
-0.011430959217250347,
0.010925398208200932,
-0.012958550825715065,
-0.013473015278577805,
0.008007093332707882,
-0.0009773452766239643,
-0.014937665313482285,
-0.012380403466522694,
-0.008038641884922981,
0.009636569768190384,
-0.00818363577127456... | [
0.11206690967082977,
-0.36818182468414307,
0.2704991400241852,
0.1483740210533142,
-0.11490515619516373,
0.1442607194185257,
-0.1935483068227768,
-0.4019640386104584,
0.142658993601799,
-0.7064530849456787,
-0.08286850899457932,
0.5313571691513062,
-0.28913965821266174,
0.24251046776771545... |
I need to round numbers that are in my table in CartoDB, so that in the Map visualization infowindows I only show results to two decimal digits rather than the long decimals that are currently showing. I tried using `SELECT round(column,10,2) FROM table` as suggested on this webpage How can I show numbers as percentages in CartoDB? but the SQL query results in no records, and I get the following error: `function round(double precision, integer) does not exist` Is there another way to reformat the numbers so they display fewer decimals in the CartoDB infowindows without editing the source data? | [
0.0128865335136652,
0.0037321620620787144,
-0.009513244032859802,
0.004778888542205095,
-0.04958787560462952,
-0.006791488267481327,
0.008301946334540844,
0.008726827800273895,
-0.0185073409229517,
-0.015587318688631058,
-0.017984339967370033,
0.0019542877562344074,
-0.026988616213202477,
... | [
-0.06402833759784698,
0.22261148691177368,
0.2595560550689697,
0.01410682499408722,
-0.17124639451503754,
0.3378269672393799,
-0.07025932520627975,
-0.33072909712791443,
-0.3906177282333374,
-0.4316800832748413,
0.275342732667923,
0.5247411131858826,
0.07194551080465317,
0.1051770672202110... |
Sakurai mentions that the propagator is a Green's function for the Schrodinger equation because it solves $$\left(H-i\hbar\frac{\partial}{\partial t}\right)K(x,t,x_0,t_0) = -i\hbar\delta^3(x-x_0)\delta(t-t_0)$$ I don't see that. First of all, I don't understand where the $-i\hbar$ comes from. And if I recall correctly, a Green's function is used to solve inhomogeneous linear equations, yet Schrodinger's equation is homogeneous $$\left(H-i\hbar\frac{\partial}{\partial t}\right)\psi(x,t) = 0$$ i.e. there is no forcing term. I do understand that the propagator can be used to solve the wave function from initial conditions (and boundary values). Doesn't that make it a kernel? And what does Sakurai's identity mean? | [
0.003008730709552765,
0.007402087561786175,
-0.0007170331664383411,
0.007739664986729622,
-0.007152036298066378,
0.0013189304154366255,
0.00535300187766552,
-0.005122354254126549,
-0.008085917681455612,
-0.007838737219572067,
-0.009456106461584568,
0.005437836982309818,
-0.01232084445655346,... | [
-0.12679949402809143,
-0.5005097985267639,
-0.18935228884220123,
0.1849018931388855,
-0.49648287892341614,
0.4148348271846771,
0.16339081525802612,
-0.6505470871925354,
-0.0032678204588592052,
-0.012020174413919449,
0.007113201078027487,
0.1047753095626831,
-0.4694521725177765,
0.413150578... |
I am developing a plugin for WordPress 3.5. My blog's default title is "Test". Here is some code of my plugin: add_filter('wp_title', 'my_replace_title'); function my_replace_title(){ return get_option('my_site_title'); } The new title is "my website title". But it's out putting the result: "my website titleTest" How to replace old title with new title? | [
0.004697361495345831,
0.013133334927260876,
-0.005847604479640722,
0.0056032887659966946,
-0.021783845499157906,
0.011995968408882618,
0.00944345910102129,
0.0022514064330607653,
-0.011083518154919147,
-0.01405584067106247,
-0.01918414980173111,
0.008784400299191475,
0.018597939983010292,
... | [
0.5449615120887756,
0.08106780797243118,
0.5940029621124268,
-0.1430996209383011,
-0.035227544605731964,
-0.3690377175807953,
0.3656253218650818,
0.07972047477960587,
0.26183685660362244,
-0.7447583079338074,
0.08235841244459152,
0.4269890785217285,
-0.11619137972593307,
0.2033287137746811... |
First, note that $4^{96}\equiv96\ (mod\ 100)$. _Mathematica_ claims that `PowerMod[96, 1/96, 100]` has no integer solutions. Even more obviously wrong, I get `{}` for Solve[4^96 == x^96, Modulus -> 100] which obviously has at least one solution (4). Is this a known bug? Can I work around it somehow? EDIT: I'm on _Mathematica_ 9.0.1, running on OSX 10.10 Yosemite developer beta. Perhaps that could be the cause of this issue for me (also, mathematica crashes whenever I try to paste.). | [
0.012308396399021149,
0.004199347924441099,
-0.014148803427815437,
0.010561224073171616,
-0.019032573327422142,
0.011258767917752266,
0.005510758142918348,
-0.007678221445530653,
-0.017728380858898163,
-0.010975977405905724,
-0.008384838700294495,
0.014928021468222141,
-0.017123993486166,
... | [
-0.14769503474235535,
0.15260836482048035,
0.37847068905830383,
0.06637450307607651,
0.11583368480205536,
-0.3162062466144562,
0.35123494267463684,
0.2457600086927414,
-0.1483820080757141,
-0.6763368844985962,
-0.14346717298030853,
0.6020472049713135,
-0.5621790289878845,
-0.06916692107915... |
I'm solving an exercise about small oscillations and I have a doubt about coordinates that I have to use. This is the text of the exercise: "A bar has mass M and lenght l. Its extremity A is hooked to a coil (with lenght at rest $l_0$), its extremity B is hooked to the point O that is the origin of axes. " I have considered three coordinates: $x$, $y$ (that are the coords of the extremity A on the x-axes and y-axes) and $\theta$ that is the angle that the bar forms with a parallel to the y-axes. I have to find the points of equilibrium. I have written the coordinates of the center of mass of the bar as: $M=(x+l/2 \sin \theta, -y-l/2 \cos \theta)$ and the potential energy as $ V=\frac{1}{2}k(x^2+y^2)-Mg(y+\frac{l}{2}\cos \theta)$ Then I have posed $gradV=0$ and I have obtained: $y=\frac{Mg}{k}$ $x=0$ $\theta=0, \pi$ My doubt is about the result of y.. I was waiting for a negative value.. could you "clarify" my ideas and tell me where I'm making a mistake? | [
0.00814978126436472,
0.012657003477215767,
-0.010268625803291798,
0.0003246520645916462,
-0.018749430775642395,
0.004859083332121372,
0.007841691374778748,
0.007138550281524658,
-0.01396896317601204,
-0.04393751174211502,
-0.004227564670145512,
0.015491830185055733,
-0.019902918487787247,
... | [
0.3351026177406311,
-0.0253087617456913,
1.05848228931427,
-0.10066613554954529,
0.003087193239480257,
0.2819926142692566,
-0.05330636724829674,
-0.3904559910297394,
-0.28331464529037476,
-0.16181612014770508,
0.2467346340417862,
0.16263563930988312,
-0.11376474797725677,
0.304492026567459... |
this is a repost from math.stackexchange (I didn't know how to migrate the post, apologies). I was informed that this would be a better place to ask. * * * TLDR: I am trying to do maximum likelihood fitting of a dataset having two mixed populations, observed over a subset of their parameter space, within it to two pdfs. I include working code with gaussian examples. * * * Hello, I am new to this site. My question (and example code) is also long, so I will host some information in pastebin. I apologize, I am neither a mathematician nor a computer scientist, so some things may be ugly or inefficient. I have a set of data which has samples of two populations within it (we have a good prior of what percentage of the data belongs to each population). Neither sample probes the full spatial extent of the data, but the probability density function (pdf) is presumed to be smooth. I use a wide variety of pdfs, but for this discussion I will utilize only Gaussians for simplicity-- however, I manually normalize the Gaussians, since my pdfs are not normalized. We can think of the data as being sampled from two different Gaussians, with some kind of cut off boundaries, beyond which the population continues, but the sampling ends. The boundaries of the two populations intersect for a range. * * * The metacode would be something along the lines of: A) Make a dataset by sampling two pdfs with known parameters, within two different given ranges. # $$ pdf = gaussian = e^{\frac{-1}{2}(\frac{x-mu}{sigma})^{2}} $$ B) Guess some parameters --1) Find the normalizing constants ($n_{1}$ and $n_{2}$) for these parameters by integrating the Gaussians of this guess analytically over the observed space. $$ n = \sum{volume*density} $$ $$ density = pdf(x) :: volume = step size $$ --2) For each item in the dataset: $$ Probability (p)= \frac{ratio}{n_{1}}*pdf_{1}(item) + \frac{1-ratio}{n_{2}}*pdf_{2}(item) $$ --3) Sum the log likelihoods for all data to find the total likelihood of the guess $$ L = \sum{log(p)} $$ C) Use some likelihood landscape explorer to converge on a best answer (here I use brute grid for simplicity) * * * I have had great success disentangling two populations if they are sampled in their entirety, but poor success disentangling two populations that have been truncated (the answers are frequently nonsensical and boundary dominated). The code (~100 lines, but not dense code) will run as is. It follows the metacode above and successfully finds the parameters of the two Gaussians sampled over (essentially) their entire area. However, changing lines 31 and 32 to: range1=[-2,6] range2=[2,12] for example, causes the algorithm to find the incorrect values. My question is why. Thank you for your time. | [
0.013476255349814892,
0.011976111680269241,
0.005154475104063749,
0.010158461518585682,
0.0025131856091320515,
-0.012989640235900879,
0.0044777123257517815,
-0.007319070864468813,
-0.01192687638103962,
-0.0023002519737929106,
-0.00812636036425829,
0.014695221558213234,
-0.009449278935790062,... | [
0.455546110868454,
0.03362084552645683,
0.11199615895748138,
0.06215885654091835,
-0.3711196780204773,
-0.07584907859563828,
0.37898972630500793,
0.18167252838611603,
-0.21119476854801178,
-0.7327685952186584,
-0.09775733202695847,
-0.10783664137125015,
-0.4098038077354431,
0.2151270508766... |
What is the "definition" of a reality constraint and why is it called that way? (I mean how it is used for example in quantum field theory and string theory) | [
-0.009530924260616302,
0.01564783789217472,
-0.0002572558878455311,
0.011872528120875359,
-0.014814460650086403,
-0.004856404848396778,
0.014797749929130077,
-0.026503508910536766,
-0.018390441313385963,
-0.008593308739364147,
-0.020976945757865906,
0.01750168576836586,
-0.009485626593232155... | [
0.32099848985671997,
-0.25403785705566406,
0.2021501213312149,
0.4400339126586914,
-0.014549055136740208,
-0.24289944767951965,
0.24131059646606445,
0.12041905522346497,
-0.6177023649215698,
-0.16520321369171143,
-0.1127789095044136,
0.5545918941497803,
-0.5458852648735046,
0.5691365599632... |
How can I automatically save site's copy each week with an option to browse saved copies? I need kind of WebArchive, but on my local computer. | [
-0.01655336283147335,
-0.0015246175462380052,
-0.04102962836623192,
0.01891240105032921,
0.005261123180389404,
0.007199310231953859,
0.012269246391952038,
0.04045495018362999,
-0.03558540716767311,
0.05086342617869377,
-0.00927073135972023,
0.021850628778338432,
-0.0032780205365270376,
0.0... | [
0.5008572340011597,
0.10746161639690399,
0.30407437682151794,
0.23336458206176758,
-0.05319869890809059,
-0.4589639902114868,
0.3275339901447296,
0.21902671456336975,
-0.2950461208820343,
-0.673549234867096,
0.16431710124015808,
0.4922647178173065,
0.14640110731124878,
0.05954136326909065,... |
I have read many online articles. I've read questions and answers on this site. I still can't get my head wrapped around the difference between `past simple` and `present perfect` I know the difference is between finished time (use `past simple`) and unfinished time (use `present perfect`). But as a non-native English speaker, this still makes no sense to me. I am **not implying any kind of time period.** > I have posted on SE > I posted on SE I just want to convey that the action completed in the past. I don't want to provide any connotation whether the action just completed, or completed at certain period in past. Maybe with **"I"** , I know the time period, but what if I am talking about someone else. > He has posted on SE > He posted on SE I don't know the time frame that **"he"** did the action. I don't know what **"he"** was planning (continuous action, or at specific time). I just know the action took place in the past. I don't know if it just finished. I don't know if there is a consequence to the action or not. It seems to me that I am **forced** to imply a time frame. So my question is: if I don't want to imply any time frame, or at least to imply as little as possible, what should I use? | [
0.016181671991944313,
0.020185362547636032,
-0.00825437344610691,
0.0064043402671813965,
0.0025124340318143368,
-0.0016042664647102356,
0.009102952666580677,
0.01712639629840851,
-0.018884869292378426,
0.028100470080971718,
-0.012857992202043533,
-0.001876896247267723,
0.025618525221943855,
... | [
0.25167348980903625,
0.08612793684005737,
0.08782770484685898,
-0.01982017047703266,
-0.05147966742515564,
0.5279781222343445,
0.6889838576316833,
0.013816436752676964,
-0.37502551078796387,
-0.5792696475982666,
-0.04308106005191803,
0.13283772766590118,
0.3095514476299286,
0.5554122924804... |
Keyword research tools like Keyword Planner seem to fulfill two basic functions: 1. Generate a list of possible keywords 2. Provide estimates (CPC, traffic, ...) to whittle down this list to the most _effective_ keywords Do I need the second step? Is there any downside in uploading a huge list with thousands of keywords and just wait and see how they perform? It's pay per click so I'm not losing money on low performing keywords. Ultimately I'm only interested in conversions and that's a metric that can't be estimated by the tools anyway. Edit: As Joshak points out, I need to remove all obviously non-converting keywords. What about other keywords that could theoretically convert but the estimates show it's unlikely. For example 0 traffic or very high cpc, so it's unlikely I will win the bid. It would be more work to remove them and there are slight changes that they bring conversions. Is there any downside in using them? | [
0.01954926736652851,
0.014601603150367737,
-0.005599222145974636,
0.01320231705904007,
-0.02186024934053421,
-0.00383282289840281,
0.0061334362253546715,
0.0006413070950657129,
-0.016787048429250717,
0.011306237429380417,
-0.013353402726352215,
0.009955257177352905,
0.001710990909487009,
0... | [
0.15204210579395294,
0.12995655834674835,
0.18403950333595276,
0.30197271704673767,
-0.2895628809928894,
0.015149963088333607,
0.06839215010404587,
-0.24524036049842834,
0.010677912272512913,
-0.5431211590766907,
0.11626775562763214,
0.6238957643508911,
-0.04583850875496864,
-0.11399543285... |
Whenever I play TankBall 2, I can't help but notice the mysterious numbers at the top left of the screen...  I figured out pretty quickly that the red bar with the time in it represents the amount of time remaining in the game, but I can't make heads or tails of the numbers above it. How should I read this interface? | [
-0.01597544178366661,
0.004425035789608955,
-0.004346491303294897,
-0.00014404703688342124,
0.00045178382424637675,
-0.0182399433106184,
0.006849988829344511,
0.010834639891982079,
-0.012377569451928139,
0.03295598924160004,
-0.02291957475244999,
0.0026588854379951954,
0.015045851469039917,
... | [
0.03675283491611481,
0.0361487977206707,
0.37884044647216797,
-0.10513971745967865,
-0.35346537828445435,
-0.28810983896255493,
-0.08440878242254257,
0.202921062707901,
-0.3062672019004822,
-0.6081598401069641,
0.12526832520961761,
0.5726766586303711,
-0.02475675567984581,
-0.0098158000037... |
Is there a quick way to disable all comments on all posts and pages. I know that you can go into the dashboard under posts, click edit, and Do Not Allow, under comments. Is that the only way? | [
0.04191264510154724,
-0.0040628486312925816,
-0.021636294201016426,
0.020037507638335228,
0.00987751130014658,
-0.001006732345558703,
0.009699980728328228,
-0.017409393563866615,
-0.03609587624669075,
-0.012381504289805889,
0.014529626816511154,
0.0014199744910001755,
-0.005509129259735346,
... | [
0.3181896209716797,
0.2282828539609909,
0.225430428981781,
0.3907988667488098,
0.2786475718021393,
-0.06846893578767776,
0.15507082641124725,
0.34879162907600403,
-0.2336570918560028,
-0.36813998222351074,
0.31621357798576355,
0.5577025413513184,
-0.12225198745727539,
-0.25516441464424133,... |
Searched all over and tried playing with httpd.conf , nothing works. The domain name can be resolved with www.domain.com but without www its does not resolve. Advice for A Records does not work. I have a domain from Registrar otwohosting.net there panel is like publicdomainregistry.com and have options for nameservers. I have pointed nameservers, but with mx toolbox, only www.domain.com looks up and fails for non www What is the probable cause ? Here is snippet from httpd.conf <VirtualHost *:80> ServerAdmin webmaster@example.com DocumentRoot /var/www/html/mydomain.com/ ServerName www.mydomain.com ServerAlias mydomain.com RewriteEngine On RewriteRule ^/(.*) http://mydomain.com/$1 [L,R=301] ErrorLog logs/dummy-host.example.com-error_log CustomLog logs/dummy-host.example.com-access_log common </VirtualHost> Its centos dedicated server, with firewall and yes port 80 is opened for tcp and udp both. Additionally any change in document root explictly for www.domain.com also dont work. As default doc root is `/var/www/html/` but i do want it to resolve it to `/var/www/html/mydomain.com/` directory . Please help thanks **EDIT** Here is the link for my dns report [Link Removed] It gives strange result saying it can fetch dns records but cant resolve domain.. i am laughing and tearing down my head at same time **Edit2** Following the below answer here is the update: 1) I do have 3 Nameserver received from my web hosting company . Although i have unmanaged dedicated server, but nameservers are not ns1.mydomain .. they are like `ns1.mywebhost.com`,`ns2.mywebhost.com` and have been feeded At Registrar panel 2) created CNAME Record for `Host:www` , `Value:mydomain.com` - didnt worked 3) I do have a dedicated IP Available which i should use for creating A record - Did not worked 4) After performing these actions , domain is still not resolving. Even with www also it failed, which was working earlier. I'm running in a big trouble UPDATE 3: Finally the solution by closetnoc worked. I used domain Registrars' Nameservers, with an Additional A Record and CNAME record to my Dedicated IP and domain name respectively. Although the DNS propagation is very slow with my registrar (otwowebhosting.net a reseller of PDR ) , and i will plan to switch to dedicated DNS hosting service soon. Thanks | [
-0.024118125438690186,
-0.00026554602663964033,
-0.006479664705693722,
0.014576191082596779,
-0.02412521094083786,
0.006403343752026558,
0.009082380682229996,
0.02205757610499859,
-0.017884083092212677,
-0.007662883494049311,
-0.004128801636397839,
0.010271942242980003,
0.0006409849738702178... | [
0.08192168921232224,
0.013190866447985172,
0.7102104425430298,
0.11654235422611237,
0.12723802030086517,
-0.03987452760338783,
0.4023156762123108,
-0.10576793551445007,
-0.3677128553390503,
-0.6281120777130127,
0.2398327738046646,
0.6656339764595032,
-0.12173003703355789,
0.606647372245788... |
USSR stands for Union of Soviet Socialist Republics. The adjective "Soviet" is formed from the noun "Soviet" which in Russian means "Council". (That was roughly the idea behind the revolution and USSR formation that the workers and peasants should rule the state by means of "councils"). So why was some analogous word not created in English? Like "Councillous" or something. Is there some explanation or this "just happened"? | [
-0.008245235309004784,
0.02032284252345562,
-0.0032980984542518854,
0.032220371067523956,
-0.0174312274903059,
0.0017577811377122998,
0.010237469337880611,
0.005075724795460701,
-0.01599355973303318,
0.0008775168098509312,
-0.02235180325806141,
0.006186127196997404,
0.027881449088454247,
0... | [
0.47248557209968567,
-0.1526908427476883,
-0.09348362684249878,
-0.03456856682896614,
-0.3094339966773987,
-0.041966695338487625,
0.08479529619216919,
0.25194939970970154,
-0.222227543592453,
-0.17191828787326813,
-0.32370075583457947,
-0.07613690942525864,
-0.6213188767433167,
0.818820834... |
Suppose I'm gambling using a strategy of doubling my bet whenever I lose to recoup my losses from previous bets. If my initial bet is 1/2048 of my capital, I can bet 10 times before I run out of money. Statistics says that the chances of this happening are ~1/718. As the law of large numbers states, though, this is in no way a guarantee that such an outcome will manifest in any certain number of iterations. Is it possible, therefore, to calculate the chances that this _doesn't_ happen? For example, is there an equation we can put x (the number of bets) into and determine the likelihood of this 1/718 chance _not_ occurring? My chances of losing within the first ten bets would be extremely small, and my chances after ten thousand would be pretty high, so is there any way to calculate the middle ground? | [
0.023775994777679443,
0.022899756208062172,
-0.00363516784273088,
0.00143605493940413,
-0.019080782309174538,
-0.014028564095497131,
0.006836888380348682,
-0.019067365676164627,
-0.011694244109094143,
-0.022043488919734955,
-0.00710958382114768,
0.017108969390392303,
-0.01352530624717474,
... | [
0.19323711097240448,
0.0057953037321567535,
0.41098588705062866,
-0.13413077592849731,
0.6167324781417847,
0.05062326788902283,
0.03390475735068321,
-0.07188377529382706,
-0.5342274904251099,
-0.6497520208358765,
0.1604393869638443,
0.4913024604320526,
-0.11620625108480453,
0.0925534814596... |
Please see the following Image.  I am using a single feature layer for road. Now I using the following code #roads { [ROAD_TYPE = 1] {line-width:8; line-color:#194536;} [ROAD_TYPE = 2] {line-width:8; line-color:#456321;} [ROAD_TYPE = 3] {line-width:8; line-color:#785632;} [ROAD_TYPE = 4] {line-width:8; line-color:#236485;} [ROAD_TYPE = 5] {line-width:8; line-color:#FFFFFF;} [ROAD_TYPE = 6] {line-width:8; line-color:#194536;} line-join: round; line-cap: round; line-smooth: 0.1; } The problem is in the Black circles the white goes under the green road layer. But in the red circle the white layer overlap the green and other major layers. How can I solve it and make it a very nice style as Google or OSM map? | [
-0.008118696510791779,
0.0014227343490347266,
-0.006978008896112442,
0.011407110840082169,
-0.0047705285251140594,
-0.00012527359649538994,
0.005741905886679888,
0.012856039218604565,
-0.011391489766538143,
-0.0008998591220006347,
0.002322528976947069,
0.0020413310267031193,
0.01688164100050... | [
0.3678050935268402,
0.471835732460022,
0.695963978767395,
0.028341762721538544,
-0.3065924048423767,
0.422209769487381,
0.3042646646499634,
-0.025374099612236023,
-0.44071391224861145,
-0.8656381964683533,
-0.10175985097885132,
0.5498936772346497,
0.2778763175010681,
0.3005821406841278,
... |
Reading this great article **Data Validation and Sanitization in Wordpress** I've noticed that in my blog, in header.php, I used `<?php echo $title ?>` in a pair of codes. According to the above article, in order to secure data, it is important to validate the data itself for data's without validation are vulnerable to hackers. As suggested by the author, I changed my initial `<?php echo $title ?>` into `<?php echo esc_html( $title ); ?>`. This is the code _before_ changes: <span class="ads"> <?php if ( is_user_logged_in() ) { echo '<span><a href="/?page_id=175" title="inserisci un annuncio gratis">Pubblica il tuo annuncio gratis: è facile e veloce!</a></span>'; } else { echo '<span><a class="simplemodal-login simplemodal-submit" href=""><?php echo $title ?>Pubblica il tuo annuncio gratis: è facile e veloce!</a> </span>'; } ?> </span> This is the code _after_ changes: <span class="ads"> <?php if ( is_user_logged_in() ) { echo '<span><a href="/?page_id=175" title="inserisci un annuncio gratis">Pubblica il tuo annuncio gratis: è facile e veloce!</a></span>'; } else { echo '<span><a class="simplemodal-login simplemodal-submit" href=""><?php echo esc_html( $title ); ?>Pubblica il tuo annuncio gratis: è facile e veloce!</a> </span>'; } ?> </span> Now my question is (I guess it's a newbie one!): the change I made is good for my blog' security? * * * | [
-0.000748336547985673,
0.022201430052518845,
-0.007906222715973854,
0.006202449090778828,
-0.007743870839476585,
0.024145154282450676,
0.007871340028941631,
-0.030144941061735153,
-0.010956401005387306,
-0.03090488165616989,
-0.02698204666376114,
0.01953093148767948,
-0.023739665746688843,
... | [
0.5146622061729431,
0.421232670545578,
0.0698469802737236,
-0.20088639855384827,
-0.29748228192329407,
-0.2091589719057083,
0.3992888927459717,
-0.23461709916591644,
-0.42577266693115234,
-0.2695848345756531,
-0.021548952907323837,
0.007627875078469515,
-0.23021957278251648,
0.162788391113... |
I saw this entry in Urban Dictionary (I know, not the best place for formal English, but it does do a pretty good job at collecting slang). > 1.hang > > short for "hang out" > > "I'm just gonna hang at Lolita's place today" My question is, do people really use "hang" for "hang out"? i.e. do people really say "Let's hang some time". I have never heard of people saying it, and it sounds awful to me. It sounds like you want to hang yourself. | [
-0.009953128173947334,
0.006356494035571814,
-0.004774170927703381,
0.0030226598028093576,
0.007593618705868721,
-0.03328147530555725,
0.008376232348382473,
-0.008000030182301998,
-0.017379796132445335,
-0.019050996750593185,
-0.0018256058683618903,
0.000049548863898962736,
0.039987929165363... | [
1.2640832662582397,
-0.455152302980423,
0.31622323393821716,
0.027996545657515526,
-0.14098788797855377,
-0.5611760020256042,
-0.08793900161981583,
0.9031969308853149,
-0.13224025070667267,
0.24882590770721436,
0.5203196406364441,
0.3882444202899933,
0.22355177998542786,
-0.386934816837310... |
I would like to insert a stripey ball in TikZ, i.e. a filled-in circle that looks like this one. Would you know a clever solution for doing that?  | [
0.016405411064624786,
0.0068844985216856,
-0.00508733419701457,
0.03409000113606453,
-0.009818993508815765,
0.009049337357282639,
0.009066599421203136,
0.004090114496648312,
-0.0248261746019125,
-0.007269879803061485,
-0.012520907446742058,
-0.006989754270762205,
-0.012870433740317822,
-0.... | [
0.6269142031669617,
-0.39456647634506226,
0.43277478218078613,
0.3415427505970001,
0.17074930667877197,
-0.17074862122535706,
-0.3557174503803253,
0.5331380367279053,
-0.4049299359321594,
-0.6687180399894714,
0.08361207693815231,
-0.13917091488838196,
-0.3556560277938843,
-0.06298292428255... |
What is the best or most popular symbol for vector/matrix transpose? I have used simply `^T`, for example `$c^T x$`. I think it is ugly, mainly because it is a little too big compared with vector variables usually denoted by lower- case characters. Can you suggest a better one? | [
0.0032162838615477085,
0.01006605289876461,
-0.03593861311674118,
0.013667544350028038,
0.013871550559997559,
-0.006543593946844339,
0.010056301951408386,
-0.021929729729890823,
-0.026089781895279884,
-0.0199620109051466,
-0.015980910509824753,
0.0001941155642271042,
-0.01276860199868679,
... | [
0.27005577087402344,
0.024599691852927208,
0.3800484240055084,
0.06451237946748734,
-0.2579876482486725,
0.30113860964775085,
-0.5911001563072205,
0.16797621548175812,
-0.3771810829639435,
-0.48841771483421326,
0.3952977657318115,
0.346489816904068,
-0.4862917363643646,
-0.0450977645814418... |
How to us possessive apostrophe with words in quotes? For example, ...a few days later I discovered that those five little boys were not that well-behaved (as I firstly thought). In fact they were quite mischievous and conniving. The broken tree next to my house and the cat with its legs tied were all their "inventions". And Lilly, as it later turned out, was able to get to Denver so fast not without those **_"little inventors'"_** help. | [
0.009740622714161873,
0.007248809561133385,
-0.021616211161017418,
0.01217233669012785,
-0.028500594198703766,
0.020595291629433632,
0.004793634172528982,
0.006759844720363617,
-0.01723949797451496,
0.002490619895979762,
-0.001957351341843605,
0.00003160829146509059,
-0.003182278946042061,
... | [
0.5706543922424316,
0.024287333711981773,
0.16623525321483612,
0.5116445422172546,
0.20899473130702972,
-0.042223960161209106,
0.654518187046051,
-0.25889691710472107,
-0.08893255889415741,
-0.13786883652210236,
0.4085000455379486,
0.31601232290267944,
-0.083844393491745,
0.287958800792694... |
There's a lot to be said and read about this, but I haven't found a clear answer to this question: Bayesian statistics are said to 'penalize' vague hypotheses with weak priors, by giving more support for the null hypothesis. Say the theory I'm interested in proving actually predicts that the null is true. I could cheat, and set up my model with a weak prior. This would unfairly bias my empirical evidence towards the null. How can this be prevented? Thanks! EDIT: I realized I was actually referring to Lindley's paradox, where as far as I understand) a precise null and an uninformative prior might bias towards the null, whereas frequentist statistics would reject the null. http://www.laeuferpaar.de/Papers/LindleyPSA.pdf | [
0.01690984144806862,
0.02438068576157093,
0.004903437569737434,
0.018783286213874817,
-0.0008758492767810822,
-0.01165121328085661,
0.008164236322045326,
-0.004252738319337368,
-0.007492461241781712,
0.01032966561615467,
-0.011843649670481682,
0.01028242614120245,
-0.03690236806869507,
0.0... | [
0.28476884961128235,
0.09269076585769653,
-0.16939960420131683,
0.3883809745311737,
-0.1666639745235443,
0.017383340746164322,
0.38408708572387695,
0.042438805103302,
-0.2424021065235138,
0.040672656148672104,
0.29663169384002686,
0.2129494547843933,
-0.10665223747491837,
0.451906114816665... |
I am a research scholar and monitoring phenological events of timber line at Himalayan region from past 4 years. During data analysis I found a research paper Estimation and comparison of flowering curve similar to my work. In this paper bbmle package was used and five parameters ($β_0, ... β_4$) describe: (i) the height; (ii) the peak date; (iii) the range; (iv) the symmetry; and (v) the peakedness of the regression curve was calculated. I also read the appendix table and followed the code to estimate these parameter but as a newbie I failed to calculate. Appendix library(bbmle) ## depends R(≥ 2.0.0) ## Collect data together as a data frame, fdat <- data.frame(Days=c(212:238,250:271), Count=c(0,2,2,6,10,18,29,39,59,75,104,130, 145,169,193,209,216,227,231,214,212,226,242, 225,214,202,211,104,90,70,55,52,45,38,29,22, 14,15,14,13,11,6,5,4,3,2,1,1,0)) ## Define function GESN <- function(x,b0,b1,b2,b3,b4) { exp(b0-abs(((x-b1)/(b2∗(1+b3∗(sign(x-b1))))))ˆ(b4)) } ## Get reasonable starting values startvals <- list(b0=log(250),b1=230,b2=15,b3=0,b4=2) Please help me to find these parameters as they are mentioned in figure number 5 (link 1). | [
-0.0172142181545496,
0.018150437623262405,
-0.005449355114251375,
0.012235794216394424,
-0.009871266782283783,
-0.005960986018180847,
0.0090427715331316,
0.001115551684051752,
-0.012789905071258545,
-0.047499604523181915,
-0.009073739871382713,
0.006349486298859119,
-0.020947987213730812,
... | [
0.24704472720623016,
-0.17919762432575226,
0.651594877243042,
-0.22487220168113708,
-0.22098985314369202,
0.1477733701467514,
0.7531712651252747,
-0.23196421563625336,
-0.5088784694671631,
0.17010459303855896,
0.20821596682071686,
-0.0699915736913681,
0.3510265350341797,
0.51280677318573,
... |
My brave Minotaur just died. Is there a way to revive him, or do I have to load an older save? Waiting doesn't seem to make him any less dead. | [
0.05505717173218727,
0.010634946636855602,
0.0005529714981094003,
-0.012861451134085655,
0.023691494017839432,
-0.02187260426580906,
0.008613212034106255,
0.027659859508275986,
-0.024136902764439583,
0.02591777592897415,
-0.0027888223994523287,
0.03145588934421539,
0.047199610620737076,
0.... | [
0.21014532446861267,
0.2050172984600067,
-0.013070238754153252,
0.2208944857120514,
-0.12746645510196686,
0.5162465572357178,
0.342648983001709,
0.0828438550233841,
-0.16845756769180298,
-0.5741339325904846,
0.014396349899470806,
0.029125342145562172,
0.2054736316204071,
0.2077477425336837... |
I am a programmer exporting 32-bit floating point GeoTIFF rasters from my own application that my users need to read into ArcGIS. Unfortunately I am not an Arc user myself and have _zero_ expertise with it, so please target comments accordingly. Arc is reading and locating my rasters properly, but I am having problems making it understand which grid cells are empty. Since all positive and negative floating point numbers are legal data values in my rasters, I am trying to use IEEE NaNs as my NoData values, which I understood to be an accepted practice. Furthermore, as my rasters are often dimensionally large but sparsely populated, I'm trying to get maximal compression out of the TIFF PackBits algorithm (which really likes to see sequences of identical bytes) by making an optimal choice with regard to my particular NaN value. Since any 32-bit float whose exponent bits are all 1 is considered a NaN, I'm using 0xffffffff (32 1-bits) as my NaN value since it's both a valid IEEE NaN and is also 4 identical bytes. Arc seems to treat my 0xffffffff NaN values as low but legal values, i.e., it renders them as black similarly to the actual low legal data values in my file. How can I get Arc to treat this particular constant (which is of course _not_ a value that can be expressed as a floating point number since it's an actual IEEE NaN) as my NoData value and not render it at all? Not being an Arc user it's quite possible that I'm just not setting the right display/import options within Arc to get what I want. If someone could explain how to do this that may well solve my problem. | [
-0.0009232573211193085,
0.009512084536254406,
-0.007015341892838478,
0.01585775800049305,
0.0026908235158771276,
-0.01554335467517376,
0.009899822063744068,
-0.0023881990928202868,
-0.013644885271787643,
-0.017572995275259018,
0.001977314706891775,
0.01022336445748806,
-0.015635870397090912,... | [
0.2038339227437973,
0.24888667464256287,
0.09698043018579483,
-0.08480183035135269,
-0.23199008405208588,
0.04049765691161156,
-0.15392951667308807,
0.04621744528412819,
-0.16318127512931824,
-0.7640800476074219,
0.04041239246726036,
0.6710062026977539,
-0.1559925079345703,
-0.095074668526... |
In LyX, I'm using the reference style apsrev.bst, which I downloaded from here: http://www.maik.ru/pub/tex/revtex4/ This reference style is the one used in Physical Review. But now when I add references, they show up as question marks in brackets (?). Why does this happen? how can I fix it? Thank you for any efforts | [
0.02010991796851158,
-0.0016219312092289329,
-0.010702114552259445,
0.02363639697432518,
0.0016762816812843084,
-0.0031903130002319813,
0.010028473101556301,
0.00984963122755289,
-0.019458021968603134,
-0.0020599288400262594,
0.001875310204923153,
-0.0015788814052939415,
-0.0082053542137146,... | [
0.6455662250518799,
0.10197252035140991,
0.4185858368873596,
-0.030814295634627342,
-0.37325558066368103,
0.05173986777663231,
0.04802935943007469,
0.1459469050168991,
-0.2473261058330536,
-0.5167511701583862,
0.08555729687213898,
0.16793100535869598,
-0.21530026197433472,
-0.1011080816388... |
I need to use very simple geometry operations (read SHP or other standard format and simple searches) within an embedded device running a simple linux based os. What is the smallest and simplest API I could use? It could be written in C/C++ or Python in order to allow the integration with my system. Thanks, Samuel | [
-0.01825011894106865,
0.010166640393435955,
-0.015718165785074234,
0.01765156351029873,
-0.0051294597797095776,
0.008514618501067162,
0.010142112150788307,
0.01717187836766243,
-0.017727645114064217,
-0.055276058614254,
-0.0038927788846194744,
0.004325549118220806,
-0.0023192293010652065,
... | [
0.37491410970687866,
0.40849918127059937,
0.07112755626440048,
0.5090090036392212,
0.115554578602314,
0.1170787662267685,
0.1911315768957138,
0.3119966983795166,
0.11237015575170517,
-0.582354724407196,
0.1903480738401413,
0.15426109731197357,
-0.1589139699935913,
0.18806900084018707,
0.... |
I might be missing something basic - but it appears that the strong law of large numbers covers the weak law. If that case, why is the weak law needed? Thanks! Liran | [
0.005741239059716463,
0.03829113021492958,
0.004441488068550825,
0.04033895954489708,
-0.009729830548167229,
0.013727202080190182,
0.010655115358531475,
-0.032445263117551804,
-0.02949734777212143,
-0.05521596223115921,
-0.030453210696578026,
0.03403644636273384,
-0.02828708291053772,
-0.0... | [
0.16395822167396545,
0.44781097769737244,
-0.4749511778354645,
0.3683692216873169,
-0.07198987901210785,
-0.12698262929916382,
-0.22229348123073578,
-0.04222186282277107,
-0.0923999473452568,
-0.5861951112747192,
-0.03985348716378212,
0.5405603051185608,
0.10409143567085266,
-0.12078261375... |
I have a init script which is poorly designed because it does not conform to the Linux Standard Base Specifications The following should have an exit code of 0 if running, and 3 if not running service foo status; echo $? However because of the way the script is designed, it always returns a 0. I can not fix the script without a significant rewrite (because service foo restart is dependent on service foo status). How could you work around the issue so that `service foo status` returns a 0 if running, and a 3 if not running? What I have so far: root@foo:/vagrant# service foo start root@foo:/vagrant# /etc/init.d/foo status | /bin/grep "up and running"|wc -l 1 root@foo:/vagrant# /etc/init.d/foo status | /bin/grep "up and running"|wc -l;echo $? 0 # <looks good so far root@foo:/vagrant# service foo stop root@foo:/vagrant# /etc/init.d/foo status | /bin/grep "up and running"|wc -l 0 root@foo:/vagrant# /etc/init.d/foo status | /bin/grep "up and running"|wc -l;echo $? 0 # <I need this to be a 3, not a 0 | [
0.00509947445243597,
0.023195169866085052,
-0.006589174270629883,
0.015479307621717453,
0.003166025970131159,
-0.0055579328909516335,
0.0098398607224226,
-0.020924601703882217,
-0.015112798660993576,
0.009615881368517876,
-0.02634112909436226,
0.010011496022343636,
-0.006998508702963591,
0... | [
0.3328899145126343,
0.16419285535812378,
0.09582463651895523,
-0.24140414595603943,
-0.15949758887290955,
-0.11655284464359283,
0.8629330992698669,
-0.081302709877491,
-0.040555570274591446,
-0.4984326958656311,
0.22987067699432373,
0.28784874081611633,
-0.3927561938762665,
0.3133392632007... |
Clear the universe of all matter except for two tennis balls. Place the two tennis balls in the same inertial frame 1 Mpc apart. * Are the tennis balls getting further apart? * Will the tennis balls remain in the same inertial frame? EDIT: Don't assume the balls are massless, but please ignore the gravitational attraction two tennis balls separated at 1 Mpc would exhibit on one another. (Previously, I had asked to assume the tennis balls were massless. I couldn't switch to a strike out font, so I just removed it entirely.) | [
0.01763877645134926,
0.017428714781999588,
-0.006081991363316774,
0.016678884625434875,
-0.03406158462166786,
-0.0297554861754179,
0.011220511980354786,
-0.01866484247148037,
-0.016111597418785095,
-0.01984662376344204,
-0.013123678043484688,
0.023544181138277054,
-0.008846557699143887,
0.... | [
0.22734108567237854,
-0.03315150365233421,
0.49715155363082886,
0.14197485148906708,
0.1257372349500656,
-0.13065555691719055,
0.06314241141080856,
0.13280488550662994,
-0.17007771134376526,
-0.3601874113082886,
-0.39695116877555847,
0.014333034865558147,
-0.5603079795837402,
0.24410526454... |
I am currently suffering of a strange bibstyle problem with line breaks. There is a missing space between bib items in the bibliography, which is generated using BibLaTeX.  Even though I am using the authoryear bibstyle, I do not get the line breaks after each bib item. \usepackage[citestyle=authoryear,bibstyle=authoryear,url=false,doi=true,natbib=true,backend=bibtex]{biblatex} \renewcommand{\nameyeardelim}{ }% \renewcommand{\multicitedelim}{, }% \DefineBibliographyStrings{ngerman}{andothers={et\addabbrvspace al\adddot}} \renewcommand{\cite}{\citep}% \addbibresource{/home/robert/Documents/franzi/diss/Literatur/library.bib} Is there any possibility to adjust the space after each bib item with a dedicated biblatex parameter? I have searched through the manual but was not finding something useful. | [
0.006574142724275589,
0.0030784434638917446,
-0.011687198653817177,
0.028097938746213913,
0.014410773292183876,
0.0009506000205874443,
0.00794556550681591,
0.026530785486102104,
-0.017323462292551994,
0.0023053823970258236,
-0.021671034395694733,
0.009351604618132114,
-0.010538173839449883,
... | [
0.16240812838077545,
0.36564499139785767,
0.6702982783317566,
-0.2082023173570633,
-0.29946237802505493,
-0.15467774868011475,
0.7039895057678223,
-0.003274106653407216,
-0.44579145312309265,
-0.500747561454773,
-0.05290248990058899,
-0.033494945615530014,
-0.12814435362815857,
0.348977446... |
Say there is a circuit with two 1.5V cells, and a 100 ohm resistor. If you connect two cells in series, then the total emf is 3V. And the current will be 3/100 = 0.03 A. (Using V = IR):  If you have the cells in parallel, then the total emf is 1.5V, as the terminals of the cells are electrically the same point. So the current will only be 0.015 A:  But if you just had one 1.5V cell, so the total emf is again 1.5V. The current will still be 0.015 A.  So what benefit does adding a second cell have? The emf and the current is the same no matter if you use 2 cells in parallel or just one cell. I'm assuming that all the cells are identical, and internal resistance is negligible. | [
0.01319105364382267,
0.021201036870479584,
-0.017541490495204926,
0.0036704884842038155,
-0.006420877296477556,
-0.00927426666021347,
0.00824061781167984,
-0.02278459444642067,
-0.015394012443721294,
-0.0240994431078434,
-0.006458078511059284,
0.014200471341609955,
-0.005065262317657471,
0... | [
0.46067994832992554,
-0.3742709457874298,
0.8450111150741577,
0.047458112239837646,
-0.06324923038482666,
0.14960743486881256,
0.12888064980506897,
-0.7516322731971741,
-0.48061704635620117,
-0.3836085796356201,
0.2244383990764618,
0.3238091766834259,
-0.2806507647037506,
0.177295342087745... |
During the summer, most plants wither and stop growing. It's also very dangerous to stay around in the base, because things keep smoldering up and you will probably not have enough ice in store to keep them cool. While roaming the world, I noticed that mushrooms still appear to be growing, so I can pick them up as crock pot ingredients. I just checked out the caves and it's still pretty hot in there, in addition to darkness for which I wasn't quite prepared and with all the grass bushes withered, I'm not sure how I can keep a stable supply of light sources. With so little food and everything just waiting to burn to ashes, it doesn't look like there is going to be much "free" time. But what else can I do when it's so hot all the time? Where can I find a safe place where the heat can't get me and my valuable items? | [
-0.005851811729371548,
0.00848902203142643,
-0.0026859017089009285,
0.006434766110032797,
0.00383460964076221,
-0.01278211921453476,
0.005628893617540598,
-0.008949479088187218,
-0.01414259523153305,
-0.027056697756052017,
-0.003163103712722659,
0.015977470204234123,
-0.0023626480251550674,
... | [
0.7340541481971741,
0.473818302154541,
0.024152636528015137,
0.19210855662822723,
0.20380689203739166,
-0.5945084691047668,
1.0948071479797363,
0.20720617473125458,
-0.5929807424545288,
-0.09534315764904022,
-0.2741929590702057,
-0.14017073810100555,
0.08180230855941772,
0.2261560112237930... |
I'm pretty well sold on the "singletons are evil" line of thought. Nevertheless, there are limited occurrences when you want to limit the creation of an object. Roy Osherove advises, > If you're planning to use a singleton in your design, separate the logic of > the singleton class and the logic that makes it a singleton (the part that > initializes a static variables, for example) into two separate classes. That > way, you can keep the single responsibility principle (SRP) and also have a > way to override singleton logic. ( _The Art of Unit Testing_ 261-262) This pattern still perpetuates the global state. However, it does result in a testable design, so it seems to me to be a good pattern for mitigating the damage of a singleton. However, Osherove does not give a name to this pattern; but naming a pattern, according to the Gang of Four, is important: > Naming a pattern immediately increases our design vocabulary. It lets us > design at a higher level of abstraction. (3) Is there a standard name for this pattern? It seems different enough from a standard singleton to deserve a separate name. _Decoupled Singleton_ , perhaps? | [
0.018607862293720245,
0.02561807446181774,
-0.010446551255881786,
-0.0009653015295043588,
-0.025342848151922226,
0.006380794569849968,
0.00552191911265254,
0.020427148789167404,
-0.01110458467155695,
0.004151396453380585,
-0.007926400750875473,
0.012486380524933338,
-0.01331263780593872,
0... | [
0.6652405858039856,
0.24618060886859894,
-0.5137736201286316,
0.07552529871463776,
-0.0857914462685585,
-0.10328792780637741,
0.2859494388103485,
-0.528721272945404,
-0.1153019517660141,
-0.2952180504798889,
0.02490406110882759,
0.6724812984466553,
-0.03107485920190811,
0.3709200322628021,... |
I'm excited about the possibilities for duplex communication with WebSockets between a web browser and web server. I see that WCF for .NET 4.5 supports a WebSockets binding now, and I know there's a JavaScript API for using WebSockets. However, what I don't understand is how these are supposed to interoperate. Yes, you can deal with raw message passing with the JS API, but there doesn't seem to be any JS library for doing SOAP-based RPC compatible with WCF. Do any of you know if there are any libraries to enable a web browser to call a WCF service using WebSockets? | [
-0.004558230750262737,
-0.0010967716807499528,
-0.0067139421589672565,
0.012801256030797958,
0.007704152725636959,
-0.013176236301660538,
0.009287280961871147,
0.009095244109630585,
-0.020894335582852364,
-0.022464724257588387,
0.009404691867530346,
0.039037905633449554,
-0.01834134012460708... | [
0.6630952954292297,
0.1347484439611435,
0.24619817733764648,
0.06268923729658127,
-0.2524756193161011,
-0.294627845287323,
0.28868579864501953,
-0.010360375978052616,
-0.17061476409435272,
-0.7814195156097412,
0.14749696850776672,
0.6163809895515442,
-0.20894411206245422,
0.349667251110076... |
I'm having trouble getting some pretty simple functions using `AstronomicalData` to perform at anywhere near the speed I need them to. For example, I have many places where I need to compute something like ParametricPlot[ {AstronomicalData[ "Mercury", {"RightAscension", DatePlus[Date[], d], {$GeoLocation[[1]] , $GeoLocation[[2]]}}], d}, {d, 0, 365}, AspectRatio -> 1/1.6] but find that it takes far to long for even this simple fragment to run. I want to put several such fragments (e.g. one for each of several astronomical objects) into a single figure, and even to dynamically change the parameters for the figure (such as location and date) inside a `Manipulate`, which would be prohibitively slow given the times I'm seeing with even the simple fragment above. Why is `AstronomicalData` so slow? Is there something I can do to speed it up so I can use it as I hope to? | [
0.009986072778701782,
0.00936851091682911,
-0.00809997133910656,
0.012048350647091866,
-0.01597546972334385,
-0.017574988305568695,
0.006093953736126423,
0.003599310526624322,
-0.013716109097003937,
-0.012027883902192116,
-0.004093465395271778,
0.008214783854782581,
-0.009464792907238007,
... | [
0.024318935349583626,
-0.0719667598605156,
0.7381317019462585,
0.13729077577590942,
0.2251097410917282,
0.2815723121166229,
-0.23879270255565643,
-0.09082357585430145,
-0.10116808116436005,
-0.5955925583839417,
0.14373944699764252,
0.40929320454597473,
0.08616035431623459,
-0.1479657888412... |
Is it possible to intercept calls to current_user_can() ? Example: `current_user_can('rate', $post_id)` There's no "rate" capability registered, but could I somehow hook into that function and do my own checks, without having to register a role capability? | [
0.012923048809170723,
0.027096301317214966,
-0.01081172190606594,
0.01991567574441433,
-0.016491709277033806,
-0.016287485137581825,
0.013305580243468285,
-0.021684017032384872,
-0.026754073798656464,
0.0101082231849432,
-0.019428675994277,
0.010132436640560627,
-0.008022383786737919,
0.02... | [
0.13075760006904602,
-0.30269962549209595,
0.3035661578178406,
0.13201741874217987,
0.36905092000961304,
-0.08113234490156174,
0.3582700490951538,
-0.09372024983167648,
0.06906093657016754,
-0.2495926320552826,
0.0814133882522583,
0.782158613204956,
-0.13817031681537628,
-0.328147560358047... |
> Related question: In sex talk, how many bases are there and what do they all > mean? There are lots of English-speaking (or English-learning) countries where baseball simply isn't played much if at all. Other sports -- soccer, rugby, polo, cricket -- are likely more common. Since language follows culture, I'm curious: **what do English-speaking nations with little interest in baseball use as their metaphor for physical relations?** Effectively, I'm asking for equivalent metaphors in other English dialects. Note that, in the baseball metaphor, there is an implicit "escalation" idea baked into the metaphor. As a "player" progresses from base to base, the conduct associated with each point represents an increase in intensity and/or intimacy. I'm hoping equivalent alternative metaphors preserve this. | [
0.007479575462639332,
0.00978054665029049,
-0.0029574038926512003,
0.021344497799873352,
0.011162323877215385,
0.010174480266869068,
0.007402448914945126,
0.01141933910548687,
-0.016866249963641167,
-0.022277425974607468,
-0.0015516022685915232,
0.01352468691766262,
0.015868499875068665,
0... | [
0.5477215051651001,
-0.021973298862576485,
-0.2689327299594879,
0.4201355278491974,
-0.44700053334236145,
-0.1967790722846985,
0.3190670609474182,
0.5963745713233948,
-0.1717626303434372,
-0.7371222972869873,
0.1654670536518097,
0.6801629662513733,
-0.21103115379810333,
-0.1363976746797561... |
I'm not a physicist. I want to understand the macroscopic Maxwell's equations. But after reading Wikipedia and other Googled stuffs, I got very confused. In particular, $D$ and $H$ have two different equations, respectively. One group is called auxiliary fields: $$B = \mu_0(H+M),$$ $$D = \epsilon_0E+P.$$ The other group is called constitutive relations: $$B = \mu H,$$ $$D = \epsilon E.$$ Which group of equations are relevant to the macroscopic Maxwell's equation if we consider the case with polarization ($P$ and $M$ present). My aim is to learn computational electromagnetic. | [
-0.017696712166070938,
0.005814903881400824,
-0.004959563259035349,
0.015762507915496826,
0.0006306052673608065,
-0.023775765672326088,
0.00839170441031456,
-0.008525794371962547,
-0.016042662784457207,
0.009871008805930614,
-0.009635774418711662,
-0.00020921544637531042,
-0.0211921371519565... | [
0.1441091001033783,
0.1637900322675705,
0.2676108479499817,
-0.04777224734425545,
-0.22587254643440247,
0.21981285512447357,
0.09286416321992874,
0.017523495480418205,
0.15632526576519012,
-0.24824850261211395,
-0.20403656363487244,
0.3437708914279938,
-0.35629528760910034,
0.2436108887195... |
I have recently downloaded and installed `NVIDIA-Linux-x86_64-319.17.run` for my Nvidia Quadro NVS 130M Graphics Card. Installation went fine after disabling the X server and running the install from a tty. After rebooting; * The Slim Login manager took considerably longer to process my initial username entry (8-10 seconds vs <1s) * Any attempts to access another tty results in a system hang, from which I have to restart. * Attempting to launch Firefox results in a system hang before displaying anything. I have done some searching around but have found nothing that has fixed my problem. I am running `archbang linux x64`, so followed the instructions detailed on their site. I am attempting `pacman -Syu` but do not think this will make much of a difference... I changed from using the nouveau drivers. I have looked through my Xorg.0.log file but didn't see anything that was glaringly obvious. If you need any more details than those provided below, please comment. | [
0.0022858106531202793,
0.00046636309707537293,
-0.006643966306000948,
0.0006529535166919231,
-0.003424553433433175,
-0.009523745626211166,
0.0081498883664608,
0.0009286999702453613,
-0.009375186637043953,
-0.024366769939661026,
-0.016595400869846344,
0.003926421049982309,
-0.0147878360003232... | [
0.5730058550834656,
0.22083300352096558,
0.5744466185569763,
-0.20178914070129395,
0.11688947677612305,
-0.2042146623134613,
0.34602755308151245,
0.099616639316082,
-0.17039282619953156,
-0.6648176312446594,
-0.029353884980082512,
0.8455834984779358,
-0.3169187307357788,
-0.164501905441284... |
Like almost everyone here, I have a handfull of scripts and software that I have developed and am enthused about. I will be looking for my first job as a software designer / coder. It seems natural that I will be eager to please my employer and use scripts or similar methods that I have developed and worked for me in the past to please my employer. It seems certain that many things that I code will look very similar to things I have coded in the past. I don't understand how to document and articulate to an employer that this code base was mine before I got here and this will continue to be mine when I leave. Surely, this is a common issue, but none of the various searches I've done on the net have produced an answer to this question. How is this situation commonly dealt with in the industry? I feel like there should be a digital version of sending myself a 'certified letter' with my code/software/scripts contained. I'm not trying to protect my code from others using it; I am trying to protect my right to continue using my code base that I have developed prior to to gaining employment with an employer. | [
0.004524116404354572,
0.014974004589021206,
-0.00046156812459230423,
-0.000684850150719285,
-0.010909777134656906,
0.003938437905162573,
0.0039072539657354355,
0.0038544118870049715,
-0.013716505840420723,
-0.007425555493682623,
-0.0008866415591910481,
0.01157675962895155,
0.0207331981509923... | [
0.9172667860984802,
0.45750972628593445,
-0.1459769755601883,
0.1556873321533203,
0.200847327709198,
-0.09600479155778885,
0.16656604409217834,
0.6235377788543701,
-0.250306636095047,
-0.8931261301040649,
0.1678394228219986,
0.5158780813217163,
0.19904132187366486,
0.28267860412597656,
0... |
> **Possible Duplicate:** > How to look up a math symbol? Could anyone tell me a conventional way to print NOT `\sqsubseteq`? | [
0.02250872179865837,
-0.0016242177225649357,
-0.0003340425610076636,
0.017643215134739876,
-0.020100915804505348,
0.021029233932495117,
0.01359457429498434,
0.03823569789528847,
-0.041920483112335205,
-0.037803780287504196,
-0.012404787354171276,
0.007138783112168312,
-0.011913727968931198,
... | [
0.1695653200149536,
0.2186693549156189,
0.0950634703040123,
0.3304532766342163,
-0.21764269471168518,
0.1608469933271408,
0.25036945939064026,
0.10618427395820618,
-0.261577844619751,
-0.6465113162994385,
0.004809861071407795,
0.12946361303329468,
-0.016850467771291733,
-0.2530808448791504... |
> **Possible Duplicate:** > Is there a (better) way to debug Visual Basic Code Block in calculate field > using ArcGIS? I am graduate ESCI student, and my project involves a lot of ArcGIS use. My project is related to assessing coastal erosion rates and causes near the Arctic coastlines of Barrow, I am currently trying to figure out the azimuth of a plolyline. I think I know how, but when I run it in the field calculator, I get an error, called 999999 : Error executing function. I tried several other calculations from Easy calculate just to see if they will work, but it give me the same error. I am using ArcMap 10, and I downloaded the EC50 (easy calculate). Can anyone help me out with this, and direct me on how to solve this error ? thanks | [
-0.002500720787793398,
0.004336856305599213,
-0.009030856192111969,
0.01810811087489128,
-0.030711626634001732,
-0.00015784148126840591,
0.008002383634448051,
0.016827045008540154,
-0.016505762934684753,
-0.002545569557696581,
0.004715047776699066,
0.015137109905481339,
-0.024556782096624374... | [
0.5518165230751038,
0.17355889081954956,
0.3409917652606964,
-0.10531293600797653,
-0.25974899530410767,
0.28238797187805176,
0.15877175331115723,
0.339140921831131,
-0.5708326697349548,
-0.6482582092285156,
0.47573938965797424,
0.23947449028491974,
0.2510327398777008,
0.19788159430027008,... |
I use data from NOAA for some analysis in R and I want to transform coordinates to EPSG:54004 or something useful. There is something what they write about their coordinates. ... POS: 29-34 GEOPHYSICAL-POINT-OBSERVATION latitude coordinate The latitude coordinate of a GEOPHYSICAL-POINT-OBSERVATION where southern hemisphere is negative. MIN: -90000 MAX: +90000 UNITS: Angular Degrees SCALING FACTOR: 1000 DOM: A general domain comprised of the numeric characters (0-9), a plus sign (+), and a minus sign (-). +99999 = Missing POS: 35-41 GEOPHYSICAL-POINT-OBSERVATION longitude coordinate The longitude coordinate of a GEOPHYSICAL-POINT-OBSERVATION where values west from 000000 to 179999 are signed negative. MIN: -179999 MAX: +180000 UNITS: Angular Degrees SCALING FACTOR: 1000 DOM: A general domain comprised of the numeric characters (0-9), a plus sign (+), and a minus sign (-). +999999 = Missing ... And my problem is, that I can't transform this coodrinates correctly. I use this R command: sc <- cbind(st$LAT, st$LON) ptransform(sc/180*pi, '+proj=latlong +ellps=sphere', '+proj=merc +lon_0=0 +k=1 +x_0=0 +y_0=0 +ellps=WGS84 +datum=WGS84 +units=m +no_defs') But coordinates are somehow wrong as you can see. The points should be in borders because they are meteostations from this countries. . | [
-0.006338257342576981,
0.0011769913835451007,
-0.008608517237007618,
0.0234242994338274,
0.006957225035876036,
0.01586250402033329,
0.009823305532336235,
0.006090834736824036,
-0.014002068899571896,
-0.01860448159277439,
-0.0038111601024866104,
0.010842992924153805,
-0.009720327332615852,
... | [
-0.12099968641996384,
0.19770501554012299,
0.8431843519210815,
0.09251917153596878,
-0.10742761194705963,
0.48930415511131287,
-0.08956696093082428,
-0.26444369554519653,
-0.1027594581246376,
-0.3849292993545532,
0.12052155286073685,
0.08347712457180023,
0.15082664787769318,
0.358031004667... |
I am a Ubuntu user, and recently I had a problem with my upgrade from an older version of Ubuntu to a newer one, so I had to wipe everything out and start back from scratch. Unfortunately, that involves setting LaTeX to work back again. The problem I have is the following : I don't remember what I was using as a LaTeX distribution before, but I know that now I am using TeXLive ; when I run pdflatex, it gives me an error line on every single line in my code, which usually looks like this for one of my average-looking files : \documentclass[a4paper,10 pt]{article} \usepackage[isolatin]{inputenc} \usepackage[T1]{fontenc} %\usepackage[francais]{babel} \usepackage[cyr]{aeguill} \usepackage{xspace} \usepackage{fullpage} %Specific to TikZ \usepackage{tikz} \usetikzlibrary{calc} \usepackage{ifthen} \usepackage{animate} .... For instance, I get told from the second line that "File isolatin.def not found" and then my packages ending in ".sty" keep appearing a lot in the errors. If I keep letting them go though (by pressing enter in my terminal, it kind of ignores the warning) I can compile and get my .pdf file, but it won't show with all my pretty packages. Any idea about why this happens? Thank you! | [
0.007704185321927071,
0.01141329761594534,
-0.013352997601032257,
0.010051066055893898,
0.008295482024550438,
0.0037395297549664974,
0.006683899089694023,
0.009342269971966743,
-0.012310492806136608,
-0.025069307535886765,
-0.0139091145247221,
0.004882380831986666,
-0.002255791798233986,
0... | [
0.33643093705177307,
0.21829524636268616,
0.40548738837242126,
-0.25494441390037537,
-0.11419238150119781,
-0.17532594501972198,
0.259665310382843,
0.3422321081161499,
0.06328922510147095,
-0.8732641339302063,
0.1062915101647377,
0.7894465923309326,
-0.36000514030456543,
0.2892588675022125... |
I'm using Linux Mint Debian edition and I have set Firefox as my default browser in my settings. But HTTP links in other apps like hotot and pidgin open with Chromium! Why is this happening is there any way to track the problem? | [
-0.014016250148415565,
-0.00854537170380354,
-0.007178027182817459,
0.0014461673563346267,
-0.00006488428334705532,
-0.057080432772636414,
0.014295319095253944,
0.008597912266850471,
-0.02362176962196827,
-0.04371063411235809,
-0.013959341682493687,
-0.003522926475852728,
-0.0025942423380911... | [
0.4279412627220154,
0.3750072121620178,
0.23653370141983032,
0.12131061404943466,
0.020080283284187317,
-0.32214829325675964,
0.42815399169921875,
0.4938673973083496,
-0.27028796076774597,
-0.6935490965843201,
-0.052502017468214035,
0.38815930485725403,
-0.3296358287334442,
0.0959702804684... |
I'm using this little snippet right now to check if the current page is the child of a particular page; $studies_parent = 5860; if ($studies_parent == $post->post_parent) { Which is working fine, although now I need to check if the current page is a child of one of 2 different ID's and I'm not sure how to go about it. Basically I need to check if the current page's parent is either 5860 or 1047. Any ideas how I'd do that? | [
0.0005154870450496674,
0.008270799182355404,
-0.012598766013979912,
0.016333680599927902,
0.013875935226678848,
-0.002262504305690527,
0.0077909850515425205,
-0.01709146425127983,
-0.014565992169082165,
0.014087771996855736,
-0.005095656029880047,
-0.0011356499744579196,
-0.01365212909877300... | [
0.3235674798488617,
0.13362622261047363,
0.4238186478614807,
0.2246944010257721,
0.24408261477947235,
0.2186947762966156,
0.000457907299278304,
0.13523989915847778,
-0.3605630397796631,
-0.8975344300270081,
0.23290415108203888,
-0.259936660528183,
-0.17677517235279083,
0.46786385774612427,... |
I was wondering about the inverse of this question: Hiding section titles when the section is empty Instead of hiding empty sections in LaTeX, can one display only empty sections or somehow indicate that a section is empty in the table of contents? | [
0.03156495466828346,
0.022272499278187752,
-0.004146388731896877,
0.04569288343191147,
0.012476959265768528,
0.001960717374458909,
0.014523851685225964,
-0.022327184677124023,
-0.02715780958533287,
0.006607121787965298,
-0.03436307609081268,
0.004498185124248266,
0.002728961640968919,
0.00... | [
0.5912659764289856,
0.215068981051445,
-0.0604700893163681,
0.3497668504714966,
-0.02623678930103779,
-0.35308679938316345,
0.07788663357496262,
0.2628021240234375,
-0.579292356967926,
-0.14250344038009644,
0.02267991378903389,
0.07793746888637543,
-0.12125197798013687,
0.547466516494751,
... |
Here's a simple script saved in the file `hello` #!/usr/local/bin/MathematicaScript -script Print["Hello world"] I can then run this file using `math -script hello` or, if the file is set to be executable, `./hello`. In Windows or OSX, run `MathKernel -script hello` It prints > "Hello world" to the terminal. How can I change this so it does not print the quotation marks? | [
-0.005696335807442665,
-0.002289380179718137,
-0.010746682062745094,
-0.00008876498759491369,
-0.029860729351639748,
0.012389786541461945,
0.009014192968606949,
0.0067318398505449295,
-0.020049141719937325,
-0.04209832847118378,
-0.011628522537648678,
-0.002558221574872732,
0.007864727638661... | [
-0.25459495186805725,
0.10816280543804169,
0.2838650941848755,
0.025562219321727753,
-0.028413979336619377,
0.06238800659775734,
0.3104899823665619,
0.4092791676521301,
-0.0472615547478199,
-0.7386300563812256,
0.031203649938106537,
0.6649376153945923,
-0.1641288697719574,
-0.0464836284518... |
I have two independent continuous random variables, $A$ and $B$. I want to find the cumulative distribution function of $A+B$. $A$ is log-logistically distributed, and $B$ is normally distributed. How can I go about finding the CDF of $A+B$? | [
-0.03260616213083267,
0.014512770809233189,
-0.030835429206490517,
0.011039486154913902,
-0.00986965000629425,
-0.04918067157268524,
0.014909560792148113,
0.0015937668504193425,
-0.036011360585689545,
-0.02390974946320057,
-0.025703057646751404,
-0.0005491197807714343,
-0.026048507541418076,... | [
0.3890375792980194,
0.06225127726793289,
0.08917546272277832,
-0.10699682682752609,
-0.12575845420360565,
0.10046388953924179,
-0.21978814899921417,
-0.1766236573457718,
-0.2229844033718109,
-0.20406165719032288,
0.1465771645307541,
0.2930659055709839,
-0.5387206077575684,
0.42575037479400... |
How should I draw pathes with multiple branches in Tikz? The `draw` command only provides 1-dimensional pathes. Is there a way to naturally merge them? **My thoughts:** Usually I would simply combine multiple `draw` commands. But in the case of e.g. dashed lines the drawing fails at the tie points. **Minimal working example:** \documentclass{article} \usepackage{tikz} \begin{document} \begin{tikzpicture} % way 1: only draw missing part \draw[dashed,->] (0,0) -- (1,0) -- (1,1); \draw[dashed] (2.5,0) -- (1,0); \end{tikzpicture} \vspace{1em} % dummy space \begin{tikzpicture} % way 2: draw full path \draw[dashed,->] (0,0) -- (1,0) -- (1,1); \draw[dashed,->] (2.5,0) -- (1,0) -- (1,1); \end{tikzpicture} \end{document} The result is looking ugly in both cases:  | [
-0.007553378585726023,
0.021483633667230606,
-0.006714982911944389,
0.019465096294879913,
-0.00036642776103690267,
-0.01905011013150215,
0.009592710062861443,
0.010219987481832504,
-0.019871635362505913,
0.006139757111668587,
-0.023301834240555763,
0.010757711715996265,
-0.0176530908793211,
... | [
0.2082260251045227,
-0.05051861330866814,
0.23358377814292908,
0.06586749851703644,
-0.10574279725551605,
-0.18538640439510345,
0.640161395072937,
-0.32857203483581543,
-0.08208593726158142,
-0.9217352271080017,
0.1200208067893982,
0.4291389286518097,
-0.20231573283672333,
-0.2783285975456... |
How can I set my bukkit server on my mac to use more than 1 core? I set my server to use 12 gigs of RAM for now, considering that I will be adding more servers soon. My iMac has these configurations: * 3.5GHz Quad-core Intel Core i7, Turbo Boost up to 3.9GHz * 32GB 1600MHz DDR3 SDRAM - 4X8GB * 1TB of PCIe-Based Flash Storage * NVIDIA GeForce GTX 780M 4GB GDDR5 It uses mavericks. Also, is it possible to use the graphics card for the server (considering that the RAM on it is MUCH faster than the standard 1600MHz RAM) when the card is not in use? Thanks so much guys! | [
0.017289934679865837,
-0.008950211107730865,
-0.0053402697667479515,
-0.003776164259761572,
0.035226840525865555,
-0.03482116013765335,
0.006404993124306202,
-0.003495458047837019,
-0.017425645142793655,
-0.009723493829369545,
0.009711059741675854,
0.0064800214022397995,
0.00588188786059618,... | [
0.03327357769012451,
0.19763630628585815,
0.6145614981651306,
0.02821594476699829,
0.03015008755028248,
0.349662184715271,
0.14292971789836884,
-0.17517009377479553,
-0.34791600704193115,
-0.7058027386665344,
-0.05246983468532562,
0.6211947202682495,
-0.02484121173620224,
-0.11960389465093... |
How do I reference controls that have already been added to a panel. I am attempting to destroy or modify the default controls associated with the editingToolbar. These changes would be done outside of the initialization function. EDIT:My intent was to change the layer that a drawFeature controls was adding features too. | [
-0.008276866748929024,
0.014246879145503044,
0.003963641822338104,
0.01632474735379219,
-0.015014867298305035,
-0.001151208532974124,
0.011851572431623936,
0.041562698781490326,
-0.02209007553756237,
0.014506136067211628,
0.002094558672979474,
0.024067316204309464,
0.002277623862028122,
0.... | [
0.3448963463306427,
0.06514068692922592,
0.4566628038883209,
-0.05429961159825325,
0.0617729052901268,
-0.2293190062046051,
0.12860095500946045,
-0.05049056187272072,
0.08991563320159912,
-0.5864859223365784,
0.05737064406275749,
0.5119932889938354,
-0.246755450963974,
0.06180751323699951,... |
I was wondering, why in Newtonian physics torque is called "torque" while in static mechanics they call it "moment"? I prefer by far the term "torque", for not only it sounds strong, but also instead of moment, the correct synonym of torque is moment of force. | [
0.009351678192615509,
0.005261886864900589,
-0.01970839686691761,
0.005506904795765877,
-0.00012464799510780722,
0.005490830168128014,
0.018364183604717255,
-0.051133301109075546,
-0.0221642404794693,
-0.027573898434638977,
-0.01880827732384205,
0.023414747789502144,
-0.005480428226292133,
... | [
-0.3155433237552643,
-0.6143640279769897,
0.1720864325761795,
0.2821018397808075,
-0.5976917743682861,
0.05180644616484642,
-0.1885765790939331,
-0.2167443037033081,
-0.3760676383972168,
-0.4674539566040039,
-0.008033647201955318,
0.3062094748020172,
0.2059999257326126,
0.5524135828018188,... |
I am trying to change the equation number style in `classicthesis` package. I followed the answers given in here but it does not affect the equation style. I want to include the chapter number in the equation number. Could someone help me? | [
0.02806830033659935,
0.020513035356998444,
-0.018150262534618378,
0.05605695769190788,
0.02322414517402649,
-0.028376299887895584,
0.009505676105618477,
-0.01151061151176691,
-0.03315731883049011,
-0.04886740446090698,
-0.012427440844476223,
0.020328469574451447,
-0.043533265590667725,
0.0... | [
0.12051025032997131,
0.2365136295557022,
0.11615785211324692,
0.28460076451301575,
0.04269867017865181,
-0.3174266517162323,
0.551982581615448,
0.10329364240169525,
-0.16576361656188965,
-0.6001439094543457,
0.17624837160110474,
0.40113911032676697,
0.16187365353107452,
0.336112380027771,
... |
With this line everything works fine - (as a **Non-Wordpress** index.php): <link href="style.css" rel="stylesheet" type="text/css"/> When I activate the file as a wp theme and replace above line with: <link href="<?php bloginfo('stylesheet_url');?>" rel="stylesheet" type="text/css"/> everything is ok, except - there are no images. For example - a div with an image: <div id="about"> <img id="thinker01" src="images/thinker01.png" width="120" height="163" /> </div> **images** folder is inside theme's folder. | [
-0.005536699667572975,
0.0011856306809931993,
0.012546790763735771,
0.029529206454753876,
0.015162006951868534,
0.004941851366311312,
0.007089610211551189,
-0.019275778904557228,
-0.0147469537332654,
-0.01437915489077568,
-0.014050798490643501,
0.005112008657306433,
-0.007157078944146633,
... | [
0.3479563891887665,
-0.245717853307724,
1.1238330602645874,
-0.1361866146326065,
0.09358227252960205,
-0.3646475076675415,
0.46991437673568726,
-0.17575351893901825,
-0.3236246705055237,
-0.32849442958831787,
-0.2842179834842682,
0.6219927668571472,
-0.2678227722644806,
-0.1184634417295456... |
> **Possible Duplicate:** > Android Software or Device to present Application on Screen > How to plug in an Android Tablet and show it's screen on PC or Laptop for > training session? > How do I project the screen of my android phone for a presentation? > How to capture video stream from Android phone screen and show it on > laptop? I am planning to make a screencast of my phone screen. I need to know a way to record the screen Please note that the Screencast & Screencapture app won't work, as I only have an armv6 processor. | [
-0.01043314952403307,
-0.008321410976350307,
-0.005717261228710413,
0.006153346039354801,
-0.06793387234210968,
-0.013203699141740799,
0.00881177932024002,
0.04639934375882149,
-0.019866641610860825,
-0.03348894417285919,
-0.009244274348020554,
0.01547551155090332,
0.011950556188821793,
0.... | [
0.32218876481056213,
0.30470502376556396,
0.40172046422958374,
0.17012977600097656,
-0.23057937622070312,
0.16581407189369202,
0.5666798949241638,
-0.18195483088493347,
-0.2725874185562134,
-0.8700006604194641,
0.23604123294353485,
0.6818488836288452,
0.1287127137184143,
0.1096615418791771... |
I would like to better understand when I have to use _regret_ and when _remorse_. In Italian we have two words: * _Rimpianto_ : used when I'm sad because I didn't do something in the past (e.g. I didn't buy a staff and now it costs the double). * _Rimorso_ : used when I'm sad because I did something wrong in the past (e.g. I offended a friend and now we don't talk anymore). I read the dictionary and it seems that the word _regret_ can be used as the world _rimorso_. > A feeling of sadness, repentance, or disappointment over something that has > happened or been done. The word _remorse_ , on the other hand, confused me because it is defined as > Deep regret or guilt for a wrong committed. So, what are the correspondence of _rimorso_ and _rimpianto_ in English? | [
0.030191784724593163,
0.007419637404382229,
-0.009963521733880043,
0.01886773109436035,
-0.00409606471657753,
-0.01543121226131916,
0.007817482575774193,
0.016243571415543556,
-0.010367238894104958,
0.026263779029250145,
-0.01177467592060566,
-0.0022719302214682102,
0.019608888775110245,
0... | [
0.1196589544415474,
0.0011944836005568504,
0.14884637296199799,
-0.404573529958725,
-0.3153223693370819,
0.36544013023376465,
0.2470891922712326,
-0.09765903651714325,
-0.4190736413002014,
-0.4627564549446106,
0.2683424651622772,
0.5067386627197266,
0.11635559052228928,
0.21991200745105743... |
What is the difference between _hypotyposis_ , _ekphrasis_ and _iconotext_? | [
0.05205318331718445,
0.01876913011074066,
-0.009252321906387806,
0.0414285808801651,
0.01463382039219141,
0.01086579728871584,
0.024290889501571655,
0.016757162287831306,
-0.009724602103233337,
0.008766810409724712,
-0.05836296081542969,
0.015254576690495014,
-0.03505741432309151,
0.025146... | [
0.5176525115966797,
0.4748650789260864,
0.05832543224096298,
-0.02160673588514328,
-0.11408963799476624,
0.564155101776123,
0.2061849981546402,
0.10835794359445572,
-0.47184163331985474,
-0.5077502727508545,
0.2768300771713257,
-0.029179038479924202,
-0.3369559049606323,
0.0804658457636833... |
I have to translate "data backend" to Russian and Hebrew without using borrowed words. I'll appreciate if someone describes what "data backend" means so I could find proper words in my language. | [
-0.023648500442504883,
0.04585949331521988,
-0.012822371907532215,
0.035990551114082336,
0.003660420188680291,
0.04987003654241562,
0.01702408492565155,
0.0254744291305542,
-0.032438766211271286,
-0.0012407655594870448,
-0.002266146009787917,
0.030312543734908104,
0.03434174135327339,
0.01... | [
0.34995245933532715,
0.5756982564926147,
-0.08094838261604309,
0.0833684653043747,
0.06118973344564438,
0.06698324531316757,
0.1839728057384491,
0.37100064754486084,
0.35127776861190796,
-0.7442325949668884,
-0.33308932185173035,
0.3384247422218323,
0.20319882035255432,
-0.0989681631326675... |
We are facing a communication problem between the public and our private network. We have a Linux based web server behind a firewall and an internet router. As part of some integration, our web portal team is trying to Telnet the public IP of an another web portal from these Linux web servers on port 80, but the connection is closed by remote host just in time it says connected. But same is successful while we try it from a Windows based web server in our network. Everything seems fine, the firewall is passing the traffic. Does it matter if one end server is running Linux and other end is a Windows server. Because the public IP application server is running Windows server. what are the key points to meet the requirement to make the communication succeed. Refer the below error logs. ]$ telnet 220.226.190.180 80 Trying 220.226.190.180... Connected to 220.226.190.180 (220.226.190.180). Escape character is '^]'. Connection closed by foreign host. | [
-0.006584148854017258,
-0.006567515432834625,
-0.005661989562213421,
0.0019011045806109905,
-0.026571113616228104,
-0.010519536212086678,
0.009187486954033375,
0.019699834287166595,
-0.011023472994565964,
-0.021793002262711525,
-0.011797482147812843,
0.01747944951057434,
-0.00423178635537624... | [
0.3365844190120697,
-0.048540834337472916,
0.36333292722702026,
0.22863233089447021,
0.12526704370975494,
-0.030641630291938782,
0.42440274357795715,
0.0689198449254036,
-0.36967507004737854,
-0.5366783738136292,
0.28498074412345886,
0.5855993628501892,
-0.15816253423690796,
0.302239716053... |
I have a page like a welcome page which should be visited only once by each subscriber. After registration, they will redirected on my welcome page. Then I want that page not be available on same user for the second time, if the will try, I want them to be redirected. Is this possible?, should I use plugin? Thanks. :) | [
-0.0006040381267666817,
0.027071570977568626,
0.0056375241838395596,
0.03175335377454758,
-0.010006992146372795,
-0.022336842492222786,
0.008274395950138569,
0.011068976484239101,
-0.02394857443869114,
0.008115259930491447,
-0.020368754863739014,
0.01574656181037426,
0.004226732067763805,
... | [
0.38874512910842896,
0.11096218973398209,
0.5496668219566345,
0.17059998214244843,
-0.009249093011021614,
-0.14655432105064392,
0.26069387793540955,
0.15994831919670105,
-0.04353895038366318,
-0.7993283271789551,
0.035911157727241516,
0.33845704793930054,
-0.3078077435493469,
0.39790150523... |
Which package contains the implementation of IPsec and which package contains the implementation of encryption algorithms that IPsec uses for encryption? I need to use custom crytographic algorithms in IPsec, so I need to edit the implementations of these packages. | [
0.02363729290664196,
-0.001666436204686761,
-0.022576602175831795,
0.03882315009832382,
-0.008016709238290787,
0.032081350684165955,
0.01696479506790638,
-0.009623047895729542,
-0.05241904407739639,
-0.021267931908369064,
-0.008131747134029865,
0.00998019054532051,
-0.012478121556341648,
-... | [
0.2278301864862442,
0.4366304874420166,
-0.18791192770004272,
0.018477780744433403,
0.050037555396556854,
-0.18577878177165985,
0.048093847930431366,
-0.11326637864112854,
0.25642475485801697,
-0.6072143316268921,
0.013041025027632713,
0.5623750686645508,
-0.4500156044960022,
-0.0728534311... |
From Wikipedia, there are three interpretations of the degrees of freedom of a statistic: > In statistics, the number of degrees of freedom is the number of values in > the **final calculation** of a statistic that are **free to vary**. > > Estimates of statistical parameters can be based upon different amounts of > information or data. The number of **independent pieces of information** > that go into the estimate of a parameter is called the degrees of freedom > (df). In general, the degrees of freedom of an estimate of a parameter is > equal to **the number of independent scores that go into the estimate** > minus **the number of parameters used as intermediate steps in the > estimation of the parameter itself** (which, in sample variance, is one, > since the sample mean is the only intermediate step). > > Mathematically, degrees of freedom is **the dimension of the domain of a > random vector** , or essentially **the number of 'free' components: how many > components need to be known before the vector is fully determined**. The bold words are what I don't quite understand. If possible, some mathematical formulations will help clarify the concept. Also do the three interpretations agree with each other? | [
0.005282479338347912,
0.015171640552580357,
-0.010309768840670586,
0.011984963901340961,
0.002956377575173974,
-0.023435290902853012,
0.01089397445321083,
-0.019852980971336365,
-0.016186028718948364,
-0.015160337090492249,
-0.0013885516673326492,
0.01131538487970829,
-0.02468126267194748,
... | [
0.12299249321222305,
-0.21069377660751343,
0.10550893843173981,
0.4381830394268036,
-0.40342235565185547,
0.07294345647096634,
-0.07562340050935745,
-0.4967735707759857,
-0.2414582222700119,
-0.6106612682342529,
-0.2500828504562378,
0.3603127896785736,
-0.013281852006912231,
0.295646876096... |
I am trying to convert a bunch of text files into structured XML files. Using string expressions I have extracted data, so I have: title={"https://www.ometz.ca/event/abiletes-dentrevue-1-678/?langID=1"};` date={2011,7,14,13,22,22.`};` content=": text/html; charset=UTF-8 Ometz A community of services for life Employment Job Seekers Job Listings Job Seeker Registration Career Counselling Employers Cocktails & Conversation Candidate Profiles Employers - Submit a Job Posting ProMontreal Entrepreneurs Fund ProMontreal Entrepreneurs Mentors Calendar You & Yours Community Assistance Counselling Services Administered Funds Orthodox Community We";` So far I have done the following to generate XML objects, for title, date, and content in turn. It is a bit messy but seems to work. a = ExportString[ XMLElement[ "title", {}, {StringReplace[ ToString[title], {"{" -> "", "}" -> ""}]}], "XML"] b = ExportString[ XMLElement[ "date", {}, {StringReplace[ ToString[date], {"{" -> "", "}" -> ""}]}], "XML"] c = ExportString[XMLElement["body", {}, {content}], "XML"]; The output is thus: <title>https://www.ometz.ca/event/abiletes-dentrevue-1-678/?langID=1</title> <date>2011, 7, 14, 13, 22, 22.</date> <body>: text/html; charset=UTF-8 Ometz A community of services for life Employment Job Seekers Job Listings Job Seeker Registration Career Counselling Employers Cocktails & Conversation Candidate Profiles Employers - Submit a Job Posting ProMontreal Entrepreneurs Fund ProMontreal Entrepreneurs Mentors Calendar You & Yours Community Assistance Counselling Services Administered Funds Orthodox Community We</body> So far so good, but I need to get it exported to an XML file, say `text.xml`. For some reason I'm hitting a wall in how to get this working, despite reading docs & some supplementary material. How would you do this? (also, any advice on the above code is appreciated) | [
-0.013037844561040401,
0.009128481149673462,
-0.004360548686236143,
0.01572340726852417,
0.00039742840453982353,
0.009358840063214302,
0.006193598732352257,
0.020471543073654175,
-0.016795378178358078,
0.00437558488920331,
0.0006137494929134846,
-0.0007062926888465881,
0.002747222315520048,
... | [
0.22910676896572113,
0.4084304869174957,
0.5829104781150818,
-0.09924311935901642,
0.21175460517406464,
0.28048044443130493,
0.41291436553001404,
-0.0062066554091870785,
-0.1965172439813614,
-0.45249396562576294,
-0.22987109422683716,
0.4948253929615021,
0.1897028386592865,
-0.015646843239... |
What are the preferred use cases for the following sets of terms: 1. Log in / Log out 2. Log on / Log off 3. Sign in / Sign out 4. Sign on / Sign off From what I can guess, "Logging in" should be used for a long-lived session (like a website), whereas "Sign in" should be for something that you will be attending to (like IM or a financial transaction). I'm a little fuzzy here... | [
-0.002980901626870036,
0.008318903855979443,
-0.012017385102808475,
0.003969541285187006,
-0.027882210910320282,
0.015222526155412197,
0.01056759338825941,
-0.006407159846276045,
-0.008599989116191864,
-0.006267750170081854,
-0.023292886093258858,
-0.006427172105759382,
0.014824592508375645,... | [
0.48211413621902466,
0.03705259785056114,
0.3162737786769867,
-0.002220937516540289,
0.2960054576396942,
-0.4319233298301697,
0.19898399710655212,
0.2239065170288086,
-0.30701515078544617,
-0.5054060816764832,
-0.32714101672172546,
0.2727815508842468,
0.007730993442237377,
0.13465236127376... |
I have a python script that I am using to create a list of all mxd files (with full pathing) in our Projects folder. The script then uses that to iterate through the list and do a findandreplaceworkspacepaths on each mxd per ESRI's how to. I am running into problems when I hit a corrupted mxd file. The have tried try/except and haven't gotten it to work. The ideal situation would be to write the corrupt filename to a file and move on so I can come back to them at the end. I'm very new with python scripting, any help would be greatly appreciated. import arcpy, os, sys, traceback, time oldpath = 'W:' newpath = 'W:\\GIS' def find(path,pattern): matches = [] for r,d,f in os.walk(path): for files in f: if files.endswith(pattern): fpath = os.path.join(r,files) matches.append(fpath) print (fpath) return matches print ("Go: ") mxdlist = (find('C:\\gis','.mxd')) print (mxdlist) print ("Starting Path Conversion") try: for mxdold in mxdlist: mxd = arcpy.mapping.MapDocument(mxdold) mxd.findAndReplaceWorkspacePaths(oldpath, newpath) time.sleep(6) mxd.save() time.sleep(6) print (mxdold) del mxd except arcpy.ExecuteError: arcpy.AddError(arcpy.GetMessages(2)) except: arcpy.AddError("Non-tool error occurred") | [
0.006760305259376764,
0.016059353947639465,
-0.013441059738397598,
0.003257542848587036,
-0.017563551664352417,
-0.0037848700303584337,
0.007035847287625074,
0.016758019104599953,
-0.016641933470964432,
-0.010062560439109802,
-0.006964144762605429,
0.00583626376464963,
0.0026766404043883085,... | [
0.2922152280807495,
0.14016807079315186,
0.22688476741313934,
0.022853460162878036,
-0.06434006989002228,
-0.28674083948135376,
0.37648627161979675,
0.2771975100040436,
-0.13467726111412048,
-0.7757849097251892,
0.20925232768058777,
0.5462682247161865,
-0.04399065300822258,
0.0266019552946... |
I am using the beamertemplate and I have one frame: \frame{ \frametitle{Title text} \begin{textblock*}{50mm}[1,1](15.9cm,9cm) \includegraphics[width=1.9cm]{thumbnail.png} \end{textblock*} \vspace{-0.5cm} \hspace{1cm} \animategraphics[scale=0.4,autoplay]{8}{movie}{000}{200} } The problem is, that the animatedgraphic has white space at the top, which is covering my title text. So how can I tell LaTex, that the title text is "more important" and that the title is in front of the animatedgraphic? So that the animatedgraphic is in the background? | [
-0.0027754371985793114,
0.01189764216542244,
0.004682624246925116,
0.004304009955376387,
0.007755254860967398,
-0.012139962986111641,
0.006413137540221214,
0.01284713763743639,
-0.006837385706603527,
0.008705880492925644,
-0.011812238022685051,
0.0041110944002866745,
0.0037934796418994665,
... | [
0.5468032956123352,
0.16815067827701569,
1.1593098640441895,
-0.30772849917411804,
-0.1309892237186432,
0.008847760036587715,
-0.24778054654598236,
-0.5692563652992249,
0.28446540236473083,
-0.5643621683120728,
0.34337857365608215,
0.7172182202339172,
-0.09700417518615723,
0.15067428350448... |
I thought the Auction House was going down March 18th... I just opened the game after updating to Diablo 3 patch version 2.x and I see no option to open the auction house... Am I missing something? | [
-0.02024219185113907,
0.018389392644166946,
-0.009563974104821682,
-0.001479126513004303,
-0.02089182287454605,
-0.033920370042324066,
0.013472044840455055,
-0.020008482038974762,
-0.03920238837599754,
0.01023755595088005,
-0.04470000043511391,
0.01078843418508768,
0.011699465103447437,
-0... | [
0.4114112854003906,
0.14977703988552094,
0.5455915927886963,
-0.05567079782485962,
-0.022572053596377373,
-0.5201632976531982,
0.6211836338043213,
0.3113320767879486,
-0.5127896070480347,
-0.3642740547657013,
0.18399982154369354,
0.5966349840164185,
0.07782059162855148,
0.2005039006471634,... |
I know this might sound like more of an engineering question that about physics, and it probably is, but bear with me: i'm still not sure if the answer to my question lies in the physics or in the engineering, or even in the politics. Most if not all of modern submarines are powered by a nuclear reactor. At sea, only the most heavyweight (like air-carriers) use a nuclear reactor. But i've never heard of nuclear-powered aircraft Is there a weight-to-power density relation making it difficult or impossible to have nuclear aircraft? or is something related to perceived safety/politics? | [
0.012778226286172867,
0.006556572858244181,
0.005109160672873259,
0.006217834539711475,
-0.0023487526923418045,
-0.005019129253923893,
0.005259597674012184,
0.011358411982655525,
-0.007435099687427282,
-0.019821876659989357,
-0.00808759219944477,
0.016608133912086487,
-0.005117731634527445,
... | [
0.6290561556816101,
-0.051731210201978683,
-0.254549115896225,
0.5446494221687317,
0.06162549555301666,
-0.10559199750423431,
-0.2920494079589844,
0.0795673057436943,
-0.6389602422714233,
-0.32206034660339355,
0.2189546823501587,
0.3132273256778717,
0.15671344101428986,
0.26101458072662354... |
I'm new to latex , but things for me are easier since i've a partner that guides me to work on my thesis, but this he doesn't know. So... i'm trying to number the pages of the appendices like: A1,A2 . so i tried to follow all the questions that already been asked and it works partially. For example i have 6 pages of appendix A and it is numbered correctly up till page 5 i.e A1,A2,A3,A4,A5 but page 6 just showing the # 6. This solution i tried: Custom page numbering for appendix this is a link to my pdf file http://www.filehosting.org/file/details/451295/main1.pdf if i have to send my code just tell me thx in advance | [
-0.0008796912152320147,
0.014132358133792877,
-0.01698964089155197,
0.02428707480430603,
-0.0029220404103398323,
0.028299637138843536,
0.0071271867491304874,
-0.0016885241493582726,
-0.02372792549431324,
-0.017215486615896225,
-0.003483228385448456,
-0.0012286924757063389,
-0.015356323681771... | [
0.41857659816741943,
0.09233880043029785,
0.5461764335632324,
-0.14039039611816406,
0.04753660410642624,
0.2929076552391052,
0.09159564971923828,
-0.30968835949897766,
-0.5828631520271301,
-0.4284569323062897,
0.21960915625095367,
0.312719464302063,
-0.23084695637226105,
0.3176869153976440... |
I have an application where I need to convert an orientation from NED - given a WGS84 latitude and longitude - to ECEF. I have yaw, pitch and roll as Tait Bryan angles (XYZ) relative to the local tangent plane but I need this in geocentric format. Initially I was reading a few white papers on ECEF coordinate conversion but all those I could find only seem to cover position and heading. I'm able to calculate both position and heading correctly but I'm not sure what is is involved to transform pitch and roll. Roll is particularly tricky as a standard 3D velocity vector doesn't include a roll component. My gut feeling is the solution is going to involve converting to quaternions or DCM and coming up with an appropriate transformation matrix but I haven't been able to find any examples of this being done. Any advice would be appreciated. | [
-0.021276388317346573,
0.012630395591259003,
-0.014747153036296368,
0.02621801570057869,
-0.007021408062428236,
0.01227632723748684,
0.010231497697532177,
0.01769608072936535,
-0.016929225996136665,
-0.018589599058032036,
0.004124040715396404,
0.01688467152416706,
-0.016741905361413956,
-0... | [
0.49150970578193665,
0.19432033598423004,
0.6794089674949646,
-0.33876320719718933,
-0.3651479482650757,
0.2555108964443207,
-0.25945785641670227,
-0.41351258754730225,
-0.028451668098568916,
-0.7423880696296692,
0.10576403886079788,
0.1863573044538498,
0.24868237972259521,
0.3201788663864... |
At some point in the game, you start getting bribes of 5 credits for every 2 immigrants you detain. Whilst slamming the detain button as soon as you see it is easy, it often takes a while to get a suspicious character out of my booth. Given that you don't NEED to detain people, doesn't it make more economic sense to always reject their visa and get 5 credits per immigrant, rather than 2.5 with detaining? | [
-0.013736986555159092,
0.024101633578538895,
-0.022098710760474205,
-0.00859332550317049,
-0.013794420287013054,
-0.00942697748541832,
0.009113519452512264,
-0.02969670668244362,
-0.018385419622063637,
0.004843798000365496,
-0.013312085531651974,
0.01788492128252983,
-0.009610706008970737,
... | [
0.37957045435905457,
-0.13890354335308075,
0.32080787420272827,
0.36477696895599365,
0.2204619199037552,
0.04234161972999573,
0.1292242705821991,
0.09400901943445206,
-0.15976600348949432,
-0.5752148628234863,
0.22630900144577026,
0.24148614704608917,
-0.020186960697174072,
0.0387410223484... |
Is there any visual LaTeX-table editor for linux? I want something where I can load existing LaTeX-tables, modify them and save it in LaTeX. | [
0.05280556529760361,
0.014549296349287033,
0.0031080760527402163,
0.00877946987748146,
0.03564026206731796,
-0.00642787991091609,
0.015170148573815823,
0.000007068569175316952,
-0.04442078247666359,
0.0198143869638443,
0.0007106824195943773,
0.00740021001547575,
0.012795933522284031,
-0.01... | [
0.08952005952596664,
-0.024233780801296234,
0.017615916207432747,
0.3168831169605255,
-0.005115277599543333,
-0.0212392657995224,
-0.3162479102611542,
0.5338804125785828,
-0.24081255495548248,
-0.7274085283279419,
0.29305723309516907,
0.7086725234985352,
-0.2608025372028351,
-0.25435662269... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.