text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
Can you Start Skyrim 'silent' without the launcher showing up? | [
0.013981688767671585,
0.041111718863248825,
0.02388928085565567,
-0.05147650092840195,
-0.07902875542640686,
-0.024252936244010925,
0.018978100270032883,
-0.096658855676651,
-0.012406823225319386,
0.029785824939608574,
-0.055056482553482056,
0.07330629229545593,
-0.008561523631215096,
0.02... | [
0.776683509349823,
0.20184673368930817,
0.11059368401765823,
0.4058402478694916,
-0.06700744479894638,
-0.48001712560653687,
0.4418674409389496,
-0.10087282955646515,
-0.023417098447680473,
-0.03426579385995865,
-0.08275619894266129,
0.3835933208465576,
0.08476847410202026,
-0.383888989686... |
I know how the lambda expresion works and I know it is an argument for `.filter()` that establish the criteria to filter with. But I don't get how `.filter()` uses the argument, in this case a lambda expression, because .filter() does't have an implementation or at least require one at compile time. I search for this unknown implementation in the Oracle's site but there are just a few words explaning how it works and no code at all. Is that implemetation hidden or is created automatically by the java compiler? Does an aggregate operation need one? double average = roster .stream() .filter(p -> p.getGender() == Person.Sex.MALE) .mapToInt(Person::getAge) .average() .getAsDouble(); `roster` is a `List<Person>` instance of `ArrayList<Person>` `Person` is a simple class that represents a person | [
-0.003475927747786045,
0.01644185185432434,
-0.008423477411270142,
0.010871783830225468,
-0.019802257418632507,
0.0018591675907373428,
0.008412521332502365,
0.016344746574759483,
-0.01877356879413128,
-0.0038755517452955246,
-0.01783289946615696,
0.01658652164041996,
0.0005189059302210808,
... | [
0.309805303812027,
0.14524723589420319,
-0.025772113353013992,
0.03274516388773918,
-0.07963421195745468,
-0.052147164940834045,
0.14423303306102753,
-0.2307351529598236,
0.10455522686243057,
-0.3306189775466919,
0.196187362074852,
0.45851653814315796,
-0.5671395659446716,
-0.0168759841471... |
Whenever I `ping` on my personal computer, nothing gets displayed other than a single statement. Upon stopping execution, I see that several packets were transmitted in the statistics. On top of all this, I'm getting a rather absurd packet loss. What can I do to make `ping` react in a sane manner? I am running **Centos 6.5 64bit**. Below is an example output when I try `ping google.com`: [root@Virus os]# ping google.com PING google.com (74.125.230.160) 56(84) bytes of data. ^C --- google.com ping statistics --- 5 packets transmitted, 0 received, 100% packet loss, time 4274ms `ifconfig -a`: [root@Virus os]# ifconfig -a eth0 Link encap:Ethernet HWaddr B4:B5:2F:29:FE:D7 UP BROADCAST MULTICAST MTU:1500 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:0 (0.0 b) TX bytes:0 (0.0 b) lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:108 errors:0 dropped:0 overruns:0 frame:0 TX packets:108 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:9081 (8.8 KiB) TX bytes:9081 (8.8 KiB) wlan0 Link encap:Ethernet HWaddr 84:4B:F5:14:9B:58 inet addr:172.20.40.55 Bcast:172.20.255.255 Mask:255.255.0.0 inet6 addr: fe80::864b:f5ff:fe14:9b58/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:56944 errors:0 dropped:0 overruns:0 frame:0 TX packets:40200 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:55308863 (52.7 MiB) TX bytes:6291284 (5.9 MiB) `iptables -L`: [root@Virus www]# iptables -L Chain INPUT (policy ACCEPT) target prot opt source destination ACCEPT all -- anywhere anywhere state RELATED,ESTABLISHED ACCEPT icmp -- anywhere anywhere ACCEPT all -- anywhere anywhere ACCEPT tcp -- anywhere anywhere state NEW tcp dpt:ssh REJECT all -- anywhere anywhere reject-with icmp-host-prohibited Chain FORWARD (policy ACCEPT) target prot opt source destination REJECT all -- anywhere anywhere reject-with icmp-host-prohibited Chain OUTPUT (policy ACCEPT) target prot opt source destination `netstat -nr`: [root@Virus www]# netstat -nr Kernel IP routing table Destination Gateway Genmask Flags MSS Window irtt Iface 172.20.0.0 0.0.0.0 255.255.0.0 U 0 0 0 wlan0 0.0.0.0 172.20.4.254 0.0.0.0 UG 0 0 0 wlan0 | [
-0.025601238012313843,
0.006133677437901497,
-0.021134579554200172,
0.013318484649062157,
-0.0037579908967018127,
-0.022964203730225563,
0.006490775849670172,
0.0032950383611023426,
-0.022911010310053825,
-0.010467657819390297,
0.002978317439556122,
0.007593954913318157,
0.011986041441559792... | [
0.3876252770423889,
0.1299130618572235,
0.2952370047569275,
-0.06580869108438492,
-0.16617579758167267,
-0.08770795166492462,
0.6870156526565552,
-0.11678794771432877,
-0.3814883232116699,
-0.642879843711853,
0.32699427008628845,
0.34615379571914673,
-0.27161461114883423,
0.268959730863571... |
I've been tasked with creating a simple MapBasic plug-in that will loop through a table in MapInfo, zoom to the location of each point, create a circle to highlight the point, and convert the map to JPEG. I've got all the steps working except the loop (I'll post a separate question for that) and the creating a circle. This is the code I use for outputting a single JPEG; levelOfZoom = 2 circRad = (levelOfZoom / 10) Select * from TabName where StructureNo = sName Into Zoomtable Set Map Center (Zoomtable.Easting, Zoomtable.Northing) east = Zoomtable.Easting north = Zoomtable.Northing Add Map layer Zoomtable Insert Into Zoomtable (obj) Values(CreateCircle(east, north, circRad)) Close Table Zoomtable Close Table Selection Set Map Zoom levelOfZoom Units "km" Save Window my_window_id As "U:\MapInfoOutputs\"+sName+" at "+levelOfZoom+"km.JPG" Type "JPEG" Width 5.20833 Units "in" Height 3.75 Units "in" Resolution 128 I'm getting a 'Zoomtable is read-only' error when I try and execute. If I change the `Insert Into Zoomtable (obj) Values(CreateCircle(east, north, circRad))` To; `Update Zoomtable set obj= CreateCircle(east, north, circRad)` It creates the circle, but replaces the original point. I need the original point to show with the circle round it to highlight it. Also, the circles need to be removed once the jpeg has been created, I'm not sure how to do this, but having a map with | [
-0.005694348365068436,
0.005512857344001532,
-0.007975542917847633,
0.007759542670100927,
-0.007981955073773861,
0.0070053208619356155,
0.0061982665210962296,
0.021383054554462433,
-0.021972648799419403,
-0.02369432896375656,
-0.0005290996050462127,
0.007577551528811455,
0.000822084955871105... | [
0.32887935638427734,
-0.13031184673309326,
0.5739970803260803,
-0.05687818303704262,
-0.03341153264045715,
-0.15051396191120148,
-0.09824256598949432,
-0.26458030939102173,
-0.2141915261745453,
-0.7410877346992493,
-0.020596913993358612,
0.3708285093307495,
-0.18700623512268066,
-0.0479247... |
I have been working on logo for my department and I have the following code. The problem is that whenever I compile the diagram seems to be larger than the actual logo. I do not know what is the problem. I know the numbers I have selected are not the best but any insights into the matter will be highly appreciated. \documentclass[letterpaper]{article} \usepackage[top=2cm,bottom=2cm,left=2cm,right=2cm]{geometry} %\usepackage{amsmath,amssymb,units} %\usepackage{enumitem,multicol} \usepackage{tikz} %\usetikzlibrary{arrows} \usepackage{lipsum} \begin{document} \lipsum[1-2] \begin{tikzpicture}[y=0.80pt, x=0.8pt,yscale=-1] \path[draw=black,fill=white] (258.9527,290.5199) .. controls (173.9885,538.4766) and (568.5860,261.2969) .. (306.5098,257.1141) .. controls (44.4337,252.9312) and (429.9845,542.5624) .. (352.9767,292.0206) .. controls (275.9689,41.4788) and (119.6549,497.6604) .. (334.1376,346.9999) .. controls (548.6203,196.3394) and (66.4622,188.6439) .. (276.0276,346.0724) .. controls (485.5930,503.5010) and (343.9169,42.5633) .. (258.9527,290.5199) -- cycle; \end{tikzpicture} \lipsum[1-2] \end{document} See image below:  | [
-0.003757933620363474,
0.0018878945847973228,
-0.007108072750270367,
0.029866017401218414,
-0.02771455980837345,
0.0008102443534880877,
0.006989150308072567,
0.006851540878415108,
-0.01119011640548706,
-0.026510901749134064,
-0.013167336583137512,
-0.0022135190665721893,
-0.00675763282924890... | [
0.37568992376327515,
0.33397260308265686,
0.6125554442405701,
-0.3057667911052704,
0.09462124109268188,
0.20625123381614685,
0.0358649305999279,
-0.13263356685638428,
-0.3490905165672302,
-1.2090524435043335,
0.12374412268400192,
0.26772066950798035,
-0.06596623361110687,
0.012274748645722... |
I have some actions like this function simple_notification() { echo '<div>This is a notification</div>'; } add_action( 'site_notices', 'simple_notification' ); function simple_notification2() { echo '<div>This is a notification</div>'; } add_action( 'site_notices', 'simple_notification2' ); function simple_notification3() { echo '<div>This is a notification</div>'; } add_action( 'site_notices', 'simple_notification3' ); Now i'm displaying those action by calling `do_action` in my template <?php do_action('site_notices'); ?> I would like to display notice count near a menu link called `notifications` Can anyone tell me how to calculate it? ## Update: As you can see, three notices displayed when i use `do_action('site_notices');` So I want to display in the menu like `notifications (3)` | [
-0.01606760174036026,
0.0027945879846811295,
0.0007800988969393075,
0.022257735952734947,
-0.007153975777328014,
-0.003522851038724184,
0.008151322603225708,
0.012242832221090794,
-0.013301778584718704,
0.0181229580193758,
-0.023147229105234146,
0.005386837758123875,
-0.006392928771674633,
... | [
0.35669803619384766,
-0.06146012619137764,
0.3339724838733673,
-0.4828457236289978,
-0.15146292746067047,
0.2828141748905182,
0.45445331931114197,
-0.6592275500297546,
-0.13032014667987823,
-0.5784000754356384,
0.17010994255542755,
0.40070369839668274,
-0.6144178509712219,
0.00216829963028... |
In preparation for my end of semester R exam, I've been working on the following CSV file which was modelled on R, which I'm currently having problems figuring out.   To investigate this issue I decided to introduce a bilinear variable (SxPR) which is the multiplication of the variable S (power plant’s capacity) and the categorical variable PR (whether LWR existed on the same site previously) and this can be seen in the excel data. My current model which I've been asked to 'add suitable terms' to in order to investigate whether there is good reason to believe that the impact of a power plant's capacity on construction costs might not be the same for plants that did have a LWR on the same site previously and those that did not is:  This gives the equation $C = -9752+ 140.3*D + 4.868*T2+ 0.4180*S- 86.49*PR+ 153.2*NE- 8.38*N$ So basically, I need to confirm that what I've currently done to solve this problem is correct or if not, what else can I do because I've had a lot of problems with this question.  I concluded that the results show the bilinear term (SxPR) has a p-value>0.05 which suggests there is no interaction between S and PR (since SxPR is not a significant predictor to the model). This causes me to believe the impact of power plant capacity on construction cost is not the same for the plants that did have a LWR existing on the same site previously and those that did not). To support this argument, the Scatter Plot below shows the relationship between the Construction Cost and Power Plant Capacity depending on whether or not a LWR existed on the same site previously. The crosses are the plots of the Power Plant Capacity where a LWR did not exist on the same site previously and the diamonds where a LWR did exist on the same site previously.   Would anyone agree I'm on the right lines. Apologies for the lengthy question. Hope it makes sense! | [
0.0043760500848293304,
0.010613225400447845,
-0.0007188485469669104,
0.01265043206512928,
-0.0005750765558332205,
-0.009335819631814957,
0.004620784893631935,
0.023155996575951576,
-0.014648842625319958,
-0.017115917056798935,
-0.001770124537870288,
0.005425277166068554,
-0.01048986427485942... | [
0.3802901804447174,
-0.20818187296390533,
0.6610513925552368,
-0.023668790236115456,
-0.48252931237220764,
0.11955636739730835,
-0.07999534159898758,
-0.29843422770500183,
-0.3438304364681244,
-0.03447890281677246,
0.026904338970780373,
0.17132854461669922,
-0.07714202255010605,
0.69835245... |
Currently, I am using a Windows machine with _Mathematica 8_. I noticed a difference in a series expansion of the function `EllipticE[]` in comparison with a result given by _Mathematica 9_ on Linux (which I was using previously). In _Mathematica 8_ on Windows the following input: Series[EllipticE[I c, m], {c, Infinity, 0}] // PowerExpand // FullSimplify produces a warning: > General::ivar: I c is not a valid variable. >> and the following output (slightly rearranged by me to better fit the browser): $$-\frac{i\left(6m+\frac{(1+m)}{\sinh^2(c)}\right)}{6\sinh(c)m^{3/2}}-\text{EllipticE}[m]+\frac{m \text{EllipticE}[\frac{1}{m}] + (1-m)\text{EllipticK}[\frac{1}{m}] }{\sqrt{m}}-$$ $$\frac{(1+m \cosh(2 c))\sqrt{\text{Limit}\big[-m\sinh^2(c)~,~i c \to 0\big]}}{\sinh^2(c)2m}$$ Same output as a code: -((I (6 m Csch[c] + (1 + m) Csch[c]^3))/(6 m^(3/2))) - EllipticE[m] + (m EllipticE[1/m] - (-1 + m) EllipticK[1/m])/Sqrt[m] - (((1 + m Cosh[2 c]) Csch[c]^2 Sqrt[Limit[-m Sinh[c]^2, I c -> 0]])/(2 m)) Now, in _Mathematica 9_ on Linux there was no such warning, and no `Limit` term appeared. I am confused about how to treat this `Limit` term, since it might just be a sign of something going terribly wrong in the guts of _Mathematica 8_. Does anyone have an advice on how to proceed? Maybe some of you can evaluate the same series expansion in a different version of _Mathematica_ so that we could compare results? ### EDIT Evaluating instead: Series[EllipticE[c, m], {c, I Infinity, 0}] // PowerExpand // FullSimplify worked as a charm without errors. | [
0.026182828471064568,
-0.004807621706277132,
-0.0252658873796463,
0.007774005644023418,
-0.015028740279376507,
0.005334497895091772,
0.008089110255241394,
0.009624410420656204,
-0.012884493917226791,
-0.028754238039255142,
-0.008020841516554356,
0.004795118235051632,
-0.020370755344629288,
... | [
-0.2791461944580078,
0.18517081439495087,
0.5758863091468811,
-0.10656195878982544,
0.08414715528488159,
-0.09685304760932922,
0.2451549768447876,
-0.21622289717197418,
-0.3235815763473511,
-0.3002857267856598,
-0.10331284999847412,
0.5506556034088135,
-0.39572954177856445,
0.3744655549526... |
If I create a custom map with the Starcraft 2 map editor, how can I specify that some unit upgrades are already researched? For example, if I want all the units to already have the weapons upgrade 1 at the beginning of the game, can I specify this in the editor? I've looked through the menus but didn't find anything that looked like it would relate to unit upgrades. | [
0.02452722005546093,
0.011951490305364132,
0.003924711607396603,
0.030341794714331627,
0.02693590708076954,
-0.014078489504754543,
0.009628145024180412,
-0.020074596628546715,
-0.02986542135477066,
0.007347457576543093,
-0.0056111449375748634,
0.0356905534863472,
-0.028622519224882126,
0.0... | [
0.21413949131965637,
0.04120124876499176,
0.13267849385738373,
0.12524540722370148,
-0.10836412012577057,
-0.09366458654403687,
-0.21589650213718414,
-0.3820017874240875,
-0.2954840362071991,
-0.43823036551475525,
0.178704172372818,
0.710736095905304,
0.01272124145179987,
-0.19303074479103... |
In ModelBuilder of ArcGIS 10.2 does there exist a method or routine for parsing data from a KML file and "print" this data in a shapefile or convert this data a to shapefile? | [
-0.008154287002980709,
0.008953696116805077,
0.003689864417538047,
0.02683829329907894,
-0.013657317496836185,
0.06274405866861343,
0.014557916671037674,
0.0016606581630185246,
-0.023357335478067398,
-0.036077700555324554,
0.0091908760368824,
0.018665587529540062,
0.0022622982505708933,
0.... | [
0.33231985569000244,
-0.11103702336549759,
0.30370792746543884,
0.2398606687784195,
0.09458117932081223,
-0.05485624819993973,
-0.020099900662899017,
-0.16222968697547913,
-0.045802488923072815,
-0.4919012784957886,
0.063486248254776,
0.2488543540239334,
-0.3367369472980499,
-0.06535285711... |
It is said in this article that for the Wii U, an external hard drive is probably going to be needed. It is said in this article that the Wii U will work best with an external hard drive that uses its own power adapter. But I wonder for some small, portable hard drive, does somebody have any succcess using it being powered by the Wii U? (such as the Western Digital Passport, the Seagate Backup Plus, or the Toshiba Canvio, or any other one.) (it is also said that USB Flash drive technically can work, but with large amount of read and write, it can fail after too many re-writes and so it is not suitable as a Wii U drive) | [
0.012132519856095314,
-0.007051106542348862,
-0.005879642907530069,
0.01704075187444687,
-0.008042072877287865,
-0.004404502920806408,
0.007745082024484873,
0.0005239639431238174,
-0.009678667411208153,
-0.01735047623515129,
-0.004727575927972794,
0.008248250931501389,
0.028199577704072,
0... | [
0.2523437738418579,
-0.21374711394309998,
0.3805382251739502,
0.6007822751998901,
0.13389952480793,
-0.4587213099002838,
-0.2088172733783722,
0.11176265776157379,
-0.1755392849445343,
-0.32631731033325195,
-0.2523196339607239,
0.859290361404419,
0.018349647521972656,
0.05084206908941269,
... |
How to set this kind of animation on beamer? > show_me_first show_me_second. I'm trying to show this sentence as described. The `show_me_first` is showed first, and on click oon the Down button, it will shows the `show_me_second`. How to deal with this, please? Thanks a lot! | [
0.006726396735757589,
0.01137442234903574,
-0.009616862051188946,
0.01099917571991682,
-0.05610053986310959,
-0.018796920776367188,
0.010541054420173168,
0.007150604389607906,
-0.015298383310437202,
0.03272268921136856,
-0.03422514721751213,
0.0008611318771727383,
0.009948470629751682,
0.0... | [
0.3441146910190582,
-0.09977859258651733,
0.4678601622581482,
0.14012755453586578,
-0.5481252670288086,
0.14816640317440033,
0.048806339502334595,
0.3569164276123047,
-0.13266704976558685,
-0.9389914274215698,
-0.0039088851772248745,
0.4011366665363312,
-0.13683874905109406,
0.161766961216... |
I'm using a plug-in ("Wordpress plugin random post slider") for Wordpress that pulls a random post from a category, and displays it in a slider for use as a shortcode or php code in a template. I've already made some modifications to the plug-in to get it to display properly on my site (http://lillypillyps.com.au), but it has some limitations that the developer will take forever to follow up on. I was wondering if someone might be able to help with the following questions in order of importance: 1. it's using some sort of truncate on the post content which is currently set to 170 characters (through the plug-in's settings page in the admin of WP). I want it to truncate to the nearest word so it doesn't cut-off mid-word. 2. is it possible to ONLY have the "... READ MORE" button display IF the content is longer than the truncation point? ie. not display if the content is less that 170 characters. 3. there is a few options for the transitions between the slides (scrollLeft/scrollRight/scrollUp/scrollDown), but the same jquery should be capable of fade-in/fade-out shouldn't it? How can I used this transition instead of the ones the plug-in "allows" me to use? | [
-0.005700854584574699,
0.0022506495006382465,
0.0019214254571124911,
0.023138631135225296,
-0.02095487155020237,
-0.011203874833881855,
0.007890146225690842,
0.014468962326645851,
-0.016595788300037384,
-0.02089853025972843,
-0.013167446479201317,
-0.0018293714383617043,
0.012172594666481018... | [
0.7656012177467346,
-0.17016369104385376,
0.10268149524927139,
0.008007551543414593,
-0.10162923485040665,
0.17672927677631378,
-0.19232723116874695,
0.025659529492259026,
-0.02374516986310482,
-0.4667779207229614,
0.48277774453163147,
0.25814661383628845,
-0.10285605490207672,
0.169275611... |
What does the emphasized part (the sentence in bold) mean? > The hackers are the technicians of this science - “It’s a term of derision > and also the ultimate compliment.” They are the ones who translate human > demands into code that the machines can understand and act on. They are > legion. Fanatics with a potent new toy. A mobile new-found elite, with its > own apparat, language and character, its own legends and humor. Those > magnificent men with their flying machines, scouting a leading edge of > technology which has an odd softness to it; outlaw country, **where rules > are not decree or routine so much as the starker demands of what’s > possible.** | [
-0.004321441985666752,
0.002060795435681939,
0.0031599001958966255,
0.006184129975736141,
-0.03388974815607071,
-0.012863440439105034,
0.004684702958911657,
-0.013620037585496902,
-0.009247834794223309,
-0.015004927292466164,
-0.009993946179747581,
0.01181633211672306,
0.015375048853456974,
... | [
0.578068196773529,
0.24073880910873413,
-0.020024996250867844,
0.572030782699585,
-0.15211471915245056,
0.18494686484336853,
0.29612433910369873,
0.1117430105805397,
-0.14313754439353943,
-0.3272837698459625,
-0.4453046917915344,
0.859095573425293,
-0.21930669248104095,
-0.0619098059833049... |
I want to design a `tikz` shape that uses different background paths for different purposes. Consider the following minimal example where a new shape named `simple rectangle` is declared. The background path used for filling the shape is different from the background path for other purposes, like drawing or clipping: filling does not include the region of the text below the base line. \documentclass[tikz]{standalone} \makeatletter \pgfdeclareshape{simple rectangle}{ \savedanchor{\upperrightcorner}{ \pgf@x=.5\wd\pgfnodeparttextbox % widht of the box \pgf@y=.5\ht\pgfnodeparttextbox % height of the box, ignoring the depth } \saveddimen{\depth}{ \pgf@x=\dp\pgfnodeparttextbox } \anchor{north east}{ \upperrightcorner } \anchor{center}{ \pgfpointorigin } \anchor{text}{ \upperrightcorner \pgf@x=-\pgf@x \pgf@y=-\pgf@y } \backgroundpath{ \tikz@mode \upperrightcorner \pgf@xa=-\pgf@x \pgf@ya=-\pgf@y \iftikz@mode@draw \advance \pgf@ya by -\depth \fi \pgfpathrectanglecorners{\upperrightcorner}{\pgfpoint{\pgf@xa}{\pgf@ya}} } } \makeatother \begin{document} \huge \begin{tikzpicture} \node at (0,3) [simple rectangle]{This is my test.}; \node at (0,2) [simple rectangle,fill=yellow!40]{This is my test.}; \node at (0,1) [simple rectangle,draw]{This is my test.}; \node at (0,0) [simple rectangle,fill=yellow!40,draw]{This is my test.}; \end{tikzpicture} \end{document}  This example works only if filling is not combined with other actions, like drawing, as can be seen. How to fix this? | [
0.011580339632928371,
0.01588609255850315,
-0.005251075141131878,
0.02814013883471489,
-0.004262627102434635,
-0.0031890214886516333,
0.008662533015012741,
0.0110856918618083,
-0.016059182584285736,
-0.015939177945256233,
-0.018322590738534927,
0.0004906873218715191,
0.004878811072558165,
... | [
0.39131972193717957,
0.008474779315292835,
0.30220866203308105,
-0.1518617421388626,
0.2459372729063034,
-0.13827332854270935,
0.041577089577913284,
-0.09894806891679764,
0.07858970761299133,
-0.7498224377632141,
-0.23855915665626526,
0.45094719529151917,
-0.3447483479976654,
0.10789981484... |
Where is it possible to download the source code of the first _open source_ browser _Netscape Navigator_ released in about 1998? I already searched for it and wasn't able to find the original version of this browser from 1998. | [
0.013944149017333984,
-0.004323347471654415,
-0.034043412655591965,
0.00037960667395964265,
-0.020130319520831108,
0.018698235973715782,
0.011462635360658169,
0.001944071613252163,
-0.04009794816374779,
-0.019248858094215393,
-0.011673137545585632,
0.03427474573254585,
0.026587670668959618,
... | [
0.3546629846096039,
0.0629049688577652,
0.024334056302905083,
0.11889264732599258,
0.13861893117427826,
-0.289311021566391,
-0.08737459033727646,
0.697695791721344,
-0.24890291690826416,
-0.6724041104316711,
0.5415944457054138,
0.4094080626964569,
-0.30478689074516296,
0.038706570863723755... |
I am new to GRASS and started training on the spearfish sample data set in GRASS 6.4.3. I selected a bicycle tour and transformed it into a polyline. My next goal is to read out the slope and elevation data from the **elevation.dem** in order to create a 3D polyline. I want to use the **v.drape** function, but although the input polyline had an attribute table containing rows for all the street segments that form part of the bicycle tour, my output 3D-polyline doesn't have an attribute table anymore. v.drape --overwrite input=tour1poly@foo rast=elevation.dem@PERMANENT output=tour1poly3D Warning: Database connection not defined for layer <0> First of all: Why is that? Going on, I split the 3D-Polyline (using **v.split** along the vertices) into over 300 segments, so I can calculate the slope of each segment with **v.to.db**. But, of course, I first of all need an attribute table for all the 300+ segments to write their slope into. Is there a function to help me write the results (coordinates, elevation, length etc. of each line or segment) from v.drape and v.split into an attribute table? So v.to.db can add the slope of each segment to said attribute table? I know there is v.db.addtable, but how can I access the necessary information to create the proper attribute table with over 300 rows that v.to.db needs? Am I missing some steps? Are you missing some information in order to help me? Thank you for your help. | [
-0.0034651686437427998,
0.012310893274843693,
-0.007991819642484188,
0.009729687124490738,
-0.025313494727015495,
-0.00663602352142334,
0.009093803353607655,
0.010897443629801273,
-0.01659361831843853,
-0.016829533502459526,
-0.004078627098351717,
0.007678499445319176,
-0.004850002005696297,... | [
0.7357414364814758,
-0.10752859711647034,
0.5220118165016174,
0.005169152282178402,
-0.34165409207344055,
0.22118185460567474,
0.10473477095365524,
-0.2210470736026764,
-0.11686884611845016,
-1.032041072845459,
0.2421080619096756,
0.31024545431137085,
0.002454211935400963,
0.11379038542509... |
I know that describing the trajectory of all planets in the solar system around the Earth is much more complicated than if we take the Sun as the reference point. But besides this, what is the simplest experiment that can prove that the Sun is a more "inertial" point of reference than the Earth? Do we have to look at the stars in the background or something like that? | [
0.017080452293157578,
0.004851662088185549,
-0.025757847353816032,
0.005724920425564051,
-0.03729640319943428,
0.0061684539541602135,
0.009105391800403595,
0.00912315770983696,
-0.02635069377720356,
-0.03646453469991684,
-0.001877819886431098,
0.020069314166903496,
-0.006956376135349274,
0... | [
0.47418200969696045,
0.2476014494895935,
0.24854081869125366,
0.30638015270233154,
0.13544383645057678,
-0.14655084908008575,
-0.2797735929489136,
0.2475663274526596,
-0.4244188666343689,
-0.4287099242210388,
0.31425613164901733,
0.439098984003067,
-0.01828179694712162,
0.30805277824401855... |
Recently I came across this site. It says, in Linux with QT, it is possible 'to highlight characters' by pressing `Shift``Ctrl` {`Left`,`Right`} or `Shift` {`Left`,`Right`}. Could somebody please explain this feature with a screenshot or something? | [
0.024771852418780327,
0.008396325632929802,
-0.014083923771977425,
0.003255903022363782,
-0.010792084969580173,
-0.006384395062923431,
0.006647439207881689,
0.023021096363663673,
-0.013812826946377754,
0.010651923716068268,
-0.01688106544315815,
0.00881017092615366,
0.0015450718346983194,
... | [
-0.04396442696452141,
0.05389944463968277,
0.0513615645468235,
0.05419142544269562,
0.05741104483604431,
0.030837271362543106,
-0.03477989137172699,
0.5025731325149536,
0.09924578666687012,
-0.5983745455741882,
0.1268794983625412,
0.17368671298027039,
0.04096948355436325,
0.108090691268444... |
If we have the following histogram:  And, want to describe it based on one of the following options, which one do you think is correct? Why? 1. positively skewed 2. symmetric 3. none of the above 4. negatively skewed My initial suggestion would be "3. none of the above". What do you think? Thanks. | [
-0.009077184833586216,
0.004788494203239679,
-0.005572661757469177,
0.033660031855106354,
-0.0014115279773250222,
-0.02893196791410446,
0.007951649837195873,
0.012885861098766327,
-0.015344380401074886,
-0.007468708790838718,
-0.01325561199337244,
-0.008773534558713436,
-0.006783120799809694... | [
0.4022674858570099,
-0.12462318688631058,
0.5603015422821045,
0.3246912658214569,
-0.07786352187395096,
-0.03891483321785927,
0.2836049199104309,
-0.378691703081131,
0.1129918098449707,
-0.6417797803878784,
-0.033672455698251724,
0.33666402101516724,
-0.04162399098277092,
-0.21139781177043... |
I was reading andrew ng's machine learning lecture notes on SVM. I came across the following equation (finding the optimal value for the intercept term $b$ in the SVM problem):  However, I have no idea how the intercept term $b$ is derived by solving the primal problem ? I believe the primal's Lagrangian is: $$\min_{w,b} \max_{\alpha} \mathcal{L}(w,b,\alpha) = \min_{w,b} \max_{\alpha} \frac{1}{2} ||w||^2 - \sum_{i=1}^m \alpha_i [y_i (w^T x_i + b) - 1]$$ But how do I solve for $b$ ? Any help will be great. Thank you very much. | [
0.0004124972620047629,
0.0014182141749188304,
-0.009524011984467506,
-0.0070609524846076965,
-0.02760554663836956,
-0.020342670381069183,
0.007612415589392185,
0.0043618520721793175,
-0.013679273426532745,
0.002635850803926587,
-0.011600775644183159,
0.01146712340414524,
-0.02265826053917408... | [
0.10914889723062515,
-0.05643414333462715,
0.3799574077129364,
-0.20914742350578308,
-0.10530643165111542,
0.2689536213874817,
-0.001908348873257637,
-0.04033378139138222,
0.06579764932394028,
-0.06437269598245621,
-0.004613125696778297,
0.47486934065818787,
-0.5527040362358093,
0.16176091... |
I have heard the claim over and over that you won't feel anything when crossing the event horizon as the curvature is not very large. But the fundamental fact remains that information cannot pass through the event horizon so you cannot feel your feet when they have passed it. Is there a way to cross the even horizon at reasonable speed in radial direction (below say 0.01*c )? So what would it really be like? | [
0.002897717524319887,
0.025753777474164963,
-0.018734030425548553,
0.0004639422695618123,
-0.012807853519916534,
-0.024088116362690926,
0.008469106629490852,
-0.008465415798127651,
-0.01885681226849556,
-0.015691012144088745,
-0.0000817339459899813,
0.029918022453784943,
0.004236685112118721... | [
0.29120805859565735,
-0.11623455584049225,
0.48234954476356506,
0.33333876729011536,
-0.0072752125561237335,
-0.31509050726890564,
0.40464967489242554,
0.31004607677459717,
-0.6150158047676086,
-0.45036450028419495,
0.2730475068092346,
0.11845438182353973,
0.22360685467720032,
0.4713060855... |
If memory serves, I thought I remembered reading somewhere that the GTS Pokemon would be filled with all the Pokemon names. But when I'm searching for one that I haven't seen before, for example a Dragonite, I don't see it in the "D" list. But there is an option you can pick to where you can type a Pokemon name and it will search for it. So can you search for Pokemon using that option that haven't been seen in the Pokedex? Also can you shorten the name to search for any pokemon with what you're looking for? Say I searched for "Drago", would that find Dragonair and Dragonite? | [
0.009927746839821339,
0.00328578008338809,
-0.00034351786598563194,
0.0064301034435629845,
0.008626185357570648,
0.009221361018717289,
0.00702612055465579,
0.012141034007072449,
-0.01795097626745701,
-0.003629758721217513,
-0.0014342379290610552,
0.01611321233212948,
0.0010151676833629608,
... | [
0.45479848980903625,
-0.04204459860920906,
-0.023969147354364395,
0.38050299882888794,
-0.2754984498023987,
-0.03764995187520981,
0.1056295707821846,
0.4153638780117035,
-0.23766055703163147,
-0.3870109021663666,
0.014929373748600483,
0.243406280875206,
-0.028547873720526695,
0.41881051659... |
I was told that the word _uppity_ has some racial connotations originating from the times of segregation in the South. I never thought of it as such. I kind of like the sound of the word but was wondering if it should be avoided in the presence of black people. | [
-0.014863618649542332,
0.021018922328948975,
0.005374329164624214,
0.04026390239596367,
-0.010389902628958225,
0.002373964060097933,
0.010799811221659184,
-0.0017905943095684052,
-0.020899178460240364,
0.029392018914222717,
-0.01581740379333496,
0.028705084696412086,
0.011748041957616806,
... | [
0.5086013674736023,
0.20164765417575836,
-0.17023010551929474,
-0.27455055713653564,
-0.018314767628908157,
-0.08459390699863434,
0.38531753420829773,
0.44263094663619995,
-0.3942335844039917,
0.0006981766200624406,
-0.00174316531047225,
0.2684977948665619,
-0.5197818279266357,
0.368289977... |
How can I configure CUPS and/or KDE so that my printer is the default when I hit print in any KDE application (as opposed to the PDF printer)? It is set as default in CUPS, but that seems to have no impact on KDE. | [
-0.013499927707016468,
-0.022364497184753418,
-0.010877314954996109,
0.023494409397244453,
-0.02168266847729683,
-0.012342787347733974,
0.01648387685418129,
-0.018054431304335594,
-0.02862444892525673,
-0.044082436710596085,
-0.011893018148839474,
0.011360084637999535,
-0.008322201669216156,... | [
0.19473211467266083,
0.05107991769909859,
0.8318254351615906,
-0.06006214767694473,
-0.2153085172176361,
-0.3199671506881714,
-0.19246071577072144,
-0.27142661809921265,
-0.2171160876750946,
-0.5826923251152039,
-0.0360710509121418,
0.5912880301475525,
-0.3065917491912842,
0.00485831033438... |
I'm creating a script that automates build a wordpress site from another CMS. I've been able to change every setting: themes, options, subdomain, site title, etc. But what is eluding me is being able to create custom menus. The code below SHOULD be able to do it for me. However it is not, and I am completely stumped on what to do. This code is not being executed in the admin panel (Its not a plugin). Its actually sitting on top of wordpress and includes wp-load and wp-admin functions pages. $mymenu = wp_get_nav_menu_object("Main Navigation Menu"); $menuID = (int) $mymenu->term_id; $itemData = array( 'menu-item-db-id' => 0, 'menu-item-object-id' => $pageId, 'menu-item-object' => 'page', 'menu-item-type' => 'post_type', 'menu-item-parent-id' => 0, 'menu-item-position' => $itemOrder, 'menu-item-title' => $pageData['title'], 'menu-item-url' => get_permalink($pageId), 'menu-item-description' => $pageData['post_content'], 'menu-item-attr-title' => $pageData['post_excerpt'], 'menu-item-status' => 'publish',<br /> 'menu-item-target' => ''<br /> ); $thisMenuItem = wp_update_nav_menu_item($menuID, 0, $itemData); $itemData is the result of data coming in from another CMS plus the result of saving it off a post in wordpress to produce the menu item that I want in my menu. Any help on getting the wp_update_nav_menu_item to save would be very helpful. Again this is creating a custom menu without the use of the admin panel.` | [
0.0028175257612019777,
0.004630780778825283,
0.0036045419983565807,
0.01083187386393547,
-0.008608121424913406,
-0.006529339123517275,
0.0054334551095962524,
0.004533383995294571,
-0.014897228218615055,
0.0035512056201696396,
-0.02013453096151352,
0.005342348478734493,
0.004115116782486439,
... | [
0.40906116366386414,
0.11242657899856567,
0.36410218477249146,
-0.043565090745687485,
-0.06759592890739441,
-0.12989243865013123,
0.100918710231781,
-0.036574337631464005,
-0.1806425154209137,
-0.9643298387527466,
0.02114824205636978,
0.5082566142082214,
-0.3630487024784088,
0.239289641380... |
I'm a developer and I often write jsfiddles for example. I use google analytics on my personal website to monitor traffic. jsfiddle.net is not my own domain, but in a way, snippets I'm publishing on belongs to me. How can I make traffic my own jsfiddles' generate counts for my own domain's one? In other words, when someone visits a jsfiddle I have written, I'd like it counts for one visits on google analytics. | [
-0.005389341618865728,
0.013122998178005219,
-0.0031863979529589415,
0.011016782373189926,
-0.01998240500688553,
-0.005934846121817827,
0.009323530830442905,
-0.009937644936144352,
-0.021257411688566208,
-0.011931383050978184,
-0.0016836420400068164,
0.005969319958239794,
0.00151714507956057... | [
0.9567031264305115,
0.4022741913795471,
0.30407294631004333,
0.035525333136320114,
-0.4383152425289154,
-0.2265251725912094,
-0.0007717473199591041,
0.07774969935417175,
-0.4317249655723572,
-0.3430439531803131,
0.5070741772651672,
0.18810196220874786,
-0.15375880897045135,
0.2630833983421... |
I've just read an excerpt of "Growing Object-Oriented Software" book which explains some reasons why mocking concrete class is not recommended. Here some sample code of a unit-test for the MusicCentre class: public class MusicCentreTest { @Test public void startsCdPlayerAtTimeRequested() { final MutableTime scheduledTime = new MutableTime(); CdPlayer player = new CdPlayer() { @Override public void scheduleToStartAt(Time startTime) { scheduledTime.set(startTime); } } MusicCentre centre = new MusicCentre(player); centre.startMediaAt(LATER); assertEquals(LATER, scheduledTime.get()); } } And his first explanation: > The problem with this approach is that it leaves the relationship between > the objects implicit. I hope we've made clear by now that the intention of > Test-Driven Development with Mock Objects is to discover relationships > between objects. If I subclass, there's nothing in the domain code to make > such a relationship visible, just methods on an object. This makes it harder > to see if the service that supports this relationship might be relevant > elsewhere and I'll have to do the analysis again next time I work with the > class. I can't figure out exactly what he means when he says: > This makes it harder to see if the service that supports this relationship > might be relevant elsewhere and I'll have to do the analysis again next time > I work with the class. I understand that the service corresponds to `MusicCentre`'s method called `startMediaAt`. What does he mean by "elsewhere"? The complete excerpt is here: http://www.mockobjects.com/2007/04/test-smell- mocking-concrete-classes.html | [
0.0037900612223893404,
0.00017348211258649826,
-0.015899937599897385,
0.00003163609653711319,
0.0010551889427006245,
-0.009583281353116035,
0.008317610248923302,
0.005821648519486189,
-0.010527403093874454,
0.003594636218622327,
-0.013852663338184357,
0.013513010926544666,
0.0185338500887155... | [
0.3092896640300751,
-0.4481627643108368,
0.8535386323928833,
-0.010077857412397861,
0.4478432834148407,
-0.34747225046157837,
0.04106532782316208,
-0.14539261162281036,
-0.2058497667312622,
-0.5740148425102234,
-0.13606351613998413,
0.6870094537734985,
-0.0048212031833827496,
0.37571051716... |
I'm performing some basic spatial analyses on a large database of traffic injuries in California between 2003 - 2011. Because the dataset is large (nearly 1 GB of points), I'd like to first cut it down to a specific geographic region using a spatial query, but I find that QGIS consistently freezes and hangs if I try to use a spatial query or filter the layer. What are some ways I can work with this dataset? Are there more efficient data formats I can use? | [
-0.013660459779202938,
-0.004126830957829952,
-0.010784496553242207,
0.01372021809220314,
0.01871911995112896,
0.004738462623208761,
0.00881018303334713,
0.020765576511621475,
-0.014827372506260872,
-0.009461699053645134,
-0.004850688390433788,
0.013164099305868149,
-0.017260810360312462,
... | [
0.22872433066368103,
0.06895525753498077,
0.3082335293292999,
0.18033668398857117,
-0.016691671684384346,
0.23757877945899963,
0.25774335861206055,
0.18291398882865906,
-0.41032761335372925,
-0.4914843738079071,
0.5568330883979797,
0.04878084361553192,
0.0957588404417038,
0.160005614161491... |
I am not able to receive 3G network (the area has excellent 3g coverage) on my HTC desire 500. I have activated 3g on my carrier. It receives only EDGE 'E' Symbol but when I select WCDMA only the symbol changes to G for few seconds and entire network is lost, I tried with two different carriers. Should I need to add more details like my carrier name , APN settings for this question ? | [
-0.011227265931665897,
0.0025415050331503153,
-0.026335373520851135,
0.008783751167356968,
-0.035911675542593,
-0.01367229875177145,
0.009935213252902031,
0.000054693809943273664,
-0.01669878140091896,
-0.047409940510988235,
-0.02651119977235794,
0.009942136704921722,
-0.026944603770971298,
... | [
0.09878792613744736,
-0.14239786565303802,
0.8211945295333862,
0.00029074092162773013,
-0.3344578146934509,
-0.11537378281354904,
0.656753420829773,
-0.1789020150899887,
-0.3187883794307709,
-0.7493389844894409,
-0.06769105792045593,
0.4190915524959564,
-0.7540723085403442,
0.0519145838916... |
For a certain journal I have to follow this structure: 1. Main text 2. References 3. Figures, one per page 4. Extra text I've managed to accomplish that by using package `endfloat` that moves the figures to the end, and `\AtEndDocument`, which allows me to put the extra text after the figures. I have now to submit a revised version, with changes marked. I intended to do this with latexdiff, but it seems it does not highlight changes in the extra text. Unfortunately this is where many changes will be located. Perhaps relevant to this is that I'm using the --flatten option of latexdiff, since the sources are split in several .tex files (some of them being included within an \AtEndDocument). I'd like a solution that does not involve separate documents so page numbering is kept, and there is only one master file. Alas, I'm stuck here. Do you know any way around this? Probably getting rid of `\AtEndDocument`. Thanks in advance. | [
-0.00038098543882369995,
0.006771867163479328,
-0.018855474889278412,
0.016785219311714172,
0.014653016813099384,
0.002242435235530138,
0.008721446618437767,
0.0045343320816755295,
-0.014962177723646164,
-0.018145093694329262,
-0.008475033566355705,
-0.007960394024848938,
-0.0239083077758550... | [
-0.036905933171510696,
0.13628043234348297,
0.7622745633125305,
0.12207875400781631,
-0.09387935698032379,
-0.07886834442615509,
-0.06805568933486938,
-0.011543779633939266,
-0.27666836977005005,
-0.5393921732902527,
-0.04706404730677605,
0.07605445384979248,
-0.15818387269973755,
-0.18676... |
I need to display the current term in my custom taxonomy in a single post. **example:** * My custom taxonomy is _products_ and term of them is _product-1_ , _product-2_ and _products-3_. * My post is assigned to _product-2_ * And I want to print the current products = _products-2_ in my post In fact, I need a function like WordPress's `the_category();` but for my taxonomy like `the_customtaxonomy();` UPDATE : in facts i know i need to get id of this becuse i need to show a icon for this in my single , for example a function like `the_category_ID();` | [
0.019015006721019745,
0.002699462929740548,
-0.007083454169332981,
0.019905105233192444,
-0.0038740658201277256,
0.0018562318291515112,
0.0065737562254071236,
0.009350545704364777,
-0.016522526741027832,
-0.012010312639176846,
-0.02127424255013466,
0.004744020756334066,
-0.004067340865731239... | [
0.33461320400238037,
0.2657509446144104,
0.7379108667373657,
-0.06290845572948456,
-0.0627632811665535,
0.3465702533721924,
-0.11691702902317047,
-0.002746157580986619,
-0.4818200469017029,
-0.5989463925361633,
-0.13549160957336426,
0.47550374269485474,
-0.025646161288022995,
0.42703193426... |
I have heard lots of anecdotal evidence that using VSS is a horrible idea because sooner or later, it will just lose/corrupt/delete all of your work. (I am even guilty of using such evidence as one of the motivating factors to switch to SVN a while back). However, though I really prefer Update/Commit to Checkout/Checkin, it is horrible over SVN and there isn't good branching support, I have never actually had a catastrophic VSS experience. Has VSS ever ruined your project/job/sanity? Old version or new? Is there truth behind the stories? | [
0.019684957340359688,
0.017985008656978607,
-0.00801045075058937,
0.014654144644737244,
0.015069622546434402,
0.002146895742043853,
0.006256863474845886,
-0.017260635271668434,
-0.011869383975863457,
0.0020047707948833704,
-0.024810822680592537,
0.014070183970034122,
-0.009755279868841171,
... | [
0.3756263554096222,
0.18887346982955933,
-0.016958696767687798,
0.04788772761821747,
-0.1450151652097702,
-0.4370032846927643,
0.5494245290756226,
0.6703891754150391,
-0.30805882811546326,
-0.5732579827308655,
0.12148496508598328,
0.8222423791885376,
-0.15421772003173828,
0.559062540531158... |
Is it possible to get paged outside of the standard WP loop? I already use this inside the loop: <?php if ( $paged >= 2 ) { ?> Some text for the 2nd page on up <?php } ?> But I'd like to be able to echo some text outside the loop on all pages two and greater. Possible? Or a better way? | [
-0.0014152589719742537,
0.02443598210811615,
-0.018751202151179314,
0.015033265575766563,
-0.0011679650051519275,
-0.012950633652508259,
0.008707478642463684,
0.007517685182392597,
-0.03483906760811806,
-0.020978422835469246,
-0.0174675565212965,
0.0023942124098539352,
-0.003956382628530264,... | [
0.4525851011276245,
-0.11586286872625351,
0.4271625280380249,
-0.02419964410364628,
-0.14179173111915588,
-0.13432838022708893,
0.5254009962081909,
-0.25436216592788696,
-0.36236658692359924,
-0.25881049036979675,
0.29350709915161133,
0.45498353242874146,
-0.2556421756744385,
-0.1627367436... |
I am trying to get the active state of an OpenLayers SelectFeature control. I create this control and add it to my map but I do NOT activate it yet. This part works fine. Here is the script to create the control and add it to the map: var highlightControl = new OpenLayers.Control.SelectFeature(lyrstands, { hover: true, highlightOnly: true, selectStyle: { fill: "#0000FF", strokeColor: "#0000FF", strokeWidth: 2, fillOpacity: 0.5 } }); map.addControl(highlightControl); Then I have a button that calls a function. Here is the code for the button: new OpenLayers.Control.Button({title: "Select Polygon", trigger: HiLite}) I know the button is calling the HiLite function because if I put a simple alert statement in the HiLite function (for example: alert("The Function Works");) it works correctly. So what I am trying to do is check to see if highlightControl has been activated by testing the highlightControl.active value, if it is already activated the function should do nothing, but, if highlightControl is NOT activated it should activate it. Unfortunately, the function does not work. Here is the function I have written: function HiLite(){ if (highlightControl.active==false) {highlightControl.activate();} else {} } If I write the function to simply activate the tool without the If Else stuff it works fine. But, for some reason when written as above I cannot get the status of the Active attribute. The OpenLayers documentations states that the Active attribute is a boolean value and I assumed I could check the status using the highlightControl.active==false statement. I would appreciate any insight on why I cannot get this function to work. Thank you!! | [
0.0016756297554820776,
0.0049856361001729965,
-0.012902932241559029,
0.005618336144834757,
0.009826874360442162,
-0.003297262592241168,
0.0076538268476724625,
0.003837193362414837,
-0.01727069541811943,
0.02553849294781685,
-0.012935912236571312,
0.023802276700735092,
-0.019845642149448395,
... | [
0.2109537273645401,
-0.3949488401412964,
0.8536718487739563,
-0.2254292368888855,
0.03258732706308365,
-0.023348938673734665,
0.24438780546188354,
-0.1513628363609314,
0.006315483245998621,
-0.9109418392181396,
-0.013710313476622105,
0.4907490611076355,
-0.28207919001579285,
-0.07807169109... |
**_Every_** **single** _time_ Origin starts I get a UAC prompt and Origin drops into the settings menu. Sometimes even does it when unlocking PC. Has anyone else seen this behaviour or knows how to stop it ? Windows 8.1 | [
-0.001912528881803155,
0.010163258761167526,
0.002939546015113592,
0.037001121789216995,
-0.039118483662605286,
-0.012634080834686756,
0.007977346889674664,
-0.00046739779645577073,
-0.02742590568959713,
0.042234692722558975,
-0.017055220901966095,
0.019680559635162354,
0.030169343575835228,... | [
0.2213687002658844,
-0.12430212646722794,
0.39373525977134705,
0.0024330278392881155,
0.2517417371273041,
-0.17357154190540314,
0.44202399253845215,
0.16089439392089844,
-0.31953030824661255,
-0.6534819006919861,
-0.2845150828361511,
0.4241335690021515,
-0.22382530570030212,
0.139559417963... |
The ArcGIS SpatialJoin online help page has a long list of match_options when joining data spatially, including HAVE_THEIR_CENTER_IN. However running the arcpy.SpatialJoin_analysis tool in Python with this option results in an error: ExecuteError: Failed to execute. Parameters are not valid. ERROR 000800: The value is not a member of INTERSECT | CONTAINS | WITHIN | CLOSEST. Failed to execute (SpatialJoin). These options are also missing from ArcToolBox:  The licensing on the help page suggests this tool is available for all versions of ArcGIS: > ArcView: Yes ArcEditor: Yes ArcInfo: Yes So is this an error in the docs, licensing on keywords, or a misunderstanding on my part? | [
0.00021921121515333652,
0.008530720137059689,
-0.021863695234060287,
0.019388660788536072,
-0.008129065856337547,
0.02310188114643097,
0.008794886991381645,
0.02226169966161251,
-0.013856192119419575,
0.005943979136645794,
-0.006366215646266937,
0.017847709357738495,
-0.025156185030937195,
... | [
-0.24286870658397675,
-0.2847892642021179,
0.5525003671646118,
0.20427566766738892,
-0.10948962718248367,
0.05706935375928879,
0.27924659848213196,
-0.35693374276161194,
0.18338096141815186,
-0.7825903296470642,
0.033005356788635254,
0.42152896523475647,
-0.4641045033931732,
-0.20277330279... |
The phrase _someone used to_ is used to talk about habits in past. > A year ago, I used to (not) do stuff. Now I was wondering about a particular kind of question. They are kind of rhetorical and they contain a negative: > Wasn't he British?! Here's an example: I read a book and 10 pages in I realize how much writing sucks. I actually like the author. I tell a friend and he or she says... > Really. Didn't he use to write really well?! * Is that correct (I suppose not)? * Would it be understood? * What could I say instead? Also... * would the following be understood as an attempted play with the language: > Usedn't he to write very well?! | [
0.0035735045094043016,
-0.0008476194925606251,
-0.005185917019844055,
0.018675671890378,
0.006701026111841202,
-0.007824121043086052,
0.004404875915497541,
-0.012947753071784973,
-0.011371473781764507,
0.006908250041306019,
-0.004536126274615526,
0.00004791503306478262,
0.029016409069299698,... | [
0.48799845576286316,
0.4080451726913452,
0.0564718097448349,
-0.18040837347507477,
0.03310436010360718,
-0.15301993489265442,
0.44565141201019287,
0.23216938972473145,
-0.29877153038978577,
-0.2998543977737427,
0.04503897950053215,
0.2075985074043274,
-0.018757525831460953,
0.1448115557432... |
We are putting together latex classes for books. We want to make a certain tree style the default for `forest`. With the example below one has to define the style (sn edges) in every tree. Is there a way to get rid of this specification and make it the default? \documentclass{minimal} \usepackage{forest} \forestset{ sn edges/.style={for tree={parent anchor=south, child anchor=north,align=center,base=bottom,where n children=0{tier=word}{}}}, background tree/.style={for tree={text opacity=0.2,draw opacity=0.2,edge={draw opacity=0.2}}} } \begin{document} \begin{forest} sn edges [S [NP [er\\he] ] [NP [Det [das\\the] ] [N [Buch\\book] ] ] [NP [Det [der\\the] ] [N [Frau\\woman] ] ] [V [gibt\\gives] ] ] \end{forest} \end{document} | [
0.006069048773497343,
0.009704675525426865,
-0.011708691716194153,
0.025938574224710464,
-0.005612592212855816,
-0.012385940179228783,
0.009353723376989365,
-0.01120462641119957,
-0.012935151346027851,
0.0055502308532595634,
-0.007995028048753738,
0.0030747298151254654,
-0.00922456756234169,... | [
0.2749473452568054,
-0.17093594372272491,
0.3188614845275879,
0.2447555661201477,
0.2996569871902466,
-0.0800088495016098,
0.3308388292789459,
-0.08492143452167511,
-0.34888193011283875,
-0.5613263249397278,
-0.13292130827903748,
0.15151679515838623,
0.0035207951441407204,
0.04044506698846... |
I am very new to LaTeX and Sweave, and I am just playing around with getting some output. That said, the "output" I am getting is very large in size (physcially on the page). For example, one output from a 2x3 crosstab takes up a whole 8.5x11 page. If I try to use the same output in a beamer presentation, maybe 30% of the output fits on 1 slide. Simply, when pushing output from R using Sweave, how can I control the size of the output? Thanks in advance! EDIT: Below is a code snipped within my .Rnw file (before Sweave processing) <<echo=FALSE>>= cat("Education by Induced") cat("\n") CrossTable(infert$education, infert$induced, expected = TRUE, format="SPSS") @ | [
-0.013397762551903725,
0.007093725260347128,
-0.013042774982750416,
0.026691250503063202,
-0.013267973437905312,
0.021702418103814125,
0.009472507983446121,
0.0055465069599449635,
-0.019933141767978668,
-0.03990557789802551,
-0.004026427399367094,
0.011140096932649612,
-0.0034753826912492514... | [
0.4866872727870941,
0.08941680938005447,
0.6260751485824585,
-0.2777016758918762,
-0.10665300488471985,
0.10570181161165237,
0.11381101608276367,
-0.3868420124053955,
-0.23851759731769562,
-0.7548825740814209,
0.13498228788375854,
0.37779608368873596,
-0.20244793593883514,
0.17346426844596... |
I have a custom meta box set up that allows users to paste the Youtube URL of a video so that it can be embedded into a post/page. The meta box can be repeated so that a user can add as few or as many URL's as they wish so I'm using this code snippet to display each as a list item; <?php $video = get_post_meta($post->ID, 'youtube-url'); foreach ($video as $vid) { echo '<li>'.$vid.'</li>'; } ?> Is there a way that I can run `the_content` filter on each individual list item so that I can make use of oEmbed that is shipped with Wordpress? Or perhaps there's a more efficient way... | [
-0.003505836008116603,
0.0014608963392674923,
0.0035397144965827465,
0.004823856987059116,
0.00818774476647377,
-0.00036730896681547165,
0.007759688422083855,
0.0076579139567911625,
-0.01591971516609192,
-0.006853570230305195,
-0.011672226712107658,
0.0057585048489272594,
0.00093909993302077... | [
1.0129424333572388,
-0.06004376709461212,
0.8278117775917053,
-0.21940737962722778,
-0.295988529920578,
0.3052041530609131,
0.39226219058036804,
-0.5046527981758118,
-0.28415706753730774,
-0.40388309955596924,
0.2666383981704712,
0.38911405205726624,
-0.23495516180992126,
0.433028757572174... |
**edit** : I am thinking that dependency injection is the best approach. I am struggling to wrap my head around being able to share an object between two classes. I want to be able to create only one instance of the object, `commonlib` in my `main` class and then have the classes, `foo1` and `foo2`, to be able to mutually share the properties of the commonlib. `commonlib` is a 3rd party class which has a property `Queries` that will be **added** to in each child class of `bar`. This is why it is vital that only one instance is created. I create two separate queries in `foo1` and `foo2`. This is my setup: abstract class bar{ //common methods } class foo1 extends bar{ //add query to commonlib } class foo2 extends bar{ //add query to commonlib } class main { public $commonlib = new commonlib(); public function start(){ //goal is to share one instance of $this->commonlib between foo1 and foo2 //so that they can both add to the properites of $this->commonlib (global //between the two) //now execute all of the queries after foo1 and foo2 add their query $this->commonlib->RunQueries(); } } | [
0.008591835387051105,
0.02076912671327591,
-0.01753191091120243,
0.001291054766625166,
-0.01185874454677105,
-0.019829655066132545,
0.0066400994546711445,
0.014925582334399223,
-0.012411542236804962,
-0.017508506774902344,
-0.003634424414485693,
0.01691512018442154,
0.007942529395222664,
0... | [
0.3432762324810028,
0.5008832216262817,
0.07626093178987503,
-0.17282681167125702,
-0.18128183484077454,
0.05237438529729843,
0.19512660801410675,
-0.008524644188582897,
0.26936104893684387,
-0.7189028263092041,
0.22556966543197632,
0.5149650573730469,
-0.48905470967292786,
0.2024432122707... |
For some reason, when I powered my computer up and opened the Steam browser, it was in offline mode. I didn't see it, so I opened Garry's Mod, and when I connected to a server, after about 10 seconds it said that it couldn't connect to VAC servers. Can anyone help me resolve this issue? | [
-0.031153617426753044,
0.007057863753288984,
0.0030391141772270203,
0.011005929671227932,
0.006123793311417103,
-0.03916022181510925,
0.008686939254403114,
-0.023475077003240585,
-0.027566945180296898,
0.026188945397734642,
-0.007925659418106079,
0.026656314730644226,
-0.009535329416394234,
... | [
0.29111939668655396,
0.23675452172756195,
0.31714746356010437,
0.024164065718650818,
-0.1392333060503006,
-0.115831159055233,
0.6347495317459106,
0.5810303092002869,
-0.3975538909435272,
-0.38197392225265503,
0.3474242389202118,
0.4301985502243042,
-0.34000518918037415,
0.356487900018692,
... |
In my document I have numbered and unnumbered sections. Unnumbered sections look like this: \section*{Introduction} I use fancyhdr and add section and subsection marks like this. \pagestyle{fancy} \fancyhf{} % clear all header and footers \renewcommand{\sectionmark}[1]{\markboth{\thesection\ #1}{}} \renewcommand{\subsectionmark}[1]{\markright{\thesubsection\ #1}} \fancyhead[L]{\rule[-0.25in]{0pt}{0.25in}\parbox{0.9\textwidth}{% \hdrfont\textbf{\leftmark\\\rightmark}}} \fancyhead[R]{\hdrfont\thepage\ /~\pageref{LastPage}} \fancyfoot[L]{\rule[-0.25in]{0pt}{0.25in}\parbox{0.9\textwidth}{% \hdrfont\textbf{\leftmark\\\rightmark}}} \fancyfoot[R]{\hdrfont\thepage\ /~\pageref{LastPage}} \renewcommand{\headrulewidth}{0.4pt} \renewcommand{\footrulewidth}{0.4pt} As a result, the unnumbered sections are ignored in headers. I tried to use this command in unnumbered sections \sectionmark{Introduction} Now, I have 0 index for these sections in headers. How to fix this? Any solutions to create unnumbered section which would be treated as numbered ones (without sectionmark, addcontentsline, ...)? MWE \documentclass[12pt]{article} \usepackage{lipsum} \usepackage{fancyhdr} % ----------------------------------------------------------------------------- \pagestyle{fancy} \fancyhf{} % clear all header and footers \renewcommand{\sectionmark}[1]{\markboth{\thesection\ #1}{}} \renewcommand{\subsectionmark}[1]{\markright{\thesubsection\ #1}} \fancyhead[L]{\rule[-0.25in]{0pt}{0.25in}\parbox{0.9\textwidth}{% \textbf{\leftmark\\\rightmark}}} \fancyhead[R]{\thepage} \fancyfoot[L]{\rule[-0.25in]{0pt}{0.25in}\parbox{0.9\textwidth}{% \textbf{\leftmark\\\rightmark}}} \fancyfoot[R]{ \thepage} \renewcommand{\headrulewidth}{0.4pt} \renewcommand{\footrulewidth}{0.4pt} \begin{document} \tableofcontents \section{Section1} \lipsum[4-8] \newpage \section*{Section2} \lipsum[4-8] \newpage \section{Section3} \lipsum[4-8] \newpage \section*{Section4} \lipsum[4-8] \end{document} | [
0.017955217510461807,
0.015140618197619915,
0.00003987271338701248,
0.03292137384414673,
-0.004800355061888695,
-0.007639647461473942,
0.009828664362430573,
-0.0015667763072997332,
-0.017230095341801643,
0.011212989687919617,
-0.013043461367487907,
-0.00513830641284585,
0.006768850609660149,... | [
0.06207487732172012,
0.3012891113758087,
0.697502613067627,
0.018798530101776123,
-0.05559084936976433,
-0.10563758760690689,
0.06770290434360504,
0.02268807962536812,
-0.2979041039943695,
-0.5227127075195312,
-0.27507108449935913,
-0.3883908987045288,
-0.13653382658958435,
-0.123610742390... |
Hi all it seems like I am having difficulties in normalizing raw map files' resolutions to openlayer's use most probably in parameters mapext and imgext. I hope you can help me with this. Using raw (.shp, .shx, .xml, .prj, .dbf) and associated raster files for landuse-related data included in an ArcGIS 10 project (.mxd) that I've obtained from a credible private source, I have used Mxd2Qgs to read .mxd project to .qgs. Then I used quantum GIS' Mapserver Export plugin to export selected layers to MAP FILE. I can view the generated map on browser but cannot do so when I use it in openlayers. The generated map file has an extent: EXTENT 272813.696194 1546212.773835 340292.781152 1587773.606956 (its quite huge). I have tried adding to openlayer's map initialization properties maxResolution, numZoomLevels, units, projection and displayProjection but there is no map rendered, although the map has been queried. Btw I have just guessed the maxResolution here, copying from other huge datasets online :) The map has never been queried when I added the map file's EXTENT to maxExtent: OpenLayers.Bounds(...). The same thing goes when I used the files included in the MS4W tutorial #1 data set. My questions would be: * How to know a given map's maxResolution parameter for openlayers? Can this be guessed? * Are there any special processing needed for map/shapefiles before they can be used in openlayers? * How should a map file be written so that it could be used in openlayers? Thank you so much in advance for your tips. EDIT: Attached MAP FILE and javascript files. | [
-0.003738982370123267,
0.011238785460591316,
-0.007084133103489876,
0.018117843195796013,
-0.01522202230989933,
0.012054692953824997,
0.009670140221714973,
0.007573771756142378,
-0.018244143575429916,
-0.038010336458683014,
0.0048131439834833145,
0.008199652656912804,
-0.00887204334139824,
... | [
0.21148855984210968,
-0.038773998618125916,
0.5680624842643738,
0.10629404336214066,
-0.5501143336296082,
-0.029135767370462418,
0.1484115868806839,
-0.09404464811086655,
-0.03400890901684761,
-0.8631659746170044,
0.011331029236316681,
0.5872066020965576,
-0.010256615467369556,
0.217647030... |
In a `ListPlot3D[ ]` I want to have the `Mesh` at the same positions as the ticks. For some reason this isn't default if both are chosen `Automatic`. So my question is if there is some function to read (or reproduce) the automatically generated `Ticks` as a list e.g. `xTicksList` and `yTicksList` and than use this like Mesh-> {xTicksList,yTicksList} Alternative solutions are of course also welcome. But no solutions with explicite values, this isn't what I want, because finally I want to use this in a more general recipe. Edit: Easy call: SetDirectory["./"] data = Import["file.dat"]; ListPlot3D[data, ViewPoint -> {0, -2, 4}] Data to work on: -1.300000000000 0.210526315789 0.980235619886 -1.163157894737 0.210526315789 0.979505821537 -1.026315789474 0.210526315789 1.097953636927 -0.889473684211 0.210526315789 1.102989687180 -0.752631578947 0.210526315789 1.098700823572 -0.615789473684 0.210526315789 1.125990693356 -0.478947368421 0.210526315789 1.211353461701 -0.342105263158 0.210526315789 1.815613507388 -0.205263157895 0.210526315789 1.045807948880 -0.068421052632 0.210526315789 0.000001971641 0.068421052632 0.210526315789 0.000001971641 0.205263157895 0.210526315789 1.045807948880 0.342105263158 0.210526315789 1.815613507388 0.478947368421 0.210526315789 1.211353461701 0.615789473684 0.210526315789 1.125990693356 0.752631578947 0.210526315789 1.098700823572 0.889473684211 0.210526315789 1.102989687180 1.026315789474 0.210526315789 1.097953636927 1.163157894737 0.210526315789 0.979505821537 1.300000000000 0.210526315789 0.980235619886 -1.300000000000 0.421052631579 0.980942994071 -1.163157894737 0.421052631579 0.981933173223 -1.026315789474 0.421052631579 1.103154056150 -0.889473684211 0.421052631579 1.107281088495 -0.752631578947 0.421052631579 1.106941920492 -0.615789473684 0.421052631579 1.146919259544 -0.478947368421 0.421052631579 1.309678535944 -0.342105263158 0.421052631579 0.584224780318 -0.205263157895 0.421052631579 0.700847812228 -0.068421052632 0.421052631579 0.000006770074 0.068421052632 0.421052631579 0.000006770074 0.205263157895 0.421052631579 0.700847812228 0.342105263158 0.421052631579 0.584224780318 0.478947368421 0.421052631579 1.309678535944 0.615789473684 0.421052631579 1.146919259544 0.752631578947 0.421052631579 1.106941920492 0.889473684211 0.421052631579 1.107281088495 1.026315789474 0.421052631579 1.103154056150 1.163157894737 0.421052631579 0.981933173223 1.300000000000 0.421052631579 0.980942994071 -1.300000000000 0.631578947368 0.982279600261 -1.163157894737 0.631578947368 0.986279051512 -1.026315789474 0.631578947368 1.111702792051 -0.889473684211 0.631578947368 1.114982129224 -0.752631578947 0.631578947368 1.122873224846 -0.615789473684 0.631578947368 1.194737734850 -0.478947368421 0.631578947368 1.930464394003 -0.342105263158 0.631578947368 0.560661235292 -0.205263157895 0.631578947368 0.615345975499 -0.068421052632 0.631578947368 0.964754106066 0.068421052632 0.631578947368 0.964754106066 0.205263157895 0.631578947368 0.615345975499 0.342105263158 0.631578947368 0.560661235292 0.478947368421 0.631578947368 1.930464394003 0.615789473684 0.631578947368 1.194737734850 0.752631578947 0.631578947368 1.122873224846 0.889473684211 0.631578947368 1.114982129224 1.026315789474 0.631578947368 1.111702792051 1.163157894737 0.631578947368 0.986279051512 1.300000000000 0.631578947368 0.982279600261 -1.300000000000 0.842105263158 0.984506863561 -1.163157894737 0.842105263158 0.993017144639 -1.026315789474 0.842105263158 1.123329136135 -0.889473684211 0.842105263158 1.127086080172 -0.752631578947 0.842105263158 1.151323601756 -0.615789473684 0.842105263158 1.316936571158 -0.478947368421 0.842105263158 0.533092973179 -0.342105263158 0.842105263158 0.546076002707 -0.205263157895 0.842105263158 0.576957097781 -0.068421052632 0.842105263158 0.684846595150 0.068421052632 0.842105263158 0.684846595150 0.205263157895 0.842105263158 0.576957097781 0.342105263158 0.842105263158 0.546076002707 0.478947368421 0.842105263158 0.533092973179 0.615789473684 0.842105263158 1.316936571158 0.752631578947 0.842105263158 1.151323601756 0.889473684211 0.842105263158 1.127086080172 1.026315789474 0.842105263158 1.123329136135 1.163157894737 0.842105263158 0.993017144639 1.300000000000 0.842105263158 0.984506863561 -1.300000000000 1.052631578947 0.988041894481 -1.163157894737 1.052631578947 1.002844293634 -1.026315789474 1.052631578947 1.137460222080 -0.889473684211 1.052631578947 1.145431494008 -0.752631578947 1.052631578947 1.204639094902 -0.615789473684 1.052631578947 2.034283267433 -0.478947368421 1.052631578947 0.527426104920 -0.342105263158 1.052631578947 0.536322515501 -0.205263157895 1.052631578947 0.555643903601 -0.068421052632 1.052631578947 1.542776620591 0.068421052632 1.052631578947 1.542776620591 0.205263157895 1.052631578947 0.555643903601 0.342105263158 1.052631578947 0.536322515501 0.478947368421 1.052631578947 0.527426104920 0.615789473684 1.052631578947 2.034283267433 0.752631578947 1.052631578947 1.204639094902 0.889473684211 1.052631578947 1.145431494008 1.026315789474 1.052631578947 1.137460222080 1.163157894737 1.052631578947 1.002844293634 1.300000000000 1.052631578947 0.988041894481 -1.300000000000 1.263157894737 0.993507698008 -1.163157894737 1.263157894737 1.016663871059 -1.026315789474 1.263157894737 1.153029591526 -0.889473684211 1.263157894737 1.173527252601 -0.752631578947 1.263157894737 1.324259264609 -0.615789473684 1.263157894737 0.520310839569 -0.478947368421 1.263157894737 0.523127565540 -0.342105263158 1.263157894737 0.529437718570 -0.205263157895 1.263157894737 0.542365083073 -0.068421052632 1.263157894737 1.251779214963 0.068421052632 1.263157894737 1.251779214963 0.205263157895 1.263157894737 0.542365083073 0.342105263158 1.263157894737 0.529437718570 0.478947368421 1.263157894737 0.523127565540 0.615789473684 1.263157894737 0.520310839569 0.752631578947 1.263157894737 1.324259264609 0.889473684211 1.263157894737 1.173527252601 1.026315789474 1.263157894737 1.153029591526 1.163157894737 1.263157894737 1.016663871059 1.300000000000 1.263157894737 0.993507698008 -1.300000000000 1.473684210526 1.001780955507 -1.163157894737 1.473684210526 1.035419641030 -1.026315789474 1.473684210526 1.168300600319 -0.889473684211 1.473684210526 1.218742366781 -0.752631578947 1.473684210526 1.820497914358 -0.615789473684 1.473684210526 0.517801652652 -0.478947368421 1.473684210526 0.519782526335 -0.342105263158 1.473684210526 0.524378305806 -0.205263157895 1.473684210526 1.497512379500 -0.068421052632 1.473684210526 1.158442502764 0.068421052632 1.473684210526 1.158442502764 0.205263157895 1.473684210526 1.497512379500 0.342105263158 1.473684210526 0.524378305806 0.478947368421 1.473684210526 0.519782526335 0.615789473684 1.473684210526 0.517801652652 0.752631578947 1.473684210526 1.820497914358 0.889473684211 1.473684210526 1.218742366781 1.026315789474 1.473684210526 1.168300600319 1.163157894737 1.473684210526 1.035419641030 1.300000000000 1.473684210526 1.001780955507 -1.300000000000 1.684210526316 1.013974303940 -1.163157894737 1.684210526316 1.059492763356 -1.026315789474 1.684210526316 1.180869891090 -0.889473684211 1.684210526316 1.299514133592 -0.752631578947 1.684210526316 0.516453133533 -0.615789473684 1.684210526316 0.515737848476 -0.478947368421 1.684210526316 0.517124685004 -0.342105263158 1.684210526316 0.520545139810 -0.205263157895 1.684210526316 1.214149849511 -0.068421052632 1.684210526316 1.111669990419 0.068421052632 1.684210526316 1.111669990419 0.205263157895 1.684210526316 1.214149849511 0.342105263158 1.684210526316 0.520545139810 0.478947368421 1.684210526316 0.517124685004 0.615789473684 1.684210526316 0.515737848476 0.752631578947 1.684210526316 0.516453133533 0.889473684211 1.684210526316 1.299514133592 1.026315789474 1.684210526316 1.180869891090 1.163157894737 1.684210526316 1.059492763356 1.300000000000 1.684210526316 1.013974303940 -1.300000000000 1.894736842105 1.031151313623 -1.163157894737 1.894736842105 1.087145840815 -1.026315789474 1.894736842105 1.188079272660 -0.889473684211 1.894736842105 1.480552497483 -0.752631578947 1.894736842105 0.514891129597 -0.615789473684 1.894736842105 0.514017825943 -0.478947368421 1.894736842105 0.514975893386 -0.342105263158 1.894736842105 1.653213278836 -0.205263157895 1.894736842105 1.132316352893 -0.068421052632 1.894736842105 1.083852929370 0.068421052632 1.894736842105 1.083852929370 0.205263157895 1.894736842105 1.132316352893 0.342105263158 1.894736842105 1.653213278836 0.478947368421 1.894736842105 0.514975893386 0.615789473684 1.894736842105 0.514017825943 0.752631578947 1.894736842105 0.514891129597 0.889473684211 1.894736842105 1.480552497483 1.026315789474 1.894736842105 1.188079272660 1.163157894737 1.894736842105 1.087145840815 1.300000000000 1.894736842105 1.031151313623 -1.300000000000 2.105263157895 1.053263797252 -1.163157894737 2.105263157895 1.112054099666 -1.026315789474 2.105263157895 1.187887756512 -0.889473684211 2.105263157895 2.414939312234 -0.752631578947 2.105263157895 0.513547775105 -0.615789473684 2.105263157895 0.512568414028 -0.478947368421 2.105263157895 0.513214034428 -0.342105263158 2.105263157895 1.234630789787 -0.205263157895 2.105263157895 1.093026423469 -0.068421052632 2.105263157895 1.065649778360 0.068421052632 2.105263157895 1.065649778360 0.205263157895 2.105263157895 1.093026423469 0.342105263158 2.105263157895 1.234630789787 0.478947368421 2.105263157895 0.513214034428 0.615789473684 2.105263157895 0.512568414028 0.752631578947 2.105263157895 0.513547775105 0.889473684211 2.105263157895 2.414939312234 1.026315789474 2.105263157895 1.187887756512 1.163157894737 2.105263157895 1.112054099666 1.300000000000 2.105263157895 1.053263797252 -1.300000000000 2.315789473684 1.076657703675 -1.163157894737 2.315789473684 1.123492475408 -1.026315789474 2.315789473684 1.179776702031 -0.889473684211 2.315789473684 0.517636004512 -0.752631578947 2.315789473684 0.512380796022 -0.615789473684 2.315789473684 0.511335258693 -0.478947368421 2.315789473684 0.519172187319 -0.342105263158 2.315789473684 1.138037117826 -0.205263157895 2.315789473684 1.070113678169 -0.068421052632 2.315789473684 1.052968436145 0.068421052632 2.315789473684 1.052968436145 0.205263157895 2.315789473684 1.070113678169 0.342105263158 2.315789473684 1.138037117826 0.478947368421 2.315789473684 0.519172187319 0.615789473684 2.315789473684 0.511335258693 0.752631578947 2.315789473684 0.512380796022 0.889473684211 2.315789473684 0.517636004512 1.026315789474 2.315789473684 1.179776702031 1.163157894737 2.315789473684 1.123492475408 1.300000000000 2.315789473684 1.076657703675 -1.300000000000 2.526315789474 1.091865153557 -1.163157894737 2.526315789474 1.114876898745 -1.026315789474 2.526315789474 1.164983245160 -0.889473684211 2.526315789474 0.516400089510 -0.752631578947 2.526315789474 0.511360095534 -0.615789473684 2.526315789474 0.510277474936 -0.478947368421 2.526315789474 1.335414688817 -0.342105263158 2.526315789474 1.094672456667 -0.205263157895 2.526315789474 1.055241265973 -0.068421052632 2.526315789474 1.043728713057 0.068421052632 2.526315789474 1.043728713057 0.205263157895 2.526315789474 1.055241265973 0.342105263158 2.526315789474 1.094672456667 0.478947368421 2.526315789474 1.335414688817 0.615789473684 2.526315789474 0.510277474936 0.752631578947 2.526315789474 0.511360095534 0.889473684211 2.526315789474 0.516400089510 1.026315789474 2.526315789474 1.164983245160 1.163157894737 2.526315789474 1.114876898745 1.300000000000 2.526315789474 1.091865153557 -1.300000000000 2.736842105263 1.089274605713 -1.163157894737 2.736842105263 1.091729061725 -1.026315789474 2.736842105263 1.145851846320 -0.889473684211 2.736842105263 0.515298422848 -0.752631578947 2.736842105263 0.510461989727 -0.615789473684 2.736842105263 0.509366151457 -0.478947368421 2.736842105263 1.174696797334 -0.342105263158 2.736842105263 1.070314416559 -0.205263157895 2.736842105263 1.044899739158 -0.068421052632 2.736842105263 1.036763511509 0.068421052632 2.736842105263 1.036763511509 0.205263157895 2.736842105263 1.044899739158 0.342105263158 2.736842105263 1.070314416559 0.478947368421 2.736842105263 1.174696797334 0.615789473684 2.736842105263 0.509366151457 0.752631578947 2.736842105263 0.510461989727 0.889473684211 2.736842105263 0.515298422848 1.026315789474 2.736842105263 1.145851846320 1.163157894737 2.736842105263 1.091729061725 1.300000000000 2.736842105263 1.089274605713 -1.300000000000 2.947368421053 1.070982617491 -1.163157894737 2.947368421053 1.064870984592 -1.026315789474 2.947368421053 1.124857239675 -0.889473684211 2.947368421053 0.514306616086 -0.752631578947 2.947368421053 0.509667542587 -0.615789473684 2.947368421053 2.571106436571 -0.478947368421 2.947368421053 1.113772367051 -0.342105263158 2.947368421053 1.054892518975 -0.205263157895 2.947368421053 1.037352203431 -0.068421052632 2.947368421053 1.031369828620 0.068421052632 2.947368421053 1.031369828620 0.205263157895 2.947368421053 1.037352203431 0.342105263158 2.947368421053 1.054892518975 0.478947368421 2.947368421053 1.113772367051 0.615789473684 2.947368421053 2.571106436571 0.752631578947 2.947368421053 0.509667542587 0.889473684211 2.947368421053 0.514306616086 1.026315789474 2.947368421053 1.124857239675 1.163157894737 2.947368421053 1.064870984592 1.300000000000 2.947368421053 1.070982617491 -1.300000000000 3.157894736842 1.047866508842 -1.163157894737 3.157894736842 1.041164829929 -1.026315789474 3.157894736842 1.103940126372 -0.889473684211 3.157894736842 0.513409767406 -0.752631578947 3.157894736842 0.508961575187 -0.615789473684 3.157894736842 1.326407622207 -0.478947368421 3.157894736842 1.082011967173 -0.342105263158 3.157894736842 1.044357195862 -0.205263157895 3.157894736842 1.031641411387 -0.068421052632 3.157894736842 1.027100715053 0.068421052632 3.157894736842 1.027100715053 0.205263157895 3.157894736842 1.031641411387 0.342105263158 3.157894736842 1.044357195862 0.478947368421 3.157894736842 1.082011967173 0.615789473684 3.157894736842 1.326407622207 0.752631578947 3.157894736842 0.508961575187 0.889473684211 3.157894736842 0.513409767406 1.026315789474 3.157894736842 1.103940126372 1.163157894737 3.157894736842 1.041164829929 1.300000000000 3.157894736842 1.047866508842 -1.300000000000 3.368421052632 1.027707541691 -1.163157894737 3.368421052632 1.022670717601 -1.026315789474 3.368421052632 1.084322269326 -0.889473684211 3.368421052632 0.512595867665 -0.752631578947 3.368421052632 0.508332439171 -0.615789473684 3.368421052632 1.179765630236 -0.478947368421 3.368421052632 1.062774142729 -0.342105263158 3.368421052632 1.036767743753 -0.205263157895 3.368421052632 1.027197425050 -0.068421052632 3.368421052632 1.023659622763 0.068421052632 3.368421052632 1.023659622763 0.205263157895 3.368421052632 1.027197425050 0.342105263158 3.368421052632 1.036767743753 0.478947368421 3.368421052632 1.062774142729 0.615789473684 3.368421052632 1.179765630236 0.752631578947 3.368421052632 0.508332439171 0.889473684211 3.368421052632 0.512595867665 1.026315789474 3.368421052632 1.084322269326 1.163157894737 3.368421052632 1.022670717601 1.300000000000 3.368421052632 1.027707541691 -1.300000000000 3.578947368421 1.012740556267 -1.163157894737 3.578947368421 1.009098441749 -1.026315789474 3.578947368421 1.066618101119 -0.889473684211 3.578947368421 0.511854868108 -0.752631578947 3.578947368421 0.507778141612 -0.615789473684 3.578947368421 1.119870247076 -0.478947368421 3.578947368421 1.050022506614 -0.342105263158 3.578947368421 1.031080838687 -0.205263157895 3.578947368421 1.023660489456 -0.068421052632 3.578947368421 1.020842841204 0.068421052632 3.578947368421 1.020842841204 0.205263157895 3.578947368421 1.023660489456 0.342105263158 3.578947368421 1.031080838687 0.478947368421 3.578947368421 1.050022506614 0.615789473684 3.578947368421 1.119870247076 0.752631578947 3.578947368421 0.507778141612 0.889473684211 3.578947368421 0.511854868108 1.026315789474 3.578947368421 1.066618101119 1.163157894737 3.578947368421 1.009098441749 1.300000000000 3.578947368421 1.012740556267 -1.300000000000 3.789473684211 1.002490308916 -1.163157894737 3.789473684211 0.999480129131 -1.026315789474 3.789473684211 1.051031597371 -0.889473684211 3.789473684211 0.511178259186 -0.752631578947 3.789473684211 2.115303346236 -0.615789473684 3.789473684211 1.087590009962 -0.478947368421 3.789473684211 1.041038314858 -0.342105263158 3.789473684211 1.026687653195 -0.205263157895 3.789473684211 1.020792862745 -0.068421052632 3.789473684211 1.018506330614 0.068421052632 3.789473684211 1.018506330614 0.205263157895 3.789473684211 1.020792862745 0.342105263158 3.789473684211 1.026687653195 0.478947368421 3.789473684211 1.041038314858 0.615789473684 3.789473684211 1.087590009962 0.752631578947 3.789473684211 2.115303346236 0.889473684211 3.789473684211 0.511178259186 1.026315789474 3.789473684211 1.051031597371 1.163157894737 3.789473684211 0.999480129131 1.300000000000 3.789473684211 1.002490308916 -1.300000000000 4.000000000000 0.995814193962 -1.163157894737 4.000000000000 0.992838080577 -1.026315789474 4.000000000000 1.037528228234 -0.889473684211 4.000000000000 0.510558796548 -0.752631578947 4.000000000000 1.344611311368 -0.615789473684 4.000000000000 1.067648040164 -0.478947368421 4.000000000000 1.034420877924 -0.342105263158 4.000000000000 1.023210238365 -0.205263157895 4.000000000000 1.018431506605 -0.068421052632 4.000000000000 1.016545718957 0.068421052632 4.000000000000 1.016545718957 0.205263157895 4.000000000000 1.018431506605 0.342105263158 4.000000000000 1.023210238365 0.478947368421 4.000000000000 1.034420877924 0.615789473684 4.000000000000 1.067648040164 0.752631578947 4.000000000000 1.344611311368 0.889473684211 4.000000000000 0.510558796548 1.026315789474 4.000000000000 1.037528228234 1.163157894737 4.000000000000 0.992838080577 1.300000000000 4.000000000000 0.995814193962 | [
0.010538235306739807,
0.001191990915685892,
-0.007852178998291492,
0.007243831641972065,
-0.009279775433242321,
0.0007145330309867859,
0.0063378289341926575,
0.01062493585050106,
-0.0132383331656456,
0.03548606485128403,
-0.00601067254319787,
0.010224265977740288,
0.003378416644409299,
0.0... | [
0.2600470781326294,
-0.1705736517906189,
0.4326382577419281,
0.16978642344474792,
-0.3456396460533142,
-0.046227242797613144,
0.10310737043619156,
-0.3198806643486023,
-0.14544115960597992,
-0.7535092830657959,
0.13397155702114105,
0.47481778264045715,
-0.6931668519973755,
-0.0881448462605... |
I have data points that include one instance of a negative y value. With this data, is it possible to generate a exponential decay function in a way that is mathematically sound? The only variable is X. | [
0.01971234381198883,
0.02198386937379837,
-0.03324479982256889,
0.004314752295613289,
0.022983932867646217,
-0.016946399584412575,
0.011354099959135056,
0.038947638124227524,
-0.03528883680701256,
-0.07831897586584091,
-0.014764276333153248,
0.012795346789062023,
-0.017705772072076797,
-0.... | [
0.18926534056663513,
0.3419831693172455,
0.26346564292907715,
0.1139814630150795,
-0.12196436524391174,
-0.044489700347185135,
-0.03061874397099018,
-0.02956366166472435,
0.29374781250953674,
-0.3098790645599365,
0.13993406295776367,
0.3055950701236725,
-0.2848575711250305,
0.4562275707721... |
This question originates in Scalar field in pgfplots . How can I draw a quiver plot in which the arrows have the same length, but their line width varies according to some scalar property of the vectors? | [
-0.015136896632611752,
0.023967327550053596,
-0.004303384106606245,
0.01858590357005596,
0.006607265677303076,
-0.014976448379456997,
0.013555767014622688,
0.025826293975114822,
-0.021614588797092438,
-0.04408036172389984,
-0.003635226283222437,
0.008348111063241959,
-0.009919476695358753,
... | [
0.06614191830158234,
-0.12704260647296906,
0.4095242917537689,
0.3323027789592743,
-0.5295013189315796,
0.07583676278591156,
0.2032034993171692,
-0.4413878917694092,
-0.43235015869140625,
-0.18164560198783875,
0.22731928527355194,
-0.138859361410141,
-0.16908445954322815,
0.022488784044981... |
I am trying to write a macro that has a different behavior when called from text mode, math mode, display math, and math tables. The first three are easy enough with `\ifmmode` and `\mathchoice`. But the last one is a bit more complicated as I have to test for various LaTeX- and AMS-environments: `array`, `array*`, `gather`, `align` etc. However, as all those environments are implemented as variants of `\halign`, the cleanest way would be to check if the macro is called from within a `\halign`. Hence my question: is there a way to test if a macro is called from `\halign` ? (this is a user-level macro that will very probably never be called from inside `\noalign`, so no need to put an extra test for this case). **Edit** : I found a quasi-answer to this, by which I mean that it works in all the above AMS environments but not in a plain `\halign`. Namely, in plain math, `\math@cr@@@` is a macro expanding to `\cr`, but it is changed in the aligned environments. So I only have to do something like \def\plain@cr{\cr} \ifx\math@cr@@@\plain@cr ... \else ... \fi | [
0.0032397471368312836,
0.017777156084775925,
-0.009733988903462887,
0.010122631676495075,
0.0063553680665791035,
-0.006005142815411091,
0.00777176721021533,
0.0069258310832083225,
-0.011848741210997105,
-0.01071815937757492,
-0.012632416561245918,
-0.007062813732773066,
-0.005964829586446285... | [
0.14477866888046265,
0.1178453266620636,
0.24210207164287567,
-0.4043383002281189,
0.10599981248378754,
0.29312339425086975,
0.10262753069400787,
-0.08020378649234772,
-0.12019866704940796,
-0.7321782112121582,
0.2497626543045044,
0.6934770345687866,
-0.4902436137199402,
-0.132104158401489... |
Say I am developing an app for sale. Does it make sense to use a public repository for this project? Doesn't default copyright protect the code from someone else using it? If so, what advantages are there in paying for a private repository? | [
0.013806805945932865,
0.008276006206870079,
0.000250644312473014,
0.015604397282004356,
0.0029615575913339853,
0.025180049240589142,
0.013888109475374222,
0.01305266935378313,
-0.017640452831983566,
-0.04484556242823601,
-0.0112125463783741,
0.02903180569410324,
0.035743001848459244,
0.029... | [
0.9851153492927551,
0.4258284568786621,
0.18941381573677063,
0.23813696205615997,
0.2688695788383484,
-0.12203286588191986,
-0.06892417371273041,
0.44767284393310547,
-0.12979114055633545,
0.004230675753206015,
0.19725976884365082,
0.2130018174648285,
-0.020865293219685555,
0.3760619759559... |
filedata={" 1 2578Hot Pink 47:21:17 9512971 ", " 2 294Pale Green 24:57:18 4437289 ", " 3 2497Light Yellow 39:40:57 4987642 ", " 4 2010Turquoise 16:04:17 3290855 "} There is an external file. You open a Stream. You Skip to a far postion. It takes a long time for a large file. How can one conceive a more direct method for a special straightforward highly regular file structure like this? We got an index and know the file size anyway. A low level file function approach? * * * ## **Update** Records in file: 100.000.000 Set StreamPosition to: 60.000.000 Time needed: 40 seconds That is: Skip and SetStreamPosition takes the same chunk of time. | [
-0.014457764104008675,
0.007385065779089928,
0.0016788048669695854,
0.011213441379368305,
0.0037728925235569477,
0.006090810988098383,
0.005739455111324787,
0.011812480166554451,
-0.010683263652026653,
-0.001505053136497736,
-0.012064275331795216,
0.006310240365564823,
0.013906712643802166,
... | [
0.2871822416782379,
0.0014766562962904572,
0.16525833308696747,
0.3618048429489136,
0.37644264101982117,
0.41444385051727295,
-0.12695081532001495,
-0.022293783724308014,
-0.19671329855918884,
-0.36542272567749023,
0.026195846498012543,
0.4312809705734253,
-0.14537101984024048,
0.203549236... |
I'd like to test last HBO GO android application to see how they implemented chromecast, but I'm in Canada and my devices are registered to play store Canada so I'm not able to install it. I have been looking for the single .apk on the web but I'm not sure if I could trust those I found. Does any body has an idea how I could get it or how to get rid of the localization limitations ? Thanks | [
-0.0011560374405235052,
0.0019787773489952087,
-0.002757942071184516,
0.014798787422478199,
-0.001218181336298585,
0.020474061369895935,
0.006581611465662718,
0.015851275995373726,
-0.023351730778813362,
0.006161133758723736,
-0.0009839660488069057,
0.013945412822067738,
0.01709173247218132,... | [
1.3518904447555542,
0.2967628240585327,
0.6552096605300903,
0.14619162678718567,
-0.0008034585043787956,
-0.27455639839172363,
0.39482617378234863,
0.311084121465683,
-0.22875866293907166,
-0.5247064828872681,
0.04594523832201958,
0.26182013750076294,
-0.11021167784929276,
-0.3256188333034... |
Just defeated the game with some help of others, now I've unlocked the computer but everytime I try adding chocolate bars, it doesn't work. Adding 10 chocolate bar(s) doesn't work, neither does adding 10 pains au chocolat(s). | [
-0.013702704571187496,
0.001752698328346014,
-0.03642166778445244,
0.00825359858572483,
-0.018785612657666206,
0.005322918761521578,
0.012873463332653046,
-0.008406577631831169,
-0.03304498642683029,
-0.03254244104027748,
-0.04440246522426605,
0.021042009815573692,
-0.05227746069431305,
0.... | [
-0.04608524218201637,
0.24848850071430206,
0.16944962739944458,
0.17729823291301727,
-0.0214215237647295,
0.5138276815414429,
0.5361154079437256,
-0.08221419900655746,
-0.3702842891216278,
-0.7349417805671692,
0.34824052453041077,
0.2947486937046051,
0.14159472286701202,
0.5871546268463135... |
I'm learning Haskell and as an exercise I'm making binary trees. Having made a regular binary tree, I want to adapt it to be self balancing. So: * Which is most efficient? * Which is easiest to implement? * Which is most often used? But crucially, which do you recommend? _I assume this belongs here because it's open to debate._ | [
0.021205877885222435,
0.02774570696055889,
-0.017187658697366714,
0.01284637302160263,
-0.028540238738059998,
-0.007940935902297497,
0.007814022712409496,
0.00907748844474554,
-0.032989077270030975,
-0.06076103448867798,
0.0033590137027204037,
0.01368188951164484,
-0.03932768851518631,
0.0... | [
0.07832667231559753,
0.19662965834140778,
-0.027907153591513634,
-0.09175980091094971,
0.009585948660969734,
0.07830843329429626,
0.23404771089553833,
0.2455509901046753,
-0.20701058208942413,
-0.7947924137115479,
-0.0766686275601387,
0.2929982542991638,
-0.1987304836511612,
-0.19993782043... |
Given two arrays x and y, both of length n, I fit a model y = a + b*x and want to calculate a 95% confidence interval for the slope. This is (b - delta, b + delta) where b is found in the usual way and delta = qt(0.975,df=n-2)*se.slope and se.slope is the standard error in the slope. One way to get the standard error of the slope from R is `summary(lm(y~x))$coef[2,2]`. Now suppose I write the likelihood of the slope given x and y, multiply this by a "flat" prior and use a MCMC technique to draw a sample _m_ from the posterior distribution. Define lims = quantile(m,c(0.025,0.975)) My question: is `(lims[[2]]-lims[[1]])/2` approximately equal to delta as defined above? **Addendum** Below is a simple JAGS model where these two seem to be different. model { for (i in 1:N) { y[i] ~ dnorm(mu[i], tau) mu[i] <- a + b * x[i] } a ~ dnorm(0, .00001) b ~ dnorm(0, .00001) tau <- pow(sigma, -2) sigma ~ dunif(0, 100) } I run the following in R: N <- 10 x <- 1:10 y <- c(30.5,40.6,20.5,59.1,52.5, 96.0,121.4,78.9,112.1,128.4) lin <- lm(y~x) #Calculate delta for a 95% confidence interval on the slope delta.lm <- qt(0.975,df=N-2)*summary(lin)$coef[2,2] library('rjags') jags <- jags.model('example.bug', data = list('x' = x,'y' = y,'N' = N), n.chains = 4,n.adapt = 100) update(jags, 1000) params <- jags.samples(jags,c('a', 'b', 'sigma'),7500) lims <- quantile(params$b,c(0.025,0.975)) delta.bayes <- (lims[[2]]-lims[[1]])/2 cat("Classical confidence region: +/-",round(delta.lm, digits=4),"\n") cat("Bayesian confidence region: +/-",round(delta.bayes,digits=4),"\n") And get: Classical confidence region: +/- 4.6939 Bayesian confidence region: +/- 5.1605 Rerunning this multiple times, the Bayesian confidence region is consistently wider than the classical one. So is this due to the priors I've chosen? | [
0.00013771833619102836,
0.011829808354377747,
-0.019080977886915207,
0.0037308582104742527,
-0.0013822605833411217,
-0.0066720787435770035,
0.006275465711951256,
-0.014448287896811962,
-0.004571537487208843,
-0.03038131259381771,
-0.0034364445600658655,
0.009388037025928497,
-0.0275815054774... | [
0.07545427978038788,
-0.2101532369852066,
0.3267722725868225,
-0.45606717467308044,
-0.1455390453338623,
0.6229670643806458,
-0.012732870876789093,
-0.4909290373325348,
0.07084690779447556,
-0.6542848348617554,
-0.03421008214354515,
0.443075954914093,
-0.0593469999730587,
0.204148724675178... |
Is “We are quit” (meaning “We’re even, no more mutual obligations”) a usage from the 18–19th centuries? Or are the examples of this on Google hits just people making it up (possibly a bad cognate from German “Wir sind quitt”)? | [
0.0009645652025938034,
0.00666545145213604,
-0.013524373061954975,
0.025617536157369614,
0.003687763586640358,
0.02237265557050705,
0.010929390788078308,
-0.004791419953107834,
-0.02159590646624565,
0.051897697150707245,
0.0017299926839768887,
0.020682133734226227,
0.025442957878112793,
-0... | [
0.7258039712905884,
-0.06302080303430557,
0.3419944941997528,
-0.2050875872373581,
-0.12541955709457397,
-0.30979499220848083,
0.7705232501029968,
0.3108697831630707,
-0.10226468741893768,
0.021738938987255096,
-0.003225428517907858,
0.5234655737876892,
-0.33599573373794556,
0.230358913540... |
Is there any way to restrict accounts to login only once at the same time? Ex. Mike bought a subscription from my website and also lent it to his friend John. What I want to achieve is preventing Mike login if John is already logged in and the opposite. Is this possible? | [
0.019472889602184296,
0.022803738713264465,
-0.00892899464815855,
0.02908319979906082,
0.007576834410429001,
0.0063225459307432175,
0.01074962504208088,
0.013145136646926403,
-0.03185627982020378,
-0.0008994402596727014,
-0.016655372455716133,
0.018094459548592567,
-0.005437816493213177,
0... | [
0.6385940909385681,
0.1255597025156021,
0.137477844953537,
0.05401381105184555,
0.3885323703289032,
-0.18642736971378326,
0.49580761790275574,
0.3739423155784607,
-0.35632598400115967,
-0.29841887950897217,
0.04755527153611183,
0.3684116005897522,
-0.43952202796936035,
0.010185367427766323... |
I'm building a website for a restaurant. They have a downloadable PDF version of their menu (which is what the "real-world" menu is printed from), but we also want to have a text-based version of the menu on the website (because it's the 21st century and you shouldn't have to read a PDF when browsing a site like this). I'm hoping to avoid -- or minimize as much as possible -- the potential problems with maintaining data in two places (one in the PDF and one on the website). In an ideal world, we'd build a custom dashboard interface for managing menu data, with a fully-normalized schema, and from that data we'd automatically generate the PDF _and_ the website versions of the menu (and come up with a very expensive and flexible visual design for the PDF menu that takes into account all future permutations of data that may be contained within it -- because the visual layout of the PDF menu is very important, since that's what the "real-world" menu is printed from). But, since the budget is not unlimited, I'm trying to think of other ways to accomplish this goal. Are there any compromise solutions or "this gets you halfway there" kind of things I could do? Obviously I'm not expecting a perfect and cheap solution, but anything that minimizes the double-data entry to the extent that it's not so much of a chore would be really helpful. | [
0.01203249953687191,
0.004812759347259998,
0.004611542914062738,
0.008432460948824883,
0.0029866881668567657,
0.0031624601688236,
0.005834454670548439,
0.020881924778223038,
-0.013857482932507992,
-0.01605597324669361,
-0.026993021368980408,
0.0033323101233690977,
0.01558007299900055,
0.00... | [
0.5725283622741699,
0.21936191618442535,
0.22525742650032043,
0.23788756132125854,
0.4395822584629059,
0.06870143115520477,
-0.32763680815696716,
0.21181368827819824,
-0.2692375183105469,
-0.6717418432235718,
0.2342531383037567,
0.3886333405971527,
0.10242120921611786,
0.024791883304715157... |
I am running this SQL in Oracle Spatial in postgis worked. In the Oracle error `[Err] ORA-00920: invalid relational operator` I need to select all geometries that touch on my streets layer, ie the corners. SELECT DISTINCT (b.cdidelog) AS Cruzamento FROM "GIVVI_EIXO_LOGR" A, ( SELECT * FROM "GIVVI_EIXO_LOGR" ) b WHERE ST_Touches (A.geom, b.geom) AND A.cdlog = 7775026 AND A.nmidelog <> b.nmidelog | [
0.010490672662854195,
0.013514072634279728,
-0.010915491729974747,
0.028301795944571495,
-0.02662486769258976,
0.02934788540005684,
0.00814063660800457,
0.01447623036801815,
-0.01070917397737503,
-0.008247369900345802,
-0.011689115315675735,
0.009335843846201897,
-0.0006820277776569128,
0.... | [
-0.010327160358428955,
0.33221927285194397,
0.8906983733177185,
-0.018591588363051414,
-0.26661109924316406,
0.18991120159626007,
0.1740952879190445,
-0.30002209544181824,
0.04493366554379463,
-0.895045280456543,
0.14416386187076569,
0.08414453268051147,
-0.2701684832572937,
0.187067285180... |
I've got large amount of data (e.g. 100K) and I want to cluster them in very fine-grained clusters (e.g. 10K). I look for an appropriate algorithm that uses the similarity function instead of whole data and also I can't compute whole similarity matrix because of data scale. Python implementation is appreciated. | [
-0.00332269468344748,
0.001074541942216456,
-0.011296667158603668,
0.021344318985939026,
-0.013632994145154953,
-0.013190018944442272,
0.010778935626149178,
-0.001977241365239024,
-0.025241810828447342,
0.02790977992117405,
-0.003887699218466878,
0.014232936315238476,
-0.010827148333191872,
... | [
-0.0035757313016802073,
0.3653242588043213,
0.03319988772273064,
0.17428041994571686,
-0.08866938948631287,
0.17435215413570404,
0.2327118068933487,
-0.1915305256843567,
-0.29354041814804077,
-0.6556549072265625,
-0.16864806413650513,
0.1725018471479416,
0.0006447214400395751,
0.1830204725... |
I would like to stay away from finance and economic examples since they are too abstract for me to understand. Are there any "real world" examples for example healthcare, exam marks, environmental science and so on? From a managerial perspective I would like a basic understanding of how VAR could potentially be used to model real world phenomenon. | [
0.009228288196027279,
0.026857411488890648,
-0.0047726863995194435,
0.021025797352194786,
0.012922835536301136,
-0.003827648237347603,
0.012409454211592674,
0.03150767832994461,
-0.022319519892334938,
0.0008414832409471273,
-0.031273048371076584,
0.0089741675183177,
0.0054934388026595116,
... | [
0.7401399612426758,
0.2940141558647156,
-0.3329179883003235,
0.15177950263023376,
-0.4036944806575775,
0.30160659551620483,
0.16170454025268555,
0.22485926747322083,
-0.26424306631088257,
-0.4549667239189148,
0.28745022416114807,
0.5530911684036255,
-0.03348151966929436,
0.2833205163478851... |
Has learning theory ever been applied in practice for search log data? If so, what are some findings about generalization/learnability from this data? I'm interested in generalization about an individual user's search behavior from a time slice or generalization about a total population of users, given a sample. Paper references are a plus! | [
-0.012025279924273491,
0.01434877049177885,
-0.013437220826745033,
0.02232437953352928,
0.015566615387797356,
0.004603908862918615,
0.011223572306334972,
0.01070175040513277,
-0.02424805238842964,
-0.01792358048260212,
0.0021100020967423916,
0.021816743537783623,
0.009240490384399891,
0.00... | [
0.35398274660110474,
-0.05995722860097885,
-0.038038741797208786,
0.2833831012248993,
-0.0848124548792839,
-0.10785523802042007,
-0.016083411872386932,
0.2143891304731369,
-0.15523870289325714,
-0.7759107947349548,
0.12273721396923065,
0.28437331318855286,
-0.04467403143644333,
0.300308138... |
In my opinion a lot of SEO information online tends to have a very "spammy" feel to it. Can anyone point me to some truly credible and legitimate SEO websites and/or seminars/webinars (other than stackexchange)? I'd like to learn more about SEO but I'm not sure where to look. | [
-0.003898975905030966,
0.00298882438801229,
0.009628853760659695,
0.02754664421081543,
0.009993545711040497,
0.02727990411221981,
0.007515911944210529,
0.01910851150751114,
-0.02408970147371292,
-0.006384474225342274,
-0.017197325825691223,
0.015974296256899834,
-0.010186961852014065,
-0.0... | [
0.6989670991897583,
0.27177366614341736,
0.08105628192424774,
0.0650641918182373,
-0.3701946437358856,
-0.35093244910240173,
0.3929249346256256,
0.2520931363105774,
-0.06191254034638405,
-0.26149702072143555,
0.09108342975378036,
0.4391328692436218,
0.2967928647994995,
-0.3404059112071991,... |
Suppose a Gaussian stationary AR(p) process is $$ X_t = \sum_{i=1}^t \phi_i X_{t-i} + a_t $$ where $a_t$ is from iid $N(0, \sigma_a^2)$. 1. For estimating its parameters from a sample path of length $n$, I feel it more convenient to maximize the conditional log likelihood $\log p(X_{p+1}, \dots, X_n \mid X_1, \dots, X_p, \phi_1, \dots, \phi_p, \sigma_a^2)$ instead of maximizing the log likelihood $\log p(X_1, \dots, X_n \mid \phi_1, \dots, \phi_p, \sigma_a^2)$ , since I don't know how to derive the log likelihood for the joint distribution of $X_1, \dots, X_p$. I know $X_1, \dots, X_p$ must have a multivariate normal distribution, but I couldn't find explicit formula for their covariance matrix. So I wonder if you know an explicit formula for the covariance matrix of $X_1, \dots, X_p$ and for the log likelihood? 2. When calculating AIC, since I don't know how to calculate log likelihood, is it acceptable to use conditional log likelihood instead of log likelihood? My concern is that the difference between log likelihood and conditional log likelihood is $\log p(X_1, \dots, X_p \mid \phi_1, \dots, \phi_p, \sigma_a^2)$, which is related to the AR order p. If I use the "conditional" AIC to choose p, I worry I don't compare the conditional AIC's of AR models with different p on the same ground, because, the bigger p is, the more l ignore from the conditional AIC for not including $\log p(X_1, \dots, X_p \mid \phi_1, \dots, \phi_p, \sigma_a^2)$. So is such conditional AIC acceptable? Thanks! | [
-0.005584777798503637,
0.01289651170372963,
-0.0004728322383016348,
0.008842646144330502,
-0.003931027837097645,
-0.0048735542222857475,
0.004910966381430626,
-0.0065996116027235985,
-0.005469691473990679,
-0.004998416639864445,
-0.006784062832593918,
0.00394408218562603,
-0.0130212865769863... | [
-0.2733977735042572,
-0.16015589237213135,
0.3060227334499359,
-0.3290227949619293,
-0.1460438072681427,
0.6786052584648132,
-0.09748036414384842,
-0.43368345499038696,
0.030144548043608665,
-0.409972608089447,
-0.15834452211856842,
0.23375245928764343,
-0.623000979423523,
0.37218156456947... |
I have Centos 6.5 final. I have chroot setup and working properly and I also installed vsftpd. I now need to create a new user named `extDeveloper` and give this user access to a specific account. I setup the user using (adding the new account to an existing account): adduser extDeveloper -d /var/www/vhosts/domain.com passwd extDeveloper I also added this user to the apache group so they could access the files. The issue I have is the extDeveloper user is not able to write new files. I thought about making the group writable but then was concerned over security for this. What I am trying to do is create a secondary account (eventually more for this purpose) that I can give access to a few domains. Currently I only need 1 domain but this user may need to access more. I need the user to modify files and push new files but I need these files associated to the proper account:group. My accounts are setup using `domain_user:apache` where apache only has read access. Isn't it bad to let apache have write access? Is the setup I want possible? I assume so as I have been able to have multiple FTP accounts on other servers and they wrote files back using the proper permissions. Rackspace told me this is not how Linux works... I am learning as I go so any assistance or push in the right direction is very much appreciated. Thanks | [
-0.009918945841491222,
0.016155043616890907,
-0.005317427217960358,
-0.008816871792078018,
-0.016442853957414627,
0.008039642125368118,
0.009015931747853756,
0.012749070301651955,
-0.010595309548079967,
-0.04038143903017044,
-0.01162744127213955,
0.01763109117746353,
-0.001066382508724928,
... | [
0.3725143373012543,
-0.027985744178295135,
0.6304233074188232,
0.0021273107267916203,
-0.19814857840538025,
-0.11905495077371597,
0.5407980680465698,
-0.3897259831428528,
0.07894149422645569,
-0.6987893581390381,
0.152131125330925,
0.534551203250885,
-0.3073086440563202,
0.1742569804191589... |
I’m planning to get the “Ultimate Edition” from PSN that includes all the DLC. However, will the save file for my retail copy work on this version? | [
0.029462486505508423,
0.021782398223876953,
-0.0021982407197356224,
0.02876760996878147,
-0.001755453646183014,
0.046524930745363235,
0.01404848974198103,
0.0026074156630784273,
-0.022863559424877167,
-0.0388822965323925,
-0.017660295590758324,
0.027769481763243675,
0.022349892184138298,
-... | [
0.39526429772377014,
-0.1510154902935028,
0.20408570766448975,
0.3228814899921417,
0.12502650916576385,
-0.5368345975875854,
0.010481640696525574,
0.13109032809734344,
-0.09287501871585846,
-0.49113982915878296,
-0.16691796481609344,
0.45697736740112305,
0.141927570104599,
-0.0830945074558... |
**Background** I work on a development team and for each project we have a different server setup to emulate the environment of our customers. It can be hard to keep track of what I was doing on each server and where I need to start. In order to alleviate this problem I have started using a terminal multiplexer (tmux/screen). My problem is on one of the servers that I work on, the project lead is constantly making mistakes and resetting the system. When that happens I lose all of my screens as well as time getting back to where I was and what I was doing. **Question** All developers are given root access to the servers and I want to know if there is anyway I can set up the server so that my project lead gets a warning of users logged on or screens running when they decide to reset the server? (If possible I'd like to include a message of what I'm running and why they should find a different solution to resetting the server. In my company this is the only project manager that resets the server on a daily basis.) | [
0.02040260285139084,
0.0003815074451267719,
-0.008376648649573326,
-0.012986751273274422,
0.0022954337764531374,
-0.0068889157846570015,
0.005933498032391071,
0.005977953784167767,
-0.010776610113680363,
-0.006082487292587757,
-0.0030693677254021168,
0.011630546301603317,
0.00232328940182924... | [
0.49581339955329895,
0.21468907594680786,
-0.10352389514446259,
-0.06440940499305725,
0.2737651765346527,
0.29758861660957336,
0.5384568572044373,
0.7426990270614624,
-0.5197646617889404,
-0.8026919960975647,
0.2319762259721756,
0.48828211426734924,
-0.17140230536460876,
0.0203125048428773... |
I want to make some conditions on roman pages and another condition on arabic numbered page. (I want to know how to remove the background from specific pages in both numbering categories). \AddEverypageHook{% \ifnum\value{page}>0 \ifodd\value{page}\relax% \backgroundsetup{% settings for odd-numbered pages angle=0, contents={\begin{tikzpicture}[remember picture,overlay] \node at ([yshift=0cm,xshift=-9.7cm]current page) {\includegraphics[scale=1.10] {Margin2.png}}; %% yshift and xshift for example only \end{tikzpicture}} }% \else \backgroundsetup{% settings for even-numbered pages angle=0, contents={\begin{tikzpicture}[remember picture,overlay] \node at ([yshift=0cm,xshift=9.7cm]current page) {\includegraphics[scale=1.10]{Margin2.png}}; %% yshift and xshift for example only \end{tikzpicture}} }% \fi% \BgMaterial \fi} EDIT: \begin{document} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \pagestyle{empty} %No headings for the first pages. %% front setting %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \input{pg} \cleardoublepage \pagenumbering{Roman} \pagestyle{empty} \tableofcontents %Table of contents \chapter*{\textbf{Introduction}} \pagenumbering{arabic} \addcontentsline{toc}{chapter}{Introduction} \thispagestyle{plain} \large BLA BLA. \end{document} I want to remove the bg from the homepage and i want to know how we can remove background from specific pages. \NoBgThispage doesn't work here | [
0.012676949612796307,
0.02443430759012699,
-0.0021636805031448603,
0.019355174154043198,
-0.004471729509532452,
0.0070829931646585464,
0.007469171192497015,
-0.003922175150364637,
-0.016993582248687744,
0.017487695440649986,
-0.008138054981827736,
0.0038334238342940807,
-0.013777298852801323... | [
-0.11157263070344925,
0.25491827726364136,
0.7197198271751404,
-0.02338268794119358,
0.2250315248966217,
-0.002500444184988737,
0.3865688145160675,
-0.14496052265167236,
-0.0058836061507463455,
-0.7563993334770203,
-0.15567132830619812,
0.3815917670726776,
-0.20258422195911407,
0.031724281... |
In condensed matter, crystal with disclination and dislocation has both curvature and torsion. I am looking for a reference in which path integral quantization of Dirac equation on manifold with curvature and torsion was discussed. | [
-0.02317229099571705,
0.008563464507460594,
-0.006764222867786884,
0.03191521018743515,
-0.021875128149986267,
-0.04492358863353729,
0.019253935664892197,
0.0063606793992221355,
-0.019629543647170067,
-0.02750861458480358,
-0.014759985730051994,
0.019261585548520088,
-0.04268528148531914,
... | [
-0.15293759107589722,
0.09623242169618607,
0.45645034313201904,
0.028517741709947586,
-0.2839135229587555,
0.14973287284374237,
-0.39720234274864197,
-0.12097761034965515,
-0.5422540903091431,
-0.4660985469818115,
-0.3793174922466278,
0.4233524203300476,
-0.1432013362646103,
0.710854232311... |
I'm looking to create a button next to "add media" ( much like gravity forms "add form) that allows you to choose a custom post and then inserts that post id into a shortcode. I've bene looking at hooking into tinymce for the longest time, but realized that that wasn't the way to go (as those buttons aren't a part of the tinymce editor). anyone have any thoughts? | [
0.003775485558435321,
0.006795709487050772,
0.004252260085195303,
0.011391079984605312,
0.003252571215853095,
0.0006931936950422823,
0.008558322675526142,
0.030981307849287987,
-0.021700134500861168,
0.00778914662078023,
-0.014616912230849266,
0.008961992338299751,
0.011631084606051445,
0.... | [
0.9673107266426086,
0.0009384228615090251,
0.4556015431880951,
0.09748402237892151,
0.19603879749774933,
0.1559222936630249,
-0.20779310166835785,
0.507016122341156,
-0.1604708880186081,
-0.7796502709388733,
0.5387234091758728,
0.1385217308998108,
-0.1869957596063614,
0.22913305461406708,
... |
using the `\documentclass{llncs}` instead of for example `\documentclass{book}` will not make the captions aligned together for the following example. How can I do to align them together: \documentclass{llncs} \usepackage{capt-of, showframe, blindtext} \usepackage[demo]{graphicx} \begin{document} \blindtext \begin{table}[t]\fbox{ \begin{minipage}[b]{.50\textwidth }% \footnotesize \begin{tabular}{ | c || c | c | c | c | } \hline \textbf{ABC} & \textbf{KO} & \textbf{Ro} & \textbf{Ov} & \textbf{Params} \\ \hline \multicolumn{5}{|c|}{\textbf{Well this is a text here}} \\ \hline \textit{A1} & 123 & 123 & 123 & 123 \\ \hline \textit{Z2} & 123 & 123 & 123 & 123 \\ \hline \multicolumn{5}{|c|}{\textbf{Well this is another text here)}} \\ \hline \textit{O1} & 123 & 123 & 123 & 123 \\ \hline \textit{U9} & 123 & 123 & 123 & 123 \\ \hline \end{tabular}% \caption{results in this table}\label{validation}% \end{minipage}}% \fbox{\begin{minipage}[b]{.45\textwidth}% \hspace*{0pt}\includegraphics[width=0.45\textwidth]{plot.pdf}% \captionof{figure}{This is a figure not a table\label{x}}% \end{minipage}} \end{table} \end{document} | [
0.014929300174117088,
0.004307800903916359,
-0.005636543966829777,
0.013862041756510735,
-0.0054792077280581,
0.011773591861128807,
0.009703932330012321,
0.02530301734805107,
-0.013472648337483406,
0.010697523131966591,
-0.029979512095451355,
0.0033684498630464077,
-0.004980429075658321,
0... | [
0.10013280063867569,
-0.02116505056619644,
0.5592449903488159,
0.16428795456886292,
0.21088193356990814,
-0.15600915253162384,
0.27490338683128357,
-0.3805265724658966,
-0.13650527596473694,
-0.7599368691444397,
-0.043034277856349945,
0.5427678823471069,
-0.35347530245780945,
-0.1113380268... |
When my Samsung Galaxy S2 ran on Gingerbread I had a unique Bluetooth name set on it to differentiate from members of my family when I wanted to receive files by Bluetooth from their devices. However, since my upgrade to Ice Cream Sandwich I cannot see how to give my phone a name so that others searching for my phone by Bluetooth can find mine without others having switch theirs off to distingiush mine as the only one with its Bluetooth switched on. Can anybody help with this? Thank you for any help offered. | [
0.021247219294309616,
0.011005435138940811,
-0.008683283813297749,
0.0014065520372241735,
0.003917095251381397,
-0.006045719143003225,
0.009150458499789238,
0.023430483415722847,
-0.015656597912311554,
-0.015355302020907402,
-0.02272276021540165,
0.017929259687662125,
-0.005552458576858044,
... | [
0.25716865062713623,
0.3395889699459076,
0.455172061920166,
0.032661017030477524,
-0.2755641043186188,
0.013533984310925007,
0.4886133372783661,
0.21241092681884766,
-0.5815331339836121,
-0.34304267168045044,
-0.15853196382522583,
0.2490459680557251,
-0.08733444660902023,
0.387092977762222... |
I want to change a description of a template position, so when I go to Admin->Extensions->Module Manager I see a different description of a module position in the position list when I edit a module. However, when I change (for instance) the template 'beez_20' and want to rename the name of the position 'debug', I change the description (TPL_BEEZ_20_POSITION_DEBUG) in the language file 'languages\en-GB\en- GB.tpl_beez_20.sys.ini' to something different, say 'Abracadabra'. However, the changes don't appear in the position list and I can find no reference whatsoever of how or when the ini files are read or maybe cached. Does anyone has a clue? | [
0.004059533588588238,
0.015252566896378994,
-0.002219946589320898,
0.010451080277562141,
-0.013462565839290619,
0.034319695085287094,
0.00831180065870285,
0.00599182303994894,
-0.0141689358279109,
0.003357456997036934,
-0.012382693588733673,
0.020767733454704285,
-0.0033844050485640764,
0.... | [
-0.0019992662128061056,
0.0021208918187767267,
0.2455560714006424,
-0.5323905944824219,
0.5277596116065979,
0.17518456280231476,
0.001570363761857152,
-0.31406769156455994,
-0.35826680064201355,
-0.6622179746627808,
-0.2487170547246933,
0.523857057094574,
-0.3936157822608948,
0.15636339783... |
I need to know whether the following sentence is grammatically correct or not. Specifically, I am unsure about _have been being investigated_ part. Do the times that I am using match each other? > The interest in researching X has been growing in recent years, and as its > applications to various fields have been being investigated, it has been > realized that X which can directly deal with Y is desired in several fields. | [
0.0009009612840600312,
0.017045602202415466,
-0.01197756640613079,
0.018382789567112923,
0.02176039293408394,
-0.007117235101759434,
0.007071423809975386,
0.020162317901849747,
-0.01329862792044878,
0.02389276772737503,
0.007183938752859831,
0.013012068346142769,
-0.003174279583618045,
0.0... | [
0.3320707082748413,
0.4604286551475525,
0.04848074913024902,
0.024873344227671623,
-0.3337613642215729,
-0.09679772704839706,
0.47765642404556274,
0.06803865730762482,
0.0951048955321312,
-0.7166407704353333,
-0.05235721915960312,
0.3362509310245514,
-0.05511708930134773,
0.065707772970199... |
I'm searching for a special bibliographystyle to meet the following requirements (I post one example): > LÜTKEPOHL H., KRÄTZIG M. (2004). > > _Applied Time Series Econometrics_. Cambridge University Press, Cambridge. So names should be in upper case; format: last name and then initials of first name (with or without a following dot), year in brackets (with or without a following dot), followed by a line-break, the titel should be displayed italic. I use the `natbib` package because I want to have author-year citation in the text (e.g. Lütkepohl and Krätzig (2004)), which is working fine already. By the way the entry in the reference list should begin directly with the last name of the author, i.e. without numbering or acronym in front (I think this will be achieved automatically by means of using the natbib package?). I'm using `\documentclass[a4paper]{article}` and JabRef as reference manager. Would be very happy if someone can help me! Thank you very much! | [
0.009315679781138897,
0.00522032892331481,
-0.00856499932706356,
0.01677045226097107,
0.009961722418665886,
0.013924572616815567,
0.007070100866258144,
0.01263390015810728,
-0.010538613423705101,
-0.007382392417639494,
-0.00972849689424038,
0.0031144521199166775,
-0.0063760485500097275,
-0... | [
0.0576142780482769,
0.20258650183677673,
0.06381428241729736,
-0.09989050030708313,
-0.0628572329878807,
0.07211118191480637,
0.21902693808078766,
-0.03812696784734726,
-0.2634444534778595,
-0.6346546411514282,
-0.1970202922821045,
0.02816891483962536,
-0.04400142654776573,
0.3795142173767... |
I have been trying to get a feel for what high or low Drift might mean for controls. I'm still unsure. * Does high Drift mean the car sticks to the road less, allowing easier drifting? * Does high Drift mean the car, on the contrary, sticks to the road? * Assuming high Drift means no adherence... does low Drift mean high adherence? Experience seems to prove otherwise: most basic cars have all low values and yet start drifting if you take too-tight turns. * Could it be strictly related to dirt roads? I've noticed cars with high Drift are sometimes also "rally" type cars. | [
-0.052526675164699554,
0.019362779334187508,
-0.010878447443246841,
0.025174206122756004,
-0.028803588822484016,
-0.029510900378227234,
0.011216852813959122,
-0.003519070101901889,
-0.01249469444155693,
0.013675190508365631,
0.0017986828461289406,
0.02611171454191208,
0.01742154359817505,
... | [
0.45713913440704346,
0.2289772480726242,
0.41533780097961426,
0.4939994215965271,
-0.3637087047100067,
-0.4777490198612213,
0.12480739504098892,
0.11724131554365158,
-0.1959424912929535,
-0.12029286473989487,
0.3973260521888733,
0.6033698320388794,
0.17646530270576477,
0.05871181935071945,... |
Let $$X_1,\dots,X_m$$ are i.i.d. with distribution function $F$ and $$Y_1,\dots,Y_n$$ are i.i.d. with distribution function $G$. Suppose that there exists an unknown function $\psi:\mathbb{R}\mapsto\mathbb{R}$ such that $\psi(X_i)\sim N(0,1)$ and $\psi(Y_j)\sim N(0,\sigma^2)$ for all $i=1,\dots,m$ and $j=1,\dots,n$. I'd like to estimate $\sigma^2$ in this problem. I have obtained the following facts: Note that $F(x)=P(X\le x)=P(\psi(X)\le \psi(x))=\Phi(\psi(x))$ where $\Phi$ is the cumulative standard normal distribution. This implies $\psi(x)=\Phi^{-1}(F(x))$. Recall that $\psi(Y_j)\sim N(0,\sigma^2)$. So, $$\check\sigma^2=\frac1n\sum_{j=1}^n\psi^2(Y_j)$$ is an optimal estimator for $\sigma^2$. Since $F$ is unknown then I replace it with its empirical distribution function $\hat F_m$ based on $X_1,\dots,X_m$. Hence, it is natural to replace $\psi$ with $\hat\psi=\Phi^{-1}(\hat F_m)$. Therefore, I conjecture that $$\hat\sigma^2=\frac1n\sum_{j=1}^n\hat\psi^2(Y_j)$$ is an optimal estimator. I have tried to show $\check\sigma^2$ and $\hat\sigma^2$ are asymptotically equivalent, but I was failed. Could anyone help me? Or does any one have another approach? | [
-0.02041204646229744,
0.006191817577928305,
-0.0033477526158094406,
0.014092929661273956,
0.0110879922285676,
-0.012699132785201073,
0.0062302397564053535,
0.005224832333624363,
-0.007393411360681057,
-0.007262261584401131,
-0.009861700236797333,
0.0039059363771229982,
-0.011430958285927773,... | [
-0.13117113709449768,
0.21006207168102264,
0.41014447808265686,
-0.31377679109573364,
0.30237042903900146,
0.339241623878479,
-0.08570506423711777,
-0.28917834162712097,
-0.0969327762722969,
-0.5557659268379211,
-0.4139799475669861,
0.2474767565727234,
-0.6631610989570618,
0.34847384691238... |
In English as Germanic language K is ignored at the beginning of word in speech. _Night_ and _knight_ have to be pronounced similar. Then how to differentiate? | [
-0.008108867332339287,
0.02690533734858036,
-0.015075852163136005,
0.022842340171337128,
-0.04443829506635666,
0.03600075840950012,
0.017767945304512978,
-0.032967280596494675,
-0.022092506289482117,
0.02387222833931446,
-0.022942285984754562,
0.0003389751655049622,
-0.0042290291748940945,
... | [
-0.22109109163284302,
0.11930233985185623,
0.11873567849397659,
-0.6547016501426697,
-0.5465483069419861,
0.24289549887180328,
0.929267942905426,
0.2956085205078125,
-0.22833852469921112,
-0.68211829662323,
-0.1514171063899994,
-0.1639261096715927,
0.01060988288372755,
0.05235037952661514,... |
I know how to execute multiple commands at same time but now what I need is to run the same command multiple times with some time delay between the command execution. ### Requirements * I don't want to use a script for this * A single one line command to accomplish this * Needs to run on Linux * Need to control the number of times the command will run | [
0.017129963263869286,
0.01759272627532482,
-0.01733877882361412,
0.023403754457831383,
-0.021698826923966408,
-0.02088783122599125,
0.008383579552173615,
0.020541325211524963,
-0.0354573130607605,
0.014347920194268227,
-0.019778070971369743,
0.004262577276676893,
-0.00021004638983868062,
0... | [
0.6444125771522522,
0.055774737149477005,
0.1886720061302185,
0.1172797903418541,
-0.09811761975288391,
-0.08238520473241806,
0.6057085990905762,
-0.035994209349155426,
-0.27948036789894104,
-0.6459862589836121,
0.2596830725669861,
0.7581695914268494,
-0.11297441273927689,
0.22962054610252... |
How should you understand the expression: "flatter to deceive"? The Oxford Dictionaries defines flatter to deceive as: > Appear promising but ultimately disappoint. Which is all nice and dandy. But I fail to understand the wordplay, nor how "flatter" is mechanically related to "deceive" in this expression. For a real-life example of usage: > The podium that Kevin Magnussen scored on his debut was ultimately a case of > ‘flattering to deceive.’ [Kevin Magnussen's car scored a good result in his first race, but bad to worse results thereafter. The car flattered in the beginning, and deceived later on.] | [
-0.00802532583475113,
0.014423790387809277,
-0.009848218411207199,
0.015980340540409088,
-0.02586274966597557,
-0.008344091475009918,
0.008938658982515335,
-0.006227175705134869,
-0.012911627069115639,
-0.008834412321448326,
-0.004268244840204716,
0.0035844636149704456,
0.001367522869259119,... | [
0.162415012717247,
0.4268289804458618,
0.2873852849006653,
0.1589815765619278,
-0.37980055809020996,
-0.054643288254737854,
0.23911768198013306,
0.07187920063734055,
0.07782906293869019,
-0.10940863192081451,
0.3933561146259308,
0.6984827518463135,
0.4472324252128601,
-0.6580832600593567,
... |
Imagine that I have a polymer (approximated as a freely diffusing, freely jointed chain with some number of subunits 'N'), and I place this polymer into a sphere of some volume 'V'. Next, I proceed to add a series of infinitely thin, immobile chords of length 'L' to the inner walls of the sphere. Because these chords are infinitely thin, they will not change the inside volume of the sphere, but should nevertheless place certain topological and geometric constraints on the behavior of the diffusing polymer. Can we quantify the change in entropy caused by the addition of these infinitely thin chords? For this calculation, what changes if we replaced the polymer with a monoatomic gas? | [
-0.0003004205645993352,
0.009883588179945946,
-0.010248899459838867,
0.018948623910546303,
-0.017882417887449265,
-0.0322825089097023,
0.008579149842262268,
0.010764033533632755,
-0.011230921372771263,
-0.009483354166150093,
-0.002877502702176571,
0.009412236511707306,
-0.01218331977725029,
... | [
0.4281585216522217,
0.156271293759346,
0.591080367565155,
-0.11289587616920471,
0.04282068461179733,
0.2424604296684265,
-0.39580586552619934,
-0.2843441367149353,
-0.6074948906898499,
-0.3258405327796936,
-0.11273755133152008,
0.2720085680484772,
-0.3541452884674072,
0.7815513014793396,
... |
I recently picked up a cheap-o Xbox 360 wired headset (one that plugs into the wireless XBox 360 controller I use). I'm not an Xbox Live Gold member, and so I don't need chat features or anything - mostly what I'm looking for is to be able to hear the game's audio with the volume on my TV turned down. However, the headset doesn't seem to emit any of the game sounds. So far I've only tried it with Fable II. Is this a result of my stinginess? If I pick up a better headset will I be able to use it to listen to the game sounds? Or is this somehow related to the game - do only some games send game audio to the controller? Or am I crazy to think that I'll be able to hear game audio through a headset connected to my XBox 360 wireless controller? Any help or advice appreciated! | [
0.00970320776104927,
-0.014225397258996964,
-0.013249484822154045,
0.0010512247681617737,
0.024630244821310043,
-0.016682330518960953,
0.0065218610689044,
-0.01948917657136917,
-0.015848297625780106,
0.004008507821708918,
-0.004734475631266832,
0.010619208216667175,
-0.015738580375909805,
... | [
0.3726910650730133,
0.10904590040445328,
0.038572102785110474,
0.2831796705722809,
0.19679726660251617,
-0.10627216100692749,
0.2844027280807495,
-0.10676984488964081,
-0.014043337665498257,
-0.5986377596855164,
0.4878455400466919,
0.7900807857513428,
0.11750011146068573,
0.014535490423440... |
I am a Hetzner level 19 shared hosting client and would like to change the A-record for my domain. When I log in to konsoleh, I see which IP-address it points to at the moment, but not where I can change it. Does anyone know how to do it? Thanks! | [
0.0015392119530588388,
0.007763281464576721,
-0.0034447291400283575,
0.019294237717986107,
-0.0012353261699900031,
0.00834950152784586,
0.011969697661697865,
0.01026670727878809,
-0.023774191737174988,
0.019622236490249634,
0.010367813520133495,
0.020162200555205345,
-0.013680024072527885,
... | [
0.5995819568634033,
0.36144912242889404,
0.4282282292842865,
0.04782455787062645,
0.06525038182735443,
-0.3356788456439972,
0.6333975195884705,
0.2979729473590851,
-0.2239968478679657,
-0.3708236515522003,
-0.03140164166688919,
0.11159291863441467,
0.18664772808551788,
0.35641542077064514,... |
Could you explain to me why a\_b {\ttfamily a\_b} \\ \chardef\_=`\_ a\_b {\ttfamily a\_b} produces  ? I would like to have the short underscore in the normal text and the long one in fixed-width font. I cannot change `{\ttfamily a\\_b}`. The other `a\\_b` I can change. Thanks for your help! * * * EDIT: I realized that `a\textunderscore b` solves my immediate problem in text. I would still like to know where the mysterious dot is comming from. In math mode, \chardef\_=`\_ $a\_b$ LaTeX stops compiling altogether on my installation (without error, it just halts). | [
0.007006514817476273,
0.014085670933127403,
-0.007467810530215502,
0.013627825304865837,
0.00020483508706092834,
-0.0063711837865412235,
0.005790018942207098,
0.00014429783914238214,
-0.012526110745966434,
-0.003239783924072981,
-0.014913113787770271,
-0.001774332602508366,
0.005913728848099... | [
0.10343001037836075,
0.2306554913520813,
0.28880465030670166,
0.1697547882795334,
0.1693105548620224,
0.29583749175071716,
0.17634263634681702,
-0.047308508306741714,
-0.28136518597602844,
-0.2976369261741638,
-0.25188344717025757,
0.5579618215560913,
-0.10627906024456024,
0.03395999222993... |
Can you still import the save? If so what happens? | [
0.043772101402282715,
0.06239636242389679,
-0.006221757270395756,
0.013135851360857487,
0.05042815953493118,
0.0012845483142882586,
0.013250664807856083,
0.007407323457300663,
-0.048476848751306534,
-0.01118299551308155,
-0.017275162041187286,
0.05423407256603241,
0.0034148497506976128,
0.... | [
0.5419642925262451,
-0.05294610559940338,
0.1912955939769745,
0.22987711429595947,
0.31945377588272095,
0.12353663146495819,
-0.1586294323205948,
0.34361761808395386,
-0.2168761044740677,
-0.5084208846092224,
0.05541970208287239,
0.26119327545166016,
-0.031821299344301224,
-0.0883452892303... |
> In a sense, 10! (ten factorial) represents an approximate dividing line > between things which are practical to compute and things which are not. This is from Knuth's TAOCP Fundamental Algorithms book (1973). Is this still a valid statement or has computing power rendered it obsolete ? | [
-0.025947190821170807,
0.013165600597858429,
-0.0045718178153038025,
0.015244358219206333,
-0.053706586360931396,
-0.033222541213035583,
0.009493879973888397,
-0.024822132661938667,
-0.011909711174666882,
-0.014759782701730728,
-0.015460272319614887,
0.00030659345793537796,
-0.01206237915903... | [
0.015133053995668888,
0.024973366409540176,
0.28503626585006714,
0.1089649349451065,
-0.23650024831295013,
0.26537367701530457,
0.46353545784950256,
-0.36090052127838135,
-0.10940629243850708,
-0.22502774000167847,
-0.030638467520475388,
0.5093399882316589,
-0.11340872198343277,
-0.2645099... |
Can somebody please show the steps of how differentiation of Shannon's entropy yields the following result? $H = -\sum_{l=0}^{L-1} p(l)\log_2[p(l)]$ The result of differentiating is $H_m = -\sum_{l=0}^{L-1} \frac{1}{L}\log_2[\frac{1}{L}]$ = $log_2 L $ | [
-0.008278774097561836,
0.010519309900701046,
-0.016368713229894638,
0.013627887703478336,
-0.02809199132025242,
-0.020249230787158012,
0.007191635202616453,
-0.010775723494589329,
-0.010818110778927803,
-0.00902499444782734,
-0.007713234052062035,
0.00021101138554513454,
-0.02964539080858230... | [
-0.21232013404369354,
-0.16677120327949524,
0.13037434220314026,
-0.20348189771175385,
0.1678009182214737,
0.01118468213826418,
0.06494326144456863,
-0.23045384883880615,
-0.1049010381102562,
-0.3148738741874695,
0.06504631787538528,
0.6757389903068542,
-0.5214406847953796,
0.3368734419345... |
I've got Motorola RAZR (XT910) and I decided to install the Google+ app to stay in touch with it. Whenever I take pictures with my phone it would upload those pictures to my Google+ account even though I have set my Instant Upload settings to OFF. How can I really disable that without uninstalling the Google+ app? I would like to hand pick photos from my phone to be uploaded to my Google+. Thanks | [
-0.018611740320920944,
-0.004601810593158007,
0.0028746554162353277,
0.031423527747392654,
0.005892473738640547,
-0.008431563153862953,
0.010458315722644329,
0.02770334854722023,
-0.022962301969528198,
-0.0278012678027153,
-0.006978853140026331,
0.010831259191036224,
0.010190608911216259,
... | [
0.5076590180397034,
0.3603809177875519,
0.41047459840774536,
-0.11528322845697403,
0.004178335890173912,
0.2786650061607361,
0.599993109703064,
-0.1476828157901764,
-0.09326718747615814,
-0.21118077635765076,
0.3135562241077423,
0.41720908880233765,
-0.027501041069626808,
0.172442346811294... |
> **Possible Duplicate:** > “Would you have liked to have been” vs. “would you have liked to be” Is "have" redundant when repeated in successive verb phrases? Well, let us read the following sentences: > I would have liked to have seen New York before the cyclone, > > I would have liked to see New York before the cyclone. Am I right in saying that one instance of "have" seems to have prompted another? Or are there subtleties and shades of meaning? | [
0.015260430052876472,
0.025910111144185066,
-0.008863291703164577,
0.012384707108139992,
0.011659783311188221,
0.007907579652965069,
0.007294595241546631,
0.006042906548827887,
-0.010798594914376736,
0.005253761541098356,
-0.014041894115507603,
0.004450595006346703,
0.009457111358642578,
-... | [
0.2943749725818634,
-0.30079853534698486,
-0.048097532242536545,
0.06447872519493103,
-0.20756131410598755,
0.39542528986930847,
0.3816854953765869,
0.1044669821858406,
-0.607133686542511,
-0.7227979898452759,
0.22561664879322052,
0.5245763063430786,
-0.20480893552303314,
-0.29380708932876... |
I will get contour lines of electromagnetic field strength around power lines. The contours will be calculated by different agencies using different models. We need to check whether the resulting zones (x meter from power line) around the power lines are similar between models/agencies. There is no 'gold standard' because it is not possible to actually measure the field strength on the ground. Furthermore, the zone is not simply a straight line because in some situations two power lines cross or approach each other and we are most interested in these deviating situations. Is there a way in GIS to compare the modelling results (polylines/contours) among each other? | [
0.015356460586190224,
0.015246005728840828,
-0.008656233549118042,
0.024909736588597298,
-0.03689410537481308,
-0.016732756048440933,
0.010598892346024513,
-0.020048661157488823,
-0.01629849523305893,
0.008985595777630806,
0.0038125202991068363,
0.028826512396335602,
0.006211890839040279,
... | [
0.7388367652893066,
-0.13830499351024628,
0.3790329098701477,
0.31062981486320496,
-0.08144158869981766,
-0.18171586096286774,
-0.028396490961313248,
-0.3807368576526642,
-0.1285952627658844,
-0.4729185998439789,
0.07409795373678207,
0.39012283086776733,
-0.05710553750395775,
0.19026751816... |
I'm writing here following http://math.stackexchange.com/questions/402407/how- to-convert-pdf-containing-math-made-from-latex-to-word and http://meta.math.stackexchange.com/questions/9715/reopening-thread-that-i- dont-think-is-off-topic. **1.** My main goal is just to use MathType in Word and to edit with Word. Therefore, what is the best way to convert an Adobe PDF containing math (equations, symbols, tables, etc.) into a Word document? **2.** Is there another way to make the PDF appear in Word? I was thinking that maybe the PDF doesn't need to be converted. Instead, each page of the PDF could be convertered/terated as a picture which could then be pasted onto a Word document. This process takes too long to do for each page. What program or recourse is available? Thank you very much! Unfortunately, it would take too much time for me to retype the PDF, because it is essentially a free textbook containing maths on which I must annotate. Also, I do not have the LaTeX source file for the PDF, only the PDF itself. I've spent at least 3 hours trying to research and resolve this question, including reading http://www.physicsforums.com/showthread.php?t=507174 and http://forums.adobe.com/message/3906332. I've tried the following PDF to Word converters, but ALL of them deform, disfigure, and distort the math, the fonts, and the formatting: Zamzar.com, Wondershare PDF to Word Converter 4.0.1, Able2Doc PDF to Word Converter v7.0.34, and UniPDF. | [
0.0032980318646878004,
0.008659242652356625,
-0.006390835158526897,
0.010746375657618046,
0.002678538439795375,
0.0033133751712739468,
0.006295252591371536,
0.0004907713737338781,
-0.02849062904715538,
-0.008421927690505981,
0.005261790007352829,
0.0009293713374063373,
0.006149638444185257,
... | [
0.10479456186294556,
0.35933777689933777,
0.37831830978393555,
-0.08370901644229889,
-0.3980634808540344,
-0.033045437186956406,
0.13203047215938568,
0.01809116266667843,
-0.16933265328407288,
-0.6966667175292969,
0.14598657190799713,
0.33730942010879517,
0.1206948384642601,
0.051072809845... |
I used to do a lot of programming in R and recently switched to _Mathematica_. One thing that bothers me is that I haven't found a way to save variables, their values and definition! Let's look at an example: idx = Partition[RandomSample[Range[20], 20], 3]; Table[Subscript[vars, idx[[i]]] = Total[idx[[i]]], {i, 1, Length[idx]}]; Subscript[vars,{9,15,8}] (*32*) Is there a way to save the variable names as well as the corresponding values. And how would I read them into a new file? So that in a new file I can access with `Subscript[vars,{9,15,8}]` the value. I tried it with `Table[...]>>values.mx` and read it in with `values=<<values.mx` but that only gives me the values, whereas I also need the definitions. I also tried: DumpSave["values.mx", Table[Subscript[vars, idx[[i]]] = Total[idx[[i]]], {i, 1, Length[idx]}]] but got an error > DumpSave::bsnosym: "Table[..] is not defined as a symbol or a context" | [
0.0185567457228899,
-0.0002481756964698434,
-0.0066250404343008995,
0.0030153333209455013,
0.0018444252200424671,
0.015146528370678425,
0.004866477567702532,
0.009742820635437965,
-0.01268829870969057,
-0.001082241302356124,
-0.009055264294147491,
0.0010373103432357311,
-0.006445930339396,
... | [
0.04053970053792,
-0.07737744599580765,
-0.008208664134144783,
0.000019987282939837314,
-0.13038738071918488,
0.22047960758209229,
0.03432172164320946,
-0.32061532139778137,
-0.35287603735923767,
-0.6800492405891418,
0.4353282153606415,
0.3129422664642334,
-0.14754553139209747,
0.402240782... |
## Background My question concerns the calculation of risk ratios and odds ratios for meta- analysis using the `metafor` package. I know it looks long, but please bear with me, as it is a simple question. I've simply provided a lot of background information. First, copy/paste the below to make my example data. # Load package library(metafor) # Create example_data example_data <- data.frame( trial = c(1, 2, 3, 4, 5, 6), author = c("Abers", "Baker", "Cook", "Dodds", "Eggers", "Fritz"), year = c(2000, 2001, 2002, 2003, 2004, 2005), tpos = c(437, NA, 51, 26, 468, 10), tneg = c(49, NA, 1, 8, 83, 1), cpos = c(50, 7, 6, 0, 36, 0), cneg = c(33, 8, 0, 0, 19, 0) ) # Show example_data example_data trial author year tpos tneg cpos cneg 1 1 Abers 2000 437 49 50 33 2 2 Baker 2001 NA NA 7 8 3 3 Cook 2002 51 1 6 0 4 4 Dodds 2003 26 8 0 0 5 5 Eggers 2004 468 83 36 19 6 6 Fritz 2005 10 1 0 0 ## More background If I were to make a two-by-two table by hand for the first study, Abers, it would look as follows: pos neg pos 437 49 neg 50 33 The relative risk would be calculated by hand as follows: RR = (437/(437 + 49)) / (50/(50 + 33)) = 1.49 log(RR) = 0.40 The odds ratio would be calculated by hand as follows: OR = (437 * 33) / (50 * 49) = 5.89 log(OR) = 1.77 Here are the forest plots for log(RR) and log(OR). If you look at the first line in each plot, you can see that `metafor` produces the same values as I did by hand for the Abers study.   ## My question Take a look at the line for Dodds on the plots. Where are these values for log(RR) and log(OR) coming from? As I see it, if I were to make a two-by-two table for the Dodds, it would look as follows: pos neg pos 26 8 neg 0 0 The relative risk would be calculated by hand as follows: RR = (26/(26 + 8)) / (0/(0 + 0)) = Undefined! As you cannot divide by zero. log(RR) = ? The odds ratio would be calculated by hand as follows: OR = (26 * 0) / (0 * 8) = Again, undefined! log(OR) = ? **So how did the`metafor` package calculate a log(RR) of 0.41 and a log(OR) of 1.14 for the Dodds study? And more generally, for all of my studies with zero values ( _e.g._ , Cook, Dodds, and Fritz), can I even include any of them in my meta-analysis?** | [
0.014210640452802181,
0.021160222589969635,
-0.0047126938588917255,
0.010990350507199764,
0.02340700477361679,
0.002521107904613018,
0.005841991864144802,
-0.02119753137230873,
-0.010075798258185387,
-0.00797454547137022,
0.002693288726732135,
0.004949008114635944,
-0.01520288921892643,
0.... | [
0.5180771946907043,
0.4432482421398163,
0.37169358134269714,
-0.18708953261375427,
0.12090249359607697,
0.337420791387558,
0.11611413955688477,
-0.5351775288581848,
0.04271776229143143,
-0.2343168705701828,
-0.05897778272628784,
0.7779066562652588,
-0.2520243525505066,
-0.1705632358789444,... |
I'm looking for a package to do truncated regression in R. I found the truncreg and truncSP packages, but they seem to allow either a lower bound or an upper bound, but not both at the same time. All my response values are bound between 0 and 1, so I need to have both bounds taken into account. Any help? Thanks a lot! | [
0.016908228397369385,
0.014085273258388042,
-0.0031672322656959295,
0.014701448380947113,
0.006629639305174351,
0.0018412205390632153,
0.010243978351354599,
0.00042640866013243794,
-0.0255177840590477,
-0.015829281881451607,
-0.01148298941552639,
0.009832216426730156,
-0.026386260986328125,
... | [
0.02121511474251747,
-0.02928033284842968,
0.005486737005412579,
0.16426219046115875,
-0.039553187787532806,
0.08380232751369476,
0.3381335735321045,
0.04543440788984299,
-0.23548874258995056,
-0.13514579832553864,
0.08579672873020172,
0.3799396753311157,
-0.14950157701969147,
0.0217290725... |
I used the following example code from **latticeExtra** to understand two-way clustering in _R_ library(latticeExtra) data(mtcars) x <- t(as.matrix(scale(mtcars))) dd.row <- as.dendrogram(hclust(dist(x))) row.ord <- order.dendrogram(dd.row) dd.col <- as.dendrogram(hclust(dist(t(x)))) col.ord <- order.dendrogram(dd.col) library(lattice) levelplot(x[row.ord, col.ord], aspect = "fill", scales = list(x = list(rot = 90)), colorkey = list(space = "left"), legend = list(right = list(fun = dendrogramGrob, args = list(x = dd.col, ord = col.ord, side = "right", size = 10)), top = list(fun = dendrogramGrob, args = list(x = dd.row, side = "top", size = 10)))) and this is what I got  Joining of both row and column entities make sense to me but I'm confused with different color shades of heatmap. **Questions** * Do the Joining of row variables also take into account the column variables and vice versa * What does mean the different colors in heatmap for different row variables clustering as well as for column variables clustering. Specifically focus on **cyl** and **disp** row variables. | [
0.006727502681314945,
0.014383588917553425,
-0.01448669470846653,
-0.0014353615697473288,
0.011818679980933666,
0.011902745813131332,
0.008368519134819508,
-0.0036530077923089266,
-0.01193489320576191,
0.0003552529960870743,
-0.014642870984971523,
-0.00002993876114487648,
-0.0142154460772871... | [
-0.0336332768201828,
0.03825485706329346,
0.4132440686225891,
-0.0114133944734931,
-0.20354054868221283,
0.2514374256134033,
0.03249771147966385,
-0.5112866163253784,
-0.23776188492774963,
-0.7333149313926697,
0.11126380413770676,
0.1646081954240799,
-0.2557678520679474,
0.1202544420957565... |
For the bar chart proposed, I would like to use different color for each single bar. How can I do this? \documentclass{article} \usepackage{pgfplots} \begin{document} \begin{tikzpicture} \begin{axis}[ title=Title, xbar, xmajorgrids = true, bar width=6mm, width=12cm, height=5.5cm, enlarge y limits=0.2, xlabel={\#number}, symbolic y coords={A,B,C,D}, ytick=data, nodes near coords, nodes near coords align={horizontal}, ] \addplot coordinates {(1,A) (7,B) (5,C)(2,D)}; \end{axis} \end{tikzpicture} \end{document} | [
0.0076407743617892265,
0.015258376486599445,
-0.004296475555747747,
0.009455511346459389,
-0.0037722475826740265,
-0.004915193188935518,
0.0072975424118340015,
0.006046067923307419,
-0.014408178627490997,
-0.025531798601150513,
-0.002789335325360298,
-0.005756518337875605,
-0.002700946293771... | [
0.25553813576698303,
0.3094445765018463,
1.0082449913024902,
-0.0505756214261055,
0.11332352459430695,
0.4244031608104706,
0.2032134234905243,
-0.3588869273662567,
-0.11248403787612915,
-0.3713683485984802,
-0.07203517109155655,
0.3293688893318176,
-0.13798373937606812,
0.1644895076751709,... |
What is the origin of the nickname "Money Making Manhattan" for the borough of Manhattan in New York City? | [
-0.01921646110713482,
0.030320079997181892,
0.0459371954202652,
0.0020476998761296272,
-0.04325610771775246,
0.05813082307577133,
0.016614243388175964,
-0.011371636763215065,
-0.021179664880037308,
-0.032239023596048355,
-0.07766410708427429,
0.06754648685455322,
-0.0031731415074318647,
-0... | [
0.14795102179050446,
-0.23088760673999786,
-0.046280913054943085,
0.3061795234680176,
0.36944717168807983,
-0.01737746223807335,
-0.4649350941181183,
0.7047564387321472,
-0.32485777139663696,
0.10319437831640244,
0.2265409678220749,
-0.21269994974136353,
0.23249971866607666,
0.360662192106... |
I am creating a website / web application where companies can do there accounting. It is sort of a alternative to the regular desktop accounting program. I want the application to have a look and feel of a normal desktop application, and not of a traditional websites. What kind of font and font-size do you think comes close to the traditional desktop application. I am styling in CSS. | [
-0.002618693048134446,
0.009308489970862865,
0.008088668808341026,
0.016168469563126564,
-0.00017666655185166746,
0.005470349453389645,
0.009395848959684372,
0.017035750672221184,
-0.020618179813027382,
-0.002303304383531213,
-0.003945613279938698,
0.008372984826564789,
0.02398306503891945,
... | [
0.6033159494400024,
0.3779078423976898,
-0.13221660256385803,
-0.056022144854068756,
-0.020412800833582878,
0.11811458319425583,
-0.2382718175649643,
0.3065940737724304,
0.0033978940919041634,
-0.8185747861862183,
0.328877329826355,
0.4502059817314148,
0.16479834914207458,
-0.0594430640339... |
I'm trying to reposition the messages in my Woocommerce theme. For example the successfully added to cart message. They appear to automatically be placed at the very top of where the woocommerce wrapper start begins. I've already found this link which allows me to change the wording of the message but what I need is to remove the message from where it is and reposition it elsewhere. I've scoured the woocommerce forum, this forum and the woocommerce template files. There is a messages.php but that only appears to contain the markup for the messages. I cannot figure out how to change where the messages appear. Any help on this would be greatly appreciated. | [
-0.005105438642203808,
0.004506909288465977,
0.00022538634948432446,
0.03371637314558029,
0.007235136814415455,
0.011616704054176807,
0.008661150932312012,
0.010696261189877987,
-0.018872946500778198,
0.033751267939805984,
-0.019906166940927505,
0.012054584920406342,
-0.0038594354409724474,
... | [
0.4490306079387665,
0.05026022717356682,
0.8268073797225952,
0.143855020403862,
-0.33880478143692017,
0.021765660494565964,
0.017071058973670006,
0.34188491106033325,
-0.5905250906944275,
-0.4736512303352356,
-0.12091562896966934,
0.31432297825813293,
-0.13703244924545288,
0.37851762771606... |
I have almost the same question as: How can I efficiently model the sum of Bernoulli random variables? But: (1) The number of random variables for summation is ~ N=20 (case 1) or N=90 (case 2). (2) $p_i$ ~ 0.13 (case 1) (3) The precision of the model based on Poisson law is not enough. (4) We need that our approx would be the good enough to model partial sums like these as well: $\sum_{i=k,N}{X_i}$, ( $k=1,N$ ) (5) We have empirical data for every $X_i$. The diagram shows that there is almost linear dependence for $Pr(X_i=1)$ for i=1,6 and then almost constant function or some small linear dependence (for i=7,20 ). Really I'm not sure in (3) since the Le Cam's inequality looks like very general... What class of model could we try? | [
0.01651550829410553,
0.0038895383477211,
-0.015334231778979301,
0.012739710509777069,
-0.025464296340942383,
-0.019408240914344788,
0.007115258369594812,
0.001708619063720107,
-0.018954843282699585,
-0.017119569703936577,
-0.005320231895893812,
0.021936222910881042,
-0.038072891533374786,
... | [
-0.4449637830257416,
-0.015132775530219078,
0.2899041473865509,
-0.02051081880927086,
-0.12613634765148163,
0.5023384094238281,
0.1369602382183075,
-0.529848039150238,
-0.24469003081321716,
-0.647597074508667,
0.18015281856060028,
0.34523653984069824,
-0.22181864082813263,
0.28554165363311... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.