text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I am referring to _goods_ as in _services and goods provided_. I came across this definition (emphasis mine): > tender document: > > A written invitation sent to potential suppliers of **a good or service** to > inform them about the information required for the buyer to choose among > them. Issuing a tender document typically begins the tender process by which > a business selects qualified and interested suppliers based on such things > as their price, availability and proposed delivery terms. The use of _good_ as a countable noun seems wrong to me. Is _good_ used correctly here, is it a countable noun? Can one provide _a good_? EDIT: Google returns 9,790,000 hits for "suppliers of a good". Does that mean I have to accept it as "correct"? Does no one else feel it is wrong?
[ -0.003589864121749997, 0.00935263093560934, 0.00226469524204731, 0.01839859038591385, -0.013998838141560555, 0.002220225054770708, 0.006633000448346138, 0.0005457953084260225, -0.01627095602452755, -0.007966925390064716, -0.003596324473619461, 0.012200336903333664, 0.004533533006906509, -0...
[ 0.36313581466674805, 0.5130253434181213, 0.30796217918395996, -0.4688142240047455, -0.36244997382164, 0.37449973821640015, 0.22344526648521423, -0.15347571671009064, -0.16285350918769836, -0.4452654719352722, -0.22960273921489716, 0.5946003794670105, 0.006714657414704561, 0.226120725274086...
I read in a paper: > One such genus is _Dinemasporium_ which ... characterised by superficial, > cupulate to discoid conidiomata with brown setae, and phialidic > conidiogenous cells that give rise to hyaline, oblong to allantoid, aseptate > conidia with an appendage at each end. In the phrase "oblong to _allantoid_ ", is the word _oblong_ valid for this context? If not, what other word would work? ** I am editing to thank you all. Although few people agreed to close this question (because they concern to the whole phrases), I should have told that I was confused for the phrase **oblong to allantoid**. It is clear now for me that oblong to allantoid means the range of _genus Dinemasporium_ is **oblong (Roughly rectangular or ellipsoidal-shaped) to allantoid (Sausage-shaped; spores that are long with rounded ends).** I also learned from this journal for the word oblong in the same context; _The conidia were subhyaline, oblong, and ellipsoid to allantoid._
[ -0.02466656267642975, 0.00643108319491148, -0.004882943816483021, 0.024051811546087265, 0.007955657318234444, -0.011771890334784985, 0.010010454803705215, 0.010701019316911697, -0.012000320479273796, -0.0184035524725914, 0.012025614269077778, 0.0016059072222560644, 0.009811434894800186, 0....
[ 0.014332161284983158, 0.2888425886631012, 0.23309141397476196, -0.09765342622995377, -0.1860756129026413, 0.14761444926261902, -0.07235526293516159, 0.2584310472011566, -0.22331061959266663, -0.7490819692611694, 0.09537266194820404, -0.25499576330184937, -0.4907713234424591, 0.584415674209...
When I use `\cite{ref1, ref2}`, I have a character "?" which appears for ref2 instead of the number of the reference. Do you know how can I cite multiple references using \cite{ref1, ref2} ?
[ 0.027295345440506935, 0.027394693344831467, -0.020207449793815613, 0.03146684914827347, -0.0067182667553424835, -0.028563983738422394, 0.009861118160188198, 0.0035760372411459684, -0.024200312793254852, -0.01923014037311077, 0.0023859122302383184, -0.004186826292425394, -0.006684672087430954...
[ 0.23455853760242462, 0.45276081562042236, 0.2575695216655731, -0.1434078812599182, -0.11874321103096008, 0.1716812252998352, 0.005106600932776928, -0.19837012887001038, -0.09993845969438553, -0.5374014377593994, 0.12136779725551605, 0.5756452679634094, -0.21024444699287415, -0.025884149596...
I have built a network using Ordnance Survey's ITN data and I can solve routes across the network. What I need is to build in grade separation topology for overpasses and underpasses on the network so that my routing solutions do not drop off bridges onto roads underneath in an attempt to make the shortest path possible. Does anyone have any pointers as to where to start? I am using PostGIS 2.0.3, pgRouting 1.0.7dev and Postgresql 9.2 on Windows.
[ -0.00482127396389842, 0.00897394772619009, -0.0042035942897200584, 0.015743188560009003, -0.012063588947057724, 0.0032337193842977285, 0.008932980708777905, 0.0004912051954306662, -0.012396395206451416, 0.008950799703598022, -0.00481769535690546, 0.012140919454395771, -0.012890645302832127, ...
[ 0.35503455996513367, 0.10319097340106964, 0.3231591284275055, -0.060586366802453995, -0.029659245163202286, 0.24615971744060516, 0.3476666212081909, 0.27415111660957336, -0.06213129311800003, -0.8713578581809998, 0.09993952512741089, 0.2193911224603653, 0.022585321217775345, 0.095775760710...
* I have an equation extending into the right margin and off the page. * I have many equations. For this equation only, I would like to shift * it to the left. MWE below. \documentclass[twoside]{book} \usepackage{amsmath} \begin{document} \begin{equation} \text{Some big equation} \end{equation} \end{document} * * * Here is an example of an equation I wish to shift left horizontally as it is too far to the right. \documentclass[a4paper,12pt,twoside]{book} \newcommand{\MyLeftRoundBracket}{(} \newcommand{\MyRightRoundBracket}{)} \newcommand{\MyLeftSquareBracket}{[} \newcommand{\MyRightSquareBracket}{]} \usepackage{amsmath} \begin{document} \begin{align}\label{eq:CIR_Expected_Return_Variance_Spread} ^{Spread}_{CIR}\sigma^2(X^L,X^U,c) &= -{(c-X^L+X^U)^2 \over {_{CIR}\theta_1}^2} {1 \over \Bigg \MyLeftSquareBracket \frac{\partial \Phi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big)-} \cdots \notag \\ &\qquad {1 \over \frac{\partial \Phi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big) -\frac{\partial \Psi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big) + } \cdots \notag \\ & \qquad\qquad {1 \over \frac{\partial \Psi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}}, \frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big) \Bigg \MyRightSquareBracket^3 } \times \notag \\ &\qquad \Bigg [ \frac{\partial \Phi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big)^2-\frac{\partial \Phi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big)^2 + \notag \\ &\qquad \frac{\partial \Phi^2}{\partial a^2}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big)-\frac{\partial \Phi^2}{\partial a^2}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big)+ \notag \\ &\qquad \frac{\partial \Psi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big)^2-\frac{\partial \Psi^2}{\partial a^2}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^L}{{_{CIR}\theta_3}}\Big)- \notag \\ &\qquad \frac{\partial \Psi}{\partial a}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big)^2+\frac{\partial \Psi^2}{\partial a^2}\Big(0,\frac{2 {_{CIR}\theta_1} {_{CIR}\theta_2}}{{_{CIR}\theta_3}},\frac{2 {_{CIR}\theta_1} X^U}{{_{CIR}\theta_3}}\Big)\Bigg ] \end{align} \end{document}
[ 0.00423281267285347, 0.004382716957479715, -0.004669624380767345, 0.026090223342180252, -0.010125266388058662, 0.0004111742600798607, 0.008444908075034618, 0.02266601473093033, -0.014115622267127037, -0.02022663876414299, -0.007354954723268747, 0.008438516408205032, -0.004891132470220327, ...
[ -0.25060760974884033, 0.07265797257423401, 0.6552011966705322, -0.19187764823436737, -0.09402767568826675, -0.022009536623954773, 0.15752649307250977, -0.023958228528499603, -0.2721326947212219, -0.8534544706344604, 0.07746659964323044, 0.22116033732891083, -0.03755141794681549, -0.1093738...
The sun is made of fire but fire needs oxygen right? So.. 1. Why can there be flames in space, while there's no oxygen? 2. Same idea as with the rocket engines of the spaceship, which also produce fire while there's no oxygen?
[ -0.002431811299175024, 0.02468441054224968, 0.01349676214158535, 0.010200857184827328, -0.038816750049591064, -0.01881931908428669, 0.012905276380479336, 0.01417622435837984, -0.02728225104510784, -0.03716597333550453, -0.023528117686510086, 0.031806882470846176, -0.005803528241813183, 0.0...
[ 0.6990451812744141, 0.5445057153701782, -0.10158786922693253, 0.31060484051704407, -0.1097990870475769, -0.20265664160251617, -0.05356062576174736, -0.17578843235969543, -0.28421714901924133, -0.3875134587287903, 0.046857189387083054, 0.21682363748550415, -0.2506961226463318, 0.12271136045...
According to Wikipedia, _sprezzatura_ means > a certain nonchalance, so as to conceal all art and make whatever one does > or says appear to be without effort and almost without any thought about it. But has the word taken on a meaning more related to just "indifference" as opposed to the one above?
[ 0.0006580694462172687, 0.004256035666912794, 0.010779191739857197, 0.028405701741576195, 0.003790251212194562, -0.006835624575614929, 0.009150121361017227, -0.00006599152402486652, -0.013693859800696373, 0.025610296055674553, 0.0018594730645418167, 0.0043053231202065945, -0.00026422168593853...
[ 0.5258122086524963, -0.28190556168556213, 0.3325268626213074, -0.0720781460404396, -0.3568978011608124, -0.33572182059288025, 0.5034052729606628, 0.10146898031234741, -0.37513047456741333, -0.2086331695318222, -0.2076939195394516, 0.4582183063030243, -0.2926926910877228, 0.1414096951484680...
I have a dataset drawn from a social network that looks Bimodal on logarithmic scales for all attributes (I'll demonstrate only one here): ![Bimodal Gaussian Distribution](http://i.stack.imgur.com/f4R4N.png) I know the variable that would give me a clean separation for the two subpopulations (e.g., gender): ![enter image description here](http://i.stack.imgur.com/UzOC6.png)![enter image description here](http://i.stack.imgur.com/CLFs7.png) However, I need to come up with the same separation of the two clusters without using **that** binary variable (I could use internal knowledge of the two sub-populations statistics such as averages, stds etc .. ) in an unsupervised manner. I tried using Gaussian Mixture Models with EM, however, it didn't perform well on this dataset. Clearly what seems normal on logscale it isn't in reality :P I am looking for a package in python, R or Matlab that would be able to assign individual points to those two clusters.
[ -0.0015147826634347439, 0.008420189842581749, -0.0010346539784222841, 0.011839572340250015, 0.0030423866119235754, -0.016704265028238297, 0.005214208271354437, 0.010171191766858101, -0.012050259858369827, -0.009100065566599369, -0.014738007448613644, 0.001451600342988968, -0.0067011318169534...
[ 0.3888483941555023, -0.109409861266613, 0.000810210476629436, 0.17702515423297882, -0.24153520166873932, 0.6107326149940491, 0.38847991824150085, -0.048296015709638596, -0.35182568430900574, -0.5217345356941223, -0.1837453693151474, 0.03082369640469551, -0.15488821268081665, 0.257987737655...
I have trapped Odahviing in Dragonsreach, and it seems that I must free him so that I can ride him to the next dungeon. The problem is, every time I free him, everyone else in the room starts attacking him. Since the Jarl (who is invincible) is in the room, and the dragon itself is invincible, this enters a neverending cycle of one of the two of them dying, crouching for a moment to recover, and resuming battle. I can't talk to Odahviing while he is fighting. What am I supposed to do?
[ 0.011582085862755775, 0.037857383489608765, -0.012765444815158844, -0.0008375991601496935, -0.012430770322680473, -0.00890428014099598, 0.008091820403933525, 0.005071282386779785, -0.014070448465645313, -0.025100573897361755, -0.02403351664543152, 0.016237331554293633, -0.010488140396773815,...
[ -0.1126960813999176, 0.2607342004776001, 0.6229023933410645, 0.19301095604896545, -0.3323855698108673, -0.25274989008903503, 0.37119024991989136, -0.3563251197338104, -0.6351687908172607, -0.21131600439548492, -0.1759316325187683, 0.12334063649177551, -0.36562255024909973, 0.25916206836700...
I have a folder with a few files in it, and some space on a web server. I want to do a bi-directional sync between the local folder and the remote one in linux, just on modification time. How do I do this? btw I cannot install anything on the server, for all intents and purposes it is just space. Note: I already have rsa key-pairs set up, so that it can happen silently.
[ 0.00354784051887691, 0.0069813174195587635, -0.0016885933000594378, 0.016305027529597282, 0.006846355274319649, -0.01472674123942852, 0.007496150676161051, 0.006017574109137058, -0.02546633593738079, -0.022888092324137688, -0.0112381000071764, 0.018743403255939484, -0.013361888937652111, 0...
[ 0.8301122784614563, 0.24846701323986053, 0.48129352927207947, -0.14022056758403778, 0.35850778222084045, -0.24872156977653503, 0.5468873977661133, 0.12293728440999985, -0.22653728723526, -0.43564289808273315, 0.012908320873975754, 0.3004019558429718, -0.32723984122276306, -0.00901230890303...
I'm using LyX to write my thesis I've encountering this weird problem. When in-paragraph citations are formatted as English in my Danish document (which I do to get the spell checker working for English, this weird break is inserted: ![enter image description here](http://i.stack.imgur.com/cpy7b.png) As opposed to the citatioin not being formatted in English: ![enter image description here](http://i.stack.imgur.com/LlrD5.png) Curiously the bug doesn't happen if I change the language of the citation to German - only when it's English. Any suggestions as to why this could be? **EDIT:** I've managed to produce this minimal example: \documentclass[english,danish]{scrbook} \usepackage{babel} \usepackage[unicode=true,pdfusetitle, bookmarks=true,bookmarksnumbered=true,bookmarksopen=true,bookmarksopenlevel=1, breaklinks=false,pdfborder={0 0 0},backref=false,colorlinks=false] {hyperref} \makeatletter \@ifundefined{date}{}{\date{}} \AtBeginDocument{\renewcommand{\ref}[1]{\mbox{\autoref{#1}}}} \addto\extrasenglish{% \renewcommand*{\equationautorefname}[1]{} \renewcommand{\sectionautorefname}{sec.\negthinspace} \renewcommand{\subsectionautorefname}{sec.\negthinspace} \renewcommand{\subsubsectionautorefname}{sec.\negthinspace} \renewcommand{\figureautorefname}{Fig.\negthinspace} \renewcommand{\tableautorefname}{Tab.\negthinspace} } \begin{document} Kuhn udvider metaforen i sin karakteristik af udviklingen af paradigmer. Om videnskabelige nyopdagelse hedder det: \foreignlanguage{english}{``Produced inadvertedly by a game played under one set of rules, their assimilation requires the elaboration of another set.''} Videnskabelige nyopdagelse, der i første omgang forsøges bortforklaret. \end{document} The problem is clearly this from the LyX thesis template: \addto\extrasenglish{% \renewcommand*{\equationautorefname}[1]{} \renewcommand{\sectionautorefname}{sec.\negthinspace} \renewcommand{\subsectionautorefname}{sec.\negthinspace} \renewcommand{\subsubsectionautorefname}{sec.\negthinspace} \renewcommand{\figureautorefname}{Fig.\negthinspace} \renewcommand{\tableautorefname}{Tab.\negthinspace} } What does that code do? What could go wrong in my working document if I delete it?
[ 0.006260086316615343, 0.009740972891449928, -0.011051025241613388, 0.03110693022608757, -0.0042773643508553505, -0.0013624136336147785, 0.006409589201211929, 0.003979603759944439, -0.013810480944812298, 0.00018955208361148834, -0.007242000661790371, -0.0009599181357771158, -0.009057918563485...
[ 0.2759539484977722, 0.18804417550563812, 0.5874804258346558, 0.11039707064628601, -0.012869726866483688, -0.4409928619861603, 0.10075988620519638, 0.4257163107395172, -0.19830265641212463, -0.32921460270881653, 0.032632868736982346, 0.1801171898841858, 0.10076060891151428, 0.01924508437514...
I was using my phone (Galaxy S3 SGH-T999), not doing anything intensive, just web browsing on Chrome, then all of a sudden my phone restarts into Safe Mode. I opened the app drawer and noticed most (possibly all, not sure) of my downloaded apps were gone. My notifications bar was informing me that some apps were being re-downloaded. Not sure of the reason, I just restarted the phone and it booted back up normally. However strange things have been happening, like: * Apps re-downloading or updating. * Facebook app authorization failing * All my contacts were gone. I use SmoothSync for Cloud Contacts. Opening the app asked me to enter my password. My contacts came back. * My iCloud calendar was gone, again because SmoothSync logged out. Logging back in was no problem. Nothing really bad has happened, but it just seems strange and random to me. Why would Safe mode be activated out of nowhere?
[ -0.004647943191230297, 0.006353987380862236, 0.01692671701312065, 0.020860398188233376, -0.004076982382684946, -0.026368197053670883, 0.004948104731738567, 0.014916997402906418, -0.009522668085992336, -0.0015282141976058483, -0.007465658709406853, 0.007904495112597942, 0.018440034240484238, ...
[ 0.16025720536708832, 0.17488403618335724, 0.597490668296814, 0.014651939272880554, 0.12693904340267181, -0.49856916069984436, 0.8253973722457886, 0.20865808427333832, -0.40326014161109924, -0.38492146134376526, -0.11859165877103806, 0.7280529737472534, -0.2538693845272064, -0.0893045365810...
I'm planning to buy World of Warcraft but i just wanted to be clear regarding the entire process. I've made a BattleNet account and it shows the current offer for all the expansions upto MoP for $59.99. * Does this include in game time ? * After i purchase the expansions, how do i get the actual game ? Do i have to download the entire game ? If yes, then how big is the download ? Are there any ways i can get the **game and the subscription** without having to download the entire game (since my bandwidth is quite limited) i.e **Discs**
[ -0.023006217554211617, -0.0012040361762046814, -0.012176325544714928, 0.005234994925558567, 0.03225775808095932, 0.0011668752413243055, 0.005462207831442356, 0.0019744185265153646, -0.0158560611307621, 0.010052233003079891, -0.004128643311560154, 0.009422017261385918, -0.0008867067517712712,...
[ 0.5433030724525452, 0.13749945163726807, 0.5694023370742798, 0.19843393564224243, 0.10961832851171494, 0.13575953245162964, -0.12886956334114075, 0.0107651362195611, -0.5371477603912354, -0.4642380177974701, -0.019592417404055595, 0.6405264139175415, 0.41807821393013, 0.16289810836315155, ...
I am trying to find the nearest centerline segment to an address point. I was hoping that I would be able to use the ShortestLineTo method, but it is not working. I have tested with Microsoft's examples and they work. I am unable to get the method to work. Here is Microsoft's Example: DECLARE @Shape1 geometry = 'POLYGON ((-20 -30, -3 -26, 14 -28, 20 -40, -20 -30))' DECLARE @Shape2 geometry = 'POLYGON ((-18 -20, 0 -10, 4 -12, 10 -20, 2 -22, -18 -20))' SELECT @Shape1 UNION ALL SELECT @Shape2 UNION ALL SELECT @Shape1.ShortestLineTo(@Shape2).STBuffer(.25) GO My Sample: Select l.ADDRESSPOINT_TAX.shape from l.ADDRESSPOINT_TAX where OBJECTID = '1' union all select l.CADCENTERLINE.SHAPE from l.CADCENTERLINE where OBJECTID = '1' union all select l.ADDRESSPOINT_TAX.Shape.ShortestLineTo(l.CADCENTERLINE.SHAPE) This is the error that I am getting back. Remote function reference 'l.ADDRESSPOINT_TAX.Shape.ShortestLineTo' is not allowed, and the column name 'l' could not be found or is ambiguous. Msg 4121, Level 16, State 1, Line 1 Cannot find either column "ADDRESSPOINT_TAX" or the user-defined function or aggregate "ADDRESSPOINT_TAX.Shape.ShortestLineTo", or the name is ambiguous. Thanks!
[ -0.0075341942720115185, -0.0025929748080670834, -0.02909158729016781, 0.00960698351264, -0.015531862154603004, 0.006416122429072857, 0.007944599725306034, 0.016492493450641632, -0.012226604856550694, -0.016939302906394005, -0.011977087706327438, 0.005099685396999121, -0.019879143685102463, ...
[ -0.017141234129667282, -0.17620646953582764, 0.3504105508327484, -0.04859710857272148, -0.12450329959392548, 0.4116300344467163, 0.04660141095519066, 0.014471437782049179, 0.1426297128200531, -0.9941256642341614, -0.031790509819984436, -0.0061332108452916145, -0.10283809155225754, 0.195987...
I have some problem to interpret the result of MLE estimation : ![enter image description here](http://i.stack.imgur.com/cDUpN.png) Is it possible to get some advise about how to interpret it? the log likelihood function : $\sum^{n}_{i=1}\log\left( \phi\left( \frac{w-\mu}{\sigma}\right)\right) -\log(\sigma P(1)) +\log\left[ \frac{\left[ \Phi^{2}\left( \frac{w-\mu}{\sigma}\right)\right] }{2}-\delta \Phi\left( \frac{w-\mu}{\sigma}\right)+\left( \frac{a^{2}}{2} + b\right) \right]. $
[ 0.002957395277917385, 0.01001962274312973, -0.003936501685529947, 0.0012306254357099533, -0.025885354727506638, -0.0016325329197570682, 0.005431388504803181, 0.012516118586063385, -0.010441616177558899, 0.016305377706885338, -0.0018864359008148313, -0.0038077037315815687, -0.0050003668293356...
[ -0.0782482698559761, -0.12066435068845749, 0.7273580431938171, -0.0810830146074295, -0.06066986173391342, 0.3430747985839844, -0.10418515652418137, -0.3024587333202362, -0.22715872526168823, -0.6091522574424744, 0.1313776671886444, 0.6969155073165894, -0.13808247447013855, 0.20696166157722...
I received two .dxf for my project that represent my area of interest. One of them contains buildings, the other one terrain elevation. They both are 3D. I can convert them to shapefiles successfully using the Dxf2Shp plugin. They also display nicely compare to sample below). ![sample of my model area](http://i.stack.imgur.com/oA1M3.png) However, I loose the height information if I do so. The only field I get in the attribute table is "myid", that simply counts the polygons. Is there a way to convert the files to .shp creating a height field in the attribute tables, that takes the z values from the .dxf? This would already be fine, as the projection is metric. Best regards
[ -0.021313954144716263, 0.01114650908857584, -0.002857908606529236, 0.005850453395396471, -0.005471204407513142, -0.016561156138777733, 0.007424195762723684, 0.020643290132284164, -0.01611463539302349, -0.005013748072087765, -0.003641245188191533, 0.013685125857591629, -0.008263730444014072, ...
[ 0.37622371315956116, 0.3817209005355835, 0.46692052483558655, 0.028850922361016273, -0.1206621304154396, 0.4123474657535553, -0.011752936989068985, -0.3012748658657074, 0.03731703385710716, -0.6886472105979919, 0.07826051115989685, 0.323653906583786, 0.2324817031621933, 0.1312253177165985,...
Say we are trying to predict occurrences of lightning bugs based only on the dates on which people reported seeing them. So climate change, weather, and other influencing factors are imaginary. So there are $n$ recorded occurrences of lightning bugs. Since many people have contributed and the recording is not standardized, likely only the days of the occurrences - and not the number - should be considered. If every day on which a lightning bug was spotted was within July and August, and are densest around July 20, then it seems simplest to return the first and second quartile. But say there was a high density of occurrences in May, and then none again until July and August. The quartile method would be false. Where should I start in writing an algorithm to return multiple date ranges?. Could a probability be assigned to the result?.
[ 0.010438009165227413, 0.019891690462827682, -0.02075270563364029, 0.006133957765996456, 0.034974560141563416, 0.00681763980537653, 0.006272695492953062, 0.025060094892978668, -0.01463068276643753, -0.01976649835705757, 0.01120341382920742, 0.025177761912345886, 0.008129173889756203, -0.004...
[ 0.6549699902534485, 0.10270952433347702, 0.16452448070049286, 0.1812676042318344, 0.07256312668323517, 0.1896258294582367, 0.5167543292045593, 0.36009907722473145, -0.751847505569458, -0.28072506189346313, 0.28302374482154846, -0.01432119868695736, -0.4469428062438965, 0.26625385880470276,...
I'm looking for a more efficient way to compute the nearest distance between features from 2 layers, in this example for each point the nearest distance to the closest polygon. Is there a more efficient way compared to below where I compute all polygon distances for each point? (Python 2.7.2, GDAL/OGR 1.9.1) nPts = pointsLayer.GetFeatureCount() nPolys = polysLayer.GetFeatureCount() for pt in range(0,nPts): minDist = 1000000 pointFeature = pointsLayer.GetFeature(pt) for poly in range(0,nPolys): polyFeature = polysLayer.GetFeature(poly) Dist = pointFeature.GetGeometryRef().Distance( polyFeature.GetGeometryRef() ) if (Dist < minDist): minDist = Dist print (pt,minDist)
[ 0.0011601769365370274, 0.012293323874473572, -0.019096393138170242, 0.008813608437776566, -0.02511109784245491, -0.015738535672426224, 0.007138138636946678, 0.006414216943085194, -0.015296912752091885, 0.023536697030067444, -0.003022489370778203, 0.01152707263827324, -0.010560423135757446, ...
[ 0.10751629620790482, -0.1838948130607605, 0.6284510493278503, 0.043375592678785324, -0.19775909185409546, 0.41297054290771484, 0.3615545630455017, -0.6358874440193176, -0.17551559209823608, -0.7793036103248596, -0.06345658004283905, 0.5090575814247131, 0.0854790061712265, -0.17697857320308...
There are two achievements related to '`Moustachio`' in Left 4 Dead 2. These are the descriptions: 1. **Gong Show** - Prove you are stronger than Moustachio. 2. **Stache Whacker** - Prove you are faster than Moustachio. Do you know how can I get them?
[ 0.0106544503942132, 0.007945177145302296, -0.00426699360832572, 0.00673530250787735, -0.002925340784713626, -0.006216926034539938, 0.010880431160330772, 0.0021843353752046824, -0.020230069756507874, 0.008035583421587944, -0.005544669460505247, 0.016669712960720062, -0.0023742178454995155, ...
[ 0.3553359806537628, 0.13462860882282257, 0.2664274275302887, 0.16462533175945282, -0.1789049357175827, 0.20358507335186005, 0.32793447375297546, -0.22154995799064636, -0.15799446403980255, -0.08967704325914383, 0.3114006519317627, 0.343403160572052, 0.35447829961776733, -0.1380104571580886...
I have a question related to this example sentence: > Preparation of Isometrics, bill of material of all lines and supporting > details of all lines below 2-1/2" except lines under IBR and get subsequent > approval. From the above: 1. What are the sizes of lines for preparation of isometrics? Is it for all lines or for lines below 2-1/2"? 2. Is getting subsequent approval for (a) all isometrics, bill of materials and also for the supporting details or (b) only supporting details?
[ 0.005632882006466389, -0.0010944829555228353, -0.013027796521782875, 0.034763261675834656, 0.0014071717159822583, 0.017227981239557266, 0.011190047487616539, 0.020598337054252625, -0.01798035204410553, -0.022082602605223656, -0.010100403800606728, 0.0074114422313869, -0.005560355260968208, ...
[ 0.3060384690761566, 0.23129670321941376, 0.27923154830932617, 0.09740682691335678, -0.0914590060710907, -0.08586087822914124, 0.025986213237047195, -0.22347469627857208, -0.27981963753700256, -0.543178915977478, -0.1272609829902649, 0.08439593762159348, 0.01869994029402733, 0.0659783482551...
I've a Samsung Galaxy Ace and recently I installed link2sd, created the partiotion and everything it needs. It works. But I'm having two issues. The first one: each time I turn on the phone I get a notification from link2sd, I open it and a message says the following (translated by me, so it is not the exact text): > Mount warning > > The 2nd partition wasn't mounted during the start, it was mounted after the > system finished starting. With that reason your linked apps aren't visible > for the system right now. To make them available again, please try "Quick > restart". The buttons are: "Quick restart" and "Cancel". After I make a quick restart all apps work good. The second problem its that the launcher (not the default one) can't load widget. At the beggining I thought this one was easy: I unlinked the launcher and the apps which made widgets. But nothing, still can't be loaded. _Note_ killink the launcher and starting it again doesn't solve the problem Please any help, thanks!
[ -0.0037328461185097694, -0.0028319954872131348, 0.0006848735501989722, 0.018177684396505356, 0.0206611305475235, -0.0010447376407682896, 0.007998576387763023, 0.03307059407234192, -0.013378020375967026, 0.008214633911848068, -0.029236406087875366, 0.010258054360747337, -0.007199594751000404,...
[ 0.03141805902123451, 0.09339231997728348, 0.7184206247329712, -0.14868327975273132, 0.13516521453857422, -0.3741697072982788, 0.5074049830436707, -0.3344952166080475, -0.47927406430244446, -0.6585351824760437, -0.023801129311323166, 0.4955958425998688, -0.37725645303726196, -0.002607968635...
As enemy bases are captured, they just seem to respawn right in the middle of friendly captured territory. This is looking like a perpetual cycle, and none of the three teams ever gets closer than to the very edge of winning the whole continent, only to lose lots of territory again. So is this by design? Is Planetside 2 eternally to be at war, with no winner ever? I was hoping it would be more like a 'match'. I admit that both are almost the same thing, but there is at least the closure of winning/losing sides this way.
[ -0.005535060539841652, 0.028483640402555466, 0.0003370695631019771, 0.012729160487651825, 0.025955546647310257, -0.007493418641388416, 0.006524680647999048, 0.013914606533944607, -0.014063932932913303, 0.017184121534228325, -0.017359534278512, 0.017610803246498108, -0.002978974487632513, 0...
[ 0.20366492867469788, -0.11858674883842468, 0.45655596256256104, 0.2123665064573288, -0.18510963022708893, -0.35358232259750366, 0.052582819014787674, 0.4988672435283661, -0.5748594403266907, -0.674355149269104, 0.06946197152137756, -0.1287698894739151, -0.17062607407569885, 0.5231411457061...
I have the following plot \documentclass{standalone} \usepackage{pgfplots} \pgfplotsset{compat=newest} \begin{document} \begin{tikzpicture} \begin{axis}[ xlabel=$x$,ylabel=$\sin x$, axis x line*=box] \addplot[blue,mark=none, domain=-10:0,samples=40] {sin(deg(x))}; \end{axis} \end{tikzpicture} \end{document} ![enter image description here](http://i.stack.imgur.com/KX2Ei.png) But i want the x axis on top: \documentclass{standalone} \usepackage{pgfplots} \pgfplotsset{width=6cm,compat=newest} \begin{document} \begin{tikzpicture} \begin{axis}[ xlabel=$x$,ylabel=$\sin x$, axis x line*=top] \addplot[blue,mark=none, domain=-10:0,samples=40 ] {sin(deg(x))}; \end{axis} \end{tikzpicture} \end{document} ![enter image description here](http://i.stack.imgur.com/FcInH.png) How do i get the line in the bottom back? ![enter image description here](http://i.stack.imgur.com/smmSO.png)
[ -0.0052059548906981945, 0.012387151829898357, 0.0029046889394521713, 0.011259668506681919, 0.015172285959124565, 0.00037147931288927794, 0.007705464959144592, 0.016664806753396988, -0.01110593881458044, -0.007053153123706579, -0.00518784299492836, -0.003130061086267233, -0.004034216515719891...
[ 0.13570672273635864, -0.12036504596471786, 0.6881375908851624, -0.07010380923748016, 0.10478705167770386, 0.27911221981048584, 0.0735543966293335, -0.3412971794605255, 0.1761954128742218, -0.6195241808891296, 0.26064446568489075, -0.08322208374738693, -0.06348606944084167, 0.39071622490882...
I have data that involves 2 groups (equal sample size in each) and data for each group over 3 time points (they are actually 3 different monetary reward conditions). I want to investigate within group differences. All time/condition points are important, I don't have a "control" time point. Any opinions on the following: 1. If I am comfortable that sphericity is assumed by Mauchly's Test not being violated, would a multivariate Lambda F test statistic be more appropriate or a sphericity assumed estimate (within-condition estimate from SPSS)? 2. If I choose to go with the understanding that despite a test result saying it is OK, assuming sphericity may still be over-confident, any thoughts on potentially reporting ALL Greenhouse-Geiser estimates regardless of Mauchly's or Lambda? I have read that this may reduce the chance of a Type-1 error without having to assume sphericity or equality of covariance matrices. Perhaps too overcautious? Or, is potentially adjusting df's more invasive than assuming sphericity?
[ -0.010236164554953575, 0.017947888001799583, -0.02052663452923298, 0.019310932606458664, 0.04729597270488739, 0.021601028740406036, 0.0097922682762146, 0.0008787792176008224, -0.01039014384150505, 0.008561758324503899, -0.011702671647071838, 0.021675964817404747, -0.027268214151263237, 0.0...
[ 0.1539265513420105, -0.08777937293052673, -0.035009268671274185, 0.11783196032047272, -0.0007405627984553576, 0.36616984009742737, 0.08321438729763031, -0.5954021215438843, -0.11206687986850739, -0.4378840923309326, 0.17480172216892242, 0.46084722876548767, -0.3512577414512634, 0.200298830...
I'm developing an ASP.NET webforms application with pages which displays information based on the authorization level of the authenticated user (very typical). I will write a simple example of the issue I'm facing: Let's pose I want to have a page with 4 buttons: 1- "OnlyForAdmins" button. 2- "OnlyForManagersAndAdmins" button 3- "Guest" button 4- "Authenticated" button Depending on the permission level of the visitor, one or several buttons must be displayed. Right now in code behind buttons are handled in the following way: public void Page_Load(object sender, EventArgs e) { if (HttpContext.Current.User.Identity.IsAuthenticated) Athenticated.Visible = true; if (HttpContext.Current.User.IsInRole("Admins")) OnlyForAdmins.Visible = true; if (HttpContext.Current.User.IsInRole("Manager") || HttpContext.Current.User.IsInRole("Admins")) OnlyForManagersAndAdmin.Visible = true; } While this works, honestly I think that this approach is rather inflexible, having to add more "ifs" for every new role I introduce. I wonder if there exist approaches to handle this kind of problems. I'd be grateful if someone can enlighten me. Thank you.
[ -0.012830599211156368, 0.020201610401272774, -0.01075233705341816, 0.015459935180842876, -0.029564358294010162, -0.0073276483453810215, 0.008674818091094494, -0.009468505159020424, -0.010915419086813927, 0.0054475488141179085, -0.010580761358141899, 0.0026823831722140312, 0.00669550802558660...
[ 0.10537662357091904, -0.04658955708146095, 0.790636420249939, 0.01426190696656704, 0.006839561276137829, -0.20958790183067322, 0.3456633388996124, -0.1482650190591812, -0.5048962831497192, -0.5083616375923157, 0.005326203070580959, 0.3678869605064392, -0.24117934703826904, 0.16749045252799...
I have been trying many applications for development, and am trying to find a good one with Auto-completion and Syntax highlighting. I know many offer Syntax Highlighting, but I am unable to really find any at all that have Auto-completion. My main focus if Web development, but I am also interested in Java development for android applications. So far, I know of a couple but they dont necessarily fit my needs right; Terminal IDE - This is really amazing. Gives you pretty much a full Bash terminal environment, with about everything you could need for Android Development and Web Development. with vim, javac, apkbuilder, and other useful stuff like htop, midnight-commander, etc. Its really unbelievable, but its a full 105Mb when unpacked. Which is pretty crazy. It takes up a ton of space, and being that I have a tiny amount of space on my internal memory, I cant always have it on my phone as its just too much. Its nice for android development, overkill for web development. Touchcode - I checked this out for a while, and its alright. but it doesn't seem to start auto-completion on 80% of the tags I type. I am not sure why. Otherwise it would be great. AIDE - (I believe it has Auto-completion, but I cant quite recall for sure.) VI iMproved Touch - its really great, I would love it, but I cant seem to get it to auto-complete during development. (Am I missing something here? Is there Syntax highlighting and Auto-completion that I just don't know about, or a certain keyboard combination to initialize it?) So is there any others that I am missing here? I am sure there have to be a couple of people here how use their android for a bit of web development here and there. Plus it is so hard to find IDE applications because when you type "HTML" all the results are mainly source viewers, "IDE" give AIDE, and a bunch of games, same with "Development". So the search function doesn't really help much on Google Play. Plus the most useful suggestions are always from others who have experienced something themselves. Any suggestions guys? Or an explanation on how to enable those features in Vi iMproved Touch? Thanks for your time guys. Take care everyone.
[ 0.01092129573225975, -0.003988851793110371, -0.008193117566406727, 0.0068571437150239944, -0.02028101310133934, 0.0032395690213888884, 0.007276457268744707, 0.03268381953239441, -0.017419438809156418, 0.005960797891020775, -0.02126249298453331, 0.013346857391297817, 0.006766977719962597, 0...
[ 0.28763774037361145, 0.34117892384529114, 0.2689884603023529, 0.02550940215587616, -0.12037575989961624, -0.09125518798828125, 0.20160028338432312, 0.15228129923343658, -0.03259133920073509, -0.9475608468055725, 0.179948091506958, 0.9547064900398254, 0.06637082993984222, -0.098055988550186...
I have a graph where I show number of boxes that a machine can service per minute. The data may not always give a whole number (for example, a machine may be able to service 2.3 boxes / per minute). I am wondering how to label the axes on this graph. I have observed that _number_ tends to be used for integer quantities (for instance, _number of people_ , or *number of potatoes), whereas _amount_ is used for uncountable quantities, that often have fractions (such as _1.5 gallons of paint_ ). Which should I use to label my graph, _number of boxes_ or _amount of boxes_?
[ 0.015573802404105663, 0.013455469161272049, -0.009309506043791771, 0.021243685856461525, -0.024833347648382187, -0.022816285490989685, 0.007510248105973005, 0.0013255737721920013, -0.01680736429989338, 0.018194470554590225, -0.00804985873401165, -0.0024659321643412113, -0.0056190029717981815...
[ 0.7307085394859314, 0.22819571197032928, 0.046131815761327744, -0.14963465929031372, -0.14534297585487366, 0.6758081912994385, 0.04683004319667816, -0.18596519529819489, -0.6412011981010437, -0.9366868138313293, 0.3465004861354828, 0.10079868137836456, -0.08277367800474167, 0.4073128402233...
First off: I don't want to use bibtex for this; I want to typeset my bibliography myself. Looking at the `source2e.pdf` document I found out about redefining `\@cite` and `\@biblabel`, and got my citation style pretty much set up with this. The problem which remains now is this: I'd like to allow for two different ways of citing a person. (This seems to be called “harvard style”.) One would be in the body of the text: > … as John Doe (1900) demonstrated … while the other one would be parenthesized: > … has already been demonstrated (John Doe 1900). Now my problem is the different handling of the year, which has to be parenthesized in one form but not in the other. Of course, I'd need two different commands for the different citations. But so far, every attempt at splitting up the citation label has failed, at least if I want to also use `hyperref`. Without that package, the following seems to satisfy my requirements: \documentclass{article} % \usepackage{hyperref} \makeatletter \def\@cite#1#2{ ({#1\if@tempswa , #2\fi})} \def\@biblabel#1{} \def\@cite@ofmt#1{\edef\my@tmp{#1}\expandafter\my@split\my@tmp} \def\my@split#1(#2){\hbox{#1#2}} \def\my@tcite#1#2{{#1\if@tempswa , #2\fi}} \def\my@tsplit#1(#2){\hbox{#1(#2)}} \newcommand\tcite[1]{{\let\@cite\my@tcite\let\my@split\my@tsplit\cite{#1}}} \makeatother \begin{document} \ldots has already been demonstrated \cite{JohnDoe}.\par \ldots as \tcite{JohnDoe} demonstrated \ldots\par \ldots has already been demonstrated \cite{JohnDoe}.\par \begin{thebibliography}{1} \bibitem[John Doe (1900)]{JohnDoe} John Doe, The famous book, 1900. \end{thebibliography} \end{document} But with `hyperref` enabled, I get this error message: ! Undefined control sequence. \hyper@@link ->\let \Hy@reserved@a \relax \@ifnextchar [{\hyper@link@ }{\hyp... Looking at the source code of `hyperref.sty`, I see that it does attempt to support compatibility with the `harvard` package. So I adjusted my example to this: \documentclass{article} \usepackage{harvard} \usepackage{hyperref} \begin{document} \ldots has already been demonstrated \cite{JohnDoe}.\par \ldots as \citeasnoun{JohnDoe} demonstrated \ldots\par \ldots has already been demonstrated \cite{JohnDoe}.\par \begin{thebibliography}{1} \harvarditem{John Doe}{1900}{JohnDoe} John Doe, The famous book, 1900.\par \end{thebibliography} \end{document} But I keep getting LaTeX Warning: Label(s) may have changed. Rerun to get cross-references right. no matter how many times I process the document. And I don't get any hyperlinks. **How can I get harvard style citations with a manually set bibliography and with working hyperlinks for my citations?** I'm using texlive 2013, which includes `hyperref.sty` v6.83m from 2012/11/06 and `harvard.sty` version 2.0.5 apparently from 1994.
[ 0.007474937941879034, 0.002727675950154662, 0.0022124380338937044, 0.0070243095979094505, 0.025247778743505478, -0.008031224831938744, 0.008612575940787792, -0.00032928655855357647, -0.012606514617800713, -0.017480570822954178, -0.0033364356495440006, -0.003126736730337143, 0.003510695183649...
[ 0.3137058913707733, 0.2618471086025238, 0.11344010382890701, 0.08459045737981796, -0.2511088252067566, -0.0427352637052536, 0.050998248159885406, -0.01642424613237381, 0.013406760059297085, -0.5081086158752441, 0.36417797207832336, 0.4153560400009155, -0.31548774242401123, -0.0602317005395...
My storage is full and there's so many nice things to pick up. Is there a safe place I can drop things off at where they will be accessable later on? What I'm looking for is something like Diablo 2's "stash"...
[ 0.000176362314959988, 0.03835909068584442, 0.011617932468652725, -0.0028948758263140917, 0.027810553088784218, 0.012236027978360653, 0.009304770268499851, 0.02073631063103676, -0.03622306510806084, 0.03701642528176308, -0.026200443506240845, 0.02526911348104477, -0.013217009603977203, 0.02...
[ 0.3084316849708557, -0.09415004402399063, -0.3914010226726532, 0.30931031703948975, 0.21656079590320587, -0.1447962373495102, -0.14495521783828735, 0.13364534080028534, -0.08336271345615387, -0.0704042986035347, 0.3267068564891815, 0.3145580589771271, 0.2038697600364685, 0.2408694475889206...
_Updated at the bottom_ I'm using Centos as a router, and just replaced the routing system at work. I've got 2 routers to send data around. For the most part it's working. On one router I have 2 network ranges that are on the same nic (eth1). We have a server on one range, and anyone in the other cannot SSH to it. But, it can be pinged. Here's a quick layout: Internet Link: eth0 - 192.168.144.50 (link between routers) eth1 - 10.10.10.254 - 255.255.254.0 (10.10.10.x and 10.10.11.x ranges) eth2 - internet Router 2: eth0 - 192.168.144.24 (link between routers) eth1 - 98.0.0.1 & 10.98.0.1 ranges (yes, I know 98.0.0.0 is not for private use, this is a legacy setup that we haven't been able to remove yet) Here is the **ip route** printout: 98.0.0.0/24 dev eth1 proto kernel scope link src 98.0.0.1 192.168.144.0/24 dev eth0 proto kernel scope link src 192.168.144.24 10.98.0.0/24 dev eth1 proto kernel scope link src 10.98.0.1 default via 192.168.144.50 dev eth0 The server is on 98.0.0.6. I have a laptop running on 10.98.0.165. I get a response from the command: ping 98.0.0.6. But, when I try to ssh to it, I get nothing. Through doing some tcpdump commands on my routers, I'm seeing that the packets for the ssh command are going from the laptop, to the internet router (192.168.144.50) and back to the correct router. My iptables list is empty. You can see my ip route output above. Any ideas? I've been staring at this all day. The way I understand it, is that the router should send the packet back through eth1 for 98.0.0.6 to pick up. Thanks. Edit: Scratched out the bouncing part. I realized that I was seeing ssh traffic from my ssh session into the router, and attributed it to my ssh request. **Update** I separated my setup. Now 98.0.0.x is on eth1 and 10.98.0.x is on eth2. What I'm seeing isn't making much sense though. When I ping the 98.0.0.6 server from my 10.98.0.165 laptop, I get a reply. The route it is taking is odd. It goes through the default route to 192.168.144.50 (internet router) and back through to 98.0.0.6. (tcpdump on eth0 of the internet router found this) One tidbit. When I started the project, I was trying to do policy routing to make 2 internet providers work. Is there something related that could be doing this? Here's my updated layout for the second router eth0 - 192.168.144.24 (router link) eth1 - 98.0.0.1 / 24 eth2 - 10.98.0.1 /24 and **ip route show** 98.0.0.0/24 dev eth1 proto kernel scope link src 98.0.0.1 192.168.144.0/24 dev eth0 proto kernel scope link src 192.168.144.24 10.98.0.0/24 dev eth2 proto kernel scope link src 10.98.0.1 default via 192.168.144.50 dev eth0
[ 0.03927020728588104, 0.018346812576055527, -0.02848956175148487, 0.009673534892499447, -0.022523649036884308, -0.009481817483901978, 0.008855318650603294, -0.00836806558072567, -0.0151927899569273, -0.013389738276600838, -0.014272789470851421, 0.008999615907669067, -0.016494154930114746, 0...
[ 0.12337642163038254, 0.07231566309928894, 0.562204122543335, 0.11431924253702164, -0.15056182444095612, -0.11081928014755249, 0.3026881515979767, 0.17045298218727112, -0.365174800157547, -0.9614907503128052, -0.008299027569591999, 0.568670928478241, -0.1507747918367386, 0.01752778142690658...
I have a `tikz` drawing that I made within the `beamer` document class. I would like to know how to produce this drawing within the `article` document class, but preserve the sans-serif font and font size used in the default `beamer` theme. Here is the `LaTeX` code using the `beamer` document class: \documentclass[notes=show,beamer,usenames,dvipsnames]{beamer} \usepackage{tikz} \setbeamertemplate{navigation symbols}{} \setbeamersize{text margin left=1em,text margin right=1em} \begin{document} \begin{frame} \begin{tikzpicture}[yscale=3.5,xscale=4.5] \draw[->] (0,0) -- (2,0) node[below, text width=2cm] {Before Tax Income $z$}; \draw[->] (0,0) -- (0,2) node[below left, text width=1.8cm] {After Tax Income $z - T(z)$}; \begin{scope} \clip (0,0) rectangle (1,2); \draw (0,0) -- (1.5,1.5); \end{scope} \begin{scope} \clip (1,0) rectangle (2,2); \draw (0,0) -- (1.5,1.5); \end{scope} \draw[rotate around={45:(1,1)},red] (.6,1.1) node[left] {$L$} parabola bend (1,1) (1.4,1.1); \draw[shift={(.125,.125)},rotate around={45:(1,1)},cyan] (.6,1.1) node[left] {$M$} parabola bend (1,1) (1.4,1.1); \draw[shift={(.25,.25)},rotate around={45:(1,1)},ForestGreen] (.6,1.1) parabola bend (1,1) (1.4,1.1) node[right] {$H$}; \draw[dashed] (1,0) node[below] {\textcolor{red}{$z^{*}$}} -- (1,2); \draw[dashed] (1.25,0) node[below] {\textcolor{ForestGreen}{$z^{*}+\triangle z^{*}$}} -- (1.25,2); \path (.15,.1) node[right] {slope = $1-t$}; \end{tikzpicture} \end{frame} \end{document} And here is the drawing within the `article` document class: \documentclass{article} \usepackage[usenames,dvipsnames]{xcolor} \usepackage{tikz} \begin{document} \begin{tikzpicture}[yscale=3.5,xscale=4.5] \draw[->] (0,0) -- (2,0) node[below, text width=2cm] {Before Tax Income $z$}; \draw[->] (0,0) -- (0,2) node[below left, text width=1.8cm] {After Tax Income $z - T(z)$}; \begin{scope} \clip (0,0) rectangle (1,2); \draw (0,0) -- (1.5,1.5); \end{scope} \begin{scope} \clip (1,0) rectangle (2,2); \draw (0,0) -- (1.5,1.5); \end{scope} \draw[rotate around={45:(1,1)},red] (.6,1.1) node[left] {$L$} parabola bend (1,1) (1.4,1.1); \draw[shift={(.125,.125)},rotate around={45:(1,1)},cyan] (.6,1.1) node[left] {$M$} parabola bend (1,1) (1.4,1.1); \draw[shift={(.25,.25)},rotate around={45:(1,1)},ForestGreen] (.6,1.1) parabola bend (1,1) (1.4,1.1) node[right] {$H$}; \draw[dashed] (1,0) node[below] {\textcolor{red}{$z^{*}$}} -- (1,2); \draw[dashed] (1.25,0) node[below] {\textcolor{ForestGreen}{$z^{*}+\triangle z^{*}$}} -- (1.25,2); \path (.15,.1) node[right] {slope = $1-t$}; \end{tikzpicture} \end{document} I would like the latter to use the same font and font size as the former, including within math mode.
[ -0.0011650954838842154, -0.0007324310718104243, -0.0033784080296754837, 0.01761646941304207, 0.0037878286093473434, 0.00648709898814559, 0.010493533685803413, 0.016360007226467133, -0.016224734485149384, -0.031180813908576965, -0.01245264895260334, -0.0064282165840268135, 0.00721733458340168...
[ 0.6088793873786926, 0.2358991652727127, 0.87471604347229, -0.2287261039018631, 0.0042580426670610905, 0.0663645938038826, -0.030299672856926918, -0.2231883406639099, 0.32348349690437317, -0.7561946511268616, 0.11601530015468597, 0.44470110535621643, -0.4433026909828186, -0.0184069983661174...
In The Cave there are 2 cans of corn that you find during your journey: the first one at the Miner's puzzle and the second hidden behind a crate on the Island. I tried everything that came to my mind with them: blowing the miner's one with dynamite (it's easy, just leave the can where it stands and the miner will throw a stick at it), jumping with the island one in the volcano, getting drowned with them, and many other creative ways to try and get this achievement. What am I missing? What am I supposed to do?
[ 0.008193383924663067, -0.006628877017647028, -0.011047666892409325, -0.0003819529665634036, -0.02686004713177681, -0.005247864406555891, 0.005304248072206974, 0.0010029298719018698, -0.01454145647585392, 0.00815674290060997, -0.014863875694572926, -0.00021962705068290234, -0.0115026254206895...
[ 0.4556727409362793, 0.8097475171089172, -0.06639113277196884, 0.30806484818458557, -0.0008749537519179285, -0.06328971683979034, 0.6942588090896606, -0.0004876448947470635, -0.2117791771888733, -0.19133087992668152, 0.18849995732307434, 0.18533319234848022, -0.044758036732673645, 0.5021079...
I run an academic network for college students to communicate at particular universities and we run Google Adsense. The site pulls in a decent amount for a side project but our CTR is horrible <0.2% and our RPM is equally low. The problem lies in the fact that Google pegs us as an education site (which we are) but shows our users ads for U of Phoenix, Devry U and other for-profit universities. All of our users are students of the more higher-caliber institutions and therefore have no use for these ads. I've known about this problem for some time but I don't know what to do to show more relevant ads instead (i.e. Spring Break, school apparel, poker, sports, etc). What would be the best way to change this?
[ -0.030679339542984962, 0.0008723605424165726, -0.011752805672585964, 0.008101561106741428, -0.03528420627117157, -0.008678393438458443, 0.008316822350025177, 0.006343496963381767, -0.012584177777171135, 0.013091225177049637, -0.005426852032542229, 0.0034737680107355118, -0.003295495873317122...
[ 0.8427007794380188, 0.6468519568443298, 0.24490362405776978, 0.4579242467880249, 0.07091476023197174, -0.26994189620018005, 0.25997331738471985, 0.6490793228149414, 0.11241469532251358, -0.3877228796482086, 0.2922532558441162, 0.3739228844642639, 0.07046754658222198, 0.32154467701911926, ...
I would like to know if a particular folder is present or not. I used the following command find /mnt/md0/ -maxdepth 1 -name 'dcn'||'DCN' I want to know if folder name is DCN or dcn . How would I do this ?
[ -0.011012573726475239, 0.017698753625154495, -0.011766244657337666, 0.005704761482775211, -0.008076325990259647, -0.03131198137998581, 0.011901339516043663, -0.011934353038668633, -0.02017507143318653, -0.009662442840635777, 0.004023431800305843, -0.0013063621008768678, 0.017926450818777084,...
[ 0.2534150779247284, -0.17965592443943024, 0.22963032126426697, 0.16394855082035065, 0.2407914400100708, -0.18960458040237427, -0.2642359733581543, 0.34226712584495544, -0.35352179408073425, -0.5408398509025574, 0.06112783029675484, 0.548467218875885, -0.42265084385871887, 0.242306739091873...
I'm trying to use DDP (Data Driven Pages) and am running into a problem. I'm using arcpy.SelectLayerByAttribute_management to select a series of records, and then trying to use the ddp.exportToPDF(pdfddp, "SELECTED") but keep running into an error (below). It shouldn't be the select by attribute function as I run it in ArcMap and it runs correctly and I can see the selected features highlighted. Also, I can successfully use the ddp.exportToPDF(pdfddp, "CURRENT") but what I need is to be able to have a user select a series of features and have those be passed. mxd = arcpy.mapping.MapDocument("C:\\temp\\Graffiti_DM4.mxd") df = arcpy.mapping.ListDataFrames(mxd, "*")[0] pdfddp="C:\\temp\\pdfddp1.pdf" arcpy.SelectLayerByAttribute_management("C:\\temp\\test.lyr","NEW_SELECTION", '"Incident_Z" =' + "'10002'") ddp.exportToPDF(pdfddp, "SELECTED") And here is the error Runtime error Traceback (most recent call last): File "<string>", line 1, in <module> File "c:\program files\arcgis\desktop10.1\arcpy\arcpy\utils.py", line 181, in fn_ return fn(*args, **kw) File "c:\program files\arcgis\desktop10.1\arcpy\arcpy\_mapping.py", line 446, in exportToPDF return convertArcObjectToPythonObject(self._arc_object.exportToPDF(*gp_fixargs((out_pdf, page_range_type, page_range_string, multiple_files, resolution, image_quality, colorspace, compress_vectors, image_compression, picture_symbol, convert_markers, embed_fonts, layers_attributes, georef_info, jpeg_compression_quality, show_selection_symbology), True))) AttributeError: PageLayoutObject: Error in exporting pages As always, any help is EXTREMELY appreciated!
[ -0.010036533698439598, 0.002224463503807783, -0.011367907747626305, 0.013473652303218842, -0.01245882548391819, -0.0007073725573718548, 0.008491741493344307, 0.026899095624685287, -0.014195827767252922, -0.030715126544237137, -0.007695968262851238, 0.012991376221179962, -0.019483957439661026...
[ 0.04391694441437721, -0.20180603861808777, 0.5952414870262146, -0.06583298742771149, -0.06445327401161194, -0.22158274054527283, 0.2424011528491974, -0.3260759115219116, -0.3093959093093872, -0.8576511144638062, -0.06801164895296097, 0.6750602722167969, -0.3544580638408661, 0.0799295306205...
My blog posts have have multiple items from my custom taxonomy attached to them. For the sake of this post lets just use Tags taxonomy... So a post may contain 4 tags, is there a way I can select one tag as the primary tag for the post?
[ 0.02011539414525032, 0.031005537137389183, -0.010077564977109432, 0.03755119442939758, 0.03704304248094559, 0.014419740997254848, 0.012171762995421886, 0.04191906750202179, -0.030935220420360565, -0.027237968519330025, -0.009164758026599884, 0.01563866063952446, -0.0032558580860495567, 0.0...
[ 0.7689348459243774, 0.07634884119033813, 0.42323368787765503, 0.29903143644332886, -0.42956188321113586, 0.027250850573182106, 0.25320637226104736, 0.0034286899026483297, -0.39453333616256714, -0.4426555633544922, 0.35175320506095886, 0.020168783143162727, -0.3371436595916748, 0.3437137603...
"Show your **support for** the XYZ (Organization name) this season" or "Show your **support of** XYZ (Organization name) this year"? I have seen _support of_ and _support for_ both being used. Is there a rule? "Show your **support of** our clients" or "show your **support for** our clients"? ( _Support_ to be used in the context of making a donation)
[ -0.0027627081144601107, 0.013841347768902779, -0.0077431704849004745, 0.040334418416023254, -0.009257176890969276, 0.006552614737302065, 0.01156909205019474, 0.043935731053352356, -0.022560395300388336, -0.027651816606521606, -0.009105856530368328, 0.009451556950807571, 0.002030260395258665,...
[ 0.38966354727745056, -0.10846048593521118, 0.5241054892539978, -0.02342691645026207, -0.1969183087348938, -0.08415491878986359, 0.03635888919234276, 0.08738179504871368, -0.6390510201454163, -0.3389033377170563, -0.040465135127305984, 0.1371864527463913, -0.03595917299389839, 0.02297917194...
I was watching a documentary last night on the first planet discovered outside our solar system. The first one apparently is a giant gas planet that orbits very close to its sun over a very fast period (~1 week per orbit), and it's so hot all metal elements vaporise and form parts of it's atmosphere (so the current theory states anyway). Out of curiosity, when they estimate the surface heat of the planet (they give big numbers), did they take into account reflection of heat from it's metallic atmosphere, and internal reflection from heat bouncing off the surface and it's atmosphere (like global warming) when estimating surface temperature of the planet?
[ -0.012880753725767136, 0.006633423268795013, -0.009562205523252487, -0.0022272702772170305, -0.00020579900592565536, -0.0136099923402071, 0.006279915105551481, -0.013600367121398449, -0.012337274849414825, -0.014232474379241467, 0.005447635892778635, 0.013157324865460396, 0.01023432519286871...
[ 0.4995938539505005, 0.3106631338596344, 0.2871321439743042, 0.27942657470703125, 0.19139891862869263, -0.5070319175720215, -0.05028771236538887, 0.30312132835388184, -0.3234170973300934, -0.23234739899635315, 0.033521633595228195, 0.042216043919324875, -0.0307341106235981, 0.25543704628944...
I used Python to make the plot shown below (saved as png). It should have a common colorbar and all data is read from a `h5` file, but I thought saving it to a txt file or something using a Python script to read in for the TikZ figure. There are 5 plots and 1 colorbar: The first plots only has one variable (with values 0,25,33,50,66,75 %), each colored square represents the corresponding output, which can be represented by an array of length 6. The four other plots have two variables (with values [0,25,33,50,66,75 %] and [0,1,2,3,4,5,6,7,8]) and can thus be represented by a 6x9 matrix. I would like to make a TikZ figure just like the PNG, but off course with non- overlapping text and maybe a title for each subplot and a better outline of the colorbar (the png was quick and dirty to be able to quickly view the output). Can anyone help me out here? (If having the values is usefull, please tell me in what format it would be easiest and I'll upload a `txt` file asap.) Edit: Adding the data in `data.txt` data.txt Edit2: I forgot to mention that I would like to be able to use \figureheight and \figurewidth, defined in the `tex` file, to adjust the dimensions (like I did in the figure from another question). ![enter image description here](http://i.stack.imgur.com/gGD3b.png)
[ 0.011754038743674755, 0.013422545045614243, -0.012194477021694183, 0.004505590535700321, -0.004897439852356911, 0.006159661337733269, 0.008003990165889263, 0.00030930526554584503, -0.02000552788376808, -0.011382773518562317, -0.013726773671805859, -0.0023453105241060257, -0.00012310669990256...
[ 0.35748347640037537, -0.2309851050376892, 0.7069978713989258, 0.139510378241539, -0.0835738256573677, 0.6261184811592102, -0.1867094784975052, -0.3386504054069519, -0.03550170734524727, -0.4899216294288635, -0.05079995468258858, 0.30342766642570496, -0.09711993485689163, 0.2570786774158478...
I'm writing an article using the REVTeX documentclass, and use hyperref. The problem is that when I add URLs to my references, the linked titles aren't broken at all. Minimal example: .tex file: \documentclass[twocolumn]{revtex4-1} \usepackage{hyperref} \begin{document} \cite{reference, reference2} \bibliography{references} \end{document} references: @TECHREPORT{reference, AUTHOR = {Firstname Lastname}, TITLE = {Extremely long and complicated title of something I want to reference several times in my very interesting article}, INSTITUTION = {Institution}, YEAR = {2013}, URL = {http://url/} } @TECHREPORT{reference2, AUTHOR = {Firstname Lastname}, TITLE = {Regular reference}, INSTITUTION = {Institution}, YEAR = {2013}, FILE = {http://url/}, } Compile using `latex`, `bibtex`, `latex`, `latex`, `dvipdf`. The result:![enter image description here](http://i.stack.imgur.com/FGvza.png) Does anyone know of a way to remedy this?
[ 0.005515894386917353, 0.009358870796859264, -0.015054134652018547, 0.02214895375072956, 0.008795433677732944, -0.0011582598090171814, 0.008647969923913479, 0.0068897344172000885, -0.012268795631825924, -0.03569060564041138, 0.00024799141101539135, -0.006987966597080231, -0.004076078534126282...
[ 0.1909293383359909, 0.35415250062942505, 0.7733744978904724, 0.21272514760494232, -0.044519733637571335, -0.2600131630897522, 0.18434683978557587, -0.11651165038347244, -0.21714290976524353, -0.8132843375205994, -0.15546752512454987, 0.6173454523086548, -0.24355193972587585, 0.380041271448...
I was pretty good with algorithms and data structures once, a long long time ago. Since then, I programmed professionally, and then went to manage a small team, which totally shot my tech skills in this field back. I've decided I want to be a developer again, and work for Google. The thing is, I'm so out of practice, that if I were to be interviewed right now I would surely flunk out in 10 minutes. What training program would you recommend for me to get back into shape? I already started this weekend by going back to the absolute basics and implementing a few sort algorithms, linked list, and hash table. Next, I think I'll read through the entire course material on the other basic data structures and graph algorithms. I want to find a focused set of practical exercises I can do in a relatively short amount of time, to juggle the old brain cells. I know this stuff - I just need to remind myself that I know it.
[ -0.0023542409762740135, 0.013665149919688702, -0.005468055605888367, -0.010203852318227291, 0.0006380248814821243, 0.007527274079620838, 0.004227502271533012, -0.004932689480483532, -0.01344902254641056, -0.015960777178406715, 0.00308926310390234, 0.012349165044724941, 0.01582127809524536, ...
[ 0.9178881645202637, 0.2701430022716522, -0.13188010454177856, -0.002278103493154049, 0.3714393675327301, -0.11913455277681351, 0.4965493679046631, 0.3871043026447296, -0.0719529464840889, -0.653077244758606, 0.3785569965839386, 0.5794515013694763, 0.1726991981267929, -0.1433081328868866, ...
I just unlocked this skill, but can't fathom how to use it in conjunction with my other skills. Since the giant toad sits there doing damage to one enemy, it seems rather useless when large crowds of monsters attack me (when I'm playing solo). My first thought is to use something like Horrify or Mass Confusion to keep everything else away while using other attacks like Acid Rain, Haunt, Zombie Dogs, or Gargantuan. However, the benefit of using the Toad of Hugeness to incapacitate one enemy doesn't seem to make a huge difference if I'm using those other attacks anyway. Is there another skill that I can currently use with Toad of Hugeness or are there other combinations of skills that I will eventually unlock that go better with Toad of Hugeness?
[ -0.031827010214328766, -0.007698539644479752, -0.0038445349782705307, -0.0022551552392542362, 0.00020376103930175304, -0.002642089733853936, 0.009359519928693771, -0.018646210432052612, -0.016038937494158745, 0.008075334131717682, -0.009985832497477531, 0.00921186339110136, -0.01662585139274...
[ 0.0007258926634676754, -0.09395137429237366, -0.1849660873413086, 0.14633594453334808, -0.3473818898200989, -0.10261096805334091, 0.4348456561565399, 0.12204766273498535, -0.43175020813941956, -0.5800678730010986, 0.3019427955150604, 0.48018431663513184, 0.10847139358520508, -0.04608292505...
I write my thesis with the memoir class and want to include an Acknowledgment in the same style as the standard abstract from the memoir class. How can this be done?
[ 0.023539971560239792, 0.010468082502484322, -0.00539833027869463, 0.039553556591272354, -0.0333317406475544, 0.029728751629590988, 0.015577691607177258, -0.011747567914426327, -0.021777378395199776, 0.004901156760752201, 0.00495551573112607, 0.010687055997550488, -0.022529713809490204, 0.0...
[ 0.4811307489871979, 0.4773055911064148, 0.09246046841144562, 0.06388621777296066, 0.030336227267980576, -0.30132198333740234, 0.03784836083650589, 0.0771666169166565, 0.4083513915538788, -0.2814934253692627, 0.2735185921192169, 0.11581163853406906, 0.6597437262535095, 0.029949214309453964,...
I have an up-to-date installation of MacTeX (2012) on my machine, but a recent upgrade of my system software seems to have removed TeX from my path. TeX Live Utility recognizes my installation and confirms that it is current, as do the settings in System Preferences, but the path is missing. Is there a TeX utility that I can run to restore it, or do I need to re- donwnload and re-run the MacTeX installer?
[ 0.0016948000993579626, 0.00497440667822957, -0.0030270868446677923, -0.0007615627837367356, 0.023487918078899384, -0.0035514214541763067, 0.007397691253572702, -0.0027774013578891754, -0.017458349466323853, -0.0335572250187397, -0.004008538089692593, 0.009208098985254765, 0.01677422225475311...
[ 0.12244714051485062, 0.24948300421237946, 0.4860302209854126, -0.11676876991987228, 0.4246175289154053, 0.1340886503458023, 0.4140566289424896, 0.17769785225391388, -0.10462525486946106, -0.6821373105049133, 0.1578463613986969, 0.8750667572021484, -0.2403593361377716, 0.016888536512851715,...
So I have a feeder service. A train drops off goods (i.e. is ordered to transfer) and a truck picks these up and makes the short journey into town. The problem I seem to be having however is that whenever my truck makes the delivery it actually loses money. I'm assuming this is because the train station is slightly further away and the train gets overpaid for the transfer. Is there a way I can fix this? Short of demolishing and moving my train station?
[ -0.017512619495391846, 0.02003772184252739, -0.0017159917624667287, 0.014705456793308258, 0.005116465035825968, -0.023683706298470497, 0.00761054502800107, 0.00882537942379713, -0.015459604561328888, -0.016451865434646606, -0.02510947547852993, 0.02159401774406433, -0.005525535903871059, 0...
[ 0.6121290326118469, -0.3635340631008148, 0.11348534375429153, 0.14000645279884338, 0.06805985420942307, 0.44243311882019043, 0.2140365093946457, 0.34201061725616455, -0.3956299126148224, -0.4681173861026764, 0.2552463412284851, 0.0839080810546875, -0.35738107562065125, 0.321058452129364, ...
I have a site that will hold roughly 1000 posts/articles from an in-house magazine dealing with Australian naval history. The articles cover a range of subjects (such as Aviation, Intelligence, Ship Histories, Personal Histories and the like) and I've created categories to suit. The names of ships in an article is very important for research and cross- referencing but there are a LOT of ships. And, how does one distinguish between a ship that is the focus of an article and a ship that is merely mentioned in passing? I have uploaded about 300 articles so far. I've used the categories to identify the subject material, and I've created "tags" for the ship names (and other keywords). But tags dilute the significance of the ship covered by an article; e.g. if the article focuses on HMAS A, but also mentions HMAS B, C and D, then 4 ships are tagged but the relevance of the article to HMAS A is no greater than any of the other three ships. In addition, there's no way to create a list of articles by ship name or to cross-reference articles to the ship name in other material. So, I am pondering whether I should instead create categories for ships that are the focus of an article and use tags only for the names of the "passing reference" ships. This sounds good in theory but it could add, maybe, 300+ extra categories; and in many cases there might not be more than one article where that ship is the focus. Perhaps I am going about it the wrong way; might it be that I need a table devoted to ship names and create a meta link between the post and the table (in much the same way as, say, product names, work in an ecommerce transaction).
[ 0.007164468057453632, 0.011624099686741829, -0.0017423774115741253, 0.026566337794065475, -0.008928962051868439, -0.0002707722596824169, 0.00792645663022995, -0.004112901631742716, -0.01663847453892231, -0.016190050169825554, 0.01209255401045084, 0.019128870218992233, -0.005437690764665604, ...
[ 0.7341358661651611, 0.44173693656921387, -0.1416904330253601, 0.03655494004487991, -0.2862217128276825, 0.2505607306957245, 0.11172215640544891, 0.2902044355869293, -0.5393075942993164, -0.40152403712272644, 0.045604050159454346, 0.3075995147228241, 0.41136178374290466, 0.42560991644859314...
With the advent of weather in 1.5, lightning strikes during rainstorms can set fire to trees, frankenstein pigs to zombie pigmen, and perhaps most insidiously of all, super-charge Creepers with blue electricity. ![enter image description here](http://i.stack.imgur.com/P77br.png) The minecraft wiki states that a supercharged creeper has a "stronger" explosion, and can kill the player much more easily. While "stronger" is unquestionable in meaning for blast resistance and raw damage, I was wondering if the explosion radius is actually increased as well? How large is the explosion radius compared to that of a regular creeper? *Note: anecdotal evidence may not be sufficient alone, as it may be difficult to determine "bigger craters because of more blast power" and "bigger craters because the blast power covers a larger area". It could even be both at once.
[ -0.010282017290592194, 0.028268614783883095, 0.0019021817715838552, 0.004029756411910057, -0.00613760482519865, -0.008654965087771416, 0.008252477273344994, -0.010467254556715488, -0.015250890515744686, 0.0034603935200721025, -0.002320358995348215, 0.022921646013855934, 0.02208586595952511, ...
[ 0.0529521107673645, -0.18360635638237, 0.10813651978969574, 0.22325807809829712, -0.31964802742004395, -0.376485675573349, 0.6578685641288757, 0.16915492713451385, -0.4610574245452881, -0.3104645311832428, -0.22151541709899902, 0.0551954060792923, -0.26453810930252075, 0.4245757758617401, ...
I'm trying to format my thesis, and I copied the formatting stuff from a friend who finished her thesis last year. Unfortunately, she only had two chapters and then subsections. I have more chapters, and since they are roman numerals, the numbers are running into my titles. This is formatted (to the best of my knowledge) correctly, but I'd like to start the title text a little more to the right: \documentclass[12pt,fleqn, letterpaper]{report} \usepackage{indentfirst} \usepackage{setspace} \usepackage{titlesec} \usepackage[hidelinks]{hyperref} \oddsidemargin 0in \textwidth 6.5 in \topmargin 0in \headheight 0in \headsep 0in \textheight 8.6in \footskip 0.4in \makeatletter \renewcommand*\l@chapter[2]{% \ifnum \c@tocdepth >\m@ne \addpenalty{-\@highpenalty}% \vskip 1.0em \@plus\p@ \setlength\@tempdima{1.5em}% \begingroup \parindent \z@ \rightskip \@pnumwidth \parfillskip -\@pnumwidth \leavevmode \bfseries \advance\leftskip\@tempdima \hskip -\leftskip #1\nobreak\normalfont\leaders\hbox{$\m@th \mkern \@dotsep mu\hbox{.}\mkern \@dotsep mu$}\hfill\nobreak\hb@xt@\@pnumwidth{\hss #2}\par \penalty\@highpenalty \endgroup \fi} \makeatother \titleformat{\chapter} {\normalfont \center}{\thechapter}{1em}{} \titlespacing*{\chapter}{0pt}{-.5 in}{0.35in} \renewcommand{\thechapter}{\Roman{chapter}} \renewcommand*\contentsname{TABLE OF CONTENTS} \begin{document} \begin{singlespace} \tableofcontents \end{singlespace} \addtocontents{toc}{\noindent CHAPTER\par} \chapter{this is chapter I} \chapter{this is chapter II} \chapter{this is chapter III} \chapter{this is chapter IV} \chapter{this is chapter V} \chapter{this is chapter VI} \chapter{this is chapter VII} \chapter{this is chapter VIII} \end{document}
[ 0.00017010001465678215, 0.017575334757566452, -0.008636586368083954, 0.024519529193639755, 0.006371264811605215, 0.004018253646790981, 0.0077410913072526455, 0.01177558209747076, -0.016402212902903557, -0.004669409245252609, -0.01184192392975092, 0.003053228836506605, -0.009908003732562065, ...
[ 0.4544409513473511, 0.3243854343891144, 0.36975371837615967, 0.020666785538196564, -0.10898865014314651, 0.000516564236022532, 0.012584904208779335, -0.18840734660625458, -0.19824491441249847, -0.6687189340591431, 0.12989650666713715, 0.05233251303434372, 0.18374769389629364, -0.1068985089...
I've got a function that takes an Openlayers mouse click and puts the click location into a SQL statement which then gets passed to a JSP script which runs the SQL statement on the PostGIS database and returns geojson. I've grown bleary eyed googling for possible reasons why the line var features = geojson.read(e.responseText) always returns features as null. I'm interested in getting an attribute/property of one of the features in the returned geojson. What am I not understanding about reading the returned geojson? function getWatershed(e) { var geojson = new OpenLayers.Format.GeoJSON(); // Get the mercator coordinates from the viewport var xy = map.getLonLatFromViewPortPx(e.xy); // Find the watershed ID. var sql = "select st_asgeojson(the_geom), id, area_sqmi "; sql += "from medford.subshed2 "; sql += "where st_contains(the_geom,st_transform(st_setsrid("; sql += "st_makepoint(" + xy.lon + "," + xy.lat + ")"; sql += ",900913),26918))"; OpenLayers.Request.GET({ async: false, url: "./04-sql-to-json.jsp", params: {sql: sql}, success: function(e){ var features = geojson.read(e.responseText); if(features) { if(features.constructor != Array) { features = [features]; } for(var i=0; i<features.length; ++i) { watid = features[0].attributes.id; } } } }); The geojson returned from the server looks like this (the coordinates were abbreviated): {"type":"FeatureCollection", "features":[{"type":"Feature","geometry":{"type":"MultiPolygon","coordinates":[[[[251083.32142683025,4381268.9660757259,0],[251139.94862688333,4381204.021947667,0]]]]}, "properties":{"area_sqmi":"344.0191", "id":"PU2_3900_3750" } } ]} What am I not understanding about reading the returned geojson? Why is features always null?
[ -0.011581962928175926, 0.0067527322098612785, 0.00036179088056087494, 0.004646774381399155, -0.00747273862361908, 0.00020925141870975494, 0.008348491042852402, 0.020671144127845764, -0.014108609408140182, -0.01966298744082451, -0.0057066697627305984, 0.02096382901072502, 0.006968976929783821...
[ 0.3372020721435547, -0.10788781940937042, 0.5891635417938232, -0.07050647586584091, 0.04288534075021744, -0.028905164450407028, 0.4363599121570587, -0.05433480814099312, -0.04836161434650421, -0.8281499147415161, -0.1459692269563675, 0.5039287209510803, -0.2784656882286072, 0.3433613777160...
I want a feature like vim's going forward or backwards command `(w,e,ge,b)` etc in the shell prompt. I am using _tcsh_ shell in _Linux_. I would like to go _'n'_ number of words backs or forward etc and do some editing(may be append or replace). How can I do that? It doesn't have to be the exact same commands as vim but how do I achieve the same results in the shell prompt?
[ 0.00966082513332367, 0.004632514901459217, -0.013744423165917397, 0.011414886452257633, -0.0032627619802951813, -0.0018481268780305982, 0.009025044739246368, -0.004594985395669937, -0.02138645388185978, 0.000848158320877701, 0.00020782658248208463, 0.0050018853507936, 0.012759143486618996, ...
[ 0.40011611580848694, -0.0021302360109984875, 0.40879881381988525, -0.12866660952568054, -0.20570465922355652, -0.17042697966098785, 0.06552199274301529, 0.22331513464450836, -0.07361772656440735, -0.7806329727172852, 0.019084367901086807, 1.0940412282943726, -0.14813362061977386, 0.2624949...
I can calculate that if you want to, for example, desalinate water, you will have to pay a free energy cost of $k_B T$ for each ion you remove. In other words, removing an ion from a volume of water requires $\log_2 e$ bits of information. Is there an intuitive reason why? Below is the calculation: Imagine a tank of water from which we are removing ions, which currently have concentration $c$. We move the ions to a reservoir of concentration $c_0$. Each ion can be considered to occupy a volume $1/c$, so the entropy change in moving an ion from our tank is $k_B\ln(c/c_0)$. As we empty the tank starting from $c = c_0$ to $c = 0$, the average entropy change per ion is $$\mathrm{d}\bar{S} = \frac{k_B}{c_0}\int_{c_0}^0 \ln(c/c_0)\mathrm{d}c =- k_B$$ Because the internal energy doesn't change, the free energy cost is $$N k_BT$$ to completely remove $N$ solutes.
[ 0.011983949691057205, 0.011692081578075886, -0.01650267466902733, 0.011145984753966331, -0.011020485311746597, -0.018324166536331177, 0.008200353011488914, -0.01066441647708416, -0.016288425773382187, 0.020377593114972115, 0.0014106965390965343, 0.013919327408075333, -0.035269446671009064, ...
[ 0.4459799528121948, 0.26946690678596497, 0.19649727642536163, 0.03018319234251976, 0.13308794796466827, 0.1407850980758667, -0.11900308728218079, -0.5870107412338257, 0.04689668118953705, -0.5582704544067383, 0.08879925310611725, 0.5191614031791687, -0.272170752286911, 0.5476095676422119, ...
Something belongs to the University. Is it the _Universities' logo_ or the _University's logo_? I somehow don't think _University's_ exists.
[ 0.012225115671753883, 0.017508048564195633, 0.008952938951551914, 0.05924662575125694, -0.05846983566880226, -0.02136852592229843, 0.013452104292809963, 0.06395803391933441, -0.028651855885982513, 0.020897142589092255, -0.01534754317253828, 0.0050435191951692104, 0.01808476261794567, -0.00...
[ 0.9424418807029724, 0.416596382856369, 0.16565276682376862, 0.3355232775211334, 0.2397751808166504, -0.04707259312272072, 0.042446915060281754, 0.5702240467071533, -0.3471628725528717, -0.46355393528938293, -0.20162004232406616, -0.2281891256570816, 0.19555123150348663, 0.6777757406234741,...
I try to represent in a UML Class Diagram the following object and features: * The Object: A System (e.g. an electronic device) that can be of Type A, B or C * If the System is of type A, it has a specific extra attribute (let say "sub-type", which values can be A1 or A2) I created a Class named "System" and: 1. Created a class "System_TypeA" that holds an extra attribute named "subSystem" that can take any value within A1 and A2. 2. Created a generalization relationship beteween the two classes ("System_TypeA" _is a_ "System") So far, so good. Now, each System can be the backup system of other Systems. If this is the case, the System should hold a collection of references of the systems it is backing-up So for this feature, I also created a Class named "Backup_System" and created generalization between the two classes, since a "Backup_System" _is a_ "System". So how to represent in the Class Diagram the fact that there can be Systems of Type A that are Backup systems (i.e. the problem of multiple inheritance)?? (For info, the implementation should be done in java, so the same problem of multi-inheritance will be faced at implementation.)
[ -0.005832017865031958, 0.01495903916656971, -0.007669843267649412, 0.002312155207619071, -0.0013329631183296442, 0.001823619706556201, 0.0067236153408885, 0.034419432282447815, -0.010000219568610191, -0.021109050139784813, -0.017290936782956123, 0.002040551509708166, -0.0028250901959836483, ...
[ 0.04978535696864128, 0.04170174151659012, 0.08971583098173141, 0.07225821167230606, -0.04153391718864441, 0.3217114806175232, 0.24988843500614166, -0.20136044919490814, 0.003575953422114253, -0.7674564719200134, -0.1383439600467682, 0.26946529746055603, -0.3074960708618164, 0.1798284202814...
I'm running the Boruta algorithm with a $179\times 36$ predictor matrix and a numerical response. Most of the variables have a score going to `-Inf`. Should I conclude there are not important or there's a problem ? bor <- Boruta(X,y) > dim(X) [1] 179 36 > bor$ImpHistory[,1] [1] -0.33842437 0.03724477 0.77207194 -0.15966460 0.73550293 -2.11845770 -0.28029409 -1.73339443 1.80120877 -1.22191877 -0.45272011 -0.71728205 [13] -0.57268506 -0.48498812 0.61710237 0.58776416 -0.38667490 0.51795751 0.32422847 -1.79565022 0.26975995 -1.10919356 0.77731727 -1.10266785 [25] 0.19157178 -1.16694014 0.53795858 0.96632680 -0.99366419 -1.22646545 -1.71817624 -0.68100802 -1.41193813 -0.18065431 -Inf -Inf [37] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [49] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [61] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [73] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [85] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [97] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [109] -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf -Inf [121] -Inf
[ 0.00016895320732146502, 0.017031632363796234, -0.014172988012433052, 0.009295279160141945, 0.0025436580181121826, 0.004024210385978222, 0.006109589710831642, -0.004222994204610586, -0.010490728542208672, -0.0008614175021648407, -0.009043371304869652, 0.007080261129885912, -0.0094675729051232...
[ -0.055106740444898605, 0.12885898351669312, 0.5631869435310364, -0.16937881708145142, -0.26427310705184937, 0.312601238489151, 0.3132311701774597, -0.26303038001060486, -0.24181543290615082, -0.09824098646640778, 0.18204164505004883, 0.6526084542274475, -0.6354271769523621, 0.0834963470697...
Mathematica has several ways of generating random numbers: Congruential, ExtendedCA, Legacy, MersenneTwister, MKL, ParallelGenerator, ParallelMersenneTwister, Rule30CA. For `Method -> "Congruential"` the page notes that this generates low-quality randomness. What is known about the others? I came across a paper today (available here and here) showing that `Method -> "Rule30CA"` has an extremely small effective size, suggesting that it is probably unsuitable for simulation work. I was a little concerned on reading this, since there was nothing in the help that suggested it should not be used. Which of the others are known to be good? In particular, is the default `Method -> "ExtendedCA"` known to pass such tests as BigCrush or Dieharder? I know of results for the Mersenne Twister, of course (does well on all but the linear complexity tests). ## Summary of Results So Far * Congruential: Low quality (per documentation) * ExtendedCA: Apparently high-quality; Lichtblau reports (in an answer below) that it passes BigCrush. * Legacy: Presumably low quality, else would not have been obsoleted * MersenneTwister: Fairly high quality, though fails some BigCrush tests * MKL: Similar to MersenneTwister? * ParallelGenerator: What is this? * ParallelMersenneTwister: Same as MersenneTwister? * Rule30CA: Low quality (Meier & Staffelbach 1991, Sipper & Tomassini 1996) _This question touches on math and computer science, but this se seemed the best fit._
[ 0.009280864149332047, 0.004310308955609798, -0.004078741651028395, 0.02023402974009514, 0.00022559473291039467, -0.0214806217700243, 0.007669254206120968, 0.0011126508470624685, -0.013609465211629868, 0.0018300781957805157, -0.016095945611596107, 0.011754622682929039, 0.006338547915220261, ...
[ -0.29250720143318176, -0.1939389854669571, 0.2076176106929779, 0.3392886817455292, -0.13499335944652557, 0.36081644892692566, -0.30591973662376404, -0.043324463069438934, -0.11009188741445541, -0.1822812855243683, 0.24010109901428223, 0.2991948425769806, -0.2962400019168854, 0.198503479361...
I've heard of hitboxes, then I heard of hurtboxes. They seem most relevant to fighting games. What do each of them mean and how do they differ?
[ -0.03557952120900154, 0.04788538068532944, 0.007695668376982212, 0.029688309878110886, 0.030191432684659958, 0.007984025403857231, 0.011869965121150017, -0.00032162442221306264, -0.03220690041780472, 0.01742423325777054, -0.0056786248460412025, 0.0321061834692955, 0.02509297803044319, 0.01...
[ 0.30044183135032654, 0.01432668138295412, -0.19188503921031952, 0.08497928082942963, -0.49011552333831787, 0.29586485028266907, -0.22907541692256927, 0.016291789710521698, 0.11500094085931778, -0.6392068266868591, 0.23419615626335144, 0.28668081760406494, -0.0036689075641334057, -0.1705519...
I'm not asking about its formal definition (a windy one here) or its general definition in English, the latter of which states: > Stocks and shares that carry no fixed interest: Instead, I'm asking how definition 1 implies or relates to definition 2, the one for finance?
[ -0.004032870288938284, 0.015055205672979355, 0.012130729854106903, 0.028466343879699707, 0.008658509701490402, -0.019860995933413506, 0.01221407763659954, -0.0005516739329323173, -0.01732558012008667, -0.013240055181086063, -0.020067337900400162, 0.006555887404829264, 0.019820567220449448, ...
[ 0.6625308394432068, -0.18270811438560486, -0.23234006762504578, 0.00854259729385376, -0.040558017790317535, -0.21054214239120483, -0.6273707747459412, -0.015518588945269585, -0.3063388168811798, -0.03894132003188133, -0.030083822086453438, 0.5236501693725586, -0.28616026043891907, 0.084957...
I am curious that why in CFA model, observed measures operate as dependent variables. In my understanding, observed measures served as data point that estimating other parameters, such as latent variables and errors. Thanks,
[ -0.001702594687230885, 0.04832984134554863, 0.01119508221745491, 0.04148675128817558, 0.0637301653623581, -0.031689099967479706, 0.01907176896929741, -0.03399931266903877, -0.032316092401742935, -0.022237956523895264, -0.024815035983920097, 0.02730836346745491, -0.029053956270217896, 0.008...
[ 0.19386987388134003, 0.07518471032381058, -0.1314876675605774, 0.18821005523204803, -0.25922778248786926, -0.0731128454208374, -0.0007509097340516746, 0.14225101470947266, -0.22228795289993286, -0.26529043912887573, 0.42043566703796387, 0.36103931069374084, 0.03945104032754898, 0.274652391...
Every time an acronym is used for the first time, the glossaries package (GP) includes the long form in the text. I don't want GP to do that (the abbreviations are clear/they can be looked up in the list). I now solved it using \hphantom{all acronyms} somewhere on the first pages, but I bet there is a better solution to this. \documentclass{article} \usepackage[nopostdot,nogroupskip,nonumberlist]{glossaries} \newacronym{OSA}{OSA}{Old South Arabian} \newacronym{MSA}{MSA}{Modern South Aramaic} \begin{document} Some text about how interesting \gls{OSA} and \gls{MSA} are. And I want them ALWAYS to be abbreviated as \gls{OSA} and \gls{MSA}. \end{document}
[ 0.011064889840781689, 0.0019419877789914608, -0.009026922285556793, 0.030481718480587006, -0.01096535287797451, 0.003234677016735077, 0.007101983763277531, 0.014718407765030861, -0.013291570357978344, -0.01802806928753853, -0.010906573385000229, -0.006060804706066847, -0.00046501157339662313...
[ 0.2750827670097351, 0.39705803990364075, 0.3725775182247162, -0.20309393107891083, 0.12329430878162384, -0.17519843578338623, 0.33304616808891296, -0.005291426088660955, -0.2846432626247406, -0.7543178796768188, -0.04349372535943985, 0.38792547583580017, -0.20102594792842865, -0.0479671210...
Suppose I am using a standard inhibition model to find biochemical parameters that fit my data. The equation is: $y = \frac{A}{{1 + \exp \left( {\ln \left[ S \right] - \ln IC_{50}} \right)}} $ where $\left[ S \right]$ is the concentration of my inhibitor and $IC_{50}$ is the concentration of my inhibitor at which the measurement (with a maximum $A$) is reduced by half. Which of these approaches should I take? 1. enter the equation into the NLR procedure as-is (FYI: I am using SPSS) and let it fit the values of $A$ and $IC_{50}$ in the same manner, i.e. with all of the assumptions of OLS regression parameters. Or... 2. enter the equation as $y = \frac{A}{{1 + \exp \left( {\ln \left[ S \right]- \ IC_{50}^* } \right)}} $ where $ IC_{50}^* = \ln IC_{50}$. This of course would require a transform on the output parameter and confidence limits of $ IC_{50}^*$ giving me asymmetric error bars. Which of these strategies is most rigorous? My instincts suggest the 2nd option, as $ IC_{50}$ is actually bound by 0 and is thus more likely to be log-normal rather than normal. Any help (direct answers, references, etc) is appreciated.
[ 0.010651406832039356, 0.0175644438713789, -0.02517796866595745, -0.0009011005749925971, -0.0013191259931772947, -0.006116379983723164, 0.006739331874996424, -0.01228083111345768, -0.009424449875950813, -0.007358993403613567, -0.009947076439857483, 0.004761579912155867, -0.011164840310811996,...
[ 0.3956931531429291, -0.08517227321863174, 0.34400326013565063, -0.48284879326820374, 0.13113288581371307, 0.12258791923522949, 0.16606557369232178, -0.720315158367157, 0.2908967435359955, -0.4051806628704071, 0.3758624196052551, 0.42490407824516296, -0.4431135356426239, 0.425027072429657, ...
I'm fretting over an email to what was a potential customer. This person announced that he will do the project I pitched on with another company. I sincerely want to wish the customer good luck with the project, but I can't find the correct phrasing. I'm afraid "Wishing you all the best with the project" might sound like the hypocritical uttering of a sore loser who actually wants the project to go flat on its belly. Any suggestions as to how one who 'has lost' can still sincerely write he wishes the very best for everyone?
[ 0.004352877847850323, -0.0016255404334515333, -0.0011006402783095837, 0.03126200661063194, 0.001980417175218463, -0.0009984662756323814, 0.00496458075940609, -0.005080311093479395, -0.011121280491352081, 0.004330157767981291, -0.007793943397700787, 0.01625492423772812, 0.008960559032857418, ...
[ 0.5249336957931519, 0.48816952109336853, 0.061901919543743134, -0.1447073519229889, 0.16635434329509735, 0.009932797402143478, 0.2137346714735031, 0.4922660291194916, -0.09093969315290451, -0.32207581400871277, 0.1453634798526764, 0.15703946352005005, 0.31151852011680603, 0.489247292280197...
I've got two addresses: `ipaddress` and `ipaddress:100`. The first one uses the default port of 80, but the second server (running GitLab) uses port 100. I have multiple IP addresses, but I am wondering if it is possible to have DNS link an address like `git.domain.com` to `ipaddress:100`? Is this possible with DNS? I have full domain and root access, so any option is available.
[ 0.00036392395850270987, 0.001176107325591147, -0.0007383712800219655, 0.0073194196447730064, -0.006720634642988443, 0.0061207665130496025, 0.008696114644408226, 0.0038611446507275105, -0.02155546098947525, -0.00004859022737946361, 0.00041585907456465065, 0.00900533888489008, -0.0008836276247...
[ 0.19929543137550354, 0.2907571792602539, 0.8510610461235046, -0.14839203655719757, -0.009051133878529072, 0.5747471451759338, -0.043471142649650574, 0.34059345722198486, -0.22049668431282043, -1.0213291645050049, 0.18835924565792084, 0.24564555287361145, -0.1287854164838791, 0.277805656194...
A damped harmonic oscillator has three cases for the damping: underdamped, critically damped and overdamped. With partial differential equations, I know the hyperbolic wave equation, the parabolic heat equation and the elliptical Laplace equation. Since underdamping means oscillation and therefore waves and the heat diffusion looks like a critically damped fluid, I was wondering, if those have anything in common.
[ -0.012194240465760231, 0.017600005492568016, -0.004560289438813925, 0.03385712206363678, 0.0014958237297832966, -0.028764594346284866, 0.01698901690542698, -0.005174650344997644, -0.015140251256525517, -0.025863055139780045, -0.010698569007217884, 0.0163082554936409, -0.014185783453285694, ...
[ -0.13098114728927612, -0.11040735989809036, 0.34551143646240234, 0.09942774474620819, -0.35645005106925964, 0.3004670739173889, -0.041059501469135284, -0.33285218477249146, -0.05647939443588257, 0.019993361085653305, 0.24262873828411102, 1.0630712509155273, -0.08980807662010193, 0.38767048...
I am familiar with how the X11 system works, where clients connect via a socket to the server process and send operations to the window server to perform certain operations on their behalf. But I do not understand (and I could not find good documents) describing how a GUI application interacts with the window system on Mac OS X. Some of the questions that I have include: * How does an app receive events from the windowing system? * Does the app need to register with the kernel, or some windowing system server? * How does the windowing system request that an app update its display? * How does an app trigger a re-display operation? * Is there a socket-based protocol, or some other RPC system to the windowing system? * Does the windowing system, or the application, have direct hardware access? * What are the operations available between client apps and the windowing system?
[ -0.010589088313281536, -0.0019626268185675144, 0.009654561057686806, -0.0011767518008127809, 0.0011545331217348576, -0.02918447181582451, 0.007582744583487511, 0.005441153421998024, -0.01873466745018959, -0.007438227068632841, -0.014514864422380924, 0.010679595172405243, 0.009525860659778118...
[ 0.38962268829345703, 0.19168062508106232, 0.3565877079963684, 0.0041994331404566765, -0.3343532979488373, 0.13880731165409088, 0.0969441831111908, -0.03658171743154526, -0.4320865869522095, -0.6090360879898071, -0.22947418689727783, 0.54906165599823, -0.30454206466674805, -0.02037581615149...
Writing CDs with `cdrecord` and making images with `genisoimage` is no problem. I want to write DVDs in a similar manner, first creating an image, then burning it to disk. This is quite beneficial, since I can inspect the image before burning it to disk. Now, all I've seen, is how to use the `growisofs` command to burn something to DVD, but all how-to's were using ISO9660 for DVDs. But I want to burn UDF images. Before I can burn them, I need to make them, but how do I create UDF images? Also, If you could explain, or link on how to burn BluRay, that would be great, too. All tools must be command line, as I need to work with it, where GUI is no option.
[ -0.016618555411696434, 0.014312728308141232, -0.017186395823955536, 0.017263071611523628, -0.02775096893310547, 0.004012172110378742, 0.008869418874382973, 0.007291147951036692, -0.013356707990169525, 0.0036865605507045984, -0.006311304401606321, 0.00026585604064166546, 0.0034153854940086603...
[ 0.43769484758377075, 0.0690871998667717, 0.4152993857860565, 0.11902355402708054, -0.03481562063097954, -0.5849000215530396, -0.07094357162714005, -0.07281800359487534, -0.32518529891967773, -0.9282012581825256, -0.03397351875901222, 1.040405511856079, -0.3219444155693054, -0.0643237456679...
I'm trying compiling this latex template example (CHI Extended Abstracts template) under linux (XeTeX, Version 3.1415926-2.2-0.9997.4 (TeX Live 2010/Arch Linux) using xelatex. This is the output I'm getting: http://sprunge.us/YaXi It seems that it's having troubles finding Helvetica, but helvetica is installed: fc-list http://sprunge.us/GChE Can you guys help? **UPDATE:** Trying to simplify the issue, I've been suggested on #latex to try with a simpler example: \documentclass{scrbook} \usepackage[english]{babel} \usepackage{blindtext} \usepackage{fontspec} \usepackage{xunicode} \defaultfontfeatures{Mapping=tex-text} \setmainfont{Helvetica} \begin{document} \Blinddocument \end{document} Which fails just as the template above. See http://sprunge.us/GJdG
[ 0.011972922831773758, -0.005307817831635475, -0.01609889790415764, 0.023349659517407417, -0.0038814672734588385, 0.014074156060814857, 0.009343850426375866, -0.010739360004663467, -0.02026216685771942, -0.02777762897312641, -0.006296797189861536, -0.002551085315644741, -0.01057683490216732, ...
[ 0.30500316619873047, 0.2778654098510742, 0.49422889947891235, 0.13097044825553894, -0.44368281960487366, -0.09160250425338745, 0.32192978262901306, 0.2008770853281021, -0.34198135137557983, -0.8407917618751526, -0.023967446759343147, 0.29002106189727783, -0.42351529002189636, -0.0436742752...
After I compile my PNAS journal document, a part of the references should be in helvetica-boldoblique style. However, it is printed simple bold (helvetica- bold). When I change the font settings in the style and class file from boldoblique to oblique, it text will be italic but not bold anymore. I have tried this solution, but it is not working: Latex font error ... I can't find 'hv' Files available at: http://www.pnas.org/site/authors/LaTex.xhtml
[ 0.016994604840874672, -0.006483328994363546, -0.013286581262946129, 0.029105452820658684, 0.007868459448218346, -0.001976668369024992, 0.010921549052000046, 0.0252678282558918, -0.013850575312972069, -0.011386295780539513, -0.012878919020295143, 0.0024753171019256115, -0.010648730210959911, ...
[ 0.2367693930864334, 0.4642404317855835, 0.968035876750946, -0.26854071021080017, -0.11380862444639206, 0.053266577422618866, 0.19337958097457886, -0.2946256697177887, 0.11618594825267792, -0.5252044796943665, 0.24734602868556976, 0.056754425168037415, -0.29489707946777344, 0.11821303516626...
Is there a probability that an electron in an atom change its orbital by emitting a quantum of gravitational radiation instead of photon?
[ 0.03103313036262989, 0.04556148126721382, 0.010529808700084686, 0.0195668563246727, -0.029484272003173828, -0.044939786195755005, 0.017379410564899445, -0.04859684407711029, -0.0038332114927470684, -0.052989840507507324, -0.017241936177015305, 0.03916522115468979, -0.01467694342136383, -0....
[ 0.4125156104564667, -0.028578082099556923, -0.23963965475559235, 0.2251552939414978, 0.1310095489025116, 0.07985479384660721, -0.14588294923305511, -0.23989075422286987, -0.2833145260810852, -0.276945561170578, -0.03026503138244152, 0.14654134213924408, -0.39284342527389526, 0.178112670779...
I want to test if expressions (mix of variables, functions and numbers) are zero valued, as fast as possible, and `PossibleZeroQ` is sometimes very slow. One solution I found was to substitute the variables for random reals and test if the value of the substituted expression is less than, say, $0.0001$. It works good, but maybe there are other solutions out there. I know it can cause some wrong answers, but what is most important is the speed, since I can check the false positive later with `PossibleZeroQ`. Can you think of an algorithm that can perform fast zero value tests in detriment of some loss of accuracy? Edit: I'll post my algorithm here: TestZeroValuedExpression[expression_,symbolslist_]:=Module[{numericvalue}, Quiet[TimeConstrained[If[Check[ numericvalue=N[Expand[expression/.Table[symbolslist[[i]]->RandomReal[{1,2}],{i,Length[symbolslist]}]]]; ,False]=!=False, If[Abs[numericvalue]>0.00001,False,PossibleZeroQ[expression]], PossibleZeroQ[expression] ],3,False]] ];
[ 0.03423134610056877, 0.014353612437844276, -0.011150076985359192, 0.00015957257710397243, -0.014766717329621315, -0.0035043905954807997, 0.007829241454601288, -0.0023436041083186865, -0.014859367161989212, -0.004837814252823591, -0.014241031371057034, 0.0064295148476958275, -0.01655561476945...
[ 0.25381770730018616, -0.07714257389307022, -0.05334676429629326, 0.3009624481201172, 0.1940177083015442, 0.21853339672088623, 0.08064961433410645, 0.05104925110936165, 0.24122868478298187, -0.5294573307037354, 0.14156955480575562, 0.5292915105819702, -0.3053783178329468, -0.042199492454528...
The question is to figure out how the energy can be derived knowing just one thing: > **There is a quantity called Energy that is conserved over time.** The goal is to get an equation that somehow implies the basic formulas for kinetic energy and potential energy, $\frac{1}{2}mv^2$ and $mgh$. So how can we get this with just a rudimentary knowledge of physics and algebra, and the first principle (stated above)?
[ 0.027110684663057327, 0.015796542167663574, 0.004858370404690504, -0.003608043072745204, -0.008605188690125942, -0.019977159798145294, 0.006248018238693476, -0.024380628019571304, -0.017904477193951607, -0.021120233461260796, 0.0011134212836623192, 0.018509306013584137, -0.009930114261806011...
[ 0.29257139563560486, 0.02835005894303322, 0.45986631512641907, 0.22795987129211426, 0.3485093414783478, -0.03571370244026184, -0.020580139011144638, -0.5548731088638306, -0.03311461582779884, -0.6097986102104187, -0.08709985017776489, 0.7186306715011597, -0.052148185670375824, 0.1837612688...
**EDIT** : By a comment below, I learned that what I'm asking is: How do I add an epigraph in the beginning of a section? **EDIT** : I managed to use the epigraph package. I'm importing it in the LaTeX preable and in my document open a LaTeX red box and write: `\epigraph{the quote}{the author}`. It looks pretty much fine. In LyX, I would like to add a famous > ``If you can't solve a problem, then there is an easier problem you can > solve: find it.'' > > George Pólya I'd like the whole thing to be slightly indented, and name "George Pólya" much more indented. Is there a standard way to do that? I tried the "Quotation" and "Quote" envionment, but they seem to slightly indent the whole paragraph, which is not exactly what I want. (and, by the way, is this a good practice, or is there another way a quote should look like?)
[ 0.018311459571123123, -0.0023418308701366186, -0.008765818551182747, 0.01205784548074007, 0.00022227782756090164, 0.009409746155142784, 0.005716881714761257, 0.009751597419381142, -0.018562868237495422, -0.005494079552590847, -0.002257810439914465, 0.0021820086985826492, -0.02069226279854774...
[ 0.3518080413341522, 0.29081764817237854, 0.5075680613517761, 0.19368545711040497, -0.25608766078948975, 0.09472569078207016, -0.11523318290710449, 0.06239066272974014, -0.3876985013484955, -0.46356749534606934, -0.1719755232334137, -0.0016137146158143878, -0.2818063497543335, -0.1608063876...
I am using this get_posts array to retrieve top 15 popular posts by comments. The problem is that it shows those all those highly commented posts from the day posts were added into blog. So, the list sometimes changes but most of the time remains same. How can I limit/show posts of the last 7 days? This would show most commented posts of the last week. Thanks Here is my code: global $post; $args = array( 'orderby' => 'comment_count', 'order' => 'DESC' , 'numberposts' => 15); $popular_posts = get_posts( $args ); foreach( $popular_posts as $post ) : if (has_post_thumbnail()) { ?> <li> <div class="widgetimg"> <a href="<?php the_permalink(); ?>"><?php the_post_thumbnail('widgetimg-thumb'); ?></a> </div> </li> <?php } endforeach; ?>
[ 0.0022551168221980333, 0.014463559724390507, -0.021959982812404633, 0.020550383254885674, 0.011090872809290886, -0.004793078638613224, 0.007813672535121441, -0.0019274083897471428, -0.012241758406162262, -0.0072829751297831535, -0.01253504492342472, 0.009096525609493256, 0.002675865544006228...
[ 0.3243721127510071, 0.06639129668474197, 0.8386215567588806, 0.08926598727703094, -0.36176028847694397, -0.15781766176223755, 0.29556623101234436, 0.25273454189300537, -0.5132051706314087, -0.6275418400764465, 0.07408389449119568, 0.4346598982810974, -0.29119619727134705, 0.397480756044387...
I see this advice given a few times on the World of Tanks forum: buy a premium tank to make silver/credits. Are premium tanks really that much more effective at farming money than other tanks of the same tier?
[ -0.00711803138256073, 0.017725301906466484, -0.011099116876721382, -0.007311407942324877, 0.00527487276121974, 0.028178613632917404, 0.0115437600761652, 0.018346242606639862, -0.024662857875227928, 0.011672638356685638, -0.008414465002715588, 0.018574580550193787, -0.022922223433852196, -0...
[ 0.9490430355072021, -0.059283968061208725, -0.20062671601772308, 0.4832649528980255, -0.154995396733284, 0.15736904740333557, -0.05386125668883324, 0.1682315617799759, -0.2887831926345825, -0.47779831290245056, 0.28000956773757935, 0.766083836555481, 0.24542400240898132, 0.0049731256440281...
Example: > The name Self-Arming Group isn't as [...] as Gun Club, but controversial > nevertheless. What word can I use in the example above?
[ -0.011025611311197281, 0.01268183346837759, -0.009253102354705334, 0.07661832869052887, -0.06289812177419662, -0.02791217900812626, 0.01781747303903103, 0.007202377077192068, -0.027231253683567047, -0.04573547840118408, -0.011677269823849201, 0.027591200545430183, -0.01580085977911949, 0.0...
[ 0.2946079969406128, 0.15113916993141174, -0.16895025968551636, -0.19600048661231995, -0.6205763220787048, -0.3220713138580322, 0.28384703397750854, -0.39556020498275757, -0.08920204639434814, -0.2318969964981079, 0.36342209577560425, 0.5818119049072266, -0.2662758231163025, -0.114377804100...
Eyal's babel incompatibility of the week... this time a question due to H.R. on the IVRITEX mailing list. The `babel` is incompatible with the `subfig` package. If you add a subfig with a caption: \documentclass[english,hebrew]{article} \usepackage{babel} \usepackage{subfig} \begin{document} \begin{figure}[tbph] \subfloat[\L{A sub-figure}]{\L{The subfigure}} \caption{\L{A figure caption}} \end{figure} \end{document} The sub-figure is numbered; the numbering uses Hebrew letters (Aleph, Bet, Gimel; in this case, Aleph - א). The numbering is enclosed in parentheses, but due to direction issues, the parentheses are in the opposite directions of where they should be: ![rendering](http://i.stack.imgur.com/h2plM.png) Is there a workaround for this problem? A minimal workaround with no additional package dependencies is preferred. Note: You don't see this problem with Arabic instead of Hebrew (and the \L's removed) since when it's Arabic, you get a left-to-right float and Latin enumeration. Maybe it manifests in some other way with Arabic.
[ 0.01411572378128767, 0.0011524672154337168, -0.0015097090508788824, 0.028621027246117592, 0.012909719720482826, 0.020024649798870087, 0.00937253050506115, 0.029121939092874527, -0.010712726041674614, 0.011924218386411667, -0.022749947383999825, 0.004436025395989418, -0.018598897382616997, ...
[ -0.2875185012817383, 0.19041699171066284, 0.5244573354721069, -0.05448663979768753, -0.7251374125480652, -0.01371776033192873, 0.17799635231494904, -0.47117871046066284, 0.037315383553504944, -0.3582943379878998, -0.35037899017333984, 0.3119209110736847, -0.4515661299228668, 0.346915543079...
My question is in regards to installing openambit which is software for Suunto's ambit watch. I followed the installation instructions here but it didn't work. When I did: $ mkdir libambit-build $ cd libambit-build $ cmake ../src/libambit $ make Optionally: $ sudo make install Everything went fine, but when I ran the `make` step: $ mkdir openambit-build $ cd openambit-build $ cmake ../src/openambit $ make Optionally: $ sudo make install I was told: dustin@dustin:~/openambit-build$ sudo cmake ~/Desktop/openambit-0.2/src/openambit/-- The C compiler identification is GNU -- The CXX compiler identification is GNU -- Check for working C compiler: /usr/bin/gcc -- Check for working C compiler: /usr/bin/gcc -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ -- Check for working CXX compiler: /usr/bin/c++ -- works -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Looking for Q_WS_X11 -- Looking for Q_WS_X11 - found -- Looking for Q_WS_WIN -- Looking for Q_WS_WIN - not found. -- Looking for Q_WS_QWS -- Looking for Q_WS_QWS - not found. -- Looking for Q_WS_MAC -- Looking for Q_WS_MAC - not found. -- Found Qt4: /usr/bin/qmake (found version "4.8.1") -- Found libambit: /usr/local/include, /usr/local/lib/libambit.so -- Found UDev: /usr/include -- Found ZLIB: /usr/lib/x86_64-linux-gnu/libz.so (found version "1.2.3.4") -- Configuring done -- Generating done -- Build files have been written to: /home/dustin/openambit-build dustin@dustin:~/openambit-build$ sudo make [ 2%] Generating movescount/moc_movescountxml.cxx [ 4%] Generating ui_mainwindow.h [ 6%] Generating ui_settingsdialog.h [ 9%] Generating ui_confirmbetadialog.h [ 11%] Generating qrc_resources.cxx [ 13%] Generating moc_mainwindow.cxx [ 16%] Generating moc_devicemanager.cxx [ 18%] Generating moc_settingsdialog.cxx [ 20%] Generating moc_settings.cxx [ 23%] Generating moc_logstore.cxx [ 25%] Generating moc_udevlistener.cxx [ 27%] Generating moc_confirmbetadialog.cxx [ 30%] Generating moc_single_application.cxx [ 32%] Generating movescount/moc_movescount.cxx [ 34%] Generating movescount/moc_movescountjson.cxx [ 37%] Generating movescount/moc_movescountlogdirentry.cxx Scanning dependencies of target openambit [ 39%] Building CXX object CMakeFiles/openambit.dir/main.cpp.o [ 41%] Building CXX object CMakeFiles/openambit.dir/mainwindow.cpp.o [ 44%] Building CXX object CMakeFiles/openambit.dir/devicemanager.cpp.o [ 46%] Building CXX object CMakeFiles/openambit.dir/settingsdialog.cpp.o [ 48%] Building CXX object CMakeFiles/openambit.dir/settings.cpp.o [ 51%] Building CXX object CMakeFiles/openambit.dir/logstore.cpp.o [ 53%] Building CXX object CMakeFiles/openambit.dir/logentry.cpp.o [ 55%] Building CXX object CMakeFiles/openambit.dir/udevlistener.cpp.o [ 58%] Building CXX object CMakeFiles/openambit.dir/confirmbetadialog.cpp.o [ 60%] Building CXX object CMakeFiles/openambit.dir/single_application.cpp.o [ 62%] Building CXX object CMakeFiles/openambit.dir/movescount/movescount.cpp.o [ 65%] Building CXX object CMakeFiles/openambit.dir/movescount/movescountjson.cpp.o /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp: In member function ‘int MovesCountJSON::generateLogData(LogEntry*, QByteArray&)’: /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:230:16: error: ‘class QJson::Serializer’ has no member named ‘setDoublePrecision’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:231:16: error: ‘class QJson::Serializer’ has no member named ‘setIndentMode’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:231:30: error: ‘IndentCompact’ is not a member of ‘QJson’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:290:47: error: no matching function for call to ‘QJson::Serializer::serialize(QVariantMap&, bool*)’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:290:47: note: candidates are: /usr/include/qjson/serializer.h:50:10: note: void QJson::Serializer::serialize(const QVariant&, QIODevice*, bool*) /usr/include/qjson/serializer.h:50:10: note: no known conversion for argument 2 from ‘bool*’ to ‘QIODevice*’ /usr/include/qjson/serializer.h:62:16: note: QByteArray QJson::Serializer::serialize(const QVariant&) /usr/include/qjson/serializer.h:62:16: note: candidate expects 1 argument, 2 provided make[2]: *** [CMakeFiles/openambit.dir/movescount/movescountjson.cpp.o] Error 1 make[1]: *** [CMakeFiles/openambit.dir/all] Error 2 make: *** [all] Error 2 dustin@dustin:~/openambit-build$ sudo make install [ 2%] Building CXX object CMakeFiles/openambit.dir/movescount/movescountjson.cpp.o /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp: In member function ‘int MovesCountJSON::generateLogData(LogEntry*, QByteArray&)’: /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:230:16: error: ‘class QJson::Serializer’ has no member named ‘setDoublePrecision’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:231:16: error: ‘class QJson::Serializer’ has no member named ‘setIndentMode’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:231:30: error: ‘IndentCompact’ is not a member of ‘QJson’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:290:47: error: no matching function for call to ‘QJson::Serializer::serialize(QVariantMap&, bool*)’ /home/dustin/Desktop/openambit-0.2/src/openambit/movescount/movescountjson.cpp:290:47: note: candidates are: /usr/include/qjson/serializer.h:50:10: note: void QJson::Serializer::serialize(const QVariant&, QIODevice*, bool*) /usr/include/qjson/serializer.h:50:10: note: no known conversion for argument 2 from ‘bool*’ to ‘QIODevice*’ /usr/include/qjson/serializer.h:62:16: note: QByteArray QJson::Serializer::serialize(const QVariant&) /usr/include/qjson/serializer.h:62:16: note: candidate expects 1 argument, 2 provided make[2]: *** [CMakeFiles/openambit.dir/movescount/movescountjson.cpp.o] Error 1 make[1]: *** [CMakeFiles/openambit.dir/all] Error 2 make: *** [all] Error 2 I am not sure what is going wrong. I can post more of the output if needed as well.
[ -0.008423780091106892, -0.005507219582796097, -0.009993648156523705, 0.003636345500126481, -0.007397332228720188, 0.006896410137414932, 0.011548967100679874, 0.015620317310094833, -0.01941070333123207, 0.0001410241238772869, -0.004415072035044432, 0.004692094400525093, 0.0006911961827427149,...
[ 0.24215660989284515, -0.10894186049699783, 0.44424962997436523, -0.2076699435710907, -0.05575507879257202, 0.056189969182014465, 0.32085710763931274, -0.38202038407325745, -0.21624699234962463, -0.3638819754123688, -0.20246724784374237, 0.7257301807403564, -0.4180622696876526, -0.058464750...
I'm trying to grey out edges of a graph to high a region of interest. For instance, I would like to make every edge expect for that labeled "interest" grey in the following MWE. \usefonttheme{serif} \usepackage{tikz} \usetikzlibrary{arrows} \definecolor{burntorange}{cmyk}{0,0.52,1,0} \begin{document} \begin{tikzpicture}[->,>=stealth',shorten >=1pt,auto,node distance=3cm, thick,main node/.style={circle,fill=blue!20,draw,font=\sffamily\Large\bfseries}] \node[main node, scale=1.4] (interest) {$i$}; \node[main node, fill=blue!10] (j) [below left of=interest] {$j$}; \node[main node, fill=blue!10] (k) [below right of=j] {$k$}; \node[main node, fill=burntorange, below right of=interest, distance = 6cm ] (input) [below right of=interest] {input $u$}; \path[every node/.style={font=\sffamily\small}] (interest) edge [bend right] node[left] {$\mathbf{M}_{ji}$} (j) edge [loop above] node {$\mathbf{M}_{ii}$} (interest) edge [bend left] node {$\mathbf{M}_{ki}$} (k) (j) edge node [right] {$\mathbf{M}_{ij}$} (interest) edge [loop left] node {$\mathbf{M}_{jj}$} (j) edge [bend right] node[left] {$\mathbf{M}_{jk}$} (k) (k) edge node [right] {$\mathbf{M}_{kj}$} (j) edge node [bend right] {$\mathbf{M}_{ik}$} (interest) edge [loop below] node {$\mathbf{M}_{kk}$} (k) (input) edge [bend right] node {$\mathbf{W}_i$} (interest) edge [bend left] node {$\mathbf{W}_k$} (k); \end{tikzpicture} \end{document}
[ 0.004102464299649, 0.010207807645201683, -0.0031309654004871845, 0.010218160226941109, -0.014519435353577137, -0.004122687503695488, 0.007485336624085903, -0.0048650880344212055, -0.013076132163405418, 0.009755615144968033, -0.0005037235678173602, -0.0017112134955823421, -0.00596646126359701...
[ 0.5549732446670532, 0.11350622773170471, 0.193795308470726, -0.11058008670806885, 0.12426220625638962, 0.011811678297817707, 0.222251757979393, -0.03758477792143822, -0.1302814483642578, -0.8274130821228027, 0.24985522031784058, 0.27121618390083313, -0.2273259460926056, 0.1028362512588501,...
I haven't bought SC2: HotS yet but from the question "What bonus experience can you get?" I can see that you now gain some experience after a battle? If so, * What effect does it have if your profile is "Level 5" or something? * Do you get transferred to a higher league? * Do you have in game bonuses for single player?
[ -0.015472352504730225, 0.01664162427186966, 0.012562700547277927, -0.0003096477594226599, 0.04031362012028694, -0.012588861398398876, 0.008733917959034443, -0.018981866538524628, -0.02669290453195572, 0.024814030155539513, -0.0043470775708556175, 0.03695495054125786, 0.0029293771367520094, ...
[ 0.2435629814863205, -0.3375926613807678, 0.21163131296634674, 0.350997656583786, -0.28081458806991577, -0.18900179862976074, 0.28790611028671265, -0.024109994992613792, -0.4888419806957245, -0.3621232509613037, 0.1440410017967224, 0.34870994091033936, 0.35299554467201233, -0.07771756500005...
I need to convert Powerpoint files to PNG in PHP. I can do the conversion with using COM in PHP but COM is only supported on Windows. I need to do this for both Windows and Linux. I am wondering if someone give me hints about that conversion.
[ -0.015705376863479614, -0.0035456439945846796, -0.023669198155403137, 0.04473724961280823, -0.04043380171060562, 0.0036298988852649927, 0.011371511965990067, 0.028102446347475052, -0.0359608493745327, 0.009197166189551353, -0.008674336597323418, 0.0028398188296705484, 0.028033703565597534, ...
[ 0.5265742540359497, 0.0020365226082503796, 0.3382892608642578, 0.11852090805768967, -0.473113089799881, 0.08772759139537811, 0.008464084006845951, 0.3384494185447693, -0.17944969236850739, -0.9430463910102844, 0.23223182559013367, 0.714938759803772, -0.059958793222904205, 0.029149644076824...
Does Darwin have any features that are specific to it? Do other Unixe(s) have features that Darwin lacks?
[ -0.03019586391746998, 0.0029842196963727474, -0.03585479408502579, 0.02945505641400814, 0.011573866941034794, -0.07191108912229538, 0.016351865604519844, 0.0692346915602684, -0.043973200023174286, -0.04458580166101456, -0.02702668122947216, 0.017874909564852715, 0.030457746237516403, 0.003...
[ 0.3729822635650635, 0.23983316123485565, -0.2295314371585846, 0.26655346155166626, -0.17895570397377014, -0.28580358624458313, 0.06751945614814758, 0.6383817195892334, 0.02803051844239235, -0.016586478799581528, 0.1857520192861557, 0.4540172219276428, -0.2705783247947693, 0.285273581743240...
Is it a bad idea to keep scripts in /root? Simple bash scripts that have the following permissions: -rw-r--r-- They execute fine when ran as root, but is this bad practice? It's a multi- user system, that needs sudo to execute the scripts in question when not logged in as root....
[ 0.02042453922331333, 0.014055585488677025, -0.00013513737940229475, 0.0215939749032259, -0.011594788171350956, -0.008658687584102154, 0.011844642460346222, 0.006032511591911316, -0.027930056676268578, -0.023272542282938957, -0.019604071974754333, 0.0047955503687262535, -0.022003961727023125,...
[ 0.3463155925273895, 0.08664219081401825, -0.16433097422122955, 0.03746222332119942, -0.28770217299461365, -0.3838679790496826, 0.5327975749969482, -0.1458442062139511, -0.32736173272132874, -0.39482808113098145, 0.09733439236879349, 0.4354945719242096, 0.04748910292983055, -0.0711031034588...
Suppose I have a 20x20x20 c"m plastic box where the top cover is open. On the top cover there is a fan ( 20x20 cm) which suppose to cool a cpu which is in the box . At the bottom there is a CPU which is getting warm. Now - in order to cool the CPU - I have 2 options : * put the fan to suck air from the box towards outside * put the fan to push air from outside towards the box. **Final goal :** Max - Cooling the cpu. **Question:** Which is the preferred approach ?
[ -0.0010963943786919117, -0.003327619284391403, -0.010812398977577686, 0.015138490125536919, -0.02521505206823349, -0.03085384890437126, 0.008965462446212769, -0.013185087591409683, -0.012646486982703209, 0.00509477686136961, -0.0008046806324273348, 0.005691739730536938, -0.002864331007003784...
[ 0.22925825417041779, 0.09869975596666336, 0.4298418462276459, -0.034787848591804504, 0.01152167096734047, 0.32128942012786865, -0.00493127666413784, -0.4658585488796234, -0.2179909348487854, -0.5961393117904663, -0.2724904716014862, 0.30124637484550476, 0.1787836253643036, 0.02792328782379...
Alexithymia refers to a state wherein a person cannot understand or describe his or her feelings, and means literally "without words for emotions". Is there a term (psychological, medical, or otherwise) for someone who is overly conscious of and articulate about his or her emotions?
[ -0.0004156071809120476, 0.020147042348980904, -0.0021395746152848005, 0.014614296145737171, -0.012990950606763363, -0.054266445338726044, 0.01329793967306614, 0.0074311960488557816, -0.011286982335150242, 0.03223177045583725, -0.026515444740653038, 0.008194886147975922, 0.0008461641846224666...
[ 0.2693493068218231, 0.37232521176338196, 0.07157547771930695, 0.06417680531740189, 0.2016979306936264, -0.02164638042449951, 0.9308660626411438, 0.030913012102246284, 0.09500978887081146, -0.2881442904472351, -0.24070464074611664, 0.17259515821933746, -0.6628077030181885, 0.464475661516189...
I am trying to learn continuations and use them to implement coroutines in Scheme. I have two procedures (coroutines) `a` and `b`, and I switch between them in the following way: ;; c is a continuation. (define (a c) ... ;; Switch to the other coroutine. (call/cc c) ... (define (b c) ... ;; Switch to the other coroutine. (call/cc c) ... I start the execution with (a b) I have tested a small example and, indeed, the execution switches back and forth between the two procedures until one of the two terminates without calling the continuation it was given as an argument. I have two questions: 1. Is my example an appropriate / idiomatic implementation of coroutines using continuations? 2. Is tail-call optimization mandatory for implementing coroutines with continuations? In my solution, switching too often between the two coroutines might cause a stack overflow if Scheme did not provide TCO. Are there other solutions that do not require TCO? **EDIT** Basile's comment seems to be sufficient to answer my question. The background of the question is that I would like this approach to implement coroutines in Python, where I do not have TCO.
[ -0.012721330858767033, 0.005677190609276295, -0.01953795552253723, 0.0158432275056839, -0.003145703347399831, -0.020027093589305878, 0.008550566621124744, -0.02873876504600048, -0.014651783742010593, -0.02805332839488983, 0.000423194607719779, -0.00663117878139019, -0.0076988255605101585, ...
[ 0.5822713971138, 0.08856489509344101, 0.13469813764095306, -0.2781500220298767, 0.24257148802280426, 0.12446307390928268, 0.23747844994068146, -0.26067230105400085, -0.2071732133626938, -0.21173661947250366, 0.15747438371181488, 0.5857998728752136, -0.4217217266559601, 0.18809619545936584,...
Is there any applications out there that can let me selective limit network access when i put a program into the background?
[ 0.04083507880568504, -0.012426615692675114, -0.009153504855930805, 0.010748513974249363, 0.0358324833214283, 0.009085525758564472, 0.013319598510861397, 0.008356287144124508, -0.026320528239011765, 0.04283928871154785, -0.0132245859131217, 0.033887967467308044, 0.014481185004115105, 0.0571...
[ 0.805957555770874, 0.20027656853199005, -0.16762347519397736, 0.3212689459323883, 0.25639593601226807, -0.09469999372959137, 0.02040822245180607, 0.30137869715690613, -0.4628376364707947, -0.20437169075012207, 0.24499927461147308, 0.4084220230579376, -0.1427999585866928, 0.0412937439978122...
Suppose $X \sim U(0,1)$, and $Y[x] = g(x)$ where $g(.)$ is some complicated function. I want to calculate/plot the density of $Y$. I can do this analytically for simple enough $g$. I can also generate some large number $N$ samples of $X$ from the stardard uniform, calculate $y_i = Y(x_i)$ for each sample $x_i$, and then compute the histogram of these $ y_i, \forall i \in (1, ... N)$. This will turn out to be inefficient if, say the details of $p_Y(y)$ come from very small regions of $X$. Is there a better way to calculate/estimate $p_Y(y)$?
[ -0.002409781562164426, 0.008621693588793278, -0.015488370321691036, 0.00664903037250042, -0.0051927724853158, -0.01785217598080635, 0.005858061835169792, -0.0008994681993499398, -0.010331178084015846, -0.011800073087215424, -0.0004529292928054929, 0.0009514722623862326, -0.007564033847302198...
[ 0.14496655762195587, -0.08245080709457397, 0.33564209938049316, -0.026678476482629776, -0.059066612273454666, 0.29967179894447327, 0.05325967073440552, -0.2205432504415512, -0.2029038369655609, -0.5848372578620911, -0.1864970326423645, 0.14796504378318787, -0.54380863904953, 0.334541738033...
I need to apply unit testing for one of my plugins. I recently re designed it into classes, so unit testing should be easier to apply now. What are some effective strategies to unit test my plugin? I need a way to use WordPress' functions used in the plugin, but I shouldn't need a running WordPress site (just test data on the DB AFAIK), right? I have a hard time figuring out how to just unit test the plugin and not the plugin in the site's context. I could use Selenium for this, but right now I just want to focus on unit testing the plugin classes functionality, so that I don't break them with each update.
[ 0.013509216718375683, 0.00322400382719934, -0.005058822687715292, 0.004636058583855629, 0.007053124252706766, 0.010196520946919918, 0.007225364446640015, 0.005514541175216436, -0.018364552408456802, -0.004239997826516628, -0.006000777240842581, 0.010899843648076057, 0.0006543949712067842, ...
[ 0.5840052366256714, 0.1715928614139557, 0.10927242785692215, 0.2513445317745209, -0.032684437930583954, 0.06930428743362427, 0.2918011248111725, -0.20979773998260498, 0.2639606297016144, -0.8610260486602783, 0.09132103621959686, 0.7382466197013855, 0.05105370283126831, 0.1442747265100479, ...
I have a custom post type Event with multiple start and end dates for each Event. What I need then is a calendar similar to the calendar widget with links to these events. I need multiple multi day timespans that are not recurring regularly. The problem is that all plugins I could find only offer parts of the solution for me: * Events that only recur in regular intervals * Events that only offer one timespan for each event, forcing me to duplicate it for other dates of the same event. Is this such a weird problem that no plugin exists for this?
[ 0.01508701778948307, 0.011988176964223385, -0.00382376741617918, 0.012484626844525337, 0.026707034558057785, 0.002211721381172538, 0.007985268719494343, 0.05226662755012512, -0.02085115946829319, 0.0029504606500267982, -0.00018505238404031843, 0.01552735548466444, 0.0013462551869452, 0.003...
[ 0.4985964596271515, -0.09742651134729385, 0.22680623829364777, 0.08260778337717056, -0.22869367897510529, 0.005347168073058128, 0.4650985598564148, 0.5358244776725769, -0.5838139057159424, -0.6686922907829285, 0.3244960904121399, 0.16949379444122314, -0.2899405062198639, 0.5216367840766907...
Why massless particle can't exceed speed of light?
[ 0.027054939419031143, 0.04084896296262741, -0.02857159450650215, -0.04126066341996193, -0.0788479596376419, -0.07541072368621826, 0.03507615998387337, -0.1366586834192276, -0.02101558819413185, -0.13738688826560974, 0.0008948961040005088, 0.12193452566862106, -0.06109131872653961, 0.041492...
[ 0.3040638864040375, -0.07377678155899048, 0.04828862100839615, 0.4086592495441437, -0.14786314964294434, -0.300559937953949, 0.02537880465388298, 0.08894708007574081, -0.4831451177597046, -0.37132519483566284, 0.11303737014532089, -0.024166954681277275, -0.3919351398944855, 0.6564857959747...
In 1 I had the answer on how to put the translation of an acronym and used both English and Spanish version. The issue now is that I have acronyms that are only Spanish, others only English without translation and English with translation. The format for the document mandates that English is written emphasized while Spanish is in normal. Is there any way of automating it? I was thinking on having 2 list, English acronyms and only Spanish, then load them, assigning a different style and then merging them so in the list of acronyms they appear in order. The other solution I have found is on my acronym's list create two styles: \newcommand{\engstyle}[1]{\emph{#1}} \newcommand{\espstyle}[1]{#1} \newacronym{ENG}{ENG}{\engstyle{English acronym}} % \newacronym{ESP}{ESP}{\espstyle{Spanish acronym}} % But then in the list of acronyms they appear with the assigned style. And I want them without any style. Any suggestion? **Edit: What I want as exit.** Within the test _English acronym_ (ENG) is displayed and the Spanish one this way, Spanish acronym (ESP). List of Acronyms ENG English acronym ESP Spanish acronym Edit 2: I didn't realize that Nicola updated the package. Now the question is: Is it possible to set multiple setacronymstyle within the same doc?
[ 0.019704971462488174, 0.017578933387994766, -0.03149612247943878, 0.02635340765118599, -0.029932070523500443, 0.008365428075194359, 0.009183745831251144, 0.0323207825422287, -0.02301991917192936, 0.0072450703009963036, -0.017567859962582588, 0.0037866351194679737, -0.0014107362367212772, 0...
[ 0.4304474890232086, 0.17077642679214478, 0.04642180725932121, -0.20193129777908325, -0.3369159400463104, -0.4076480567455292, 0.18028008937835693, -0.02495172619819641, 0.1071692556142807, -0.5688588619232178, 0.05171316862106323, 0.46104496717453003, -0.6285292506217957, -0.53847777843475...
I wanted to print out a nomenclature with symbols (i.e. VARIABLE - DEFINITION - SYMBOL form). Below is the MWE \documentclass[11pt,onecolumn,draft,a4paper]{book} \usepackage{longtable} \usepackage{makeidx} \usepackage[acronym, toc, section, shortcuts, nopostdot, nogroupskip]{glossaries} \usepackage{glossary-long} \newglossary[slg]{symbols}{syi}{syg}{Nomenclature} \newglossary[ilg]{indices}{iyi}{iyg}{List of indices} \makeglossaries \newacronym{ECD}{ECD}{equivalent circuit diagram} \newacronym{RES}{RES}{renewable energy source} \newglossaryentry{Rx1} { name={\ensuremath{R_{x1}}}, description={Equivalent series resistance}, symbol={\ensuremath{\Ohm}}, sort=Rx1, type=symbols } \newglossaryentry{ref} { name={*}, description={Indicates reference value}, sort=ref, type=indices } \begin{document} \glsaddall \printglossary[type=\acronymtype, style=long, title=List of Abbreviations and Acronyms] \vspace{40pt} \printglossary[type=symbols, style=long4col, title=Nomenclature] \vspace{40pt} \printglossary[type=indices, style=long, title=List of indices] \end{document} The error is: ! Undefined control sequence. <argument> \Ohm l.5 ...etentrycounter[]{page}\glsnumberformat{1}}} % The control sequence at the end of the top line of your error message was never \def'ed... BLA-BLA-BLA It occurs only with 4cols styles one of which I need for printing out the symbols column. I also totally don't understand the error. Btw, it doen't occur with *3col styles.
[ 0.005314799956977367, 0.003124518785625696, 0.0014725547516718507, 0.008532976731657982, -0.024557892233133316, 0.01858275569975376, 0.007803899236023426, 0.024310050532221794, -0.010536188259720802, -0.009972753003239632, -0.011720515787601471, -0.0015319836093112826, 0.0019104904495179653,...
[ 0.453162282705307, 0.01433863490819931, 0.18024228513240814, 0.07713573426008224, 0.24080738425254822, 0.00027415560907684267, -0.14958615601062775, -0.2432863563299179, -0.07974197715520859, -0.526148796081543, 0.4268045127391815, 0.3415343463420868, -0.20507191121578217, 0.39124912023544...
I am creating a poster using `tikzposter`. But it seems to have a bug with `algorithm` packages. I use the following code but the borders of algorithm block span to the whole width of the poster: \documentclass[a0paper, landscape]{tikzposter} \usepackage{graphicx} \usepackage{color} \usepackage{tikz} \usepackage{amsmath,amsfonts,amsthm} \usepackage{algorithm} \usepackage{algorithmic} \title{Title} \usetheme{Autumn} \begin{document} \maketitle \begin{columns} \column{0.25} \block[titlecenter]{1. Introduction}{ \begin{algorithm}[H] \caption{Algorithm} \begin{algorithmic}[1] \STATE get the input \STATE do the computation \STATE output the results \end{algorithmic} \end{algorithm} } \column{0.50} \block{2. Some graphic}{ some graphic } \column{0.25} \block[titlecenter]{3. Efficiency Comparison}{ some results } \end{columns} \end{document} I tried `\begin{frame}[fragile also tried] \end{frame}` but didn't work. Is there a solution? I used the packages in this zip file that I got from fancytikzposter project page: https://bitbucket.org/surmann/tikzposter/downloads
[ 0.005734459962695837, 0.005689407233148813, 0.001164073240943253, 0.031961292028427124, -0.012655476108193398, 0.00925888679921627, 0.00975497905164957, 0.006351638585329056, -0.012415016070008278, -0.017154643312096596, -0.015744363889098167, 0.0013024737127125263, 0.0009248367277905345, ...
[ 0.7053658366203308, 0.09866247326135635, 0.4975563883781433, -0.12511853873729706, -0.08284471929073334, 0.24200811982154846, -0.1269041746854782, -0.07815565168857574, -0.21132436394691467, -1.1035925149917603, 0.42260754108428955, -0.030651606619358063, -0.27175429463386536, 0.1003810390...
In an earlier question More economic and robust command for abstract at the beginning of chapters, David Carlisle told that one should never use directly some TeX primitives. How can I distinguish which primitives are safe to use?
[ 0.004534980747848749, 0.04565930366516113, -0.017692742869257927, 0.023200679570436478, -0.0361129567027092, 0.035773489624261856, 0.01556801050901413, -0.0013415382709354162, -0.029766123741865158, -0.03129248321056366, -0.019120831042528152, 0.00666201813146472, -0.016386909410357475, 0....
[ 0.43867671489715576, 0.3754442036151886, -0.0742853507399559, -0.07254329323768616, -0.06542950868606567, -0.09580335021018982, 0.3742567300796509, -0.5048387050628662, 0.43163126707077026, -0.038595765829086304, 0.007324738893657923, 0.6049979329109192, -0.23803164064884186, -0.0817632675...
I recently started using the Amazon App store, and noticed that it copies the apps I already purchased or downloaded to other Android devices if I sign in using the same account. Is there any specified limit on the number of devices that this is supported on? If I have 10 Android devices can I buy an app once and install it on all of them?
[ -0.009087812155485153, 0.009570014663040638, 0.0013959718635305762, 0.022358357906341553, -0.005106633063405752, 0.007935972884297371, 0.009497423656284809, 0.032572224736213684, -0.02262086048722267, -0.022415902465581894, -0.0027092339005321264, 0.02353544346988201, 0.0194208025932312, 0...
[ 0.8326123356819153, 0.18597230315208435, 0.3716413676738739, 0.10117513686418533, 0.2745656371116638, 0.39373546838760376, 0.3130470812320709, 0.06649410724639893, -0.4406224489212036, -0.7507836222648621, 0.2044719159603119, 0.5853657126426697, -0.20271658897399902, -0.06240659952163696, ...
Suppose we have weekly data for some attribute (e.g. housing prices). Say that we have $500$ weeks worth of housing price data. Suppose some major event happened on week $256$. If we want to detect any significant changes in housing prices after week $256$ versus before week $256$, would it be better to use a longer time scale? Maybe convert weeks into months? What other models would you suggest?
[ -0.0038171301130205393, 0.03061525709927082, -0.02019997499883175, 0.01481406856328249, 0.03811235725879669, 0.009701904840767384, 0.009979575872421265, -0.02380324900150299, -0.014501146972179413, -0.0017927007284015417, -0.005362905096262693, 0.010614504106342793, -0.0009035876137204468, ...
[ 0.40474119782447815, -0.25089335441589355, 0.3188263177871704, 0.24378953874111176, 0.034132521599531174, 0.33366313576698303, -0.03262792155146599, 0.06910933554172516, -0.6844437122344971, -0.46600058674812317, -0.020324083045125008, 0.3963148891925812, -0.18007822334766388, 0.4203889369...
If you go to the developer tools menu in Settings, then tap Launch Tools and open Bad Behavior, there's an option to "Report a WTF Condition". What is a WTF Condition? Why would an app report one?
[ -0.02662479132413864, 0.019674096256494522, 0.01876678690314293, 0.01027629990130663, 0.02444487251341343, -0.005358985625207424, 0.015632495284080505, 0.034931108355522156, -0.025250433012843132, 0.014443072490394115, -0.02090187929570675, 0.025321586057543755, -0.0018481273436918855, 0.0...
[ 0.40475502610206604, 0.05388922244310379, 0.08680698275566101, 0.06091117486357689, 0.05717404559254646, -0.2052602767944336, 0.6822535991668701, -0.019029753282666206, -0.1764233112335205, -0.4885968267917633, 0.011222421191632748, 0.6504100561141968, -0.5071768760681152, -0.0726581960916...
I am not sure which one is correct when using `\intertext` or `\shortintertext`. Adding `\\\` before them or not. \documentclass[preview,border=12pt]{standalone} \usepackage{amsmath} \begin{document} \section*{Without a line break } \begin{gather*} E = mc^2 \intertext{Multiply both side by 2, we have} 2E = 2mc^2 \end{gather*} \section*{With a line break} \begin{gather*} E = mc^2\\ \intertext{Multiply both side by 2, we have} 2E = 2mc^2 \end{gather*} \end{document} ![enter image description here](http://i.stack.imgur.com/Nujuh.png) I don't see any differences. Or there is something wrong with my eyes. So the question is: _Is it necessary to give a`\\\` before `\intertext` or `\shortintertext`?_
[ 0.003199410391971469, 0.014794869348406792, -0.0140244010835886, 0.004939987789839506, -0.002865247428417206, 0.006135871168226004, 0.009319260716438293, -0.006563453935086727, -0.012787781655788422, -0.01934691146016121, -0.005812634713947773, 0.00761817954480648, -0.004562757443636656, 0...
[ 0.08264049142599106, 0.0942796915769577, 0.539034903049469, 0.017979376018047333, 0.20503614842891693, -0.2650284469127655, 0.3454863131046295, -0.4858195185661316, -0.07860281318426132, -0.7890686392784119, -0.1602293998003006, 0.492951363325119, 0.00829638633877039, -0.19963975250720978,...
Here's the deal I have an apple aluminium usb keyboard connected to my laptop with "windows" keyboard. And almost everything works great on my apple keyboard except three annoying problems. 1. Each keyboard must have a different layout, can I specify a layout for each one automatically or do I need to change manually every time ? 2. On the apple keyboard the key with the ">" and "<" is switched the key with "\" and "|". I know that I can remap keys with **xmodmap** , but is it possible to remap only to the apple keyboard since the "windows" keyboard is perfectly fine? 3. This is connected with the second, if can change cmd with alt, but only for the apple keyboard I'm using linux mint 17 with mate **Edit** I've managed to fix the second problem from the list, it seems it's a known bug from ubuntu https://bugs.launchpad.net/ubuntu/+source/linux/+bug/214786 After reading the comments on launchpad and this page to fix that I had to this: 1. Append the configuration line to the file /etc/modprobe.d/hid_apple.conf creating it if necessary: echo options hid_apple iso_layout=0 | sudo tee -a /etc/modprobe.d/hid_apple.conf 2. Trigger copying the configuration into the initramfs bootfile. sudo update-initramfs -u -k all **Edit 2** So this is the info I get with `setxkbmap -print` xkb_keymap { xkb_keycodes { include "evdev+aliases(qwerty)" }; xkb_types { include "complete" }; xkb_compat { include "complete" }; xkb_symbols { include "pc+pt(mac)+pt:2+inet(evdev)" }; xkb_geometry { include "pc(pc105)" }; }; If I try to set the macintosh layout only to my apple keyboard with this command: setxkbmap -device 11 -v pt -variant mac This is the result I get: xkb_keymap { xkb_keycodes { include "evdev+aliases(qwerty)" }; xkb_types { include "complete" }; xkb_compat { include "complete" }; xkb_symbols { include "pc+pt(mac)+inet(evdev)" }; xkb_geometry { include "pc(pc105)" }; }; If you notice on the line **xkb_symbols** I have this: include "pc+pt(mac)+inet(evdev)" Instead of this: include "pc+pt(mac)+pt:2+inet(evdev)" If I find a way of adding the option `pt:2` I could create a bash script that would run on startup to set the layout for my apple keayboard solving my first issue from the list.
[ 0.0051321047358214855, 0.009950448758900166, -0.009236792102456093, 0.014606735669076443, -0.028631413355469704, -0.010296545922756195, 0.006830668542534113, -0.008376747369766235, -0.017392948269844055, 0.023406341671943665, -0.007822244428098202, 0.0013922349316999316, 0.000160681316629052...
[ 0.2673502564430237, 0.04584270343184471, 0.6994141340255737, -0.12034342437982559, 0.20911665260791779, -0.07738198339939117, -0.04151696339249611, -0.27766138315200806, -0.16952629387378693, -0.5400280356407166, 0.17447160184383392, 1.1113717555999756, -0.16212353110313416, 0.111563220620...
I want a macro that somehow replaces whitespace (spaces, newlines, tabs?) from the start of a string. Something like mystring.replace(/^\s*/g, "") in JavaScript. Is that possible?
[ -0.008700519800186157, 0.020907729864120483, -0.002243120688945055, 0.0040012262761592865, 0.016823042184114456, -0.03289852663874626, 0.010904014110565186, 0.016994956880807877, -0.02536625973880291, 0.004717393312603235, 0.0013973555760458112, 0.0064650955609977245, 0.012197182513773441, ...
[ 0.2114376723766327, -0.09475978463888168, 0.2827881872653961, 0.01726066879928112, 0.37901797890663147, 0.281106561422348, -0.26668581366539, 0.20029768347740173, -0.20068106055259705, -0.6806125044822693, -0.048669684678316116, 0.48390069603919983, -0.6051976084709167, -0.1393843442201614...
I'm making a LaTeX presentation using beamer, and I'd like to prepare some slides for possible questions I can be asked for. I find how to ignore slides numbering in this post, but I don't like the result, which is roughly the following: > Slide 1/n Slide 2/n...Slide n/n [from here questions slide] Slide (n+1)/n > Slide (n+2)/n What I would achieve is roughly the following: > Slide 1/n Slide 2/n...Slide n/n [from here questions slide] Slide I Slide II How can I achieve it?
[ -0.00487799197435379, 0.01509794220328331, -0.01690845564007759, 0.009920037351548672, -0.009705125354230404, -0.00835866667330265, 0.010686993598937988, -0.009250173345208168, -0.014441221952438354, -0.012260498479008675, -0.006740120705217123, -0.015794355422258377, -0.01290555763989687, ...
[ -0.08903378993272781, 0.10260429233312607, 0.6256728768348694, -0.13938961923122406, -0.15026792883872986, 0.13834378123283386, 0.25807124376296997, -0.49682608246803284, -0.2616664469242096, -0.6923167109489441, 0.16727228462696075, 0.4167773127555847, 0.13490359485149384, 0.2984777390956...