text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I have two tables, `boundary` polygon table and `location` point table. I need to find all `boundary` polygons that contain at least 1 `location` point. What is the proper way to do this in PostGIS? I tried: SELECT * FROM boundary, location WHERE ST_Intersects(boundary.geom, location.geom) But this returns each combination of `boundary` and `location` that intersect, when all I want is the `boundary` returned once. For example, I got 76 results, but only have 10 `boundary` entries. Do you know a better way to do this?
[ 0.02106635272502899, 0.009510861709713936, -0.009330708533525467, 0.012098582461476326, 0.0011085402220487595, 0.021824335679411888, 0.008962469175457954, 0.025818996131420135, -0.01689661480486393, 0.012750274501740932, 0.00015814200742170215, 0.01392708346247673, -0.01040638517588377, 0....
[ -0.24720048904418945, 0.2891261875629425, 0.4725731909275055, 0.054415781050920486, -0.24577166140079498, 0.30984869599342346, 0.09218411892652512, -0.2143360823392868, -0.16271521151065826, -0.9461190700531006, 0.17538735270500183, 0.17211350798606873, -0.060346633195877075, 0.28646972775...
> In 2014, the organization of the “media-capital” prize in the business > journalism field is the main project focused on improvement the financial > literacy of the population. Or can you help me to structure this sentence correctly?
[ -0.00621999055147171, -0.0047766980715096, 0.010638270527124405, 0.02268124558031559, -0.005324556026607752, -0.011215392500162125, 0.01579534076154232, -0.0163580309599638, -0.021342918276786804, -0.0009494715486653149, -0.03660215809941292, 0.024127239361405373, -0.02420872449874878, -0....
[ 0.21327804028987885, 0.2580977976322174, 0.29495370388031006, 0.28852126002311707, -0.39761993288993835, 0.05863162875175476, 0.5404845476150513, -0.0032287687063217163, -0.17558768391609192, -0.4366013705730438, 0.12603628635406494, 0.5219705700874329, 0.04223591461777687, 0.0328829102218...
Open tank is partially filled with water and is divided into two equal parts by the wall. The wall has pipe that connects both sections. Obviously, water levels are equal from both sides (Fig.1). Then, water is pumped from right section of the tank to the left one with constant flow rate (Fig.2). Q1: Is it possible to calculate the difference in water levels knowing all values (volume, flow rate, pump capacity, etc.)? Q2: How the water levels difference will change if the tank will double its size (Fig.3)? Or it will be the same? ![enter image description here](http://i.stack.imgur.com/QJcRC.png)
[ -0.01871386170387268, 0.01917787827551365, -0.00006907235365360975, 0.012977932579815388, -0.0005534454248845577, -0.007018104195594788, 0.009139498695731163, 0.017057862132787704, -0.018682017922401428, -0.0036701131612062454, -0.004428869113326073, 0.017260778695344925, -0.0223062671720981...
[ 0.3735821545124054, -0.09099013358354568, 0.4033302068710327, -0.04237907752394676, -0.01010227669030428, 0.48023805022239685, -0.15672744810581207, -0.12847265601158142, -0.3183252513408661, -0.5608974695205688, -0.31981441378593445, 0.30606240034103394, -0.33278951048851013, 0.4584172666...
I looked on similar questions but in vain. The map is working fine but has a center of (0,0) as returned in the Firebug console not (-8, 32) that I want. I did the projection transformation.. Here is my code : ![enter image description here](http://i.stack.imgur.com/uy51X.png)
[ -0.014064338058233261, 0.007254267577081919, 0.0014570927014574409, 0.007786863017827272, -0.012515279464423656, -0.01880849152803421, 0.004746570251882076, 0.007758614607155323, -0.019678018987178802, -0.0028378483839333057, -0.005697100423276424, 0.010692504234611988, -0.019797101616859436...
[ 0.08905598521232605, 0.02971121296286583, 0.6212188601493835, 0.07494127005338669, 0.1495164930820465, 0.1455417424440384, 0.29613858461380005, -0.2992752194404602, -0.30323469638824463, -0.846916913986206, 0.009243562817573547, 0.522984504699707, 0.0625150054693222, 0.09494481980800629, ...
I was looking for a quick and dirty method to output a breadcrumb navigation on a WP site, without requiring the installation of a plugin and also leveraging the built-in WP menu. Here's what I came up with. I'm curious to know if there is a better solution I am missing, and/or if there is anything obviously wrong with doing it this way. So far in my testing it works quite well. class BreadcrumbWalker extends Walker_Nav_Menu { private $i_current_page_id; private $i_depth; private $a_output; function __construct() { // sets our current page so we know when to exit $this->i_current_page_id = get_queried_object()->ID; } function start_lvl(&$output, $depth=0, $args=array()) { // increment the depth every time a new ul is entered $this->i_depth++; } function end_lvl(&$output, $depth=0, $args=array()) { // decrement the depth when we exit a ul $this->i_depth--; } function start_el(&$output, $item, $depth=0, $args=array()) { // if this value is zero, we're starting a new branch if($item->menu_item_parent == 0) { // reset the output array and depth counters $this->a_output = array(); $this->i_depth = 0; } // if we haven't set the representative menu item for this depth, do so if(!isset($this->a_output[$this->i_depth])) { $this->a_output[$this->i_depth] = '<a href="' . get_permalink($item->object_id) . '">' . $item->title . '</a>'; } } function end_el(&$output, $item, $depth=0, $args=array()) { if($this->i_current_page_id == $item->object_id) { // check to see if this is our last item, if so display the breadcrumb if($this->i_depth > 0) { // but only show it if we actually have a breadcrumb trail $this->display_breadcrumb(); } } else { // if not, unset the item for this depth since this isn't what we're going to display unset($this->a_output[$this->i_depth]); } } function display_breadcrumb() { // implode our array into a string echo implode(' &raquo; ', $this->a_output); } } Basically what I'm doing is using the Walker_Nav_Menu methods to set an output array that uses the depth as its key, and once the end_el method is called, if it's not the ID of the page that's the end of the breadcrumb, it unsets the item at that depth and keeps iterating over the array. If it _is_ the correct ID, it calls the display_breadcrumb method to output the menu. Thanks in advance for any input. I feel like there's probably a better way to do this, but at the moment am stumped as to what that might be. Cheers.
[ -0.0035388499964028597, 0.017819644883275032, -0.0017940583638846874, 0.017577312886714935, -0.021811535581946373, -0.0008249660022556782, 0.006395355332642794, -0.00828627310693264, -0.020962903276085854, -0.009787999093532562, -0.007795192766934633, 0.019451312720775604, 0.0112121282145380...
[ -0.09234267473220825, -0.18630889058113098, 0.08842162787914276, -0.3370057940483093, 0.14989952743053436, 0.4497794806957245, 0.2277834415435791, -0.10273845493793488, -0.1765027791261673, -0.8473042845726013, 0.2465701848268509, 0.24222376942634583, -0.5496535897254944, -0.22954176366329...
I have a 9x9 grid and a polygon of municipalities of an extensive area of Brazil. My goal is to have a field indicating the biggest municipality per grid cell. I've been trying with intercept without good results. I know there are similar questions but i cant figure it out how to handle it in QGIS. Here is a screenshot of what i'm talking about: ![enter image description here](http://i.stack.imgur.com/ULVk3.png) I was thinking that maybe with intersecton and then spatial union joining with maximum but it didnt work. I want to get the final grid with a field indicating the biggest municipality inside each grid cell. Thank you all in advance Sergio Now
[ 0.00008870937745086849, -0.0027603055350482464, -0.004741173703223467, 0.016371097415685654, -0.03553033992648125, -0.012735903263092041, 0.0069289430975914, 0.024705076590180397, -0.01317250169813633, 0.001543708611279726, 0.0037802783772349358, 0.011579230427742004, -0.01451765839010477, ...
[ 0.13015994429588318, -0.2196619063615799, 0.7397069931030273, 0.2721906006336212, -0.4321311116218567, 0.06761784106492996, -0.07847071439027786, 0.17797023057937622, -0.2258235514163971, -0.9778071045875549, 0.38574373722076416, 0.2754330635070801, -0.13739746809005737, 0.0325175598263740...
If I'm going to use the free tier, why do I need to enter my credit card information? How can I be sure I won't be charged for it while I'm using it?
[ 0.01759052649140358, 0.016100876033306122, 0.0026740620378404856, -0.015491398982703686, 0.000021212479623500258, 0.025669114664196968, 0.010722016915678978, -0.020517468452453613, -0.015780214220285416, -0.00502013461664319, -0.008567530661821365, 0.010689714923501015, -0.008734660223126411...
[ 0.4728700518608093, 0.28194060921669006, 0.2716708481311798, 0.28918829560279846, 0.41584712266921997, -0.23785284161567688, 0.19219006597995758, 0.009456495754420757, -0.2780279815196991, -0.22798795998096466, 0.40789148211479187, 0.4378912150859833, 0.1061018779873848, -0.008477252908051...
in Google adsense, if a user from any territory revisits page impression is getting updated but that eCPM is not getting updated its showing zero only, does it work only for unique visitors? Thanks Sneha
[ -0.0025093834847211838, 0.029500378295779228, -0.006839663255959749, 0.04596206918358803, -0.035355620086193085, -0.0015391363995149732, 0.013977611437439919, -0.02715137228369713, -0.026184750720858574, -0.04370125010609627, -0.014054709114134312, 0.025726819410920143, 0.02138729952275753, ...
[ 0.14218126237392426, -0.07262187451124191, 0.026762105524539948, 0.45515045523643494, -0.49361106753349304, -0.14352093636989594, 0.23613803088665009, -0.05741244927048683, -0.05696715787053108, -0.510683000087738, 0.2660011947154999, 0.22048397362232208, -0.47619467973709106, -0.162100300...
I use the following code in LaTeX to input the code Matlab there but I want to change the color of the background: \usepackage{textcomp} \usepackage{listings} \lstdefinestyle{customc}{ belowcaptionskip=1\baselineskip, breaklines=true, frame=L, xleftmargin=\parindent, language=Matlab, showstringspaces=false, basicstyle=\footnotesize\ttfamily, keywordstyle=\bfseries\color{green!40!black}, commentstyle=\itshape\color{purple!40!black}, identifierstyle=\color{blue}, stringstyle=\color{orange}, } \lstdefinestyle{customasm}{ belowcaptionskip=1\baselineskip, frame=L, xleftmargin=\parindent, language=[x86masm]Assembler, basicstyle=\footnotesize\ttfamily, commentstyle=\itshape\color{purple!40!black}, } \lstset{escapechar=@,style=customc} How can I do that to look like that in LaTeX report? ![output](http://i.stack.imgur.com/5Ilun.jpg)
[ 0.024515874683856964, 0.008108233101665974, -0.0057303118519485, -0.001220080186612904, 0.005023118108510971, 0.01588650792837143, 0.00770600838586688, 0.005436757579445839, -0.01194386463612318, -0.008085982874035835, -0.01107783429324627, -0.005330451764166355, 0.009209150448441505, 0.01...
[ 0.1963852047920227, 0.05519624799489975, 0.5313405990600586, -0.2980384826660156, 0.1663837879896164, 0.17669861018657684, 0.22788184881210327, -0.132968470454216, -0.2566308379173279, -0.6214966177940369, -0.25638294219970703, 0.3852570950984955, -0.2679249346256256, -0.18828582763671875,...
I unioned 2 layers into 1 layer, and it created sliver polygons. I used the elimination tool to merge them with neighboring polygons. As as result, it merges all sliver polygons to only one polygon (selected in the picture). What I need, is that, every sliver polygon to be added to its neighboring polygon. How that could be achieved? ![enter image description here](http://i.stack.imgur.com/U98oC.png)
[ 0.0011091084452345967, 0.011477919295430183, -0.007875140756368637, 0.01823335327208042, -0.005320006515830755, -0.019611524417996407, 0.007083368021994829, 0.013483664952218533, -0.018347911536693573, 0.010775561444461346, -0.005287283565849066, 0.013417287729680538, -0.01257513090968132, ...
[ 0.3104376196861267, -0.14445734024047852, 0.2256215214729309, 0.06704031676054001, 0.10101927816867828, 0.3020480275154114, -0.10649557411670685, -0.3086574971675873, -0.4876094162464142, -0.7263490557670593, 0.023228254169225693, 0.014915774576365948, -0.4381566047668457, 0.26945230364799...
What is a good app for blocking incoming SMS? I want it to be easy and able to add numbers from contacts and not require to many privileges (probably only some storage and intercepting SMS).
[ 0.023194046691060066, -0.00595506839454174, -0.02294892817735672, 0.017992136999964714, 0.0361250676214695, -0.01510756928473711, 0.012830541469156742, 0.05166175588965416, -0.044878214597702026, -0.026183607056736946, -0.03014136478304863, 0.016280589625239372, 0.010433575138449669, 0.030...
[ 0.15276125073432922, 0.19666743278503418, 0.2890062630176544, 0.22504079341888428, 0.008537263609468937, 0.11695001274347305, 0.09361583739519119, 0.1916772723197937, -0.1408541351556778, -0.26391616463661194, 0.20451031625270844, 0.5609972476959229, -0.4644259214401245, -0.237307429313659...
Perhaps this is just an overly futurist version of those "does programming have a future" topics, but the questions been on my mind a lot lately. Maybe you've seen this future timeline or things like it, and one of the predictions it makes is that by the year 4000, computer science will be dead as a science. That is to say, that the hardware will have reached the limits of physics and that all the problems of algorithms and software will be solved. Is that a thing that can happen? I could see how, theoretically, physics might reach an end point where you've found all the natural laws there are. But I'd always considered computer science to in some ways be more like economics, a field about decisions and trade-offs. Or could we someday discover all the algorithms, and then it's just a matter of software engineering to select the most appropriate one to use?
[ -0.0022672810591757298, 0.008526379242539406, -0.007211789488792419, -0.0022066570818424225, -0.01145070232450962, -0.019993219524621964, 0.0037946198135614395, -0.002122994977980852, -0.006220749579370022, -0.019011052325367928, -0.00864991545677185, 0.007265964522957802, 0.0097800474613904...
[ 0.6303459405899048, 0.08570031821727753, 0.24911928176879883, 0.41932427883148193, 0.3144663870334625, -0.04129704087972641, 0.07207714021205902, 0.46601980924606323, -0.5876535177230835, -0.26914963126182556, 0.16761013865470886, 0.32926082611083984, 0.020768405869603157, 0.51182287931442...
Is there any way to make _Mathematica_ stop for some miliseconds? I know about the `Pause[]` but it seems it only works with seconds.
[ 0.00743948994204402, 0.011879168450832367, -0.030497126281261444, 0.02259872294962406, 0.007532728835940361, 0.031423572450876236, 0.010973406955599785, 0.016809629276394844, -0.032663341611623764, 0.029317421838641167, -0.017447777092456818, 0.01974625140428543, -0.016427071765065193, 0.0...
[ 0.21836280822753906, 0.14886613190174103, 0.20795927941799164, 0.379498690366745, 0.026259643957018852, -0.14123782515525818, 0.19164416193962097, 0.4060233533382416, -0.12638279795646667, -0.13477583229541779, 0.344023734331131, 0.22546426951885223, 0.3294495940208435, -0.0785579606890678...
Obviously in hardcore mode there is no HUD, so I cannot check fire mode there. But are there any visual indicators on the weapons themselves which would hint that?
[ -0.025689512491226196, 0.005499252118170261, 0.0008868096047081053, 0.005131544079631567, -0.033646952360868454, -0.024240348488092422, 0.01283927820622921, -0.033439718186855316, -0.031947482377290726, 0.027926841750741005, -0.017701325938105583, 0.02071339450776577, -0.016167713329195976, ...
[ 0.8359001278877258, 0.26840662956237793, -0.14825740456581116, 0.0860980898141861, -0.1273757517337799, -0.15822207927703857, 0.6552976965904236, -0.160733163356781, -0.3197152316570282, 0.10314896702766418, 0.19686412811279297, 0.6566482186317444, 0.09376320987939835, -0.4275933504104614,...
So ive been trying to download these big files that are around 500mb-2gb but after like 10 minutes it says download failed. Im using Wi-Fi not data. Im trying to download the file in google chrome. I have dolphin browser should I try using that.
[ -0.03457991033792496, -0.01846284791827202, -0.01738625206053257, 0.0003739739186130464, 0.021038560196757317, -0.03841767460107803, 0.010523946024477482, -0.02509612962603569, -0.02860989235341549, -0.027686353772878647, -0.01467886008322239, 0.0019599315710365772, 0.03420869633555412, 0....
[ 0.42713016271591187, -0.019004523754119873, 0.3734051287174225, -0.13698820769786835, -0.16701722145080566, -0.12486965954303741, 0.07191221415996552, 0.180107980966568, -0.24487291276454926, -0.7396363019943237, 0.2121722251176834, 0.6857547163963318, -0.04789075627923012, 0.1557462811470...
I coloned Geoadmin repo from github. I tried tu run this localy but no sucess. I have my own rc_ file but still it's not working. Anyone can help me?
[ -0.025258779525756836, 0.008548998273909092, -0.003174584126099944, 0.026598453521728516, -0.0023831308353692293, -0.02082039788365364, 0.011870183050632477, -0.022049929946660995, -0.03365402668714523, -0.0025344255845993757, -0.012827340513467789, 0.013783496804535389, 0.002815520390868187...
[ 0.552118182182312, 0.4291575253009796, 0.38418081402778625, 0.2157503068447113, 0.05824409797787666, -0.04308764263987541, 0.22460433840751648, 0.6944807171821594, -0.2277614027261734, -0.48974600434303284, 0.0725996345281601, 0.5867803692817688, -0.1929747313261032, 0.6425516605377197, ...
I've accepted two side quests in the game; in both cases asking about rewards was a bit too douchebaggey for me but… if there's something in it for me in addition to karma… I wouldn't _mind_ it, you know what I mean? Do I give up on rewards, or part of it, if I ask about rewards?
[ -0.0034401509910821915, 0.015424528159201145, 0.016464537009596825, 0.001458080019801855, 0.014080499298870564, 0.011419848538935184, 0.006349772214889526, 0.0060636671259999275, -0.027981633320450783, 0.0046925307251513, -0.002020459156483412, 0.019251659512519836, -0.018609384074807167, ...
[ 0.17765218019485474, -0.06779605895280838, -0.18254703283309937, 0.018477175384759903, -0.29017987847328186, -0.04070398211479187, 0.4029064178466797, 0.10485398769378662, -0.32227301597595215, -0.15487363934516907, 0.3774496614933014, 0.3746495544910431, 0.08383844047784805, -0.0724774673...
I've come across a *.apk where permissions are needed to be set to the *.db file. But I don't know what's it meaning. Help?
[ -0.025531355291604996, 0.03632139042019844, 0.02550712786614895, 0.03829018026590347, -0.01611896976828575, 0.05160195380449295, 0.011523066088557243, 0.02416260540485382, -0.02765478566288948, 0.013413473963737488, -0.022777890786528587, 0.021609768271446228, 0.015240035951137543, 0.02413...
[ 0.4088906943798065, 0.1125577986240387, 0.058109696954488754, 0.2807278037071228, 0.09287304431200027, 0.19527223706245422, 0.4762811064720154, -0.2840200364589691, -0.15115119516849518, -0.1731826514005661, -0.1749740093946457, 0.27031317353248596, -0.1687844842672348, 0.21534326672554016...
Should I use the second "will" in constructions like this one: "it will definitely help you and **will** make the text more readable" And should I write "to" before every infinitive in enumeration, or only before the first one, e.g. "it helps to develop and test" or "it helps to develop and **to** test"? Could this example be influenced by the fact that "help" can go with to-invinitive and bare infinitive, and with other verbs the rule would be different?
[ 0.015642477199435234, 0.010484742932021618, -0.01919625885784626, 0.009937671944499016, -0.03558121621608734, 0.015817448496818542, 0.0076462323777377605, 0.010251197963953018, -0.01204544585198164, 0.013753511942923069, -0.0178742166608572, 0.0032437548507004976, -0.010264584794640541, 0....
[ 0.05566907301545143, -0.16647446155548096, 0.1374971568584442, -0.24076952040195465, -0.22280442714691162, 0.03511277586221695, 0.44528064131736755, -0.381195604801178, 0.18345679342746735, -0.7272415161132812, 0.015135449357330799, 0.806698739528656, 0.13647645711898804, -0.57629162073135...
I have been told that the a freezer will freeze food slower if: a) the product load is greater then the capacity b) the capacity is greater then then the load. In a) I think this means if a freezer can freeze 2kg of food in 12 hours, using more food will cause the freezer to work slower and so freeze slower. Is this the case? In b) surely if the capacity is greater then the load that means it will freeze fast or at a normal rate? Or do you think it is referring to a much larger capacity i.e. if 1kg of food is placed inside a deep freezer, then it will freeze slower then if 1kg of food was placed inside a smaller freezer. If this is the case, why? Not sure what is right. Can somebody please explain. Thanks
[ 0.014399927109479904, 0.014228619635105133, -0.019812967628240585, 0.022750195115804672, -0.008816427551209927, -0.01991618052124977, 0.00847691297531128, 0.0016700215637683868, -0.010741513222455978, -0.002678965451195836, -0.011837362311780453, 0.013013279996812344, 0.00762682082131505, ...
[ 0.48083093762397766, -0.07599709928035736, 0.051392730325460434, -0.08993830531835556, -0.12764351069927216, -0.0042993901297450066, 0.26336896419525146, 0.0830259695649147, -0.7564402222633362, -0.48593011498451233, -0.09426160156726837, 0.08737090975046158, -0.017909105867147446, 0.12443...
`$user->user_registered` displays "Registered since: 2014-08-18 07:25:22" I just want to show the date, not the hour. How can I hide the hour and just show the date? thank you
[ -0.00013775283878203481, 0.0074876113794744015, -0.017147650942206383, 0.015258003026247025, -0.0518275685608387, -0.01862371526658535, 0.012158839963376522, -0.017010021954774857, -0.03590526059269905, 0.017298299819231033, -0.024881724268198013, 0.0019879706669598818, -0.001287881867028772...
[ 0.6921013593673706, 0.35909005999565125, 0.7758195400238037, 0.2029760181903839, -0.04041695222258568, -0.1040458232164383, 0.6233379244804382, 0.36451974511146545, -0.5279237627983093, -0.35705289244651794, 0.12077518552541733, 0.17138546705245972, 0.18347179889678955, 0.4437810182571411,...
I need to support a legacy system that's running CentOS 5.4. There are errors coming up during the install and walking through the admin on the other end is quite tedious and inefficient. Does anyone know if there's a way where I can take over the install session remotely somehow with vnc/teamviewer or ssh? The method chosen will need to support some sort of authentication as well. I don't mind doing the installation in text mode either.
[ -0.004430396016687155, -0.0013535928446799517, -0.013371267355978489, 0.004596538841724396, -0.036749180406332016, -0.0114352498203516, 0.007783873938024044, -0.006331239361315966, -0.01810956373810768, -0.006643961183726788, -0.021635690703988075, 0.0050720339640975, 0.012310628779232502, ...
[ 0.49892309308052063, 0.2659015357494354, 0.25333601236343384, 0.026108458638191223, 0.0989413782954216, -0.23413436114788055, 0.5001823306083679, 0.05357978492975235, -0.07496121525764465, -0.5579731464385986, 0.054483115673065186, 0.6530624628067017, 0.08929160237312317, -0.10089020431041...
I have a 3.0.1 site with `MULTISITE` enabled and would like one of the sites to live at `/blog`, but when I try to create a new site with that path, I get this error: The following words are reserved for use by WordPress functions and cannot be used as blog names: page, comments, blog, files, feed How can I get a site at `/blog`?
[ -0.0030406720470637083, 0.007648388389497995, -0.005652841180562973, 0.029711294919252396, -0.015344375744462013, 0.01997831091284752, 0.009098042733967304, 0.023716481402516365, -0.01695152558386326, -0.03326372429728508, -0.018199512735009193, 0.014371803030371666, 0.001566727296449244, ...
[ 0.5526913404464722, 0.09368475526571274, 0.3959047794342041, 0.09194359183311462, 0.1193840503692627, 0.12836498022079468, 0.5626351237297058, 0.3464626371860504, 0.017801035195589066, -0.6435362696647644, 0.038998305797576904, 0.20186881721019745, -0.28569936752319336, 0.4698927700519562,...
I'm looking to perform a binary classification using random forests, but I do not quite understand how to minimize the entropy of the data / what tests I should run on the nodes to do so. I'm fairly new to the field, so I'd greatly appreciate any pointers or resources that could lead me in the correct direction. Thanks!
[ 0.022478414699435234, 0.007941772229969501, -0.0044911447912454605, 0.014170875772833824, -0.039215270429849625, 0.002356158336624503, 0.0061009787023067474, 0.031027063727378845, -0.026607342064380646, 0.008233766071498394, -0.01162628922611475, 0.014676818624138832, -0.014165925793349743, ...
[ 0.375620573759079, 0.09423362463712692, -0.10030528903007507, 0.31907784938812256, 0.2456750124692917, 0.09358664602041245, 0.290993332862854, 0.32775425910949707, -0.1476031094789505, -0.4247392416000366, 0.22037747502326965, -0.06057833880186081, 0.12593874335289001, 0.19381126761436462,...
How can I mount a Windows partition so that the files within it don't have execution permission? I mount a Windows partition using: sudo mount /dev/sda3 win `win` is a folder in my home dir. This of course works. But files in the mounted partition are given execute permission, or to be specific, `777`. How to mount the partition so that files are given `666` or other permission?
[ -0.01028390135616064, 0.001021869946271181, -0.012663712725043297, 0.011800709180533886, 0.0017364880768582225, -0.0038106071297079325, 0.011970282532274723, -0.003172062337398529, -0.020873503759503365, -0.017803581431508064, -0.014237994328141212, -0.0010423398343846202, -0.015158185735344...
[ 0.17655624449253082, 0.09885920584201813, 0.3102826774120331, -0.26183101534843445, -0.020417461171746254, -0.28663620352745056, 0.3345172107219696, -0.39135077595710754, 0.03078492544591427, -0.7917547821998596, -0.12569554150104523, 0.8971633315086365, -0.06206921860575676, 0.05011528730...
How can I get the IP address, MAC address, etc. of a user connected to a Wi-Fi hotspot? ( Using Karbonn A21 Phone running Android 4.0.4 with minor changes to the skin)
[ -0.005702248774468899, 0.01193283125758171, -0.00888501014560461, -0.013040968216955662, -0.05179182440042496, 0.033222585916519165, 0.011991286650300026, 0.0027904438320547342, -0.0164810698479414, 0.018577633425593376, -0.006274787243455648, 0.023107657209038734, -0.03179672732949257, 0....
[ 0.26855456829071045, -0.14420953392982483, 0.27607011795043945, 0.41111940145492554, 0.30835625529289246, -0.05989740043878555, 0.6272267699241638, -0.1388084888458252, 0.10156416147947311, -0.6285005211830139, -0.14466780424118042, 0.07907276600599289, -0.10449836403131485, 0.020402874797...
I am working with some data that shows land management. There are many places where corners "touch". I would like to dissolve my dataset so that features are grouped (dissolved) with those that they share a touching corner. Each section (square) is a unique feature (row). In the image below I've circled the features that I would like to group in this dissolve. I am using ArcGIS and have an editor and SA license. ![enter image description here](http://i.stack.imgur.com/mk9YM.jpg)
[ 0.00916407722979784, 0.0007529156864620745, -0.005552536342293024, 0.0320824570953846, -0.024143563583493233, -0.0027087561320513487, 0.006580754648894072, 0.017275067046284676, -0.015622591599822044, 0.004692394752055407, -0.0034002731554210186, 0.009210804477334023, -0.013517527841031551, ...
[ 0.19680596888065338, 0.04706110060214996, 0.45228084921836853, -0.05073822662234306, -0.020999187603592873, -0.16242985427379608, 0.22022531926631927, -0.31452882289886475, -0.2829820215702057, -0.7114282250404358, 0.021012065932154655, 0.01789071597158909, 0.06038679927587509, 0.290196090...
How is it possible that a higgs at ~125 GeV can decay into 2 W bosons @ ~ 80 GeV a piece (for example)? Shouldn't a particle only be allowed to decay to lighter particles + energy? Diagram copied from this question ![enter link description here](http://i.stack.imgur.com/eljaD.jpg)
[ -0.003989985212683678, 0.014698442071676254, -0.02290944755077362, 0.01074125524610281, -0.004451057408004999, -0.02931051515042782, 0.009404248557984829, -0.006165515165776014, -0.02082221955060959, -0.017924843356013298, 0.001058238442055881, 0.00860755518078804, -0.006935654208064079, 0...
[ 0.22493195533752441, -0.07937705516815186, 0.41670721769332886, -0.31166884303092957, -0.10669268667697906, -0.16098381578922272, 0.12129681557416916, -0.6441418528556824, -0.4942377507686615, -0.13784460723400116, 0.014318124391138554, 0.1905408650636673, -0.2999112010002136, 0.4184203743...
I need to change this word "Dashboard" in the top menu in wp-admin If anyone can supply the code for the functions.php file or point me in the direction I would greatly appreciate it!
[ 0.004839660134166479, -0.003164596389979124, -0.006343884859234095, 0.029478508979082108, 0.003906471189111471, -0.03316783159971237, 0.012635196559131145, 0.030321309342980385, -0.019776204600930214, 0.017743416130542755, -0.011503026820719242, 0.010240110568702221, -0.0007122213137336075, ...
[ 0.1831417679786682, 0.26726025342941284, 0.42985999584198, 0.40352147817611694, 0.17568151652812958, -0.19072513282299042, 0.34572359919548035, 0.7038188576698303, 0.13341237604618073, -0.899041473865509, 0.39757639169692993, 0.6601760983467102, 0.3453608751296997, -0.15221327543258667, ...
I'm using the wrapfigure environment. When the wrapfigure goes to the bottom of the page, there is blank space included on the next page where the figure would be. The problem is illustrated in the attached figure and mwe. How do I fix this? ![The problem: the caption goes to the end of the page \(fine\), but blank space continues onto the next page \(not cool\)](http://i.stack.imgur.com/n4jZz.png) I have prepared a mwe, but it's not very minimal (needed to be multiple pages to show the problem). If you want to compile this, you'll need to replace the .pdf file with some figure of your choice. \documentclass[11pt]{article} \usepackage{placeins} \usepackage{wrapfig} \usepackage{graphicx} \begin{document} "Neque porro quisquam est qui dolorem ipsum quia dolor sit amet, consectetur, adipisci velit..." "There is no one who loves pain itself, who seeks after it and wants to have it, simply because it is pain..." Lorem ipsum dolor sit amet, consectetur adipiscing elit. Quisque et dui quis velit sollicitudin egestas. Phasellus dictum, libero sed pellentesque pulvinar, lectus arcu fermentum libero, vel ultricies augue eros vel ligula. Proin tempor velit id dolor sollicitudin pharetra. Duis consectetur metus et quam congue commodo vitae et velit. Quisque laoreet nibh ac lorem porta id vulputate lorem vestibulum. In condimentum nulla convallis orci commodo sit amet scelerisque tortor dapibus. Nam rhoncus, ipsum ut adipiscing ullamcorper, libero odio vehicula neque, ut sagittis turpis velit vitae velit. Ut sit amet ipsum ac leo cursus dictum vel eget augue. Proin non sapien felis. Aenean nec volutpat est. Morbi feugiat velit sed leo laoreet mattis. Etiam blandit metus ut dui posuere vehicula. Donec id eros at enim vestibulum feugiat. Duis quam est, malesuada sagittis porttitor pretium, lacinia vel augue. Aliquam at eros purus, eu tempus sem. Ut iaculis, augue et accumsan tempus, sem tortor rhoncus massa, at imperdiet tellus elit eget urna. Pellentesque risus leo, congue in imperdiet id, interdum id ante. Pellentesque convallis varius mattis. Suspendisse potenti. In auctor faucibus rhoncus. Vestibulum scelerisque malesuada metus, eu ultricies urna posuere vel. Etiam in sem eu mi semper condimentum at eget quam. Nullam libero dolor, tristique eget tristique id, gravida ut nisi. Nam ornare mi et est hendrerit gravida fringilla sapien suscipit. Etiam erat purus, porttitor nec bibendum venenatis, elementum ac urna. Praesent pulvinar imperdiet ipsum in placerat. Ut euismod purus quis libero iaculis egestas. Integer vel erat eros, tincidunt viverra orci. Integer ante risus, consectetur quis lacinia at, varius ut lorem. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Vestibulum porttitor orci in ligula imperdiet id lobortis nulla viverra. Nulla eget enim vitae purus placerat malesuada. Maecenas tincidunt mattis magna, sit amet gravida mi feugiat eu. Vestibulum sed diam nulla. Proin velit augue, posuere vitae cursus non, lobortis sed nibh. Nullam aliquet molestie ante, at dictum libero posuere eget. Fusce a libero ac odio vestibulum malesuada. Curabitur luctus nunc in leo sagittis ut pretium leo egestas. Morbi vel ligula ligula. Donec nulla ipsum, placerat id posuere ut, condimentum sit amet augue. Nunc vitae venenatis felis. Ut orci eros, commodo vitae rutrum vel, ultrices id orci. Aliquam egestas velit eget tortor dignissim eu aliquam neque viverra. Morbi sodales rhoncus feugiat. Ut ac erat et nisl adipiscing ornare porttitor fringilla est. Donec eleifend metus et nulla ultricies ultricies suscipit quam euismod. Suspendisse potenti. Nullam eget tristique erat. Fusce sed vestibulum quam. Vivamus aliquam, urna vel mattis dapibus, nisi lacus hendrerit est, vitae venenatis ipsum nisi vitae lectus. Integer laoreet enim non metus aliquet sagittis. Nam semper consequat eros, id ultricies risus placerat non. Donec risus nunc, porttitor id placerat sit amet, varius nec libero. Cras vel dolor magna. Aenean dui ante, posuere in vehicula sed, mattis viverra mi. Integer quis orci et eros gravida consequat nec quis metus. Donec dictum augue quam. Praesent neque urna, laoreet eget lobortis quis, pretium sit amet neque. Donec egestas molestie neque, vel egestas enim luctus eu. Nullam diam magna, suscipit ac ornare vel, vehicula et lectus. Ut eu mi mi. Morbi blandit, leo in tincidunt posuere, ante nisl dictum justo, vitae feugiat erat sapien quis dui. Duis lacus libero, fermentum nec malesuada sed, elementum eu nisl. Integer aliquam metus a lacus tincidunt condimentum. Pellentesque pulvinar porttitor tristique. Cras ante dui, volutpat vel lacinia iaculis, accumsan a tellus. Etiam risus quam, sodales volutpat feugiat vel, euismod ut odio. Duis porttitor lacus urna, vel blandit turpis. Vivamus gravida ultrices nulla in viverra. Aliquam vulputate posuere ipsum, ut porta quam consequat eget. Etiam pulvinar, quam sed auctor semper, metus sapien consequat dolor, id interdum ante nisi id tellus. Nulla iaculis augue quis lacus hendrerit quis euismod mauris gravida. Fusce eu neque et tellus aliquet vulputate nec ut magna. Aenean pharetra molestie lacus nec aliquam. Cras at ultrices augue. Sed venenatis, risus vel gravida accumsan, metus dui blandit enim, nec sollicitudin magna ligula at nibh. Sed elementum ipsum vitae odio facilisis feugiat. Vivamus ac felis est. Etiam massa nunc, imperdiet in porttitor non, placerat sit amet felis. Sed elit urna, scelerisque non rhoncus id, consequat id tortor. Quisque adipiscing fermentum dui, in vehicula nibh facilisis convallis. Cras condimentum adipiscing eros, nec tempor justo consectetur placerat. Sed accumsan, lacus id rhoncus consectetur, libero orci euismod erat, non sollicitudin arcu orci quis velit. Curabitur vitae sollicitudin diam. Duis hendrerit augue sit amet orci dictum in mattis augue volutpat. Vivamus nec sagittis tellus. Etiam felis felis, pretium sed luctus ac, euismod vitae nisl. Phasellus vestibulum, odio sed condimentum ultrices, enim ante interdum nisi, id sodales augue lectus vitae massa. Curabitur non convallis velit. Etiam eget diam et nisi euismod vehicula. Donec dignissim, nisi vitae porta scelerisque, felis leo sodales mauris, quis condimentum purus turpis dapibus arcu. Mauris molestie fermentum risus sit amet lacinia. Maecenas sollicitudin mollis purus ut consectetur. Maecenas et ligula quam. Morbi eget diam nisi, sed luctus arcu. Praesent id libero ante, nec dapibus dolor. Vivamus egestas, nunc id malesuada mollis, nisl neque condimentum lorem, non accumsan enim lorem ac neque. Morbi in ligula magna, vehicula interdum metus. Sed eget lorem nec nibh ornare suscipit eu nec nunc. Nulla est diam, tempus vitae egestas vitae, egestas vel dui. Integer ac libero diam. Mauris sit amet arcu eu neque feugiat iaculis. Aenean non nulla eu erat malesuada fringilla sed non justo. Curabitur non libero orci, et egestas elit. Sed gravida est sit amet ligula pretium non volutpat purus dictum. Quisque eros leo, consectetur vulputate hendrerit vitae, semper sed est. Sed malesuada luctus massa sed posuere. Quisque blandit orci nec libero placerat varius. Donec luctus nulla eu sapien consectetur sit amet scelerisque turpis hendrerit. Fusce sed mi augue. Donec tristique, ante nec rhoncus auctor, sapien lacus varius metus, eu vulputate metus sem ac justo. Pellentesque arcu orci, tempor consequat luctus pretium, elementum quis augue. Fusce tincidunt varius molestie. Nam ac velit ut est fermentum porttitor. Etiam vel lobortis justo. In aliquet consequat elit eget rutrum. Mauris non eros ligula, eget ultrices sapien. Quisque egestas, libero eu venenatis faucibus, orci massa molestie ante, et aliquet enim nisi vitae elit. Suspendisse ultricies sagittis mauris. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; Nam quis turpis at est imperdiet accumsan. Aenean sed venenatis magna. Morbi viverra dui sed lacus semper sed convallis tortor eleifend. Vestibulum at erat sapien. Donec eget sagittis neque. Vestibulum in interdum urna. \begin{wrapfigure}{l}{0.5\textwidth} {\includegraphics[scale=0.4,angle=0,width=3.5in]{MassiveStarHR.pdf}} \caption{blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah } \label{fig:hr} \end{wrapfigure} Nullam cursus bibendum magna. Aliquam id nisi nisi. In lobortis ultrices sapien id suscipit. Vestibulum ut arcu id lectus pretium eleifend. Duis dui urna, aliquam id malesuada eget, dictum ut ligula. Phasellus venenatis odio a ligula dictum ut pellentesque tellus tincidunt. Cras sed feugiat nulla. Curabitur varius vestibulum nisi. Nulla sit amet purus sed eros euismod vulputate. Suspendisse faucibus risus turpis, ut aliquet purus. Aenean a erat at mi mollis elementum semper sit amet magna. Donec pretium mauris at lorem pulvinar elementum. Suspendisse in quam ut turpis cursus imperdiet eu et justo. Curabitur iaculis nulla ac magna semper luctus. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Etiam nec eros orci, et egestas tortor. Nullam ac odio sit amet est volutpat lacinia. Suspendisse leo nibh, fringilla ac lacinia sit amet, tincidunt sed lorem. Mauris eu lacus et lectus pellentesque blandit in eget purus. Proin vel dui erat, ut aliquet quam. Pellentesque vitae elit vel mi venenatis tempus. Vivamus id augue diam. Sed consequat aliquam risus, id accumsan turpis suscipit a. Pellentesque habitant morbi tristique senectus et netus et malesuada fames ac turpis egestas. Maecenas fringilla odio at dui accumsan consectetur. Maecenas non risus vel mi pulvinar bibendum quis ac magna. Mauris et hendrerit enim. Sed et sem id felis eleifend ornare. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Etiam a sagittis enim. Donec ac erat ac nulla mollis mollis. Curabitur urna turpis, porta nec elementum sed, ornare nec leo. In facilisis ipsum nunc, eget suscipit odio. Vestibulum tincidunt velit nec magna sodales fringilla faucibus eget eros. Curabitur pulvinar ante interdum turpis dignissim dapibus. Donec ultricies suscipit enim ac blandit. Nunc eu lectus vel metus volutpat bibendum. Sed a purus est. Mauris faucibus nulla eu mi convallis vel suscipit nibh dignissim. Nulla porttitor iaculis sodales. Praesent convallis arcu quis magna suscipit suscipit. Curabitur congue tortor eget ligula consequat in commodo tellus pharetra. Vestibulum neque erat, gravida non faucibus non, ultricies a nibh. Mauris sem lacus, varius in imperdiet et, sollicitudin ut nisl. Sed porttitor, velit ut posuere mattis, odio purus condimentum urna, id aliquet nisl nulla ut mi. Sed vulputate porttitor purus et pretium. Fusce rutrum convallis nibh, varius fringilla lectus ullamcorper a. Morbi purus odio, tristique aliquet vestibulum sed, iaculis vel quam. Quisque dictum, nunc quis aliquam iaculis, lorem est fermentum nibh, eget luctus nibh lorem sed sapien. Sed in commodo ligula. Suspendisse tempus nisi ac massa aliquam blandit. Vestibulum libero augue, scelerisque eget rutrum fringilla, euismod non nibh. Nullam venenatis velit non erat tempus sit amet pulvinar ligula tristique. In vulputate dapibus est a convallis. \end{document}
[ 0.003151284996420145, -0.005831323564052582, -0.0005505082663148642, 0.023355595767498016, 0.005278124939650297, 0.01440921239554882, 0.006731214001774788, 0.005540414713323116, -0.014038232155144215, 0.023210879415273666, -0.029155641794204712, 0.006988417357206345, 0.008676666766405106, ...
[ 0.35675323009490967, 0.06800299137830734, 0.5646957159042358, -0.017063479870557785, 0.22661593556404114, -0.1338195502758026, 0.3928718864917755, -0.17083317041397095, -0.5681261420249939, -0.7773534655570984, 0.04212428256869316, 0.6430818438529968, -0.3254086971282959, 0.086231961846351...
I have two paragraphs in my paper, the previous paragraph ends with > This shows that TF-IDF is still an important feature for text analysis task and my next paragraph starts with > **On the other hand** , various studies suggested that LDA may not work on > short documents due to insufficient context Is there another way of begin my next paragraph instead of saying _On the other hand_ but keep the meaning of contrasting?
[ 0.006249268539249897, 0.03296424821019173, -0.027593424543738365, 0.03549020737409592, -0.014270054176449776, 0.025189561769366264, 0.011520616710186005, 0.014862491749227047, -0.0209647323936224, -0.046916667371988297, -0.0120440274477005, -0.008416353724896908, -0.01255313865840435, 0.02...
[ -0.07092106342315674, -0.01279500499367714, 0.31093934178352356, -0.20981630682945251, -0.6751761436462402, -0.024059880524873734, 0.27750399708747864, -0.18634849786758423, -0.04097859933972359, -0.7372992038726807, 0.2306746542453766, 0.4550638198852539, -0.0785251334309578, 0.0071094506...
If a person wants to do something but he is not able to do that due to some personal reason, what is the exact word for that reason?
[ 0.006097020115703344, 0.02068459987640381, 0.00456143356859684, 0.03345068171620369, 0.008502887561917305, 0.009724339470267296, 0.0140983946621418, 0.03289123252034187, -0.018299032002687454, 0.030243761837482452, -0.042588744312524796, -0.004773635417222977, 0.02096293866634369, 0.024405...
[ 0.7452062368392944, 0.27216944098472595, -0.1022820696234703, 0.1679617166519165, 0.1423604041337967, -0.17534561455249786, 0.4188525378704071, -0.0360189750790596, -0.31018614768981934, -0.1664721816778183, -0.12740720808506012, 0.25582876801490784, -0.11593939363956451, 0.039994608610868...
i wanna generate euclodian distance map for my road data. I 'll use it for multi criteria analysis. I can get it by using Arcgis Euclodian Distance tool but i wanna generate it in QGIS environment. I got Sextante plugin and enabled Grass and Saga tools. But i dont know which tool i must use ??? There is a link about ouput of this anaylsis (below image). ESRI The Euclidean distance example
[ -0.004541241563856602, 0.005318556912243366, -0.009377273730933666, 0.012366375885903835, -0.019497448578476906, -0.013815348036587238, 0.013246515765786171, 0.012994208373129368, -0.023281404748558998, -0.0230094064027071, 0.0054903109557926655, 0.01099487766623497, -0.028086772188544273, ...
[ 0.20022615790367126, 0.3384113907814026, 0.4603240489959717, 0.18850959837436676, -0.3099217414855957, 0.17699451744556427, 0.15507838129997253, 0.16803978383541107, 0.1711146980524063, -1.121297001838684, 0.09482742100954056, 0.6346049308776855, 0.16748879849910736, 0.26051217317581177, ...
On my template I have this code <div class="blog_block"> <?php wp_reset_postdata(); ?> <?php $nbmax= 9; ?> <?php $tarali_query = new WP_Query('posts_per_page='.$nbmax.'&ignore_sticky_posts=1&paged='.$paged); ?> <?php while ($tarali_query -> have_posts()) : $tarali_query -> the_post(); ?> <?php get_template_part( 'content-blog', get_post_format() ); ?> <?php endwhile; ?> <?php tarali_paging_nav(); ?> </div><!-- blog_block --> Which render **ALL** recent posts of my blog. How can I modify this code to query ONLY the posts which belongs to **Portfolio Category**?
[ -0.0019365530461072922, 0.020656954497098923, 0.005991694051772356, 0.011523902416229248, -0.002124984283000231, 0.013075264170765877, 0.007840634323656559, -0.0004963470855727792, -0.007912473753094673, -0.008232626132667065, -0.005931186489760876, 0.009775194339454174, -0.01528452150523662...
[ 0.6413732767105103, -0.06560004502534866, 0.6085017919540405, -0.3095848560333252, -0.13099102675914764, 0.3304271399974823, 0.2279973328113556, -0.43685463070869446, -0.24898463487625122, -0.8321532607078552, 0.2348761111497879, 0.20820704102516174, -0.566205620765686, 0.286325603723526, ...
I am trying to use SAGA to triangulate a set of .las files I have. I have tiled them using lastools, since the single file was too big for my computer to handle. I wonder, can I use the saga_cmd grid_gridding "Triangulation" for multiple files, or do I have to process each individually? And I want tif files based on the Z-value, where do I specify the Z value? Is it the -FIELD Attribute? I have found very little information about the cli scripting of SAGA, are there any book available on the subject? Thanks in advance!
[ -0.003253921400755644, 0.00890292227268219, -0.00020244729239493608, 0.016650637611746788, 0.03210638090968132, 0.016355548053979874, 0.00969712994992733, -0.00023357849568128586, -0.02125735953450203, -0.006303990259766579, -0.002558652777224779, -0.0010133425239473581, -0.01336769573390483...
[ 0.41733506321907043, 0.1996457427740097, 0.021822074428200722, 0.1585666835308075, -0.21491564810276031, -0.03621125966310501, -0.16876709461212158, -0.011253084056079388, -0.1432998925447464, -0.5763906240463257, 0.28558099269866943, 0.662673830986023, 0.02198570780456066, 0.2795850038528...
I am trying to draw a simple vertical tree to represent a hierarchical partition of a graph in communities. I have written some code to draw it in tikz, but I am not very satisfied of the result. My code is: \begin{tikzpicture}[shorten >=1pt, auto, node distance=3cm, ultra thick, node_style/.style={font=\sffamily\Large\bfseries,minimum size=0.7cm}, edge_style/.style={draw=blue, ultra thick}, community_label_style/.style= {font=\sffamily\Large\bfseries,minimum size=0.7cm,text height=1.5ex,text depth=.25ex,}] \node[community_label_style] (v15) at (0,-4) {\emph{a}}; \node[community_label_style] (v16) at (1,-4) {b}; \node[community_label_style] (v17) at (3,-4) {c}; \node[community_label_style] (v18) at (4,-4) {d}; \node[community_label_style] (v19) at (8,-4) {e}; \node[community_label_style] (v20) at (9,-4) {f}; \node[community_label_style] (v21) at (11,-4) {g}; \node[community_label_style] (v22) at (12,-4) {h}; \node (v1) at (6,1) {}; \node[community_label_style] (v13) at (1,-2) {$C2$}; \node[community_label_style] (v7) at (9,-2) {$C6$}; \node (v2) at (6,0) {}; \draw (v1) edge (v2); \node[community_label_style] (v3) at (2,0) {$C1_2$}; \node[community_label_style] (v4) at (10,0) {$C2_2$}; \node[community_label_style] (v11) at (4,-2) {$C4$}; \draw (v2) edge (v3); \draw (v2) edge (v4); \node (v5) at (2,-2) {}; \node[community_label_style] (v12) at (3,-2) {$C3$}; \node[community_label_style] (v14) at (0,-2) {$C1$}; \node (v6) at (10,-2) {}; \node[community_label_style] (v9) at (11,-2) {$C7$}; \node[community_label_style] (v10) at (12,-2) {$C8$}; \node[community_label_style] (v8) at (8,-2) {$C5$}; \draw (v3) edge (v5); \draw (v4) edge (v6); \draw (v6) edge (v7); \draw (v7) edge (v8); \draw (v6) edge (v9); \draw (v9) edge (v10); \draw (v11) edge (v12); \draw (v12) edge (v5); \draw (v5) edge (v13); \draw (v13) edge (v14); \draw (v15) edge (v14); \draw (v16) edge (v13); \draw (v17) edge (v12); \draw (v18) edge (v11); \draw (v19) edge (v8); \draw (v20) edge (v7); \draw (v21) edge (v9); \draw (v22) edge (v10); \draw (v13) edge (v12); \draw (v7) edge (v9); \draw (v4) edge (v3); \end{tikzpicture} The result is: ![enter image description here](http://i.stack.imgur.com/7hLER.jpg) I don't like that lines are separed by white dots. I would like to draw a more elegant representation. Is it possible? How could I improve my drawing? Thank you for your help.
[ -0.0010852685663849115, 0.010183373466134071, -0.006088778376579285, 0.012648219242691994, -0.025302940979599953, -0.0130492327734828, 0.007276787888258696, 0.008531064726412296, -0.0159913282841444, 0.0031077703461050987, -0.021023213863372803, -0.007969779893755913, -0.0017811038997024298,...
[ 0.45496854186058044, 0.007847590371966362, 0.6016566753387451, -0.4204193949699402, 0.10796286165714264, 0.2477993220090866, 0.06782713532447815, 0.0441267266869545, -0.1132466271519661, -0.6186528205871582, 0.20483532547950745, 0.216976597905159, -0.15861962735652924, 0.06117821857333183,...
When I su to root, any control character that I type is echoed by the Android shell. For example, when I type "ls /sd" followed by a tab, instead of displaying the default tab completion of "/sdcard", the shell prints out "ls /sd" followed by the tab space. The tab completion appears to work behind the scenes, however. When I press the Enter key, the contents of "/sdcard" are listed. As a more obvious example, when I press the "Ctrl" key followed by an R (Ctrl-R) or an E (Ctrl-E), the shell displays ^R and ^E, respectively. This happens under both Jack Palevich's Android Terminal Emulator and the localhost connection mode of ConnectBot (I haven't tested with any remote host), and under both the default CyanogenMod shell and bash. My attempt to work around the problem by changing command prompt settings ($PS1) had no positive effect. I also tried the stty command but the obvious options like "-echo" don't work for me either. Is there some command that will prevent the shell from echoing the control characters? Or better, is there a shell configuration option that I need to explicitly set? At the moment, I have to type blindly when using tab completion and the Ctrl key as root.
[ 0.0031435934361070395, 0.005982773844152689, -0.01949792355298996, 0.005545947700738907, -0.033310819417238235, 0.021264148876070976, 0.009630915708839893, -0.0006092693656682968, -0.010168579407036304, -0.002854662947356701, -0.02085695043206215, -0.0025723064318299294, -0.00178247317671775...
[ -0.23852726817131042, -0.27798399329185486, 0.582695484161377, -0.34966760873794556, 0.22819489240646362, 0.29445934295654297, 0.4945814311504364, -0.19776949286460876, 0.33363115787506104, -0.7653014659881592, -0.25242847204208374, 0.7033679485321045, -0.0189292561262846, 0.14476720988750...
> **Possible Duplicate:** > How to set the font for a \section title (and chapter etc) How to edit the properties of `chapters`' and `sections`' headings (size of font, positions, indents) in `report` style?
[ 0.005993073806166649, -0.007177297491580248, -0.0010575761552900076, 0.037270452827215195, 0.013018147088587284, -0.010616703890264034, 0.00962323509156704, 0.022783080115914345, -0.030797451734542847, -0.010195833630859852, -0.009936113841831684, 0.01033443957567215, 0.012950833886861801, ...
[ 0.2666395604610443, 0.40788471698760986, 0.38937392830848694, 0.2545153796672821, 0.052002910524606705, 0.02165958844125271, 0.3261880576610565, -0.24003183841705322, -0.05350107327103615, -0.6951601505279541, -0.22048023343086243, 0.3041982352733612, -0.16257983446121216, -0.3271097242832...
My book about quantum mechanics states that the hamiltonian, defined as $H=i\hbar\frac{\partial}{\partial t}$ is a hermitian operator. But i don't really see how I have to interpret this. First of all: from which to which space is this operator working? They are defining a vectorspace called the "wavefunctionspace $F$" which contains all square-integrable functions that are continious and infinite differentiable (and everywhere defined). But it looks to me, that if the hamiltonian acts on this space, it's not necessary true that the image of a random vector of $F$ is again in $F$. I think in fact, that there are some vectors of $F$ so that the hamiltonian of those vectors is not an element of $F$ (so that it's not an endomorphism on $F$). And if the hamiltonian has to be hermitian, it has to be an endomorphism on some space. If we define instead the vectorspace $V$, which is the same space as $F$ but where functions don't have to be square-integrable, the hamiltonian will be an endomorphism (so at first I thought this was the solution). But now the inner product on functions $<f,g> = \int_{-\infty}^\infty{f^*g}$ which was defined well on $F$ because the integral will always exist if $f$ and $g$ are function of $F$, is no longer properly defined. I hope someone can clarify how I have to interpet this operator (the same question holds in fact for some other operators).
[ -0.030091511085629463, 0.019965682178735733, 0.006734195630997419, 0.008290362544357777, -0.005017470568418503, -0.02783782407641411, 0.010560214519500732, 0.002194080501794815, -0.011958485469222069, -0.0011988482438027859, -0.026409892365336418, 0.006195164751261473, -0.010548188351094723,...
[ 0.1363036036491394, 0.09060261398553848, 0.22091776132583618, -0.24497562646865845, -0.07868267595767975, -0.035962898284196854, -0.10682225972414017, -0.17340007424354553, -0.27178433537483215, -0.3619002103805542, -0.14820453524589539, 0.5818378925323486, -0.4557873010635376, 0.368931651...
I am currently binding number key 1 to an Iron Sword, and number key 2 to a bow. However, when I press #2 to use the Bow, and press #1 to use the Iron Sword again, the binding to bring up the bow disappears. How can I quickly switch between bow and sword/shield?
[ 0.0317394845187664, 0.030169237405061722, -0.03518802672624588, 0.005248166620731354, -0.03259691596031189, -0.007209017872810364, 0.01209309697151184, -0.0026580337435007095, -0.018840258941054344, 0.03708691522479057, -0.015551736578345299, 0.0176254715770483, -0.04905709624290466, 0.033...
[ -0.019359009340405464, 0.10218393057584763, 0.7463594675064087, -0.024334879592061043, -0.6314277648925781, 0.05532315745949745, 0.1732240915298462, -0.8392391204833984, -0.019182024523615837, -0.7598811388015747, -0.11051517724990845, 0.28425347805023193, 0.1616201102733612, 0.15650820732...
I've been thinking to this for the last two hours and haven't been able to come with a solution. **Problem.** A mole of gas initially at pressure $P_A = 2 \text { atm}$ and occupying a volume $V_A=20\,\text {L}$, passes through an irreversible transformation in which it absorbs $Q=1200J$ as heat and ends with pressure $P_B=3/2 P_A$. Then it expands reversibly and in adiabatic conditions, until its pressure is again $P_A$ ($P_C=P_A$). Finally, it has an isobaric transformation that brings the volume back to $V_A$. Given that the absolute values $|W_{CA}|=300J$ and $|\Delta U _{CA}|=450J$, determine the thermodynamic coordinates $P,V,T$ in the three states $A,B,C$. I think that the only way to find the coordinates is to work in reverse: finding $V_C$ from $$W_{CA}=p_A(V_A-V_C),$$ and then $V_B$ using the adiabatic's $P|V$ law. Now, the information on the absolute values tells us that $$Q_{\text {tot}}=Q_{AB}+Q_{CA}>0,$$ since $Q_{CA}\geq-750J$ and so $W_{\text{tot}}>0$. Having noted this, I don't see how does it put a constraint on the sign of $W_{CA}$. Since the transformation $A\to B$ is totally unspecified, I think that the volume $V_B$ could be everything without any contradiction, and this affects the sign of the work in last transformation. Have any idea?
[ -0.003079742193222046, 0.013439200818538666, -0.012623313814401627, 0.00804133526980877, -0.015796707943081856, 0.007174230646342039, 0.006349905859678984, -0.0017714904388412833, -0.011466857977211475, -0.011035231873393059, -0.0005785486428067088, 0.002575232880190015, -0.01996960118412971...
[ 0.15932567417621613, 0.2894645929336548, 0.4507804214954376, -0.3314802348613739, 0.5212241411209106, 0.21348753571510315, 0.3599575459957123, -0.2693457305431366, -0.33092331886291504, -0.3238884210586548, -0.10615015774965286, 0.24270202219486237, -0.454877108335495, 0.5243957042694092, ...
I use Google Analytics to monitor one of my Websites. However, I have a problem and I cannot monitor the stats with the way I should. What I mean... I have a page which a user can use a link with get parameters to autofill a form. Let's say that I have a page like this one: `www.mywebsite.com/mypage` On the history (and on realtime stats) I have different stats for: www.mywebsite.com/mypage www.mywebsite.com/mypage?param=123 www.mywebsite.com/mypage?param=234 www.mywebsite.com/mypage?param=345 ... ... There are several parameters and I cannot check them one by one. I would expect a stats page which will present stats for `www.mywebsite.com/mypage` and for any same link with parameters as one page. Is there any way to do this?
[ -0.00011784990783780813, 0.007375925779342651, 0.00005501159466803074, 0.0029675080440938473, 0.003854813752695918, -0.00021567358635365963, 0.007444282062351704, -0.012521423399448395, -0.017303194850683212, -0.003980016335844994, -0.0023141964338719845, 0.0009331518085673451, 0.00722221517...
[ 0.8595457077026367, 0.2714282274246216, 0.2980312407016754, -0.12128596752882004, -0.10612308233976364, -0.07322879880666733, 0.37760162353515625, 0.10566290467977524, -0.4896240830421448, -0.9486339688301086, 0.48789989948272705, 0.15417222678661346, 0.09763990342617035, 0.243458300828933...
I am using `glmmLasso` for variable selection. In my case, `n` is slightly less than `p` and `p` are bioclimatic variables for different time periods, so are highly correlated. How do I choose the right values for the arguments: `lambda` and `control`. I have tried with different values of `lambda`, `maxIter` (even to 10,000) and `control` (`start` and `steps`). But, the algorithm never converges. The p.values of all the selected variables are 0 and that makes me wonder if the non-convergence is the cause. What factor could help achieve convergence in a reasonable number of iterations? Among the variables selected, since p.values are all 0 (the rest of the variables are NA), can I estimate the relative variable importance based on the `StdErr`? Also, the `lambda` resulting in the lowest `BIC` values selects too many variables as being significant. Would it be OK in this case,(as I am only doing a rough variable selection for modelling) to not worry about the BIC, but choose the lambda that gives me a reasonable number of variables that also make sense given my data. What is the `start` arguments in `control`, anyway? All zeros are not acceptable, and so are vectors below a certain length. What do the values and vector length depend on? Apologies, if that shows ignorance of the mathematics behind mixed-models... For details on the data, please refer to Variable selection using mixed-models (lme4) I went through the `GMMBoost`, also by Groll, but did not find something to guide me in this case. Has anyone used glmmLasso for analysis and faced similar situations? Glad to hear any suggestions
[ 0.005072568543255329, 0.016664788126945496, -0.021174568682909012, -0.004228830803185701, 0.003259182907640934, 0.01205266360193491, 0.008985085412859917, 0.012097178027033806, -0.00761671457439661, -0.019621267914772034, -0.0034174902830272913, 0.004691934213042259, -0.024626757949590683, ...
[ -0.041727930307388306, -0.11217790842056274, 0.27405938506126404, 0.2891060411930084, 0.09651371836662292, 0.5944753289222717, -0.19556882977485657, -0.2954784333705902, -0.07569864392280579, -0.2253132462501526, 0.2885110378265381, 0.5777630805969238, -0.3105137050151825, 0.36394691467285...
I got really confused about the graph of the relationship between wavelength and intensity of black body radiation. What does the peak stand for? And what does the graph tell us?How can we analyze it?
[ 0.02814124897122383, 0.014879354275763035, -0.004977475386112928, 0.01870962791144848, -0.05371378734707832, -0.02993221953511238, 0.01608002558350563, 0.007454483304172754, -0.039899807423353195, -0.05628625303506851, -0.019798627123236656, 0.021415766328573227, 0.00012133543350500986, -0...
[ 0.7466610670089722, 0.07067963480949402, 0.05226980894804001, 0.24184000492095947, -0.3012032210826874, -0.47610318660736084, 0.28368350863456726, -0.014912212267518044, -0.13812541961669922, -0.08408305794000626, 0.2709531784057617, 0.543325662612915, -0.32401564717292786, 0.4034062623977...
I am trying to quantify the effect of a future random shocks on my seasonal ARIMA model. If I have understood the theory correctly, the easiest way is to express my seasonal ARIMA model in its "random shock" form, and calculate the corresponding psi weights. Is there a way to do this in R? There is ARMAtoMA, but I think this only works for ARMA models, and not seasonal ARIMA models. Thank you for your help. * * * UPDATE: Apologies, I'll post the question about R onto stack overflow. It would be good to get confirmation that this is the correct method to quantify the effect of future random shocks to a seasonal ARIMA model.
[ 0.02305224910378456, 0.01206941343843937, -0.011325530707836151, 0.02698381617665291, 0.010133437812328339, -0.007146853022277355, 0.00872756727039814, 0.01159660704433918, -0.016944143921136856, -0.0014863531105220318, -0.005540697835385799, 0.019297253340482712, -0.01661638170480728, 0.0...
[ 0.260771781206131, -0.08269702643156052, 0.22711181640625, 0.08587296307086945, -0.7281003594398499, -0.13776391744613647, 0.33580297231674194, -0.1745828092098236, -0.34476539492607117, -0.1601741909980774, -0.05954417958855629, 0.3598479628562927, 0.11449051648378372, 0.23885247111320496...
I know that I could delete the last three chars with: echo -ne '\b\b\b' But how can I delete a full line? I mean I dont want to use: echo -ne '\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b' ...etc... to delete a long line..
[ 0.008762309327721596, 0.010578266344964504, -0.02615761011838913, 0.001597035676240921, -0.019731972366571426, 0.011726138181984425, 0.004840194247663021, -0.004790535196661949, -0.014661378227174282, 0.000004855915904045105, -0.007181569933891296, -0.0010816368740051985, -0.0163972172886133...
[ -0.07931949943304062, 0.2221660315990448, 0.04074912518262863, -0.09699008613824844, -0.2634720802307129, 0.28922033309936523, 0.8089056611061096, 0.12893399596214294, -0.21380645036697388, -0.04130614921450615, -0.004001829773187637, 0.08228655904531479, -0.5453755855560303, 0.12811616063...
The legend entries created by a more complex math formula as in this example are overlapping and the alignment to the lines is wrong. \documentclass[]{scrbook} \usepackage{pgfplots} \usepackage{amsmath} \begin{document} \pgfplotsset{width=0.8\textwidth, height=0.6\textwidth} \centering \begin{tikzpicture} \begin{axis}[scale only axis,samples=2000, /pgfplots/enlargelimits=false, legend style={legend pos=north west}] \addplot[domain=0:30] gnuplot{5*exp(-((x-5*pi)/(2.5*pi))**2)*sin(2*x)+x}; \addplot[domain=0:30] gnuplot{x}; \legend{$f(x) = 5\exp\left(-\left(\dfrac{x-5\pi}{2.5\pi}\right)^2\right) \sin(2x) + x$, $f(x)_\text{fit} = x$} \end{axis} \end{tikzpicture} \end{document} ![enter image description here](http://i.stack.imgur.com/EfesP.png)
[ -0.006780806463211775, 0.013081456534564495, 0.0011265373323112726, 0.009657732211053371, 0.007675710134208202, 0.006270878948271275, 0.00794969405978918, 0.008864262141287327, -0.012742219492793083, -0.00851508043706417, -0.0011157260742038488, 0.0006385128945112228, -0.003157250117510557, ...
[ -0.0738769993185997, -0.0921962708234787, 0.43447503447532654, 0.12943001091480255, 0.05183771252632141, 0.20878511667251587, 0.5679095983505249, -0.3045641779899597, -0.32370665669441223, -0.4370482861995697, -0.1463632881641388, -0.07743042707443237, 0.25275272130966187, 0.46432399749755...
I've been having trouble defeating this little bugger. EVen with lots of buff potions, health potions, high tier armor (Molten), and ranged weapons with plenty of ammo (Minishark), I can't seem to kill this. Worse, when I get it down to about 2000 hp, even if I'm farily far away, it'll suck me in and continually rise towards the skybox, which doesn't allow me to do anything for some reason.
[ 0.0021390863694250584, 0.004884843714535236, -0.01316229160875082, -0.010237030684947968, -0.03679569438099861, -0.0034738413523882627, 0.004998689517378807, 0.0060570272617042065, -0.017421526834368706, 0.00784209929406643, -0.018879707902669907, 0.013883243314921856, -0.033673226833343506,...
[ 0.183889240026474, 0.3875422179698944, 0.1849527806043625, 0.2725018262863159, -0.34648752212524414, 0.14963816106319427, 0.6419841051101685, -0.22912076115608215, -0.3200713098049164, -0.31412172317504883, 0.2546645700931549, 0.33480092883110046, 0.2693190276622772, 0.21257607638835907, ...
i am trying to make a QGIS plugin ready for the new QGIS API in the current development branch (for the upcoming QGIS 2.0 release). http://hub.qgis.org/wiki/quantum-gis/Python_plugin_API_changes_from_18_to_20 However i struggle a bit on how to add new attributes to the attribute table. Up to now i have added new attributes like this: # Working in QGIS Lisboa layer = ... # QGsVectorLayer object name = "newAttribute" provider = layer.dataProvider() caps = provider.capabilities() # Check if attribute is already there, return "-1" if not ind = provider.fieldNameIndex(name) try: if ind == -1: if caps & QgsVectorDataProvider.AddAttributes: res = provider.addAttributes( [ QgsField(name,double) ] ) except: return False In the recent QGIS dev. version these steps aren't working anymore (the attribute column isn't added) and i was not able to find the solution in the above mentioned wiki-page. Can someone give me hint?
[ 0.003169435076415539, 0.009406580589711666, 0.00195584399625659, 0.010806859470903873, 0.02007204107940197, 0.005003894679248333, 0.007535351440310478, 0.00883436668664217, -0.018424034118652344, -0.009992677718400955, -0.00024365086574107409, 0.008130315691232681, -0.01894892007112503, -0...
[ 0.5234888195991516, -0.19158002734184265, 0.7481060028076172, -0.23474635183811188, -0.37004566192626953, -0.14670579135417938, 0.2598925828933716, 0.011031672358512878, -0.20727713406085968, -0.937193751335144, 0.22107566893100739, 0.4345644414424896, -0.07031752914190292, 0.3146679401397...
I was playing Minecraft on a server and a weird thing happened when I started placing netherrack on the roof of my house. I heard ominous sounds I hadn't heard before so I decided to run inside for safety. When I got to the door it started opening and shutting repeatedly. When I got inside the house a stone block appeared in the room and slowly broke down to nothing and the music stopped. I went back outside and found a burnt looking music disk on the ground, C418-11. I played it and confirmed it was the sounds I was hearing. So, I'm wondering is this a normal way to get C418-11, is there a spooky mod that is doing this, or did someone with special privileges go invisible, play the music, and then prank me? I've tried placing more netherrack around with no effect.
[ -0.02336779236793518, 0.020551595836877823, 0.005496347788721323, 0.023231256753206253, -0.020743584260344505, -0.004110836423933506, 0.007838004268705845, 0.00196515629068017, -0.018792729824781418, 0.017249833792448044, -0.016365792602300644, 0.012844879180192947, 0.03206322714686394, 0....
[ 0.2190898060798645, 0.3580135107040405, -0.09833038598299026, 0.3588300049304962, -0.2816864550113678, -0.41011202335357666, 0.030130518600344658, -0.06460307538509369, -0.16442081332206726, -0.6704622507095337, 0.5686658620834351, 0.06088960915803909, 0.15603457391262054, 0.70442998409271...
Thanks so much to all involved with the us-atlas project. I am learning how to use d3, leaflet and other tools to make maps and this is really helpful. I have grabbed files from the us-atlas project on github (https://github.com/mbostock/us-atlas) and have had success doing the basic make for US Counties. However, the make for US Zip codes seems to hang. I have tried running in background with & and also using nice to increase the priority but it still seems not to finish. In response to Brad's question, the process has run in background for upwards of 10-11 hours overnight. The top command line utility at that point shows that node has 0% of CPU (whereas at the start I monitor and see it is taking 100% of CPU for the first hour). Could it just be that the OS falls into sleep mode and it kills the processing? Is there anything I can do to get OS X to keep the make (and subsequent node process) a priority and keep it going? How long should this take with a 2.6GHz Intel Core i7 and 8MB RAM? As an alternative I am mostly interested in MD Zip codes. Is there a way to restrict a TopoJSON build to just those zip codes?
[ -0.0012446590699255466, 0.001834643306210637, -0.005487602204084396, 0.006082618143409491, 0.0031393805984407663, 0.0013162835966795683, 0.005619935225695372, -0.0014440808445215225, -0.018807141110301018, 0.016867617145180702, 0.008610799908638, 0.006480161100625992, -0.010186314582824707, ...
[ 0.7529152035713196, 0.1992781013250351, 0.5039616227149963, 0.3064150810241699, 0.10674189031124115, -0.2934603989124298, 0.47439202666282654, 0.07194255292415619, -0.07270047068595886, -0.6539055109024048, 0.09687332063913345, -0.15620897710323334, 0.1595982015132904, 0.21411843597888947,...
I installed Kitkat 4.4.2 following this video https://www.youtube.com/watch?v=9YGzh3r9Q58 After installing, it only reboots into the recovery mode. :/ I need help
[ 0.013637755066156387, -0.006289729382842779, 0.006790418643504381, 0.03421350196003914, -0.06813503801822662, -0.021855471655726433, 0.009807527996599674, -0.038328684866428375, -0.03250628337264061, 0.02147011272609234, -0.03730735927820206, 0.015222851186990738, -0.006990326568484306, 0....
[ 0.30676698684692383, 0.15692612528800964, 0.539254367351532, 0.03766404464840889, -0.2683747708797455, -0.1820334643125534, 0.6107621192932129, -0.22754555940628052, -0.09939417988061905, -0.16645842790603638, -0.13765668869018555, 0.7650678157806396, -0.19762936234474182, -0.1502449959516...
I have three techniques, called A, B and C. Each can be used independently when trying to perform four related tasks (Tasks 1, 2, 3 and 4). I have run lots of tests, trying all combinations of each technique being on or off. My results look something like this. Each number represents how many times the task was completed successfully when attempted, using the given combination of techniques, 100000 times. So, the higher the number, the better. $$ \begin{array}{l|r|r|r|r|r|r|r|r|r|} \mbox{Technique $A$} & - & - & - & - & X & X & X & X \\ \mbox{Technique $B$} & - & - & X & X & - & - & X & X \\ \mbox{Technique $C$} & - & X & - & X & - & X & - & X \\ \hline \mbox{Task $1$} & 433 & 277 & 911 & 492 & 686 & 4211 & 3775 & {\bf 9732}\\ \mbox{Task $2$} & 149 & 1063 & 5562 & {\bf 6035} & 3 & 58 & 1391 & 1708\\ \mbox{Task $3$} & 220 & 1278 & 7014 & {\bf 7018} & 10 & 97 & 2083 & 4452\\ \mbox{Task $4$} & 218 & 1255 & 6142 & {\bf 8656} & 1 & 73 & 1087 & 2056\\ \end{array} $$ I've highlighted the largest number in each row. Looking at the numbers, it seems that $B+C$ is good for Tasks 2, 3 and 4, and that adding $A$ as well is best for Task 1. But I want to say a bit more. I'd like to be quantitative if I can. My question is: can I deduce _anything_ quantitative from this data? Or do I really need some measure of the variance of the observations? That is, I suspect the numbers might be different if I ran all the tests again. * * * **Edit.** I originally presented my question pretending there were 3 techniques, just in order to save space. There are in fact 4. I'm mentioning this now because it might affect how the results are analysed. In particular: note now that the `intercept' is zero, i.e. there is no success at all when none of the techniques are used. $$ \begin{array}{l|r|r|r|r|r|r|r|r|r|} \mbox{Technique $Z$} & - & - & - & - & - & - & - & - & X & X & X & X & X & X & X & X \\ \mbox{Technique $A$} & - & - & - & - & X & X & X & X & - & - & - & - & X & X & X & X \\ \mbox{Technique $B$} & - & - & X & X & - & - & X & X & - & - & X & X & - & - & X & X \\ \mbox{Technique $C$} & - & X & - & X & - & X & - & X & - & X & - & X & - & X & - & X \\ \hline \mbox{Task $1$} & 0 & 0 & 0 & 0 & 0 & 2700 & 1 & 10113 & 433 & 277 & 911 & 492 & 686 & 4211 & 3775 & {\bf 9732}\\ \mbox{Task $2$} & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 149 & 1063 & 5562 & {\bf 6035} & 3 & 58 & 1391 & 1708\\ \mbox{Task $3$} & 0 & 0 & 0 & 0 & 0 & 664 & 0 & 3043 & 220 & 1278 & 7014 & {\bf 7018} & 10 & 97 & 2083 & 4452\\ \mbox{Task $4$} & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 218 & 1255 & 6142 & {\bf 8656} & 1 & 73 & 1087 & 2056\\ \end{array} $$ Here are those numbers in by-column form (for pasting into `R`): 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2700, 0, 664, 0, 1, 0, 0, 0, 10113,0, 3043, 0, 433, 149, 220, 218, 277, 1063, 1278, 1255, 911, 5562, 7014, 6142, 492, 6034, 7018, 8656, 686, 3, 10, 1, 4211, 48, 97, 73, 3775, 1391, 2083, 1087, 9732, 1708, 4452, 2056
[ -0.0027308862190693617, 0.02014244720339775, -0.0172725822776556, 0.0028965130914002657, -0.01781976781785488, -0.0001420709304511547, 0.006916092708706856, -0.03440609201788902, -0.014233862049877644, 0.00007040705531835556, -0.017232278361916542, 0.0035344052594155073, -0.01036174222826957...
[ 0.24622100591659546, 0.22906894981861115, 0.33760982751846313, -0.23061031103134155, 0.13634087145328522, 0.5184292793273926, 0.47626903653144836, -0.6246817111968994, -0.527484118938446, -0.37995126843452454, 0.07280957698822021, 0.5638054609298706, 0.014357119798660278, 0.078893221914768...
I have the following data: data = {{0, 0}, {20, 1.4}, {25, 9.8}, {30, 32.2}, {35, 38.2}, {40, 15.6}, {45, 2.7}, {50, 0.1}}; where the second coordinates are frequencies measured as percent of total (adding, therefore, to $100$). I would like to produce a cumulative frequency plot. What would be the neatest way to do so?
[ -0.0021169891115278006, 0.004973696544766426, -0.018243683502078056, 0.007332686800509691, 0.004052960313856602, -0.01502396259456873, 0.005283130798488855, -0.011710545979440212, -0.013353319838643074, -0.00480148009955883, 0.003949983511120081, 0.0012443093582987785, -0.016853945329785347,...
[ 0.4033697247505188, -0.040273960679769516, 0.7123944759368896, 0.09223701059818268, 0.28163209557533264, 0.23045872151851654, -0.19438911974430084, -0.3126070499420166, 0.14030608534812927, -0.5445998311042786, 0.45869573950767517, 0.4003351926803589, 0.06694234907627106, 0.497480273246765...
I'm Converting an excel with 3000 rows and 10 columns. My Problem is while exporting excel to a shapefile resulting that " **The Maximum Record length has been exceeded** ". Plz Suggest..
[ -0.019010821357369423, 0.006833323277533054, -0.007215314079076052, 0.039009299129247665, -0.032628946006298065, 0.04600498825311661, 0.013392878696322441, 0.01974865421652794, -0.01827225461602211, -0.054851654917001724, -0.0031668064184486866, 0.023483870550990105, 0.02095840312540531, 0...
[ 0.14835800230503082, 0.5252947807312012, 0.2621179521083832, 0.01303570345044136, -0.15357241034507751, 0.21647857129573822, 0.37846067547798157, -0.1892751157283783, -0.26067036390304565, -0.8430832624435425, -0.008550003170967102, 0.17551450431346893, -0.05790431424975395, 0.050853088498...
I am having difficulty getting ERDAS or ENVI to correctly read and display cloud mask attributes from MOD 35 HDF files. They files are displayed as 0-255, whereas I expected some kind of code 0, 1, 00, 01 etc indicating clear and cloud pixels. I believe someone may have used this data before.
[ -0.018790915608406067, 0.014120640233159065, -0.021266667172312737, 0.024673521518707275, -0.004366370849311352, -0.004455073270946741, 0.01379738561809063, -0.0034812097437679768, -0.01823258586227894, -0.00634160079061985, -0.017945216968655586, 0.009188107214868069, -0.01982693374156952, ...
[ 0.1521560251712799, 0.1493593454360962, 0.40850088000297546, -0.045896269381046295, -0.012843634001910686, -0.13954457640647888, 0.2383907437324524, -0.1629822850227356, -0.03203551471233368, -0.7230783104896545, -0.13627542555332184, 0.5999966263771057, 0.05002157762646675, -0.12755876779...
I am using this snippet: \documentclass{beamer} \usepackage{pgfplots} \begin{document} \begin{tikzpicture} \begin{axis}[ xlabel=$x$, ylabel={$f(x) = x^{-1}$} ] \addplot {x^{-1}}; % this was previously: x^2 \end{axis} \end{tikzpicture} \end{document} That worked until I changed `x^2` to `x^{-1}`. Now I am getting the following error (there are also lots of errors after it): > PGF Math: Sorry, an internal routine of the floating point unit near > '2Y1.0e0]'. (in 'x^{-1}'). \end{frame} **Research:** I could not find many search results except these two questions from Tex SE: * Package PGF Math Error: Sorry, an internal routine of the floating point unit * Create Table for a Family of Curves Unfortunately, they seem to be unrelated since their problem is caused by a foreach construct which I don't even have in my code. **My system** : * MiKTeX-pdfTeX 2.9.4535 (1.40.13) (MiKTeX 2.9) * PdfLaTeX * Document class: Beamer
[ -0.00403923774138093, 0.00771324010565877, -0.0028765627648681402, 0.010638687759637833, 0.00893557257950306, -0.003654393134638667, 0.007177717983722687, 0.011020704172551632, -0.008424031548202038, -0.009879005141556263, -0.0033204976934939623, -0.0015473848907276988, -0.012690113857388496...
[ 0.13445979356765747, 0.04487523064017296, 0.6666840314865112, -0.18715716898441315, 0.23619167506694794, 0.025433819741010666, 0.02326064370572567, 0.21172453463077545, -0.03243096172809601, -0.7539791464805603, 0.08313052356243134, 0.3987557291984558, -0.37843218445777893, 0.0604340173304...
![enter image description here](http://i.stack.imgur.com/jfJ0C.png) The above text is produced by the following command: \documentclass[a4paper]{article} \usepackage[margin=0.8in]{geometry} \usepackage{ifxetex} \RequireXeTeX \usepackage{fontspec,multicol} \newcommand\dn{\catcode`\~=12 \fontspec[Script=Devanagari,Mapping=velthuis-sanskrit]{Sanskrit 2003}} \setlength{\columnseprule}{1.5pt} \setlength{\fboxrule}{2pt} \begin{document} \title{\fbox{{\dn\Huge lalitaa tri"satii stotram}}\vspace*{-1.8cm}} \date{} \maketitle \begin{flushleft} {\dn\Large asya "sriilalitaatri"satii stotra mahaamantrasya | bhagavaan hayagriiva .r.si.h | anu.s.tup chanda.h | "sriilalitaamahaatripurasundarii devataa | ai.m biijam | sau.h "sakti.h | klii.m kiilakam |} \begin{multicols}{2} {\dn\large kakaararuupaa kalyaa.nii kalyaa.nagu.na"saalinii |\\ kalyaa.na"sailanilayaa kamaniiyaa kalaavatii ||1|| kamalaak.sii kalma.saghnii karu.naam.rta saagaraa | \\ kadambakaananaavaasaa kadamba kusumapriyaa ||2|| \end{document} Now my question is the width between the text in the left hand side column and the column seperator is clearly more than what we see in the right hand side column. How to rectify this.
[ -0.001817829441279173, 0.0002167068887501955, 0.005038741044700146, 0.00884364265948534, 0.006584815680980682, -0.0007525463588535786, 0.0056329877115786076, -0.0012887229677289724, -0.013826476410031319, 0.006907155737280846, -0.015196477994322777, -0.002355108270421624, 0.00265907286666333...
[ 0.18792971968650818, 0.04959998279809952, 0.7059859037399292, -0.16792269051074982, -0.3086046874523163, 0.22069039940834045, 0.25768062472343445, -0.6058158278465271, -0.3859063684940338, -0.47080540657043457, -0.32542815804481506, 0.5068739652633667, -0.3660416305065155, -0.0083322534337...
I am having some trouble with Search Engine Optimization. When I look up "site:example.com", the first result is the index page. The third result that pops up says the title is: "Read More". I want it to display like the 2nd result does. The 2nd result displays: "Suma Karandiar, M.A., LCPC - Counselors In Association" but I can't find any differences between the 2nd page and the first. What am I supposed to do to tell Google that I want a different title than what they are displaying? This is not the same as Title tag different from title appearing in Google? because that question is about Google rewriting titles based on searches. In this case "Read More" is the title used for all searches.
[ -0.006814677268266678, -0.00006547779776155949, -0.003929983824491501, 0.00995614007115364, -0.030417777597904205, 0.013532343320548534, 0.007746292278170586, -0.0002129790373146534, -0.015082977712154388, 0.0038634915836155415, -0.010630136355757713, 0.008296716958284378, 0.0075502311810851...
[ 0.2358081042766571, 0.31095221638679504, 0.2801484763622284, 0.17611603438854218, -0.058856260031461716, 0.021613236516714096, 0.1980423927307129, 0.05597301200032234, -0.743660032749176, -0.35415974259376526, 0.07848897576332092, 0.5427067279815674, 0.06943873316049576, 0.5150899887084961...
I am currently storing some data in a custom field like this: $get_ref = $_SERVER['HTTP_REFERER']; $ref = get_post_meta($post->ID, 'page_ref', true ); $ref[] = $get_ref; update_post_meta($post->ID,'page_ref', $ref); The above updates the post_meta but doesnt overide it. I would like to save additional data to the same value that is saved above. The above produces this: Array ( [0] => http://test.com [1] => http://website.com ) I would like something like this: Array ( [0] => array(http://test.com, 14/04/2014) [1] => array(http://website.com, 10/01/2014)); Any ideas how i can achieve this?
[ 0.006969799287617207, 0.015839150175452232, 0.0023385374806821346, 0.012899743393063545, 0.04445099085569382, 0.01016139518469572, 0.007077474147081375, -0.004811754450201988, -0.013770543970167637, -0.013109122402966022, 0.0038440031930804253, 0.007881938479840755, -0.004864945542067289, ...
[ 0.2508396506309509, 0.18956202268600464, 0.6351093053817749, -0.3652576506137848, -0.24644134938716888, 0.13337895274162292, 0.23879168927669525, -0.5611016154289246, -0.1879768669605255, -0.5757522583007812, 0.08199220895767212, 0.5679349303245544, -0.17218557000160217, 0.1364182382822036...
We are developing a web GIS using Flexviewer and we want to allow users to upload a shp file and insert it into a ArcSDE geodatabase. I think there are two steps at least. 1. upload the shp file (zip) to a directory on the server 2. call some function to load it to the ArcSDE geodatabase I am not sure what is the best practice for this requirement. using ArcObjects or Geoprocessing service (ArcPy)? Any advise is appreciated.
[ -0.02894071862101555, 0.012981117703020573, -0.011721888557076454, 0.024194566532969475, -0.010145161300897598, -0.021528100594878197, 0.011841800063848495, 0.031400471925735474, -0.02002870850265026, -0.04374635964632034, 0.001874288427643478, 0.015490991063416004, -0.025014111772179604, ...
[ 0.6291685104370117, -0.194494366645813, 0.3707200586795807, 0.1930420845746994, -0.2464255839586258, -0.2096852958202362, -0.27362456917762756, -0.14527258276939392, -0.0693579763174057, -0.8524606227874756, -0.09171342849731445, 0.4295620024204254, -0.1854444295167923, -0.1648644208908081...
I need to install MySQL 5.5 from source in Ubuntu 10.04. So I'm trying to find tar.gz source from here. But when choosing "Linux Generic" from the drop down menu, it is showing an RPM extension I wonder why? Can we install from source using RPM package in Ubuntu now? It's like a joke but this prevents me from installing it now.
[ 0.02380327321588993, 0.01196000725030899, 0.002040555002167821, 0.01999029703438282, 0.00045017246156930923, -0.017572008073329926, 0.010508541017770767, -0.013988726772367954, -0.02533351071178913, -0.027569973841309547, -0.014325817115604877, 0.015316677279770374, 0.0005152492667548358, ...
[ 0.5877276062965393, 0.2267065942287445, 0.30701544880867004, -0.0005848673754371703, -0.28608062863349915, -0.4622268080711365, 0.20368601381778717, 0.5647954344749451, -0.241981640458107, -0.6613393425941467, 0.01879124902188778, 0.8541430830955505, -0.1675446480512619, 0.3696609437465668...
I have an equation like \documentclass[12pt,a4paper]{report} \usepackage[latin1]{inputenc} \usepackage{amsmath} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsthm} \begin{document} \begin{align*} a+b+c+d+e+f&=g\\ \Leftrightarrow a &=g-a-b-c-d-e-f \end{align*} \end{document} I'd like to have the arrow left to both equation, how do I do this?
[ 0.007670313119888306, 0.01254013366997242, -0.00849857646971941, 0.008840840309858322, -0.020444286987185478, -0.005077943671494722, 0.006320097018033266, 0.008928922936320305, -0.010032651014626026, -0.002349823247641325, -0.011167122982442379, 0.002068513073027134, -0.004451322834938765, ...
[ -0.0094183050096035, 0.3231581449508667, 0.36593207716941833, -0.17816874384880066, 0.18550826609134674, 0.0822833701968193, 0.5950349569320679, -0.24065232276916504, 0.20543262362480164, -0.7857180833816528, 0.25191691517829895, 0.2123187929391861, -0.10744356364011765, 0.1378888785839080...
This is a question about evaporative cooling as used in residential evaporative cooling appliances. This type of cooling uses the heat in the ambient outside air to evaporate water and remove the heat from the air, then push the cooled air inside. The equation to predict the temperature of the resulting air after it's given up its heat to evaporate the water is as follows: $$T_{output} = T_{dry} - (T_{dry} - T_{wet}) * \epsilon$$ where $T_{output}$ is the output air temperature, $T_{dry}$ is the air temperature of the dry bulb, $T_{wet}$ is the air temperature of the wet bulb, and $\epsilon$ is the cooling efficiency. For example, on a very dry summer day (dry bulb 95 degrees, wet bulb 60 degrees) my evaporative cooler with 90% efficient media is capable of cooling the air to 63.5 degrees. However, this equation does not seem to take into account the temperature of the water itself. Does it matter? Intuitively, it would seem to make sense to me that hotter water would be easier to evaporate, since it's closer to its boiling point. Or maybe colder water is better because it will absorb more heat from the air? Or maybe it's a wash because the same amount of heat is required, but with hotter water, more is needed because it will evaporate faster? Help me understand this.
[ 0.006856975611299276, 0.010339930653572083, -0.014119090512394905, 0.008901961147785187, -0.01960030570626259, -0.010294279083609581, 0.008579034358263016, -0.025227446109056473, -0.012256151996552944, -0.019011640921235085, -0.0043215020559728146, 0.010897992178797722, -0.013951574452221394...
[ 0.10206484794616699, 0.04993148520588875, 0.6188392043113708, -0.18516264855861664, -0.08270612359046936, 0.24515576660633087, 0.13728824257850647, -0.7832274436950684, 0.32240399718284607, -0.39899173378944397, -0.35371530055999756, 0.26758575439453125, 0.04777711629867554, 0.411630034446...
I'm going crazy! This mirror has been broken for ages, I've e-mailed the landlord about it multiple times, but they keep telling me off. However, when I took a quick sneak peak at their reception computer, they have an e-mail sitting in their inbox from the mirror supplier asking them to come pick it up since it's been waiting for over a month. How do I get them to fix my mirror?
[ 0.01526301447302103, 0.004850518424063921, -0.006171772256493568, 0.01650254987180233, 0.011038010939955711, 0.005316208116710186, 0.004666596185415983, 0.0021122447215020657, -0.017216820269823074, -0.03070702590048313, -0.01125740259885788, 0.01992177404463291, -0.0013784188777208328, -0...
[ 0.8575794100761414, 0.3215997815132141, 0.5647788643836975, -0.04340196028351784, 0.2893323004245758, 0.10437450557947159, 0.8850098252296448, 0.2986680269241333, -0.5962660312652588, -0.05348251387476921, 0.22798196971416473, 0.2859227955341339, 0.009653119370341301, 0.8599140048027039, ...
I'm trying to connect a usb garmin forerunner device to my system, but the usb device is not detected. I have loaded the following mods: garmin_gps 11544 0 usbserial 25449 1 garmin_gps But I never see any device detected in neither lsusb nor dmesg when I connect the device. Other USB devices like disks and mp3 players seems to work fine. I'm unsure how I can continue debugging to find the issue and wonder if someone can guide me?
[ -0.01708701252937317, -0.0065653459168970585, -0.007815939374268055, 0.017641151323914528, 0.010754916816949844, -0.022518455982208252, 0.009478077292442322, -0.009290081448853016, -0.018115315586328506, -0.020191185176372528, 0.0015986160142347217, 0.005978242494165897, -0.01054888777434825...
[ 0.17597153782844543, 0.3105964660644531, 0.24078255891799927, 0.14027343690395355, 0.0030643686186522245, 0.008294658735394478, 0.24483689665794373, 0.06527061015367508, -0.2461681365966797, -1.0016173124313354, 0.25590112805366516, 0.6030328273773193, -0.2625190317630768, 0.05340810120105...
I have a bunch of tarball backups which I just restored onto my new Windows 8.1 + Cygwin system using GNU tar: `zsh$ for file in **/*.tgz; do tar xvzf $file; done` To my surprise a lot of these extracted files were corrupt. I tried replacing GNU tar with BSD tar and repeated the process, but the same files were still corrupt. Then I tried extracting them with WinRAR, and they turned up just fine. Does anybody know what's going on?
[ 0.005358872003853321, 0.014352452009916306, -0.009014183655381203, 0.01777489483356476, -0.003978892229497433, -0.014316236600279808, 0.007737654261291027, 0.0024572601541876793, -0.017943715676665306, -0.03664788976311684, -0.007644077762961388, 0.009663276374340057, -0.008227052167057991, ...
[ 0.4595907926559448, 0.3611071705818176, -0.07696161419153214, 0.184776172041893, -0.11942886561155319, -0.3024759292602539, 0.3930049240589142, 0.37781253457069397, -0.22221167385578156, -0.3696472644805908, -0.08871085941791534, 1.044123888015747, -0.30460768938064575, 0.23642566800117493...
![curve](http://i.stack.imgur.com/5JCmh.gif) In the black body radiation curve, why is the wavelength directly proportional with the intensity before the peak and inversely proportional after the peak.
[ 0.02143339067697525, 0.010540514253079891, -0.002549707656726241, 0.02401644177734852, -0.015906838700175285, -0.024174829944968224, 0.014200418256223202, 0.009332379326224327, -0.027023836970329285, -0.08999785035848618, -0.01946547068655491, 0.016013585031032562, -0.022164296358823776, 0...
[ 0.5033144354820251, 0.23518534004688263, 0.32464009523391724, -0.16183432936668396, -0.09118577837944031, -0.33271241188049316, 0.40884634852409363, -0.6841961741447449, -0.3964010179042816, -0.2389291375875473, -0.12199445813894272, 0.6716203093528748, -0.16305096447467804, 0.194735407829...
I have a blog. (just started) I am having a confusion whether I should keep the blog focused on only one subject (for instance, technology) or not? If I post different posts on various topics (with proper categorization), will it affect my traffic? i.e. If I write about technology, photography, personal experiences and some more, will it affect my readers' mindset? Also, can I use a regional language for some posts?
[ -0.0005293671856634319, 0.030480187386274338, 0.0021163499914109707, 0.029501168057322502, 0.02377765066921711, 0.0024675617460161448, 0.009090948849916458, 0.0230474304407835, -0.011059850454330444, -0.018015215173363686, -0.017922418192029, 0.016095440834760666, 0.0032937093637883663, 0....
[ 0.8693431615829468, 0.49320581555366516, 0.346638023853302, -0.06427808105945587, -0.10908165574073792, -0.35552582144737244, 0.3194189965724945, 0.4671495258808136, -0.2573099136352539, -0.45058077573776245, -0.0022659231908619404, 0.16445334255695343, 0.20252397656440735, 0.3250814676284...
I have a website with web hosting I am planning on switching away from (Windows Server). So I got a hosting plan with a linux server and I was going to build a site there using Perch CMS. Right now I have a temporary domain (minus a real domain name) at http://gator1784.hostgator.com/~elaine/perch/ However Perch does not recommend using a temporary domain with a tilde (~elaine), what should I do? What are my options? I would just like to build this site on the new server and not have too many issues when I switch the domain over. The current website has to remain live for now (on the other server which is windows) while I build the current site on a linux server.
[ -0.0006582671776413918, 0.0036209465470165014, -0.007278991863131523, 0.014646776020526886, 0.0014376535546034575, -0.004090567585080862, 0.010839506983757019, -0.01682441309094429, -0.01946776546537876, -0.035742320120334625, -0.019486572593450546, -0.0012865231838077307, 0.0003325557336211...
[ 0.49753764271736145, -0.04110180586576462, 0.6271350979804993, -0.02777453139424324, -0.18115296959877014, -0.23253077268600464, 0.09989824146032333, 0.4749894440174103, -0.6142726540565491, -0.6984339952468872, 0.6533552408218384, 0.053611498326063156, 0.2415100336074829, 0.53483814001083...
Manipulate[{u*v}, Row[{Control[{u, 0, 1}], Control[{v, 0, 1}]}], ControlType -> VerticalSlider, ControlPlacement -> Up] Q1.- How to select the operation *, +, - or /? in the first parameter of Manipulate Q2.- AND how to force Manipulate/MAthematica to WAIT that I modify various paramenters ( u , v and operation) to execute the operation, AND NOT make the operation in every change I made in u, and in every change of v, and in every change of "+-*/" ? . Iwant to select u, v and operation, and THEN (only whe I push a buttom or similar) execute with the parameters.
[ 0.003068163525313139, 0.004082459956407547, -0.003040284151211381, 0.013499731197953224, -0.00523705780506134, -0.004564398899674416, 0.005195125937461853, 0.008520403876900673, -0.01484671887010336, -0.01055170502513647, -0.02154206484556198, 0.00175857066642493, -0.015742914751172066, 0....
[ -0.041910503059625626, 0.13040441274642944, 0.3041626214981079, 0.10778871178627014, 0.025296470150351524, 0.6006671190261841, -0.11435189098119736, -0.5997137427330017, -0.031170539557933807, -0.028339149430394173, 0.3203631043434143, 0.441052109003067, -0.06690533459186554, -0.2387624084...
> **Possible Duplicate:** > Remove ugly borders around clickable cross-references and hyperlinks Let's say we have document with url: \documentclass[a4paper,10pt]{article} \usepackage{hyperref} \begin{document} \url{http://tex.stackexchange.com/} \end{document} How to get rid of box around url (and ensure it's black) ? (pdflatex)
[ 0.0029740820173174143, 0.0020910317543894053, 0.004095207434147596, 0.01922641322016716, -0.003151259385049343, -0.010118677280843258, 0.0070878262631595135, 0.010462290607392788, -0.015444139018654823, -0.007781302556395531, -0.014666520059108734, 0.0003966711228713393, -0.00484857475385069...
[ 0.28320056200027466, 0.052223362028598785, 0.42228230834007263, 0.05174720659852028, 0.08964517712593079, -0.1389487087726593, 0.5269006490707397, -0.3449004292488098, -0.1875525265932083, -0.750837504863739, -0.18209049105644226, 0.38233044743537903, -0.4431194067001343, 0.077595755457878...
I decided to be sensible and scout a level 9 Wilderness before attacking it and found out it had at least 5000 Stenches: ![enter image description here](http://i.stack.imgur.com/OubIT.png) What size of army from my side would I need to defeat them and how can I estimate what army I'll be up against in the future? Any suggestions on upgrades and unit composition would be welcome!
[ -0.00016376910207327455, 0.02621397376060486, -0.0013176738284528255, 0.01649555191397667, -0.0012876172550022602, -0.005497807636857033, 0.006069457624107599, -0.009864763356745243, -0.023650364950299263, -0.0028354439418762922, 0.01146283932030201, 0.015413864515721798, -0.0044607203453779...
[ 0.6540817618370056, 0.22877277433872223, -0.1428474485874176, -0.03875313326716423, -0.4469951391220093, 0.22521783411502838, 0.8107011914253235, -0.09323635697364807, -0.012930553406476974, -0.5936180353164673, 0.07416355609893799, 0.3554512560367584, -0.1973133385181427, 0.10740644484758...
This question points out that in the book class, as soon as the `\backmatter` command appears, the chapter numbering ceases to work properly. Thus, one is essentially forced to put the bibliography after the appendices. In principle, I agree with this ordering. However, my university's administration requires that the contents of my thesis appear in the following order: * Intro and so forth * Body * Bibliography * Appendices It is causing me a major headache to get the numbering of the appendix chapters to appear at all. Any suggestions for a workaround?
[ 0.0019979269709438086, 0.0194347333163023, -0.015653889626264572, 0.02675260230898857, -0.016563478857278824, 0.02152763493359089, 0.009676877409219742, 0.01810387894511223, -0.019387438893318176, -0.03568682447075844, -0.020700031891465187, 0.0008381897350773215, -0.015936896204948425, 0....
[ 0.19414278864860535, 0.4313548505306244, 0.5444733500480652, -0.05839722231030464, 0.031629979610443115, -0.18156732618808746, 0.039550766348838806, 0.1296018660068512, -0.14312048256397247, -0.46850940585136414, -0.2960187792778015, 0.49759209156036377, 0.10972073674201965, 0.237432867288...
Which preposition is correct in the phrase " _proficient in/at/with English_ "?
[ -0.00983542762696743, 0.030119480565190315, -0.028106534853577614, 0.017157888039946556, -0.024256659671664238, 0.06381488591432571, 0.01877194456756115, -0.039839062839746475, -0.006244027521461248, 0.037306226789951324, -0.011147887445986271, 0.024226203560829163, 0.010785261169075966, 0...
[ -0.17783710360527039, 0.31271690130233765, -0.04209728538990021, -0.21495355665683746, -0.30246663093566895, 0.2643367648124695, 0.413225919008255, 0.29773810505867004, -0.12730778753757477, -0.27369481325149536, -0.2564212381839752, 0.7789473533630371, 0.13011890649795532, -0.610623300075...
When I put in the preamble. It has several warning as > Token not allowed in a PDF string (Unicode) Because I use the `$^1$` in the title. Then when I put in the document. the warning appears. How do I solve it? Thanks > Option `pdfauthor' has already been used My MWE \documentclass[presentation]{beamer} \let\Tiny\tiny \usepackage{hyperref} \usetheme{Berkeley} \begin{document} \renewcommand{\thefootnote}{$\aleph$} \title[Cross]{\large Experimental Studies % \thanks{\scriptsize Project supported by foundation: Supported by the National Nature Science Foundation}} \author[A,B,C]{A$^1$, B$^1$ and C$^2$} \institute[Fluid Mechanics]{$^1$Department of Mechanics \& Engineering Science\\ $^2$College of Science} \date[The second presentation, 2013]{The second presentation, 2013} %\logo{\includegraphics[height=16pt]{picture/Fudanlog.PNG}} \date[\initclock\tdtime]{\today} \begin{frame} \titlepage \end{frame} \section{A section} \subsection{A subsection} \begin{frame} \frametitle{Frame title1} \framesubtitle{frame subtitle1} Some text s \end{frame} \end{document}
[ -0.008377447724342346, 0.004290154669433832, 0.005235678981989622, 0.02154267020523548, 0.006501664407551289, 0.015565150417387486, 0.007909162901341915, 0.009235006757080555, -0.01143583469092846, -0.01927904412150383, -0.009593651629984379, -0.0011473078047856688, -0.00627104053273797, 0...
[ 0.009141451679170132, 0.43073228001594543, 0.9293594360351562, 0.12548814713954926, 0.19178101420402527, -0.10421505570411682, 0.41330304741859436, -0.31613633036613464, -0.11285261064767838, -0.5782102942466736, -0.04477497190237045, 0.505190372467041, -0.17989729344844818, 0.078754097223...
I have an svm-related question. I have an unbalanced dataset, meaning classA could be 1/10 to 1/35 of classB. Well I am interested in getting a _linear_ svm which would separate the data and would achieve the greatest possible recall of classA. Let me show some **examples** to be more specific: * If 1000 samples are mapped to classA by the hyperline I want to have 80%-90% of classA. This does not mean that classA would have only 1000 samples, it could have 2000 or 10000. I am not interested in the accuracy of classA, i.e. the standard criterion to get the more samples of classA by svm out of the total samples belonging to classA. * If I can get 100 samples of classA and no of classB it would be preferable than getting 500 samples of classA and 500 of classB. My question is how do I choose this hyperplane. I have tried the following: 1. I used unbalanced weights for the samples (inverting the class cardinalities) but I can get high classA accuracy but low recall. 2. I tries to use the default weights, i.e. for all being 1, but I got peculiar results: when run on _train set_ for all samples it gets an accuracy of 95% (as expected since dataset is unbalanced). **BUT** if checked in classA members it gives 0%! So have you got any ideas besides playing around with weights? P.S.1 I am using liblinear (in matlab if that matters). P.S.2 I know svm optimization criterion basically does not account for what I am looking for since margin maximization maximizes accuracy but I thought I would give it a try.
[ 0.009900346398353577, 0.01607280783355236, -0.013930251821875572, 0.021411150693893433, 0.008476309478282928, 0.004473501816391945, 0.007873909547924995, 0.01314455084502697, -0.011851240880787373, -0.006454250775277615, -0.005716152489185333, 0.018245363608002663, -0.009446581825613976, 0...
[ 0.0791189968585968, -0.13617514073848724, 0.27830901741981506, 0.25217288732528687, -0.23274697363376617, 0.2782350778579712, 0.005648489575833082, -0.07671629637479782, -0.2637523412704468, -0.7296996712684631, 0.1975499987602234, 0.029439834877848625, -0.08060887455940247, 0.429543048143...
I have downloaded the sampe dataset under "my documents". When I extract the contents of the zip file, the only folders that are actually installed, are "gml", "gps" and "shapefiles". Thus, "climate", "csv", "grassdata" and "raster" do not get extracted correctly. Does anyone have an idea of what is going here?
[ -0.018006103113293648, -0.005647978745400906, 0.006081961095333099, 0.024499885737895966, -0.0013575056800618768, 0.012669258750975132, 0.008345128037035465, -0.010531107895076275, -0.01943170092999935, -0.03380908444523811, 0.007443327456712723, 0.002418848453089595, -0.01878412999212742, ...
[ 0.47259363532066345, 0.21859021484851837, 0.6496918201446533, 0.021630339324474335, 0.5065471529960632, -0.2232276350259781, 0.497171550989151, 0.2078218162059784, -0.39405331015586853, -0.3760833442211151, -0.32025259733200073, 0.08137448132038116, -0.16503044962882996, -0.113734677433967...
How do I convert (in a clear and concise way) from ESRI.ArcGIS.ADF.ArcGISServer.LayerDescription.SelectionFeatures to ESRI.ArcGIS.ADF.Web.Display.Graphics.GraphicsLayer I want the selected features and the attributes. _side note: I got the the selection features via FIDSet returned from a SOAP based MapServerProxy.QueryFeatureIDs query using a buffer geometry to pull intersecting features._
[ -0.01179911196231842, 0.005387464538216591, -0.0031773054506629705, 0.015410944819450378, -0.009157502092421055, -0.01697050780057907, 0.00973881222307682, 0.02947137877345085, -0.016761554405093193, 0.010107558220624924, -0.010379569604992867, 0.019650757312774658, -0.016752541065216064, ...
[ 0.2511001527309418, -0.1338091641664505, 0.6665344834327698, -0.12999297678470612, -0.07030560076236725, 0.14534878730773926, -0.08428735285997391, -0.6600834727287292, -0.014175992459058762, -1.125474214553833, -0.15945330262184143, 0.49548888206481934, 0.053678445518016815, -0.0329791493...
The problem is that the code works but when i make a post sticky it doesn't appear at first. The function of the code: It only shows posts that have a thumbnail or a slideshow image. <?php $args = array( 'numberposts' => 5, 'meta_query' => array( 'relation' => 'OR', array( 'key' => '_thumbnail_id', 'compare' => '!=', 'value' => '' ), array( 'key' => 'slideshow_image', 'compare' => '!=', 'value' => '' ) ) ); $my_posts = get_posts( $args ); global $post; foreach( $my_posts as $post ) : setup_postdata($post); ?> <?php the_post_thumbnail('thumb-small'); ?> <a href="<?php the_permalink(); ?>"> <?php the_title(); ?> </a> <?php endforeach; wp_reset_postdata(); ?>
[ -0.010236754082143307, 0.007710292469710112, 0.007887251675128937, 0.010984394699335098, -0.005158613435924053, -0.007830711081624031, 0.005444766022264957, 0.0032281456515192986, -0.012368621304631233, 0.008810802362859249, -0.022309351712465286, -0.0029319205787032843, -0.00391875393688678...
[ 0.1041940227150917, 0.2512942850589752, 0.8341108560562134, -0.06128020957112312, -0.02995561622083187, -0.09312102198600769, 0.353657603263855, 0.0035800328478217125, -0.13679008185863495, -0.5153836011886597, 0.2578335702419281, 0.266067236661911, -0.4213056266307831, 0.28618890047073364...
I'm looking at purchasing a refurbished Xbox 360. The one I am currently looking at is on Gamestop here. It doesn't seem to list how large the hard drive is however. Is there a way to tell? Or can I add another hard drive (20Gb) to it?
[ -0.003498917678371072, -0.00708877295255661, -0.01845882274210453, 0.009521249681711197, 0.06213914602994919, -0.018215153366327286, 0.00909520499408245, -0.0006253154133446515, -0.025924868881702423, -0.030102066695690155, 0.012928266078233719, 0.010821440257132053, 0.016463465988636017, ...
[ 0.7254371643066406, -0.02025018073618412, 0.3319360911846161, 0.3818730413913727, 0.1695007085800171, 0.24364711344242096, -0.21530930697917938, 0.42538782954216003, -0.49248605966567993, -0.4596826136112213, 0.1680171638727188, 0.39755961298942566, 0.3035759925842285, 0.27097928524017334,...
I have an app using OpenLayers and OpenStreetMap tiles. There are more zoom levels for the capital and less zoom levels for the entire country. Let's say the user is in the capital, in his car, in traffic, with localisation active. The app is showing the map at the maximum zoom level. Them map is following him like a GPS, keeping current position in the center and marking it with a graphic symbol. When the car reaches the bounds of the city the map have no tiles and show pink color. I need to change the zoom level before the pink missing tiles appears. I can calculate an area inside the bounds of the capital and when the current position is inside this area the app will zoom out to the maximum zoom level available for the country. Is this the best solution? The same problem exists when panning map with the finger. I thought to restrictExtend to the capital but then the user will be captive inside and the single chance to see the country again is to zoom out. Do you know an algorithm or an OL function/event to help me find the best solution? Thank you, Bogdan
[ -0.013212520629167557, 0.005983677692711353, -0.01672574318945408, 0.00853804312646389, -0.013961649499833584, 0.019273199141025543, 0.010120267048478127, 0.007534143514931202, -0.012597497552633286, -0.022654077038168907, -0.009929575026035309, 0.012267583049833775, 0.0032499046064913273, ...
[ 0.2742850184440613, -0.04683688282966614, 0.7624971866607666, 0.3280312418937683, -0.16724462807178497, 0.04046420380473137, 0.2622801959514618, 0.0012660193024203181, -0.45943745970726013, -0.8398870229721069, -0.02085372805595398, 0.21953463554382324, -0.2101459950208664, -0.045432198792...
I have an excel file with missing values encoded as blanks and I want this excel file to be converted into DBF but with the conversion process, the blanks (missing values) become 0. The reason I am doing this is because I want to perform interpolation but because of the missing values becoming 0 the interpolation becomes wrong. My question is how to treat missing values as without deleting the point shapefile and still proceed with interpolation.
[ 0.00224914588034153, 0.024310780689120293, -0.010424732230603695, 0.009854859672486782, -0.016526632010936737, 0.0031484125647693872, 0.012144193984568119, 0.007500194478780031, -0.014918413013219833, -0.025817910209298134, 0.0006791276391595602, 0.018254749476909637, -0.019708743318915367, ...
[ 0.14103154838085175, 0.4334565997123718, -0.07251010835170746, -0.21868056058883667, -0.2566203474998474, 0.04663510620594025, 0.16332386434078217, -0.3599376082420349, 0.3455773591995239, -0.8725961446762085, 0.1707964539527893, 0.30849263072013855, -0.2340313047170639, 0.4137825667858124...
I'm writing a small webpage that will allow the user to download small tools to run locally, I'm thinking what might be a suitable choice of language for the small tools. There will be a rather large amount of fairly specific small tools, so I'm interested in cutting down the development time for each one as much as possible. Requirements: * The tools will have graphics, nothing fancy mainly 2D-shapes. * Can be run offline. * Preferably no installation at all, just run. Some of the users will not be very technically at all. * Cross-platform, primarily Windows and OSX but also a bit of Linux. My current thoughts: * Python. Easy to write in but can be a hassle for the user to get to run if using libraries like PyGame. I can't count on the user already having python installed on windows or osx. * I'm discounting .NET and Mono for more or less the same reason as Python, they'd need installers, or is there a way to ship everything needed in the same folder? * C++. Takes a slightly longer time to write things in, but does support the requirements. I thought about using a scripting language(Lua, Angelscript or similar) on top of SDL to get the benefits of both c++ and shorten the development time once the middleware is well developed enough. Do you have any other suggestions? EDIT: Please leave a comment as to why if you vote negatively.
[ -0.013316282071173191, 0.015388278290629387, 0.003547930158674717, 0.006170911714434624, -0.012820020318031311, -0.0021897293627262115, 0.005616367794573307, 0.016859911382198334, -0.01601359061896801, -0.004199893679469824, -0.012547768652439117, 0.004893030971288681, 0.02016930654644966, ...
[ 0.5841259956359863, 0.29700586199760437, 0.06190767511725426, 0.23625300824642181, 0.008662259206175804, 0.08237241953611374, 0.08440007269382477, 0.18969425559043884, -0.009533194825053215, -0.9369583129882812, -0.21543988585472107, 0.700751006603241, 0.06509938091039658, -0.0561519600450...
I'm developing a new ArcGIS Server JavaScript application, and I'd like to show the progress to the client, so he can provide feedback as I develop it. The application consists of an Index.html file, a *.JS file containing my code, a *.CSS file for the styles and a *.JSON file containing the map configuration. I don't have access to a public-facing web server in my organisation. **Are there any sites (preferably free) where I can upload my files, so he can run and test the application?** I looked at github but from what I can gather it's a _repository_ , rather than a place where my application will actually _run_ (eg, test the Identify tool and verify that it works) - please let me know if that's not correct. Thanks
[ -0.01564532145857811, 0.006334793753921986, 0.006429185159504414, 0.008960872888565063, 0.006448070518672466, 0.007050009910017252, 0.008571288548409939, 0.0018508201465010643, -0.021799664944410324, -0.02792777493596077, 0.006018956191837788, 0.008078176528215408, 0.0005895440699532628, 0...
[ 0.5686716437339783, 0.2656972408294678, 0.40464717149734497, 0.005657556001096964, -0.27813050150871277, -0.023103665560483932, 0.22529390454292297, -0.22722314298152924, -0.04418030381202698, -0.7763220071792603, 0.16045351326465607, 0.208144411444664, 0.034273501485586166, -0.31201475858...
I am marking some student work and one of the sentences was > The author is by Katherine Patterson. What is the term for the error in this sentence?
[ -0.005637084133923054, 0.008363284170627594, -0.007757204584777355, 0.059327248483896255, -0.0035786982625722885, 0.03338881582021713, 0.01568593643605709, -0.05603349208831787, -0.03459079936146736, 0.004671080969274044, -0.03282158449292183, -0.005656373221427202, 0.007544690743088722, -...
[ 0.4980833828449249, -0.019560599699616432, 0.009131831116974354, 0.057482101023197174, -0.125905379652977, 0.10690245032310486, 0.23255209624767303, 0.30543941259384155, -0.13389483094215393, -0.6996210217475891, 0.33865779638290405, -0.29856887459754944, 0.10670860856771469, 0.05857601389...
I have a minted code listing, displayed in a multicol environment with a separating border between the columns: ![A minted code listing, displayed in a multicol environment with a separating border. The border is too close to some of the content.](http://i.stack.imgur.com/EZC1A.png) The border is too close to some of the content, like `stracka(1,sg,ka).` and `stracka(3,sg,ka).`. **What's a good way to fix it?** _test.tex_ : RequirePackage[l2tabu, orthodox]{nag} \documentclass[a4paper,12pt,oneside]{scrartcl} \usepackage{multicol} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage[swedish]{babel} \usepackage{lmodern} \usepackage[a4paper]{geometry} \usepackage{microtype} \usepackage{color} \usepackage[all]{xy} \usepackage{minted} \usepackage{caption} \setkomafont{disposition}{\normalfont\bfseries} \renewcommand\listingscaption{Kodlistning} \begin{document} \setlength{\columnseprule}{0.5pt} \begin{multicols}{4} \inputminted[]{prolog}{someprolog.pl} \end{multicols} \label{Blah} \captionof{listing}{Beskrivande text} \end{document} _someprolog.pl_ : stracka(1,c,sg). stracka(1,sg,ka). stracka(1,ka,o). stracka(1,o,t). stracka(2,c,sg). stracka(2,sg,sm). stracka(2,sm,r). stracka(2,r,h). stracka(3,c,sg). stracka(3,sg,ka). stracka(3,ka,o). stracka(3,o,kr). stracka(4,c,sg). stracka(4,sg,ka). stracka(4,ka,o). stracka(4,o,j). stracka(5,c,sg). stracka(5,sg,sm). stracka(5,sm,r). stracka(5,r,n). stracka(6,c,sg). stracka(6,sg,sm). stracka(6,sm,r). stracka(6,r,h). stracka(7,c,st).
[ 0.004992990754544735, -0.0016263193683698773, -0.00531410239636898, 0.020250491797924042, 0.01776280626654625, 0.009685280732810497, 0.007550331298261881, 0.010779478587210178, -0.010560188442468643, 0.03227695822715759, -0.01184244453907013, 0.0052663725800812244, -0.005378221161663532, 0...
[ 0.23080945014953613, 0.4059118628501892, 0.23594489693641663, 0.14337579905986786, -0.08170782029628754, 0.020750172436237335, 0.2997797727584839, -0.10226212441921234, -0.1915186047554016, -0.4380705952644348, 0.040671318769454956, 0.05675739794969559, -0.2296166568994522, 0.3523583412170...
How can I tell when a given process exits? Like when it's done running and stuff. For instance: # Command 1 wget http://releases.ubuntu.com/14.04/ubuntu-14.04-desktop-amd64.iso # Command 2 echo "I'm a command" How would I schedule to have `Command 2` run when `Command 1` exits?
[ -0.002617407590150833, 0.005357064306735992, -0.009247387759387493, 0.01941787637770176, -0.020427949726581573, -0.011219681240618229, 0.008704342879354954, -0.0008764818776398897, -0.024415384978055954, 0.02873176336288452, -0.021687101572752, -0.0037749893963336945, 0.019673632457852364, ...
[ 0.5696549415588379, 0.09028073400259018, 0.09548813104629517, -0.3266249895095825, 0.04714633896946907, -0.20412081480026245, 0.26509422063827515, -0.23807328939437866, -0.4121074378490448, -0.4685741364955902, -0.06148139759898186, 0.2927134931087494, -0.012101618573069572, 0.172399923205...
Hi I have created a custom post type and custom taxonomy.The custom post type page is named page-portfolio.php.I have create a loop in it in witch I am trying to link to each posts single page.For that I have created a template single-portfolio.php but for some reason when I click the permalinks I get sent to the 404.php page.Here is my entire structure code: //custom post type and taxonomy declaration <?php add_action('init' , 'portfolio_manager'); function portfolio_manager(){ $args = array( 'label' => 'Portfolio Manager', 'singular_label' => 'Portfolio Manager', 'public' => true, 'show_ui' => true, 'capability_type' => 'post', 'hierarchical' => false, 'supports' => array('title' , 'editor' , 'thumbnail'), 'rewrite' => array( 'slug' => 'portfolio', 'with_front' =>false ) ); register_post_type('portfolio' , $args); } add_action('init' , 'portfolio_taxonomy'); function portfolio_taxonomy(){ register_taxonomy("portfolio-type", array("portfolio"), array("hierarchical" => true, "label" => "Portfolio Types", "singular_label" => "Portfolip Type", "rewrite" => true, "slug" => 'portfolio-type' ) ); } ?> //page-portfolio.php loop $portfolio = new WP_Query(array( 'post_type'=>'portfolio', 'posts_per_page'=>'-1' )); while($portfolio->have_posts()): $portfolio->the_post(); global $post; ?> <li class ="item" data-type="<?php echo str_replace(" " , "" , strip_tags(get_the_term_list($post->ID, 'portfolio-type' , '' , '' , ''))); ?>" data-id="id-<?php the_ID(); ?>"> <span><a href="<?php the_permalink(); ?>">Details</a></span> <?php endif ?> </li> <?php endwhile; ?> If single-portfolio.php is not the page template I should be creating for this page then what is the corect template? If single-portfolio.php is the corect template then what am I doing wrong here.Why is the permalink not sending me to single.portfolio.php?
[ -0.005543837323784828, 0.01284349150955677, 0.018653087317943573, 0.02978193387389183, 0.03782288730144501, -0.010663297027349472, 0.009379906579852104, 0.013680297881364822, -0.011921822093427181, -0.00952399242669344, -0.008789347484707832, 0.0040767136961221695, -0.014421498402953148, 0...
[ 0.21533797681331635, -0.1183333471417427, 0.7455990314483643, -0.27486366033554077, -0.31726089119911194, 0.24401184916496277, 0.27254754304885864, -0.6522907018661499, -0.21809564530849457, -0.3362029790878296, 0.31946229934692383, 0.4226965606212616, -0.32645630836486816, -0.019219979643...
I had trouble accessing wp-login.php, so I changed the name of the theme I'm developing to force WP to switch to default. I then turned on debugging and received the errors below. I'm a little puzzled as to why these have come about because they were not showing when I was last working on the theme a few days ago, and I've used get_bloginfo in other themes without issue. Any and all help re: how to fix these errors is greatly appreciated. Notice: get_bloginfo was called with an argument that is deprecated since version 2.2! The siteurl option is deprecated for the family of bloginfo() functions. Use the url option instead. in /home3/dyluxept/public_html/wp- includes/functions.php on line 3551 Warning: Cannot modify header information - headers already sent by (output started at /home3/dyluxept/public_html/wp-includes/functions.php:3551) in /home3/dyluxept/public_html/wp-content/themes/smallbiz/functions.php on line 141 Warning: Cannot modify header information - headers already sent by (output started at /home3/dyluxept/public_html/wp-includes/functions.php:3551) in /home3/dyluxept/public_html/wp-includes/pluggable.php on line 866
[ 0.002721050288528204, 0.008001514710485935, -0.001158679835498333, 0.02228141389787197, -0.018497569486498833, 0.013956273905932903, 0.006969298701733351, 0.016662422567605972, -0.013077951036393642, -0.031770650297403336, -0.007765254005789757, 0.01806303858757019, -0.00944594107568264, 0...
[ 0.28998109698295593, 0.22034022212028503, 0.45973411202430725, -0.15604931116104126, 0.017693351954221725, -0.018554918467998505, 0.35612037777900696, 0.13188371062278748, -0.30552226305007935, -0.5449912548065186, -0.10803873091936111, 0.7086103558540344, -0.4203946888446808, 0.1344177126...
I'm working on building out a custom theme and have been struggling with this one for a bit. I am trying to modify the HTML output of `the_post_thumbnail();` function. I need to do this because I am trying to support retina images on my site and would rather bake the functionality into my theme than load a plugin on the backend. By default, `the_post_thumbnail();` merely calls `get_the_post_thumbnail();` which I found here. My first thought was to plug into the **'post_thumbnail_html** ' filter, but I can't seem to get it working. So... This is how I'm calling my post thumbnails in the loop: <?php the_post_thumbnail('custom-thumbnail-size', array('class' => 'unique-class-here', 'title' => 'unique-title-here')); ?> This is the code I need output when I call `the_post_thumbnail();`... <img src="" alt="" data-src="image.png" data-alt="Alt text" class="retina unique-class-here" /> And the below code is what I currently have in my functions.php file: <?php function modify_post_thumbnail_html($html, $post_id, $post_thumbnail_id, $size, $attr) { $src = wp_get_attachment_image_src(get_post_thumbnail_id(), $size); $html = '<img src="" alt="" data-src="' . $src['0'] . '" data-alt="" class="retina" />'; return $html; } add_filter('post_thumbnail_html', 'modify_post_thumbnail_html', 99, 5); ?> A couple things to note. I am not sure how to pass in the appropriate metadata text in the 'data-alt' attribute. Also, I need to be able to pass the specific post_thumbnail size I need since I use custom post_thumbnail sizes throughout my theme. Lastly, you can see the attribute array needs to pass in classes in addition to the default 'retina' class as well as any other attributes in that array. Thanks in advance for any help. There's really not much online about this topic and I figured my question was different enough to warrant an additional post to this community. Please let me know if you have any thoughts, solutions and/or need any clarifications.
[ -0.003135578939691186, 0.012117292732000351, 0.004328806418925524, 0.01418619230389595, -0.011604668572545052, -0.00795462541282177, 0.007840869948267937, 0.026838114485144615, -0.013729787431657314, -0.01808532141149044, -0.018169216811656952, 0.003539392491802573, -0.0015217266045510769, ...
[ 0.7867546081542969, 0.10626137256622314, 0.8860021233558655, -0.3383471965789795, -0.026322495192289352, -0.1922965794801712, 0.16341929137706757, 0.1838214099407196, -0.12158544361591339, -0.6384457945823669, -0.07848811894655228, 0.5949984192848206, -0.32483696937561035, 0.42283269762992...
AFAIK all the celestial objects have a spin motion around its axis. What is the reason for this? If it must rotate by some theory, what decides it's direction and speed of rotation? Is there any object that does not rotate about its axis?
[ 0.00560629740357399, 0.0036207332741469145, -0.01359378732740879, 0.03217991068959236, -0.01643749512732029, -0.0036499483976513147, 0.011440658941864967, 0.019794000312685966, -0.027509227395057678, -0.0036063524894416332, 0.0027942101005464792, 0.020670447498559952, -0.005750644952058792, ...
[ 0.33464178442955017, 0.4816453754901886, 0.5092136263847351, -0.1396222710609436, -0.30234822630882263, -0.10169000178575516, -0.02264038473367691, 0.14170733094215393, -0.4580113887786865, 0.11786864697933197, 0.17734338343143463, -0.00122413644567132, -0.6697444319725037, 0.4440703988075...
Is there any agreement on when to reduce data dimension before clustering in order to avoid curse of dimensionality? My intuition is that if I have say 1000 points and data dimension is 10 then it is OK to cluster. But if dimension is 50 then it is not OK because data points become sparse and hard to cluster (as a result expect to obtain "too much" clusters).
[ -0.018901245668530464, 0.029767081141471863, -0.01977192610502243, 0.010116667486727238, 0.0478094145655632, 0.008305052295327187, 0.010944933630526066, -0.01915840245783329, -0.014088907279074192, -0.022982453927397728, -0.002642019186168909, 0.02216349169611931, -0.013826221227645874, 0....
[ -0.28665173053741455, 0.15430372953414917, 0.44924426078796387, 0.15340009331703186, -0.05616461858153343, 0.015985745936632156, 0.08351100236177444, -0.14734211564064026, -0.21305575966835022, -1.0142306089401245, 0.1633477658033371, 0.11774896085262299, 0.12553560733795166, 0.27565330266...
First of all, I can't use Trac, which is what the WP dev community uses to track bugs. However, my consulting company is using WordPress as a CMS for just about everything - product information, sales portal, news site, etc. I'd like to set it up to integrate our plug-in listing with a bug tracker ... Right now, I have Flyspray set up as a standalone bug tracker. It works alright, but it's a separate domain (http://bt.jumping-duck.com) so that it remains separate from WordPress. But I'd like to be able to pull in numbers of tickets onto the WP page listing our plug-ins. Ideally, users could also submit bug reports directly from the plug-in info page. Does anyone have any experience integrating WP with a bug tracking system like this? Should I stick with keeping the two systems separate?
[ -0.005204495973885059, 0.002986153122037649, -0.003126381663605571, 0.00466884532943368, 0.011564001441001892, 0.012151634320616722, 0.00676568690687418, 0.005973425228148699, -0.015593505464494228, -0.030290961265563965, -0.004332302138209343, 0.011851165443658829, -0.004549826495349407, ...
[ 0.9906503558158875, -0.12759916484355927, -0.02628212235867977, 0.09872213751077652, -0.07262835651636124, -0.43406200408935547, -0.009315772913396358, 0.09946316480636597, -0.3548966944217682, -0.2572663426399231, 0.34944137930870056, 0.47631338238716125, -0.6785461902618408, -0.028432995...
Here's the situation, I like playing QE Invoker mid, but I have troubles when I play against Pudge. I will always crush him in lane. I always harrass him so he never comes in for any last hits. I will keep creeps in between us so he cannot hook me. Basically, the laning phase is very easy. But once Pudge hits level 6+, (if he's good) he will start to nonstop roam around the map and gank. I am then presented with these problems: * If he's good, he will get smoke, and wards become useless. I will never really know where he is on the map. * Because I do not know where he is I can't position myself to block a hook. When it reaches that point in the game where I don't know where Pudge is, what do I do to safely continue farming gold/exp? It seems there is nothing I can do to stop the risk of being helplessly hooked by a smoked-up Pudge. PS. This question works for any hero really, but if you have some tips that could be applied to Invoker specifically that would be nice.
[ -0.011445770040154457, 0.011161165311932564, -0.007185103837400675, -0.0007840310572646558, 0.019910871982574463, 0.018756048753857613, 0.00710568530485034, -0.012081708759069443, -0.015282807871699333, -0.0038942801766097546, -0.001337229274213314, 0.0078373271971941, -0.015043569728732109,...
[ 0.13793182373046875, -0.16488017141819, 0.7369362711906433, 0.025768060237169266, -0.1416773945093155, -0.4928010106086731, 0.47398462891578674, -0.3265800178050995, -0.08297320455312729, -0.46336862444877625, 0.11087074875831604, 0.30371972918510437, -0.006887250579893589, -0.163269877433...
Is there any open-source library to aggregate detached polygons? I know that there is a tool in ArcInfo, but it is commercial.
[ -0.013483911752700806, 0.0072970641776919365, -0.017359986901283264, 0.035593196749687195, 0.04457647725939751, -0.00029102800181135535, 0.013711964711546898, 0.0014511706540361047, -0.03711995109915733, 0.01675666682422161, 0.02437944896519184, 0.026595549657940865, 0.009220692329108715, ...
[ 0.47245296835899353, 0.1744331568479538, -0.10253038257360458, 0.4092549681663513, -0.11157846450805664, 0.015488345175981522, -0.4010012149810791, 0.42212334275245667, -0.30040961503982544, -0.43711790442466736, 0.26371124386787415, 0.23066017031669617, 0.024551665410399437, 0.05527960881...
My website is here but when I search term "ghiasi" in google, in text preview of my site, Google shows the menu content. But it is not useful! How to let google know where is the main content, So google shows parts of the content for users?
[ -0.0186468418687582, -0.021244529634714127, 0.025693610310554504, 0.029471611604094505, -0.013180561363697052, 0.02027682401239872, 0.014659370295703411, -0.005450398661196232, -0.029536129906773567, -0.02905024029314518, -0.02283257059752941, 0.0274409968405962, 0.0061193606816232204, 0.0...
[ 0.2967521548271179, 0.1696469634771347, 0.2063213586807251, -0.004108085297048092, -0.061663515865802765, -0.0646442323923111, 0.1389302909374237, 0.5376505851745605, -0.11012670397758484, -0.2882179319858551, -0.21140635013580322, 0.21755695343017578, -0.6405797600746155, 0.28722244501113...
Whenever I have used the Fisher's Exact test on tables larger than 2x2 contingency tables, SPSS has produced a test statistic that I can quote in my work. However, when I do the same with 2x2 tables, no test statistic is produced. I want the way in which I report my results to appear standard throughout my thesis. Therefore do I (a) omit the test statistic that is provided for the larger tables or (b) find some way of calculating the test statistic? A similar issue was noted in a question entitled "Fisher's Exact test value" but the answers didn't address this issue specifically.
[ 0.03406932204961777, 0.00898205116391182, -0.013713850639760494, 0.01707368716597557, 0.02317815274000168, 0.02273198775947094, 0.009493082761764526, -0.0023602372966706753, -0.017056670039892197, -0.045043036341667175, -0.005315701477229595, 0.0032728048972785473, -0.009799661114811897, 0...
[ 0.18747703731060028, 0.12324409186840057, -0.08328268676996231, 0.05668359622359276, -0.15115702152252197, 0.10750400274991989, -0.022421427071094513, -0.2526612877845764, -0.06830527633428574, -0.24558645486831665, 0.12825940549373627, 0.5292360186576843, -0.10866225510835648, 0.198339506...
Sorry for the basic question, but I can't track down a solution (or maybe it's just hard to search for). I'm creating my own infinite scroll system for a wordpress site I'm working on. For ease of development it would be beneficial for the 'paginated' content to be on a separate template (so that I can just output the content, and ignore the rest of the template). Can this be done in wordpress? i.e. `domain.tld/content-page` should use `content-page.php` `domain.tld/content-page/page/1` should use `content-page-paginated.php` I suppose the obvious alternative is having logic in content-page.php that checks to see if it's paged content and outputs differently if that's the case.
[ -0.002518050605431199, 0.013640044257044792, 0.002044729422777891, 0.013462565839290619, 0.0008365372195839882, 0.014360478147864342, 0.004835029598325491, 0.000435730442404747, -0.015242980793118477, 0.001851267646998167, -0.009484205394983292, 0.00874103419482708, -0.023850377649068832, ...
[ 0.4264799654483795, -0.13771507143974304, 0.41839849948883057, 0.1673831045627594, -0.03902008756995201, -0.21903842687606812, 0.2852441072463989, 0.07281433045864105, -0.24041707813739777, -0.6175076365470886, 0.030543912202119827, 0.4101516008377075, -0.37791305780410767, 0.0298641715198...
I was in correspondence with the author of a LaTeX package in which I volunteered code similar to the following (I have anonymised it so that it is not obvious what package): \makeatletter \newcommand{\savemacrostatewithname}[1]{% \expandafter\let% \csname saved@macro@state@named@#1\endcsname% \package@internal@macro% } \newcommand{\restoremacrostatewithname}[1]{% \ifcsname saved@macro@state@named@#1\endcsname% \else% \errmessage{Saved macro state named "#1" doesn't exist}% \fi% \expandafter\let% \expandafter\package@internal@macro% \csname saved@macro@state@named@#1\endcsname% } \makeatother In their reply, one thing they said was that _instead of writing_ `#1` _in your macros, you should write_ `\detokenize{#1}`, _as it is safer_. Why is it safer? What might go wrong if I don't?
[ 0.02402019500732422, 0.023578312247991562, 0.006743438076227903, 0.013851813971996307, 0.047652170062065125, 0.011300615034997463, 0.00891321711242199, -0.013549573719501495, -0.013748999685049057, 0.009890414774417877, -0.0014361846260726452, 0.006855670362710953, 0.00438936660066247, 0.0...
[ 0.030606253072619438, 0.14339441061019897, 0.12917371094226837, -0.07311344146728516, -0.028331516310572624, 0.2767728865146637, 0.3446916341781616, -0.3247302770614624, -0.10512193292379379, -0.45331376791000366, 0.08364783972501755, 0.4111618995666504, -0.1583528071641922, 0.075085088610...