text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
Say I have two vectors: Action.Taken = c(0,1,0,0,1,1,0,1,0) Success = c(0,0,0,1,0,1,0,1,0) The first tells me whether or not a specific action was taken in a trial and the second tells me whether or not that trial succeeded. How would I go analyzing these two vectors to answer the following question: Does taking action (`Action.Taken = 1`) affect whether or not success is had (`Success = 1`)? I'd like some measure of significance as an regression/hypothesis testing. I'm looking for an answer that I can implement using R. I am also quite new to stats, so it would be nice if someone could give me a relatively simple, straightforward answer/example. Thanks!
[ -0.01802447997033596, 0.028420904651284218, -0.013953091576695442, 0.014961998909711838, -0.0075295669957995415, 0.023903369903564453, 0.00913580134510994, -0.0007342789322137833, -0.010320131666958332, 0.026691555976867676, -0.023167220875620842, 0.012793166562914848, -0.002723671030253172,...
[ 0.22827522456645966, 0.06512416154146194, 0.17649327218532562, -0.014056110754609108, -0.2605205178260803, 0.6145632863044739, -0.14642474055290222, -0.5039922595024109, -0.2539408802986145, -0.216736301779747, 0.23673193156719208, 0.22003154456615448, -0.16287438571453094, 0.2396277189254...
I am using KOMA script and the `subfig` package like this \documentclass[11pt,a4paper,BCOR10mm,DIV11,toc=listof]{scrbook} \usepackage[caption=false, font=footnotesize, justification=RaggedRight]{subfig} The `caption=false` option is added in order to retain caption handling from `scrbook`. Now I would like to prevent entries for some figures from appearing in my LOF. a) Because I have some `\ContinuedFloat` figures and b) because I don't want figures from the appendices to appear in the LOF. Now I know that I can suppress a LOF entry by leaving the optional caption argument empty like this \caption[]{Bla bla} but only IF I load the `caption` package or remove the `caption=false` option from the `subfig` package. Now doing this unfortunately messes up some of my subfloat captions where space is tight. I could try fixing these manually, but maybe there is an alternative. So here are my two questions (the second one corresponding to the manually editing option..): 1. Is there a way to suppress selected figures from appearing in the LOF without using the `caption` package while still numbering the figures? 2. What would be the command(s) from the `caption` package that I could use to set the indent for selected (!) subfloat captions to zero, i.e. such that the second line starts without any indent? * * * EDIT: Add minimal example for the second question. Ok, so here is a minimal example of how I would like to change the indentation of a certain subfloat caption. Normally it's all indented such that the subfloat label (a) stands on its on and that's fine. Just for some certain subfloats I'd like to save space and start the second line completely left aligned. \documentclass[11pt,a4paper,BCOR10mm,DIV11,toc=listof]{scrbook} \usepackage[font=footnotesize, justification=RaggedRight]{subfig} \begin{document} \begin{figure}[p] \centering \subfloat[aaa aaa aaa aaa aaa aaa \newline This should be left aligned with no indent.]% {% \label{fig:a}% \rule{0.48\linewidth}{0.48\linewidth} % \includegraphics[width=0.48\linewidth]{}% }% \subfloat[bbb bbb bbb bbb bbb bbb \newline This should be left aligned with no indent.]% {% \label{fig:b}% \rule{0.48\linewidth}{0.48\linewidth} % \includegraphics[width=0.48\linewidth]{\rule{1in}{1cm}}% }% \caption[Bla]{Bla Bla} \label{fig:example} \end{figure} \end{document}
[ 0.014968983829021454, -0.0007934913737699389, -0.0005909528117626905, 0.03091433271765709, 0.0125131132081151, 0.015809405595064163, 0.009283563122153282, 0.009934190660715103, -0.014159184880554676, 0.0029566772282123566, -0.028494613245129585, -0.00036592595279216766, -0.000379611388780176...
[ 0.13609477877616882, 0.11216314882040024, 0.4362556040287018, 0.14073695242404938, 0.1207866221666336, 0.0950058326125145, 0.24727460741996765, -0.1880284994840622, -0.1939314901828766, -0.1866399198770523, 0.011332902126014233, 0.3278065621852875, -0.15368179976940155, -0.0399137511849403...
I'm trying to write a program in Python that exports data into PDF. Unfortunately, the well known reportlab module is not available on Python 3. (What a shame.) So I have decided to generate TEX files from the program and compile them with TeX. I know that my operating system have True Type fonts that can handle all of the characters that I need. So all I need is to use such a font for any UTF8 character in the document. I also need to create some tables and figures. It is that simple. I was trying to do it with pdfLaTex and also with XeLaTex. I'm not sure what is the difference between them. Here is an example generated file for pdfLaTex: \documentclass[a4paper,11pt]{article} \usepackage[utf8]{inputenc} \usepackage[hungarian]{babel} \author{Nagy László} \title{Elemzés} \begin{document} \maketitle \tableofcontents \section{Esetleírás} \subsection{A vizsgálat tárgya} \begin{itemize} \item árvíztűrő tükörfórógép \item ÁRVÍZTŰRŐ TÜKÖRFÚRÓGÉP \item 80g/m² \end{itemize} \end{document} This throws the error: ! Package inputenc Error: Unicode char \u8:² not set up for use with LaTeX. Another guy told me to use XeLaTex instead, because it has better unicode support. I don't understand the details, but I have tried that too. With XeLaTex, I got this error: Runaway argument? ! Paragraph ended before \UTFviii@three@octets was complete. <to be read again> \par l.7 \maketitle All right, I have no idea what it means, but somebody told me that I should try putting %%!TEX encoding = UTF-8 Unicode at the beginning of the file. I'm not sure what this means, because I could not find anything about a special `%% macro (???)` in the "not too short introduction into latex2e" documentation. It did not help anyway. So I removed `\maketitle` and now I got this: ! Package inputenc Error: Unicode char \u8:írá not set up for use with LaTeX. Can somebody help me where to start? I don't understand the error message, and I don't know how to fix it.
[ -0.000348023371770978, 0.009503062814474106, -0.007344346027821302, 0.001772546675056219, -0.01850542426109314, 0.005393940955400467, 0.008119367994368076, 0.01617884263396263, -0.01720980368554592, -0.04166760295629501, -0.011865154840052128, 0.0010115541517734528, 0.01708754152059555, 0....
[ 0.30271828174591064, 0.27449706196784973, 0.2876700162887573, 0.08054578304290771, -0.3197743594646454, 0.04627813771367073, 0.08799047768115997, 0.051191575825214386, 0.029753901064395905, -0.7836543321609497, 0.06217406317591667, 0.525631308555603, -0.3657001852989197, -0.197025701403617...
I have simple factory class with differently named methods which create the same object but differently. These created objects are persisted to db. These are then retrieved from the db elsewhere and modified differently based on different conditions. I want to leave the conditional logic inside the service class and centralize this object modification, just like I did with object creation. Is it valid to add update* methods inside a simple factory ? Is there another pattern I should be using instead ? In an ideal world I could add these update methods to the Dao itself, but I am using JPA repositories for Dao. So it is not straight forward to add the update methods to Dao. These Message object updating methods have to live somewhere else and I cannot decide where. public class Message { // has some properties } public class Messagefactory { public Message createMessageForConditionA() {} public Message createMessageForConditionB() {} public Message createMessageForConditionC() {} public Message createMessageForConditionD() {} // Is this okay ? public Message updateMessageForConditionE() {} public Message updateMessageForConditionF() {} } public class MessageService { @Autowired private MessageDao messageDao; public Message createMessage(String condition){ Message message = null; if(condition.equals(A)) { message = createMessageForConditionA(); } else if(condition.equals(B)) { message = createMessageForConditionB(); } else if(condition.equals(C)) { message = createMessageForConditionC(); } else if(condition.equals(D)) { message = createMessageForConditionD(); } messageDao.save(message); return message; } public Message updateMessage(string messageKey, String condition) { Message message = messageDao.findByMessageId(messageKey); if(condition.equals(E)) { message = updateMessageForConditionE(); } else if(condition.equals(F)) { message = updateMessageForConditionF(); } messageDao.save(message); return message; } }
[ -0.0012205790262669325, 0.036580875515937805, -0.0006260629743337631, 0.008271611295640469, -0.0056353723630309105, 0.01577022112905979, 0.007065678481012583, 0.010583960451185703, -0.015395494177937508, -0.014809845946729183, -0.0075487117283046246, 0.022056758403778076, 0.00203220359981060...
[ 0.33836790919303894, 0.19246645271778107, -0.07131034135818481, -0.1264333575963974, -0.19876475632190704, 0.2555113732814789, 0.44358643889427185, -0.34704792499542236, -0.07827924191951752, -0.4434721767902374, -0.05606740340590477, 0.4140886962413788, -0.3105575740337372, 0.099345475435...
I have an internet connection with a public, static IPv4 address (the IP address has a PTR record for foo.org domain, and I own foo.org domain, and have nameservers for it elsewhere, mx record is set to use the static ip). I have an OpenBSD i386 installed on a machine that is the only thing that uses this internet connection. If I nmap all the port of it from the internet side, I can't see any open ports. Great. Since no service is running on the OpenBSD machine that is faced to the internet. I need to install a mail server on it, so people could use it for sending/receiving e-mails securely. So I googled and I will choose to install: OpenSMPTd for SMTPS (using port 465) popa3d for POP3S (using port 995) **Question:** is this enough for a mail server? **Will I be able to send/receive e-mails?** (only port 465 will be open for the world/internet, pop3s will be only allowed from localhost, because we will use an ssh tunnel /with port knock/ to this openbsd machine to receive mail), so nmaping all the ports of the openbsd machine only outputs that port 465 is open. UPDATE: so it turn out I don't really need a pop3 server, because I just need to log in to the openbsd machine, and there I can see the mails locally(afaik thunderbird can be set to see local mails.).
[ 0.012830935418605804, 0.0012113640550523996, -0.016621295362710953, -0.00042415730422362685, -0.004071934148669243, 0.0035920683294534683, 0.010136056691408157, 0.0046499487943947315, -0.012919063679873943, -0.0034112920984625816, -0.00011492334306240082, 0.01618778705596924, -0.009106558747...
[ 0.31660494208335876, 0.5281537175178528, 0.1822209656238556, 0.11891621351242065, -0.014702013693749905, -0.09802412241697311, 0.32049915194511414, 0.22726581990718842, -0.13488614559173584, -0.6739622950553894, -0.060798779129981995, 0.3042753338813782, -0.2947138845920563, 0.555902123451...
In Deus Ex: Human Revolution, if I upgrade a weapon using one of the upgrade items, such as a suppressor or an ammo-capacity upgrade, etc., then I either sell the weapon, or drop it. Do all weapons of the same type subsequently have the upgrade? For possible bonus points: Do all weapon drops of the same type from enemies have the same attributes?
[ 0.013587369583547115, 0.010648587718605995, 0.005131775047630072, 0.005523367319256067, -0.03142649680376053, -0.01412659976631403, 0.01062850747257471, -0.0030621669720858335, -0.018842702731490135, -0.038334134966135025, -0.008991585113108158, 0.023753521963953972, -0.011581985279917717, ...
[ 0.179192453622818, -0.544996440410614, 0.5217779874801636, 0.16956034302711487, 0.038205474615097046, 0.2709634602069855, -0.056647684425115585, -0.26310843229293823, 0.0030318074859678745, -0.3305193781852722, -0.30312061309814453, 0.7448852062225342, -0.2867583930492401, 0.22291024029254...
Okay basically, I want to know some details about how much it would cost to buy a web server for what I want to do. The site we're making is going to accept uploads from users. A LOT of uploads. How much is this really going to cost in server space? Videos, Pictures, Audio, everything. And how easy/expensive is it to increase server space/speed to handle all of these concurrent uploads happening. I really have no idea where to start with this, haven't been a webmaster before =/
[ -0.009870342910289764, -0.0002110899076797068, -0.003918289206922054, 0.013920960016548634, -0.005814422853291035, -0.005166322458535433, 0.00797948520630598, 0.026074469089508057, -0.02687159925699234, 0.00710742874071002, 0.005090220831334591, 0.025262420997023582, 0.016437150537967682, ...
[ 0.9464437961578369, 0.1646287888288498, 0.1626783311367035, 0.41110339760780334, 0.13170881569385529, -0.10922306030988693, -0.10051466524600983, 0.26671263575553894, -0.25267642736434937, -0.4179455637931824, 0.6108285784721375, 0.5318387150764465, -0.03140348941087723, 0.3294163346290588...
How can I crop unneeded chars, and only output what I need? For example: input: ASDFQWER output: DFQ I mean in this case I need only the 3.-4.-5. chars of the string. How do I do that?
[ 0.009898887015879154, 0.017748448997735977, -0.016554582864046097, 0.01769823394715786, -0.050020672380924225, 0.020854013040661812, 0.01152835413813591, 0.02349979430437088, -0.03589843958616257, -0.032125744968652725, -0.017143061384558678, -0.0028804787434637547, 0.0068875448778271675, ...
[ 0.3926994800567627, -0.014116022735834122, -0.03698267787694931, -0.1734686642885208, -0.17980515956878662, 0.18218956887722015, 0.21655374765396118, -0.3726435899734497, -0.06695354729890823, -0.34364452958106995, 0.002472849562764168, 0.5989495515823364, -0.19305846095085144, 0.075643986...
I was wondering what the general best practice is for when you have multiple unit test fixtures testing different things that use the same file paths. My test files are in a directory tree with certain rules so currently I am reusing directory macros in each file, i.e. #define TYPE1FILE "Type1Files/" #define TYPE2FILE "Type2Files/" #define GENERIC_FILE "generic/" #define SPECIAL_FILE "special/" #define BAR1 TYPE1FILE GENERIC_FILE "/foo.txt" #define BAR2 TYPE1FILE SPECIAL_FILE "/foo2.txt" #define BAR3 TYPE2FILE GENERIC_FILE "/foo.txt" #define BAR4 TYPE2FILE SPECIAL_FILE "/foo2.txt" However, this is starting to cause duplication in multiple files because multiple test fixtures will use the same: #define BAR1 TYPE1FILE GENERIC_FILE "/foo.txt" if they both use the same file. Also, it seems like it may be better to give them some scope by using variables inside the fixture or at least within the file. I like being able to see a list of what files are being used for each test, but I am also considering moving the BAR defines all into a shared header file. I am not crazy about that idea because it will end up being a giant list of files without any indication of what is used where. Also it couples the test cases together more. Does anyone have any suggestions?
[ 0.012741182930767536, 0.02699417620897293, -0.0004600163083523512, 0.0054891472682356834, 0.03710586950182915, 0.01000845804810524, 0.009146130643785, 0.012247756123542786, -0.015257332473993301, -0.013925634324550629, 0.003787146881222725, -0.00006021664012223482, 0.02190779708325863, 0.0...
[ 0.660936713218689, 0.4172761142253876, 0.17983242869377136, -0.2977820932865143, -0.17829911410808563, 0.1723160296678543, 0.35165145993232727, -0.43501901626586914, -0.20177187025547028, -0.6351631283760071, 0.215074822306633, 0.46678274869918823, -0.12053851038217545, 0.2406540811061859,...
In Link's Awakening, sometimes I come across Secret Seashells, and the text that pops up says that if I collect a lot of them, something good is bound to happen. what will happen if I collect enough? (How many is enough, anyhow?)
[ 0.019017457962036133, 0.015963826328516006, 0.0073938872665166855, 0.02761649712920189, 0.004496762994676828, 0.018402766436338425, 0.010481667704880238, 0.014690355397760868, -0.02353544533252716, -0.019389979541301727, -0.009325293824076653, 0.026504632085561752, -0.011612867936491966, 0...
[ 0.35425034165382385, -0.3803722560405731, -0.06265420466661453, 0.32929736375808716, -0.023696405813097954, -0.013047933578491211, 0.25164490938186646, 0.5031625032424927, -0.2569906711578369, -0.16633620858192444, 0.025729792192578316, 0.10628120601177216, -0.3076086640357971, 0.435441941...
\begin{tikzpicture} \def\nuPi{3.1459265} \foreach \a in {0,6,12,18,24}{ \foreach \x in {0,3}{ \foreach \y in {0,2*sqrt(3),4*sqrt(3),6*sqrt(3)}{ \foreach \i in {0,...,5}{ \draw [ultra thick,blue]({\x + \a+2*cos(360*\i/6)},{\y+sqrt(\x)+2*sin(360*\i/6)}) -- ({\x + \a+2*cos(360*(\i+1)/6)},{\y+sqrt(\x)+2*sin(360*(\i+1)/6)}); \shade[ball color=red] ({\x + \a+2*cos(360*\i/6)},{\y+sqrt(\x)+2*sin(360*\i/6)}) circle(0.45); } } } } \end{tikzpicture} How can I put this picture of graphene on an inclined plane? All hexagons should be deformed, how can I make this deformations in TikZ?
[ -0.0010401406325399876, 0.020592235028743744, -0.008205016143620014, 0.017995983362197876, 0.006661614403128624, -0.02264796569943428, 0.005773944780230522, -0.009721215814352036, -0.007083291187882423, -0.0034233713522553444, -0.012405887246131897, -0.007182270288467407, -0.0258843079209327...
[ 0.2033393681049347, -0.4686751961708069, 0.5268837213516235, -0.15694734454154968, 0.2690098285675049, 0.1125124841928482, 0.024074316024780273, 0.12468588352203369, -0.07734755426645279, -0.5363975763320923, -0.1633763313293457, 0.4816611111164093, -0.13178326189517975, 0.1317893862724304...
I am in the planning phases of building a web based application which will be served on IOS devices through a custom application with HTML container. The functionality will be to display a number of icon markers as well as polygons. The core functionality is fairly simple which is to render markers and polygons. The issue I am having is with a feature which I have agreed to which is dynamically adding and removing polygons. I can certainly do this with client side on google maps. However performance of polygons on client side is very poor. There is also a limit of ~10 KML files at any one time to be added to a google map. The polygons need to be in separate KML files so that I can add and remove them. My question is does ESRI GIS maps handle large amounts of polygons being added/removed better than google maps apiv3?
[ -0.010192936286330223, 0.007317546289414167, -0.0049832649528980255, 0.003778572427108884, -0.0019652938935905695, -0.028004569932818413, 0.006783832795917988, 0.025088036432862282, -0.013968357816338539, 0.0022027078084647655, -0.00493225920945406, 0.012250514701008797, -0.00239939382299780...
[ 0.068028025329113, 0.25068241357803345, 0.77714604139328, 0.05775578320026398, -0.1314016431570053, 0.2847144901752472, 0.09755027294158936, 0.052858006209135056, -0.28520554304122925, -1.00711989402771, 0.2567494511604309, 0.22370067238807678, -0.3176765739917755, -0.1630343347787857, 0...
I get the energy transfer between capacitor and an inductor in a LC circuit but I am not sure what are the practical application of it.? Can someone pls help.?
[ 0.04609265178442001, 0.009223917499184608, 0.010664134286344051, 0.020473774522542953, -0.013083701953291893, -0.008502287790179253, 0.014263823628425598, -0.030420536175370216, -0.030589360743761063, -0.022312406450510025, -0.01754315197467804, 0.03451288491487503, -0.023339059203863144, ...
[ 0.6059849858283997, -0.12954534590244293, 0.32757896184921265, 0.36797380447387695, 0.04537701979279518, -0.23863454163074493, -0.384114533662796, -0.13299918174743652, -0.26905936002731323, -0.16060423851013184, 0.1516183316707611, 0.4272991716861725, -0.06863908469676971, 0.3886256217956...
My brother is an artist and he told me his paintings look more natural when he slightly change the hue (color) of the surface near the transition between shadow and light. Here's three examples: (1) without this effect, looks artificial, (2) with this effect, looks natural, (3) with overuse of this effect, looks strange. Assume pure white light (this hue change is only caused by surface properties). ![Hue change example](http://i.stack.imgur.com/9WXaN.jpg) What would be the physical explanation for this effect? The hypothesis I got is that when an object is painted (semi-transparent surface over pure reflective surface), if the rays reach the surface in a near parallel direction, then we get some refraction. If the reflectance of the surface is not a pure wavelength, then we can see some hue drift due to that refraction. Does this make sense? Is there a simpler explanation?
[ -0.0038053872995078564, 0.0003351673949509859, -0.013712793588638306, 0.011925144121050835, -0.01577744074165821, -0.02344619296491146, 0.008678024634718895, -0.032535672187805176, -0.0100413728505373, -0.010811339132487774, -0.012797150760889053, -0.0026278235018253326, 0.005618590861558914...
[ 0.5897789597511292, 0.02079060673713684, 0.1952868402004242, 0.10003821551799774, 0.2372531294822693, -0.0817737877368927, 0.44982439279556274, -0.10766944289207458, -0.42589572072029114, -0.3130790591239929, 0.05806373804807663, 0.4848695397377014, -0.15149758756160736, 0.1500905305147171...
I started on Normal. I always start on Normal. Now that I have played some of the game and have the mechanics down, can I switch difficulties for my current character?
[ 0.0006952044204808772, 0.04321720823645592, -0.010139191523194313, -0.026027511805295944, -0.009503137320280075, -0.014426463283598423, 0.01118730753660202, -0.017961984500288963, -0.036970105022192, 0.029116017743945122, -0.003307246370241046, 0.02700468897819519, 0.007018287666141987, 0....
[ 0.0071272701025009155, -0.25552070140838623, 0.035301968455314636, 0.22002874314785004, 0.07982378453016281, 0.21302102506160736, 0.04975283518433571, 0.2236231416463852, -0.30326563119888306, -0.2406325489282608, 0.055170394480228424, 0.7245853543281555, -0.25244924426078796, -0.189858108...
I am trying to understand ordinary kriging. Say I have 3 elevation measurements: Z1, Z2, and Z3 taken at X positions: X1, X2 and X3. I am also assuming some semivariogram: g(h) and that the process is stationary. From what I understand the kriging weights in this case should be ?: $ L = M^{-1}*Gs $ Where: $L=\ \left( \begin{array}{ccc} l_1 \\\ l_2 \\\ l_3 \\\ \mu\end{array} \right) $ $\ M = \left( \begin{array}{ccc} g(0) && g(X2-X1) && g(X3-X1) && 1 \\\ g(X2-X1) && g(0) && g(X3-X2) && 1 \\\ g(X3-X1) && g(X3-X2) && g(0) && 1 \\\ 1 && 1 && 1 && 0 \end{array} \right)$ $Gs=\ \left( \begin{array}{ccc} g(|X1 - Xs|) \\\ g(|X2 - Xs|) \\\ g(|X3 - Xs|) \\\ 1 \end{array} \right) $ And the elevation Zs at x position Xs should be estimated by ?: $\ Z(Xs)=\sum_{i=1}^3 l_i*Zi$ Will the inverse of the matrix pose a problem? I found a Matlab script on the internet that was using "Moore-Penrose pseudoinverse" instead. The reason I am asking this question is that I coded a Matlab script on this which drew random X and Z; and it seemed to work just fine most times, but every now and then it returned a "crazy" estimate for Z(Xs). So I am just trying to understand if I have misunderstood the theory, if the inverse of the matrix sometimes become "ill posed" or if I have just tripped up in my Matlab coding. Here is my Matlab script: % kriging.m n = 3; X = 1000*rand(1,n); Z = 1000 + 100*randn(1,n); stem(X,Z,'b'); hold on; a0 = 300; % range c0 = 4; % scale M = ones(n+1,n+1); for r = 1 : n for c = 1 : n Xr = X(r); Xc = X(c); h = abs(Xr - Xc); M(r,c) = c0*(1-exp( -(h/a0)^2 )); end end M(n+1,n+1) = 0 MI = inv(M) Xs = 500; hs = abs(X - Xs); Gs = c0*(1-exp( -(hs/a0).^2 )); Gs = [Gs 1]' L = MI*Gs; L = L(1:n) Zs = Z*L stem(Xs,Zs,'r') And here is an example of an crazy result: X = 117.4177 296.6759 318.7783 Z = 1.0e+003 * 0.9679 1.0012 0.6971 M = 0 1.2010 1.4508 1.0000 1.2010 0 0.0217 1.0000 1.4508 0.0217 0 1.0000 1.0000 1.0000 1.0000 0 MI = -0.8331 5.2214 -4.3883 1.0956 5.2214 -55.8175 50.5960 -6.3666 -4.3883 50.5960 -46.2077 6.2710 1.0956 -6.3666 6.2710 -1.4516 Gs = 3.2134 1.4732 1.2229 1.0000 L = 0.7441 -9.9423 10.1982 Zs = -2.1255e+003 ![enter image description here](http://i.stack.imgur.com/bJjIB.png)
[ 0.0015259948559105396, 0.013371629640460014, -0.004007396288216114, 0.01744169555604458, -0.01984223909676075, 0.002446202328428626, 0.008493788540363312, -0.017589036375284195, -0.011449355632066727, -0.006666736211627722, 0.004886263981461525, 0.006151504348963499, -0.019971515983343124, ...
[ 0.013774762861430645, 0.3954436779022217, 0.6237392425537109, -0.24465614557266235, 0.0046626548282802105, 0.399385541677475, 0.37594446539878845, -0.5439816117286682, -0.1725977063179016, -0.2914421260356903, -0.019351202994585037, -0.048897624015808105, -0.016358288004994392, 0.133178338...
In German there is the word _Spielraum_ which literally translated means "space available for limited movement" and is used for example to describe a limited location/angle tolerance in mechanical bearings or joints and the like, more often however it's used figuratively to describe the range of tolerable actions an individual can take in order to reach certain goals. The dictionary translates _elbowroom_ (among other things) to _Spielraum_ and I'd like to know whether it is fine to use it in the same way for describing the range of actions someone is able to take.
[ -0.013250227086246014, 0.01071260217577219, -0.004053276963531971, 0.01821606606245041, 0.0007033976726233959, 0.00271917087957263, 0.00800013542175293, -0.005144551862031221, -0.008990034461021423, 0.0056308358907699585, -0.01226081419736147, 0.017598841339349747, 0.0032103359699249268, 0...
[ 0.32636767625808716, 0.2568140923976898, 0.17812372744083405, -0.16922242939472198, -0.09563010185956955, 0.006305464543402195, 0.27700576186180115, -0.15101183950901031, -0.4220903813838959, -0.30709728598594666, -0.22816580533981323, 0.196604922413826, 0.43592149019241333, -0.27259549498...
I have a domain, registered at company A. In the domain manager panel at this company I can edit DNS records of my domain, including MX records. That's understandable. I changed them. But, the website which uses this domain is at company B. This comapany also has DNS records manager! It looks like it overwrites whatever I set at company A. How is that possible?
[ 0.014325994998216629, -0.0017434004694223404, -0.01629188470542431, 0.008289013057947159, 0.014282920397818089, 0.01745636574923992, 0.010321519337594509, 0.027930570766329765, -0.019787343218922615, 0.024532245472073555, -0.0017172925872728229, 0.02029096893966198, 0.00986749492585659, 0....
[ 0.6396311521530151, 0.37160030007362366, 0.48064014315605164, 0.06273343414068222, 0.12817977368831635, -0.2902953624725342, 0.19095024466514587, 0.22764861583709717, -0.5028109550476074, -0.35328546166419983, 0.18118834495544434, 0.05985119193792343, -0.11685586720705032, 0.74684172868728...
I am going to try and sum up what I am trying to achieve... Basically I have a custom post type called 'marketing_list' which will have custom fields holding information relating to marketing lists available for purchase. Each marketing list will belong to a category (custom taxonomy). So currently my URL's look like this: * `/marketing-lists` -> archive page * `/marketing-lists/post-slug` -> single marketing lists entry * `/marketing-lists/category/category-slug` -> custom taxonomy What I want to achieve is: * `/marketing-lists` -> archive page but editable within the admin in case the client wants to update text, like a Page. * `/marketing-lists/category-slug` -> custom taxonomy * `/marketing-lists/category-slug/post-slug` -> single entry On top of that I would like to create some additional pages which have a base of `/marketing-lists` i.e * `/marketing-lists/consumer-marketing-lists` -> single page * `/marketing-lists/business-marketing-lists` -> single page Is this at all possible? Perhaps their is a better way to achieve this?
[ 0.005966648459434509, 0.0004865701775997877, 0.006558489054441452, 0.018737750127911568, 0.015524162910878658, 0.011975033208727837, 0.00653794826939702, 0.024571692571043968, -0.010633455589413643, 0.0031346846371889114, -0.000029726652428507805, 0.00021729181753471494, 0.002419784897938370...
[ 0.9432925581932068, 0.26900720596313477, 0.44096267223358154, -0.03043070249259472, -0.2582840919494629, -0.0928296446800232, -0.2103119194507599, 0.17355579137802124, -0.40938055515289307, -0.5719218254089355, 0.29826322197914124, 0.42301470041275024, 0.034675233066082, 0.3047237098217010...
I came here to understand why latex had this "odd" habit to set the outer margin wider than the inner in two sided documents. As many have asked before me, I expected the inner margin to be wider to take into account the binding space. From the answer to those questions I discovere that the outer margin should be wider that the inner and that the inner margin does NOT takes into account the binding space (since it's not universal but depends on the binding type). NOW… I need to give the print shop a PDF version of the thesis ready to be printed. When I went there today he told me that usually the inner margin is wider to take into account the binding and to be equal to the outer once it's bond. I accept the "outer margin is wider" standard, but I need to add a bit of binding space to my pages. How do I do it? It's the first time I print a two sided document. If I told something wrong feel free to teach me how to setup (maybe also outside of latex) correctly a two sided document. Thank you very much
[ 0.006511249579489231, 0.014371376484632492, -0.02274572290480137, 0.028402045369148254, 0.0026481340173631907, 0.017466016113758087, 0.01006784662604332, 0.00842694565653801, -0.011607949621975422, -0.017299335449934006, -0.0077176825143396854, 0.004521505907177925, -0.011591818183660507, ...
[ 0.2777302861213684, 0.1417694389820099, 0.4223501980304718, -0.010211201384663582, 0.0951121374964714, -0.02509971335530281, -0.35370731353759766, -0.02200697548687458, -0.24383971095085144, -0.496408075094223, 0.1272580623626709, 0.12231236696243286, 0.18532082438468933, 0.348049610853195...
I want to measure biomass on a coastal mudflat. I can only access points inside the polygon. Are there any methods available that would allow me to estimate values of the points outside of the polygon, based on the values of points inside the polygon? set.seed(5) x <- rnorm(50, -1.841, 0.01) set.seed(50) y <- rnorm(50, 55.663, 0.01) xy <- data.frame(x,y, values=rnorm(50)) coordinates(xy) <- c("x", "y") proj4string(xy) <- CRS("+proj=longlat +ellps=WGS84 +datum=WGS84") plot(xy) makePolygons <- function(coordsx, coordsy){ coords <- matrix(c(c(coordsx, coordsy)), ncol=2) p <- Polygon(coords) p <- Polygons(list(p), ID = "p") myPoly <- SpatialPolygons(list(p)) spdf = SpatialPolygonsDataFrame(myPoly, data.frame(variable1 = c(2), variable2 = c(3), row.names = c("p"))) proj4string(spdf) <- CRS("+proj=longlat +datum=WGS84 +ellps=WGS84 +towgs84=0,0,0") print("polygon is in longlat!!!") spdf } myPoly <- makePolygons(coordsx=c(-1.841960, -1.843464, -1.888623, -1.841960), coordsy=c(55.633696, 55.68178, 55.63841, 55.633696)) plot(myPoly, add=T) ![enter image description here](http://i.stack.imgur.com/yF44s.jpg)
[ 0.007363812066614628, 0.013144338503479958, -0.00712085235863924, 0.020250696688890457, -0.0027375873178243637, -0.02034047804772854, 0.007615869864821434, 0.006976198870688677, -0.01297718845307827, 0.017539789900183678, 0.005418776068836451, 0.00926426611840725, -0.021587209776043892, 0....
[ 0.5128417611122131, 0.12488918751478195, 0.24462631344795227, 0.10995108634233475, 0.3643445372581482, 0.5346311330795288, -0.21633759140968323, -0.5170980095863342, 0.0784028172492981, -0.557835042476654, 0.27732738852500916, 0.10646958649158478, 0.19061309099197388, 0.0797099620103836, ...
Can someone please tell me how to make definitions, theorems etc. in the appendix look like **Definition A1** or **Definition 1A** (an "A" should be somewhere), if the appendix is "numbered" with **A** (meaning latex shows **A Appendix** , if I compile the document). Right now all the definitions, theorems etc.are numbered from 1 to n, where n is the last definition, theorem etc. of my document (sections don't change the numbering), even if this last one is in the appendix. Equation numbering is ok though; the first numbered equation is "A.1". The thing is, I'm using LyX, so I would prefer it, if there were some command that I could use to reset the counter and globally change the look of definitions, theorems etc. in the appendix, so that I 1) don't have to remove all the LyX-created environments (like for example the LyX-created definition environment) and replace them by hand-coded ones in order to maybe pass some additional arguments to each them, because that looks horrible in LyX. 2) don't have to use a different way to insert the appendix other then the one LyX provides and that is via \appendix (not via the appendix package). FYI: I'm using LyX 2; the document class is article; my appendix is situated after several sections (I don't used any other divisioning of my document like subsections etc. - if that makes any difference).
[ -0.00684815738350153, 0.022071676328778267, -0.004047206602990627, 0.024827148765325546, -0.008489742875099182, 0.01436302438378334, 0.0088321752846241, 0.0069078728556632996, -0.01175563596189022, 0.009964898228645325, -0.0032013216987252235, 0.0016076461179181933, 0.0064697712659835815, ...
[ -0.07794832438230515, 0.33011293411254883, 0.3808833658695221, 0.25610747933387756, 0.0773899108171463, 0.008971921168267727, 0.08837185800075531, -0.05516034737229347, -0.37298446893692017, -0.5963274240493774, 0.005479453597217798, 0.27757149934768677, 0.06377611309289932, 0.433831810951...
Is there a hypernym for the following terms? * _Local business_ * _Company_ * _Organization_ or _institution_ ## Edit I am building a website where one can create a profile that will either be a Local Business, a company, an organization or an institution. At the moment I am using the general term `Profiles` but it does not sound very professional. I would like to stand away from the term Profiles if possible. I was also thinking about the term `Entity` but it's not very nice either.. Facebook is using the general term `Pages` (see here) but for them, a page can also be a brand, a cause, a community, etc...which is a bit different.
[ 0.006418971810489893, 0.012705960310995579, 0.01069757342338562, 0.018127065151929855, 0.004324125126004219, 0.0075463782995939255, 0.007952068001031876, 0.021381234750151634, -0.011556653305888176, -0.005919029004871845, -0.014799299649894238, 0.007036812603473663, 0.023560162633657455, 0...
[ 0.8775482773780823, 0.13725128769874573, 0.4219115674495697, 0.26343825459480286, -0.3132306635379791, -0.3225238025188446, -0.016062898561358452, 0.5841982960700989, -0.5220526456832886, -0.733735978603363, 0.1533079743385315, 0.28291767835617065, -0.13445597887039185, 0.1116623803973198,...
Is it possible to have macros with the same name, but different “syntax”? E.g. we have the primitive `\font` which could be seen as having 3 arguments(a. without size information; b. giving the fontsize with `at`; c. giving scaling information with `scaled`). Is there a way to provide similar functionality for own macros? I already tried the following (stupid) experiment: \def\print#1at#2,#3{The text ``#1'' shall be printed at $#2 \abovewithdelims()0pt #3$} \def\print#1by#2,#3{The text ``#1'' shall be printed by #2 and #3 Doe} \print{Testtext}at 2,3 \print{Testtext}by{Jon},{Jane} \bye Trying to compile it (using `tex -interaction=nonstopmode def.tex`) yields following error: This is TeX, Version 3.14159265 (TeX Live 2014) (preloaded format=tex) (./def.tex Runaway argument? {Testtext}at 2,3 ! Paragraph ended before \print was complete. <to be read again> \par l.5 [1] ) (see the transcript file for additional information) Output written on def.dvi (1 page, 280 bytes). Transcript written on def.log.
[ -0.0054627712815999985, 0.025461267679929733, 0.00877059530466795, 0.01645292341709137, 0.008281184360384941, 0.0033943764865398407, 0.006118071265518665, 0.001962529495358467, -0.019841846078634262, -0.04081821069121361, -0.008666548877954483, 0.005345610436052084, -0.0033181938342750072, ...
[ 0.14191174507141113, 0.11047482490539551, 0.38340815901756287, -0.2542288601398468, 0.09769535809755325, 0.29718244075775146, 0.13714170455932617, -0.17225050926208496, -0.27991411089897156, -0.5496537089347839, 0.1422692835330963, 0.646331250667572, -0.5421255826950073, -0.234723240137100...
For an observable $A$ and a Hamiltonian $H$, Wikipedia gives the time evolution equation for $A(t) = e^{iHt/\hbar} A e^{-iHt/\hbar}$ in the Heisenberg picture as $$\frac{d}{dt} A(t) = \frac{i}{\hbar} [H, A] + \frac{\partial A}{\partial t}.$$ From their derivation it sure looks like $\frac{\partial A}{\partial t}$ is supposed to be the derivative of the original operator $A$ with respect to $t$ and $\frac{d}{dt} A(t)$ is the derivative of the transformed operator. However, the Wikipedia derivation then goes on to say that $\frac{\partial A}{\partial t}$ is the derivative with respect to time of the _transformed_ operator. But if that's true, then what does $\frac{d}{dt} A(t)$ mean? Or is that just a mistake? (I need to know which term to get rid of if $A$ is time-independent in the Schrodinger picture. I think it's $\frac{\partial A}{\partial t}$ but you can never be too sure of these things.)
[ -0.017976395785808563, 0.015423765406012535, -0.014398747123777866, 0.0036269871052354574, 0.007573904935270548, -0.017283109948039055, 0.005484344437718391, 0.004599285777658224, -0.008298015221953392, 0.007353691384196281, -0.006023211404681206, 0.007438722066581249, -0.018708819523453712,...
[ -0.3965494632720947, 0.013209897093474865, 0.4254719614982605, -0.12933500111103058, -0.029443273320794106, 0.10516376793384552, 0.03455781191587448, -0.18163102865219116, -0.22453169524669647, -0.32615646719932556, -0.24393783509731293, 0.343811571598053, -0.7639607191085815, 0.5180544853...
I'm looking for a word that essentially means "to cook" but sounds fancy and sophisticated. What synonyms are there? **Edit:** As for context, I'm thinking of a restaurant setting.
[ 0.015518052503466606, 0.022315705195069313, 0.010726754553616047, 0.044795893132686615, -0.003274251241236925, 0.000019599781808210537, 0.010466145351529121, 0.00255970424041152, -0.04146434739232063, 0.017611928284168243, -0.005218873731791973, 0.011129502207040787, 0.04335227981209755, 0...
[ 0.7727086544036865, 0.2629043757915497, -0.3287312388420105, -0.07118670642375946, 0.317170649766922, 0.23028527200222015, 0.14914679527282715, 0.4226701259613037, -0.1364569067955017, -0.5928459763526917, -0.25759992003440857, 0.1736306995153427, 0.014658618718385696, -0.01691995374858379...
I've searched for multiple forums for this questions and have seen different answers but so far, nothing have worked yet even from the ones I've seen here. I want to extract a single file from a tar to a different directory. Ive tried this: tar xvf file.tar -C /home/dir/ filename or this: tar -x filename -f file.tar -C /home/dir But got this errors respectively: x filename, 14826 bytes, 29 media blocks. File -C not present in the archive. File /home/dir not present in the archive. and: tar: /dev/rmt0: A file or directory in the path name does not exist. The first extracted the file but in the same directory, not to the folder I wanted.
[ 0.0036986456252634525, 0.013566533103585243, -0.00970898475497961, 0.010123135522007942, 0.003849801840260625, -0.012725064530968666, 0.006223778706043959, 0.005643258802592754, -0.022183720022439957, -0.01458356250077486, -0.004405261017382145, -0.0013465178199112415, -0.0037255927454680204...
[ 0.32959333062171936, 0.09770692884922028, -0.24100148677825928, 0.09196518361568451, -0.1409616768360138, 0.3824412226676941, 0.06963919848203659, -0.27446281909942627, -0.32482460141181946, -0.6334196329116821, 0.1351691335439682, 0.6701202392578125, -0.09253248572349548, 0.46653792262077...
I'm using the BigRational class from bcl.codeplex.com. It will be opensourced when it's finished, and it's modified. However, I don't get legal jargon, so even though it's quite short, could someone explain to me what it does/doesn't allow you to do?
[ 0.010544421151280403, 0.005496648605912924, -0.0032468854915350676, 0.015934446826577187, 0.005952422972768545, -0.009010094217956066, 0.006045872811228037, 0.020722996443510056, -0.01907380484044552, -0.042403604835271835, -0.00044370489194989204, 0.008241754956543446, 0.00481041194871068, ...
[ 0.7693375945091248, 0.026764892041683197, 0.17030872404575348, 0.1273338347673416, 0.36155644059181213, -0.5612936615943909, 0.33155184984207153, -0.05612320452928543, -0.18234939873218536, -0.5291473865509033, 0.00841899961233139, 0.22111180424690247, 0.03996478393673897, -0.1861215233802...
I want to create the following layout: ------------ ------------- | IMAGE 1 | | IMAGE 2 | |____________| |_____________| Notice that the two images are relatively small in size, but I want them to be placed at the right/leftmost edge of the document. I tried it using minipages like this: \begin{minipage}{4cm} \includegraphics[width=4cm]{image1.eps} \end{minipage} \begin{minipage}{4cm} \begin{flushright} \includegraphics[width=4cm]{image2.eps} \end{flushright} \end{minipage} But then, the two images are placed like this: ------------ | IMAGE 1 | |____________| ------------- | IMAGE 2 | |_____________| I fooled around with the values of minipage and image-width, but nothing seems to help... Edit: Also, I need the images to be centered to each other (horizontally speaking), so they are nicely aligned if one is bigger than the other.
[ 0.0004085776163265109, 0.0005354578024707735, -0.008536731824278831, 0.014455094002187252, -0.003130157943814993, 0.008255599066615105, 0.00582033209502697, 0.026976410299539566, -0.018399640917778015, 0.012239809148013592, -0.020762689411640167, -0.00029992428608238697, -0.01121227163821458...
[ 0.09983683377504349, -0.05147514119744301, 0.757490873336792, 0.016214469447731972, 0.1535993367433548, 0.2992888391017914, -0.10019602626562119, -0.2065860629081726, -0.20414727926254272, -0.8994081616401672, 0.18186092376708984, 0.40978455543518066, -0.19118118286132812, 0.09939516335725...
By default it appears that the guns open fire whenever a target is in range, but at the same time, often during planning stage I would see guns that are fully loaded, pointing at a target - shouldn't this never happen since they should fire the moment they are ready? Also, how is autofire aim compared to manual aiming? Naturally with manual aim you easily miss since in the delay between round start and gun firing, the enemy ship could have pulled off an unexpected maneuver. Autofire on the other hand, seems to always fire in the general direction of the enemy.
[ -0.02205643244087696, 0.014403409324586391, -0.002828506752848625, 0.004069785587489605, -0.01116548664867878, -0.027134858071804047, 0.0080060800537467, 0.011597254313528538, -0.011060756631195545, 0.017601164057850838, -0.027431532740592957, 0.010025468654930592, 0.0032379033509641886, 0...
[ -0.1756456047296524, -0.01586809754371643, 0.3036095201969147, 0.01801176182925701, -0.3706177771091461, -0.18447797000408173, 0.6676338911056519, -0.2578909993171692, -0.3457721471786499, -0.24940721690654755, -0.2750333845615387, 1.2063305377960205, -0.1706678569316864, -0.51464205980300...
I am developing a processing algorithm in ArcMap ModelBuilder. When I run it everything seems to work fine and running to completion. However when I go check the output directories I only find Info files and Logs, but no output rasters although I have these specified in the model. Does anyone know why this may be?
[ -0.006236874032765627, 0.01223065983504057, -0.0004917349433526397, 0.01949292980134487, -0.02068244479596615, -0.0033550714142620564, 0.009238021448254585, 0.01399307232350111, -0.0215054489672184, -0.014464049600064754, 0.007678298745304346, 0.02540166676044464, -0.012779485434293747, 0....
[ 0.4609510004520416, 0.17958402633666992, 0.23921114206314087, 0.09627635031938553, 0.06973663717508316, -0.033457741141319275, 0.22397677600383759, 0.2590039372444153, -0.18655845522880554, -0.6955761313438416, 0.11016618460416794, 0.45581257343292236, -0.13326877355575562, 0.0302308890968...
Is it possible to connect to same server from Mac/PC and PlayStation 3 (PS4,PS Vita in future)
[ -0.020131399855017662, 0.02455374412238598, 0.005909949541091919, 0.016388991847634315, 0.008233686909079552, -0.03732546046376228, 0.016142556443810463, -0.001234513008967042, -0.02293722704052925, -0.00403655506670475, -0.006865893490612507, 0.04553145170211792, 0.03896058350801468, 0.01...
[ 0.07597735524177551, -0.2984400987625122, 0.5980471968650818, 0.537302553653717, 0.07817763835191727, 0.1994016021490097, -0.08709704130887985, 0.29746970534324646, -0.3144122064113617, -0.6963897943496704, 0.1272149235010147, 0.20309428870677948, -0.08313881605863571, -0.07302234321832657...
Is it possible to know specifically who liked my website? I already know the numbers, I need to know the users who actually liked my website so I can contact them and enter them into a contest. Thank you!
[ 0.014243118464946747, -0.006749200634658337, 0.011785460636019707, 0.03844298794865608, -0.03312249854207039, 0.04425332322716713, 0.012852495536208153, -0.047672394663095474, -0.04105430096387863, -0.03391169756650925, 0.011445880867540836, 0.033420685678720474, 0.022883668541908264, 0.03...
[ 0.3693960905075073, 0.3410874605178833, 0.44114065170288086, 0.38551655411720276, -0.14915616810321808, -0.1199539452791214, 0.28279542922973633, 0.45977655053138733, -0.09649989753961563, -0.2705994248390198, 0.2055453658103943, 0.49455004930496216, 0.704257071018219, 0.2143138349056244, ...
What is the most general form of the metric for a homogeneous, isotropic and static space-time? For the first 2 criteria, the Robertson-Walker metric springs to mind. (I shall adopt the (-+++) signature) $$ds^2=dt^2+a^2(t)g_{ij}(\vec x)dx^idx^j$$ Now the static condition. If I'm not mistaken, it means that the metric must be time-independent and invariant under time reversal $t\to -t$. So does that mean that the most general metric that satisfies all these 3 criteria is $$ds^2=dt^2+g_{ij}(\vec x)dx^idx^j$$ for some spatial metric $g_{ij}(\vec x)$? Thank you. (My apologies for the mistaken definition. I may have misunderstood what's been said in class...)
[ 0.0011328572873026133, 0.0069081285037100315, -0.012541206553578377, 0.012900356203317642, 0.007467803079634905, -0.0023568999022245407, 0.0067770760506391525, 0.0020235886331647635, -0.01252114586532116, 0.0012311756145209074, -0.00762513279914856, 0.012764409184455872, 0.001143279019743204...
[ -0.21425779163837433, -0.12974612414836884, 0.6047307848930359, -0.034718796610832214, 0.015712765976786613, 0.22971124947071075, -0.2133723944425583, -0.19908785820007324, -0.3316776752471924, -0.42820873856544495, -0.24308696389198303, 0.39185261726379395, -0.30829453468322754, 0.4366887...
### Experiment Let say I have a black box full of numbered balls and I draw them with replacement - so the propability of drawing one specific number in N draws is described by binomial distribution? Problem is that my black box has some intristic structure so the balls are not draw with the same p. For random box (experiment) I should observed random points draw from the box. So if you put this balls id numbers on the axis you should obtain random localization of points - yes? However my experiment generates appart from random localization also points wich should be cluster in groups (localized close to eachother). ### Question * How can I find those regions where points (which are ids of balls) are localized in a non random fashion? My box has around 10^8 balls and I draw 10^5 balls with replacement in one experiment. I would be very gratefull for any comments and sugestion, what statistical tool would be the best for my purpose. I thought about running simulation - simulate my experiment as a random one, and then look into distribution of distances between draw points. Then this random distribution somehow compare to mine to see which distances are very rare in random distribution and can be hints of my non-ranomdnes behaviour. But I am not quite sure how to proceed after simulation of random distribuiton...
[ 0.0023522924166172743, 0.00809499341994524, -0.003284052712842822, 0.02544987201690674, -0.012525226920843124, -0.027127079665660858, 0.010110840201377869, 0.019822176545858383, -0.017841195687651634, -0.00928942859172821, -0.01987452805042267, 0.021621059626340866, -0.01666085049510002, 0...
[ 0.04289036616683006, -0.1823168247938156, 0.4074031412601471, 0.1798187494277954, -0.7765439748764038, 0.0551774725317955, 0.020948292687535286, -0.39414072036743164, -0.3058042526245117, -0.7192256450653076, 0.1014852523803711, 0.2602464258670807, -0.48464012145996094, 0.301759272813797, ...
When you google "latitude/longitude" you get 10 times more results than "longitude/latitude". This I find very confusing as "x/y" seems 20 times more common than "y/x". But on a map Latitude is on the Y axis and longitude on the X. Maybe I'm just venting, but maybe I'm missing a clue that makes all this madness logical. Any explanations?
[ -0.021676823496818542, 0.01488663163036108, -0.02626424841582775, -0.003970192279666662, -0.011174784041941166, -0.012061294168233871, 0.012355568818747997, 0.010190917178988457, -0.025173470377922058, -0.026524214074015617, 0.0020458376966416836, 0.002580370521172881, -0.026250122115015984,...
[ 0.41033899784088135, 0.22825945913791656, 0.41446980834007263, 0.5551893711090088, 0.012109183706343174, -0.16687706112861633, 0.5150578618049622, 0.3013470768928528, -0.3771215081214905, -0.531289279460907, 0.2119796723127365, -0.04914570972323418, -0.1073099747300148, 0.6176131963729858,...
What is the best way to draw checker board in tikz? I'm interested in what is the best way to draw many circular pieces on board. (the checker background is not important)
[ 0.025969356298446655, 0.005206921137869358, -0.031113402917981148, 0.04112166538834572, -0.01479842234402895, -0.008348431438207626, 0.015916986390948296, 0.008715907111763954, -0.058552809059619904, -0.0342528261244297, -0.022046305239200592, 0.004197939299046993, -0.01463922206312418, 0....
[ 0.5506715178489685, 0.09908134490251541, -0.016143981367349625, 0.15614192187786102, -0.2591662108898163, -0.09020878374576569, -0.1906895935535431, 0.3003355860710144, -0.35033267736434937, -0.47226253151893616, 0.34305539727211, 0.24811412394046783, -0.12624475359916687, -0.0365349203348...
What is the rational/reason for neglecting masses at asymptotic non- exceptional space-like momenta. I have come across this as a first fix for being able to extract information from the Callan-symanzik and Gell-Mann-Low methods. Is it purely physical, or is it both physical and mathematically sound reason? All responses are welcome.
[ 0.017398081719875336, 0.018837755545973778, 0.006241960916668177, 0.027403265237808228, -0.0012825754238292575, -0.01620606705546379, 0.00948232039809227, 0.0057670194655656815, -0.028353195637464523, -0.008402381092309952, -0.015495584346354008, 0.016382833942770958, -0.003775443648919463, ...
[ 0.11087139695882797, 0.01943625509738922, 0.08888854086399078, 0.21397797763347626, 0.030042024329304695, 0.3829212188720703, -0.18811027705669403, 0.11902033537626266, -0.0907345712184906, -0.5091837644577026, -0.020209666341543198, 0.366929829120636, 0.0762735903263092, 0.246800512075424...
When I first met the love of my life, he insisted that we use WhatsApp to communicate. He lives in Brasil and uses WhatsApp on his Windows Phone. I'm in the United States and use WhatsApp for Android on my Motorola Photon Q. Over the past few months, our ability to communicate through this app has deteriorated. He says he receives my messages, and that he replies to them, but apparently I don't see them. This is frustrating, to say the least, and puts a lot of unneeded stress on both of us. I've worried myself sick over it. He says that I'm the only one this happens to, and that communication with his Brazilian contacts works just fine. One of my friends in the Philippines just got WhatsApp for Android and I can communicate with him perfectly. I'm really clueless. I've scoured the Internet for answers, but I can't seem to find any. I can send messages, and in the rare time that I do receive the messages from my boyfriend, it is late at night. I never receive any during the day. I've uninstalled and reinstalled the app countless times, and my WiFi connection at my house is perfect. But this seems to only be getting worse. He says he sends me several messages a day, but I only get some of the messages about every other day--usually every two days. I'm not a tech by any means, but it seems like there might be something wrong with the different versions of WhatsApp. Android communicating with the Windows Phone. So if this is the case, then my question here is unneeded. I've contacted WhatsApp Inc. with my concerns and have yet to receive a reply. This sucks, to say the least. UPDATE: I have contacted one of my boyfriend's friends in Brasil. I can communicate with them perfectly. They said y boyfriend said there was an issue with his phone. But I'm not sure that's the case. I uninstalled WhatsApp and downloaded it through WhatsApp's website. I can now see when my boyfriend is online and when he types. But I do not receive the messages he sends.
[ 0.001651065074838698, -0.007437983527779579, 0.0028232031036168337, 0.014348877593874931, -0.01639644056558609, -0.00408058101311326, 0.006806118413805962, 0.021011225879192352, -0.01362950261682272, -0.03502359241247177, -0.010283797979354858, 0.009794914163649082, 0.006858658045530319, 0...
[ 0.7258108258247375, 0.31417250633239746, 0.19984623789787292, 0.13442042469978333, 0.38403087854385376, 0.22232437133789062, 0.30188947916030884, 0.25011005997657776, -0.07259764522314072, -0.3638295531272888, 0.3188246488571167, 0.03678109869360924, -0.4534587562084198, -0.102355875074863...
I am using a plugin the updates the option table using update_option command. When I update content with hebrew characters it turns into gibrish. My database does support hebrew (or any other utf8 chars). Is there any workaround? Thank you!
[ 0.0007570591405965388, 0.028001930564641953, 0.004447052255272865, 0.018802791833877563, -0.06451436132192612, 0.013987190090119839, 0.011038077995181084, 0.017991479486227036, -0.0322100929915905, -0.015183837153017521, -0.023784659802913666, 0.01706637628376484, 0.010936661623418331, 0.0...
[ 0.35501569509506226, 0.6444873213768005, 0.02054641582071781, -0.4894542098045349, -0.21603059768676758, 0.019691839814186096, 0.26419615745544434, 0.4073326289653778, 0.04198259860277176, -0.9762527942657471, -0.36841896176338196, 0.3842719793319702, -0.016534943133592606, 0.2430075109004...
I have problem with path management. Let me explain you: I work with subfiles that are arranged in this way: REPORT --Main.tex --lab1 ----Matlab ------CodeLab1.m ----Report ------Report_P1_1.tex --lab2 ----Matlab ------CodeLab2.m ----Report ------Report_P2_1.tex I want to be able to include the code "CodeLab1.m" in the "Report_P1_1.tex" file and "CodeLab2.m" in the "Report_P2_1.tex" and to be able to compile them either separately or from the main (as I use subfile its possible). The trick I use is this: Main.tex: \documentclass[]{article} \usepackage{xcolor, listings, subfiles} \definecolor{grey}{rgb}{0.96,0.96,0.96} \lstset{backgroundcolor=\color{grey}} \newcommand{\myPathLstONE}[0]{ \ifx\mainCmp\undefined \renewcommand{\myPathLstONE}{../Matlab} \else \renewcommand{\myPathLstONE}{lab1/Matlab} \fi} \newcommand{\myPathLstTWO}[0]{ \ifx\mainCmp\undefined \renewcommand{\myPathLstTWO}{../Matlab} \else \renewcommand{\myPathLstTWO}{lab2/Matlab} \fi} \begin{document} \def\mainCmp{} \part{Labo 1} \subfile{lab1/Report/Report_P1_1.tex} \part{Labo 2} \subfile{lab2/Report/Report_P2_1.tex} \end{document} Report_P1_1.tex: \documentclass[../../Main.tex]{subfiles} \begin{document} \section{Part 1} There's a listing: \lstinputlisting[]{\myPathLstONE/CodeLab1.m} \end{document} Report_P2_1.tex: \documentclass[../../Main.tex]{subfiles} \begin{document} \section{Part 2} There's a listing: \lstinputlisting[]{\myPathLstTWO/CodeLab2.m} \end{document} It works quite well, the only problem (exept that its not really practical to define every time these command) is that in the output I got, there is always the name of the code file as you can see. I didn't find a way to avoid this. ![output of Main.pdf](http://i.stack.imgur.com/SXZ2Z.png) As its possible to define multiple path for graphics, the inclusion of listings is my only problem working with this architecture. Can somebody help or propose an other way to do this? Thanks a lot!
[ 0.005750884301960468, 0.015108158811926842, -0.005843454971909523, 0.015209175646305084, 0.03300965949892998, 0.023237012326717377, 0.007196187041699886, 0.019829154014587402, -0.015585853718221188, -0.01876137964427471, -0.010271077044308186, 0.004969296045601368, -0.00992019847035408, 0....
[ 0.055926352739334106, 0.0287576112896204, 0.15579573810100555, 0.07013183832168579, 0.10690495371818542, -0.19255022704601288, 0.32492321729660034, -0.14362920820713043, -0.13236665725708008, -0.5998720526695251, 0.09729497879743576, 0.6802974939346313, -0.5676178336143494, -0.370295405387...
When using the class `scrbook` and customizing the font size prior the usage of `\printbibliography` I got these warning (and in big documents I got them several times): > Class scrbook Warning: Using fallback calculation to setup font sizes > (scrbook) for basic size `12pt' on input line 32. MWE: \documentclass{scrbook} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage[backend = biber]{biblatex} \renewcommand*{\bibfont}{\normalfont\changefontsizes{12pt}} \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @online{GLS:datasheet, title = "GLS Datasheet", organization = "Optoelectronics Research Centre", howpublished = "Website", date = "2004-09", urldate = "2013-07-01", url = "http://www.southampton.XXXX", address = "Southampton, United Kingdom" } \end{filecontents} \addbibresource{\jobname.bib} \begin{document} \KOMAoptions{fontsize=12pt} \nocite{*} \printbibliography \end{document} **Do I have to expect unexpected results?** And it that is the case, what can I do against it? Otherwise how could I possibly silent these warnings?
[ -0.01640266552567482, -0.005663146264851093, -0.008352103643119335, 0.037679076194763184, 0.030077673494815826, 0.0008770740823820233, 0.011297723278403282, 0.008323770016431808, -0.01580534130334854, -0.022435221821069717, -0.0015952185494825244, 0.0003129098331555724, 0.014483993873000145,...
[ -0.07588720321655273, 0.07936306297779083, 0.1661556363105774, -0.1445862501859665, -0.0058052269741892815, -0.14991579949855804, 0.8029389381408691, -0.1490085870027542, -0.1528722196817398, -0.577184796333313, -0.18410006165504456, 0.4612094461917877, -0.5985814332962036, -0.054097622632...
I have a directory full of albums I've ripped to FLAC format. Folders are named "Artist - Album [flac]" I need to create matching directories named "Artist - Album [mp3]" but only for certain folders... So "ls | grep Pantera" gives me the list of directories I want to work with... How can do I do this via command line?
[ -0.0009976924629881978, 0.003944653086364269, -0.0018737276550382376, 0.008724620565772057, 0.0224388986825943, 0.018987450748682022, 0.010361794382333755, 0.005969182122498751, -0.023529281839728355, 0.001353311468847096, -0.002809462370350957, -0.0011163445888087153, 0.007713310420513153, ...
[ 0.2766414284706116, 0.19637107849121094, -0.05839629843831062, -0.12028566002845764, 0.26958367228507996, -0.0014497787924483418, -0.04665135592222214, 0.05170032009482384, -0.1650356948375702, -0.4188360869884491, -0.0009761936380527914, 0.79172682762146, -0.20108924806118011, -0.03042778...
I'm playing Cities XL 2012 since a while. I noticed that too oftenly I end up having large areas for offices that don't develop. Clicking on those shows that "it's hard to obtain Tech industry". Alright, I say, let's develop the tech. But then I find undeveloped tech areas that require office services. Buying from OmniCorp is usury (I'm currently -27 office coins). Checking the population, I see I have enough executives and qualified workers for the office, so I don't see why they don't develop. Buying a few tech from OmniCorp doesn't solve. Also, I don't think I have hotel problems because otherwise I would read it. What is the best way to develop offices in a city? I would also like to export those offices to other towns if possible (I have a very high people capacity).
[ 0.012826578691601753, 0.010973850265145302, -0.00856069941073656, 0.0038579937536269426, -0.028054049238562584, -0.026614470407366753, 0.006608814932405949, 0.02405901625752449, -0.013630591332912445, -0.014466802589595318, -0.02246970869600773, 0.01918104477226734, -0.0046121831983327866, ...
[ 0.6121699810028076, -0.052309807389974594, 0.4800373911857605, 0.36921656131744385, -0.1569993495941162, -0.12625253200531006, -0.12635688483715057, 0.1822161078453064, -0.17049266397953033, -0.7902551889419556, 0.3026128113269806, 0.7164410948753357, -0.004163283389061689, 0.0506041720509...
Samsung phones, Galaxy Note running ICS 4.0.3 in particular has System Storage, USB Storage, and SD Card. I wanted to know if I encrypt my device by going to Settings - Security - Encrypt Device. Will it encrypt the "USB Storage" as well aside from the System Storage?
[ -0.008815208449959755, -0.011674165725708008, 0.0007094779284670949, 0.03021882474422455, 0.002246796153485775, 0.020970940589904785, 0.012381276115775108, 0.01794089935719967, -0.03542521968483925, -0.03576021268963814, -0.020680375397205353, 0.005786147899925709, -0.021511582657694817, 0...
[ -0.14981292188167572, 0.17164142429828644, 0.7995777130126953, 0.014993296004831791, 0.14723001420497894, -0.0655268132686615, 0.11608965694904327, -0.05388181284070015, -0.1989307999610901, -0.3882482051849365, -0.15371249616146088, 0.4724668562412262, -0.3032933473587036, -0.131609961390...
I decided to move my `/home` directory to a separate parition. But, since I am using more than one linux distro, I want to separate the `/home` directories as follows: /dev/sdax/ubuntu /dev/sdax/debian /dev/sdax/suse After copying all the neccessary files to this folder, I changed the `/etc/fstab` to mount the new `/home` location as the default home drive /dev/sdax /homedrive ext4 nodev,nosuid 0 2 /homedrive/debian /home bind default,bind 0 0 Now, as far as I can see while starting up debian, it is mounting the folder properly, but while I try to login it gives an error: > Could not update ICEauthority file /home/user-name/.ICEauthority And after closing that message I get: > There is problem with the configuration server (/usr/libexec/gconf-sanity- > check-2 ... What's going wrong?
[ -0.01250708382576704, 0.01776895858347416, -0.008558675646781921, 0.008213371969759464, -0.006769487634301186, -0.00977393053472042, 0.008877763524651527, -0.005980100482702255, -0.014050525613129139, -0.015010069124400616, -0.01088650617748499, 0.006373636890202761, -0.011827655136585236, ...
[ 0.08239716291427612, 0.1453896164894104, 0.3773578703403473, -0.33674585819244385, 0.3640146255493164, 0.3074508011341095, -0.22564491629600525, 0.12986405193805695, 0.04555381089448929, -0.752068817615509, -0.2664254605770111, 0.3434915244579315, -0.19989484548568726, 0.46579352021217346,...
What is the relationship between brightness perception and physical properties of light, such as light intensity?
[ -0.017547300085425377, 0.0091874273493886, 0.004484057426452637, 0.013202754780650139, -0.12985412776470184, -0.0003247072745580226, 0.025456801056861877, -0.023066336289048195, 0.00128271768335253, 0.02838008664548397, -0.027565423399209976, 0.05386866256594658, -0.0010624935384839773, -0...
[ 0.21755294501781464, 0.002967884996905923, -0.23719267547130585, 0.13140137493610382, 0.10110841691493988, 0.14783859252929688, 0.009372048079967499, -0.12485568970441818, -0.15846021473407745, -0.22092172503471375, 0.10639657080173492, 0.4807809293270111, 0.16132895648479462, 0.1843237876...
I encountered the following problem while preparing my CV. There are a few links prepared by hyperref that are not broken in several lines. I am wondering if there is a way to fix this problem. This is a minimal working example: \documentclass{article} \usepackage[utf8]{inputenc} \usepackage[left=1.8in,right=1.8in,top=0.6in,bottom=0.6in]{geometry} \usepackage[usenames,dvipsnames,svgnames,table]{xcolor} \usepackage{hyperref} \definecolor{linkcolour}{rgb}{0,0.2,0.6} \hypersetup{colorlinks=true, breaklinks=true, urlcolor=linkcolour,linkcolor=linkcolour, citecolor=NavyBlue} \begin{document} \begin{itemize} \item A.~M.~Aragón, S.~Soghrati, and P.~H.~Geubelle. In-plane deformation effect on the cohesive failure of heterogeneous adhesives: a multi-scale analysis based on a 3D generalized finite element scheme, \emph{Journal of the Mechanics and Physics of Solids}, Published online; doi: \href{http://dx.doi.org/10.1016/j.jmps.2013.03.003}{http://dx.doi.org/10.1016/j.jmps.2013.03.003} \item A.~M.~Aragón, S.~Soghrati, and P.~H.~Geubelle. In-plane deformation effect on the cohesive failure of heterogeneous adhesives: a multi-scale analysis based on a 3D generalized finite element scheme, \emph{Journal of the Mechanics and Physics of Solids}, Published online; doi: \url{http://dx.doi.org/10.1016/j.jmps.2013.03.003} \end{itemize} \end{document} The result of this code is as follows: ![enter image description here](http://i.stack.imgur.com/DqZId.png) Also, is there a way to use bibtex entries in the middle of a document as they show up in the end of an article? I need to put a list of publications and the only way I managed to do this is just writing regular text to mimic the bibtex entries.
[ -0.001055491273291409, 0.013737642206251621, -0.007703797426074743, 0.032209575176239014, 0.010216696187853813, -0.0029557738453149796, 0.007110930047929287, -0.00589210819453001, -0.011072688736021519, -0.016818996518850327, -0.006021920591592789, 0.00028101622592657804, -0.0010009650141000...
[ 0.25037938356399536, 0.12052475661039352, 0.7155542373657227, -0.2115924060344696, 0.16723893582820892, -0.09946702420711517, 0.31580254435539246, -0.1495632529258728, -0.10151426494121552, -0.635701060295105, -0.031472641974687576, 0.8405058979988098, -0.0948914960026741, 0.15261036157608...
I have to use WordPress as some meta templating system and include some `JSP` snipptes which will be evaluated later with Tomcat after rendering the page statically. But naturally the `JSP` Code could not be interpreted by the php parser. This results in plain text strings mixed with the interpreted template parts. **Is there any way to hook into the (pre) rendering of a WordPress Template?** I need way get access the rendered template so i can strip out the `JSP`-Tags. Maybe wrapped between some unique tags or something else. But this ist not part of this question. ## Example <?php get_header(); ?> <render:jsp> <%= request.getParameter("title") %> </render:jsp> <?php get_footer(); ?>
[ 0.009241259656846523, 0.004419779404997826, 0.012522817589342594, 0.02141309157013893, 0.014957059174776077, 0.023791763931512833, 0.008990421891212463, 0.003989767283201218, -0.020574407652020454, -0.02393493987619877, -0.011038814671337605, 0.013626210391521454, -0.005922081880271435, 0....
[ 0.12926650047302246, 0.07624403387308121, 0.10557935386896133, 0.2220373898744583, -0.27675026655197144, -0.18079105019569397, 0.5165519118309021, -0.017134210094809532, 0.16525427997112274, -0.7364555597305298, 0.04080076143145561, 0.4374570846557617, -0.4303446114063263, -0.1472001373767...
Is it possible to get an executable to execute by default? What I mean is this. I have an `.sh` file, which if I click on twice, it will show me this: ![enter image description here](http://i.stack.imgur.com/XeqQw.png) If I then click `Execute`, it does the right thing. Is it possible to get it to `Execute` without being shown the `Execute File` dialog box? So simply by double clicking the `.sh` file, it should do its thing without showing me the `Execute File` dialog box. I am using Lubuntu/PCManFM if that info is needed.
[ -0.0019212120678275824, 0.0016917167231440544, -0.0026556074153631926, 0.013303391635417938, -0.021710867062211037, -0.012900291010737419, 0.00549774756655097, -0.0006393776275217533, -0.020184941589832306, -0.017652198672294617, -0.013079097494482994, 0.010457209311425686, 0.003990822937339...
[ 0.41492971777915955, 0.04325143247842789, 0.19707152247428894, -0.03186360001564026, 0.1638123244047165, -0.24080274999141693, 0.532170832157135, -0.11541826277971268, -0.390474796295166, -0.4404909610748291, -0.1882510930299759, 0.7567663788795471, -0.27889707684516907, -0.212722495198249...
I'm using Windows 7 and just installed Texmaker. Is it possible to configure it in such a way that I can run Latex -> BibTex -> Latex -> Latex? I looked at the configuration options here: ![enter image description here](http://i.stack.imgur.com/ZE0fH.png) I tried replacing the `bibtex %` line with `pdflatex -synctex=1 -interaction=nonstopmode %.tex bibtex % pdflatex -synctex=1 -interaction=nonstopmode %.tex` but that did not work. What should I change to be able to run this command?
[ -0.001875240821391344, 0.00828905962407589, -0.009698780253529549, 0.01867831125855446, 0.004765940830111504, -0.0034493543207645416, 0.005725416354835033, 0.010212225839495659, -0.021679546684026718, -0.02390408329665661, -0.002719176933169365, 0.01052175648510456, -0.00997097697108984, 0...
[ 0.06545297056436539, -0.2949618995189667, 0.8882433772087097, -0.0065574790351092815, 0.2502977252006531, -0.12898868322372437, 0.44911566376686096, 0.17720834910869598, -0.43825390934944153, -0.4181172251701355, 0.0019173499895259738, 0.8525547981262207, -0.4558713436126709, -0.2040186971...
Inspired by the recent question Alternatives to LaTeX (currently 58 upvotes) on http://tex.stackexchange.com/: Are there any paid-for or open source alternatives to Mathematica which produce equal or even better functionality, specifically with regard to solving, manipulating and visualising algebraic expressions. Note: a similar question is Best open-source Mathematica equivalent, however this question is targeted at users of Mathematica.SE (and is five years in the future).
[ 0.015194389037787914, 0.02062864601612091, -0.004568240139633417, 0.009923478588461876, 0.019528187811374664, -0.01005653478205204, 0.008426759392023087, -0.0128040025010705, -0.013495133258402348, -0.02681570313870907, -0.013974283821880817, 0.013752634637057781, 0.0076880729757249355, 0....
[ 0.18520011007785797, -0.30301281809806824, 0.3309466540813446, 0.41738301515579224, 0.15740656852722168, 0.03473760560154915, -0.09117373079061508, -0.11947143822908401, -0.36804690957069397, -0.44376620650291443, 0.14363257586956024, 0.20364119112491608, 0.09269814193248749, 0.02053595334...
I was told once that the best place to work as a developer is a company where the software you write is what makes the company money, whether it be software production or software services like consulting. This is opposed to a company where the software you write is just to support some other part of the business that makes the money, like manufacturing or finance. I know there are always exceptions, but in general, are employees treated better if they are on the front lines of profit generation, as opposed to being just another cost center? **cost center (n.)** - A cost center is part of an organization that does not produce direct profit and adds to the cost of running a company. Examples of cost centers include research and development departments, marketing departments, help desks and customer service/contact centers.
[ -0.010310008190572262, 0.0056286463513970375, -0.005831075832247734, -0.007650002837181091, -0.0011777342297136784, -0.006274037063121796, 0.008644133806228638, 0.0002192154061049223, -0.011518135666847229, -0.004080253653228283, -0.01628444343805313, 0.011625096201896667, 0.0249183662235736...
[ 0.6763078570365906, 0.2660207748413086, -0.252984881401062, 0.3849993050098419, 0.38557857275009155, 0.01678839512169361, -0.21919967234134674, 0.2717537581920624, -0.11388528347015381, -0.38023698329925537, -0.18842914700508118, 0.6032673716545105, 0.3810616731643677, 0.20107683539390564,...
I'm currently writing in markdown and converting using Pandoc, but there are additional replacements that I want to do, as well as fixing some of what pandoc does that I don't like. Does anyone know the best approach to writing a script or somehow automating the process of find+replace using regex? I'm not asking for someone to do this for me, but it would be great just to get a hint as to what kind of thing I need (e.g. which programming language, if I need to go down that route).
[ -0.0016404930502176285, 0.019128471612930298, -0.008266312070190907, 0.00365849188528955, -0.008260099217295647, -0.014828471466898918, 0.004210289102047682, 0.012705725617706776, -0.019307389855384827, -0.00474992860108614, 0.00011426381388446316, 0.008295021951198578, 0.007314702495932579,...
[ 0.45077604055404663, 0.2982810139656067, -0.042765773832798004, -0.011467200703918934, 0.02880195342004299, 0.06626302003860474, 0.35142782330513, -0.015867341309785843, -0.12185829132795334, -0.8157970905303955, 0.18649205565452576, 0.5556904077529907, 0.05571671202778816, 0.0566400326788...
Hi I am following the instructions from this article and I cannot get it to work, it gives me a blank screen. I think there is something wrong with the opening and closing php tags between the div "view view-first" and the div "mask". When I remove these tags I get the site is displayed again but instead of images the is php code displayed. I did get the code to work briefly but now it is broken again. Thanks for your help. RT <?php get_header(); ?> <div > <?php query_posts(array('category__in' => array(5), 'posts_per_page' => 10000)); ?> <?php if (have_posts()) : while (have_posts()) : the_post(); ?> <?php $currentid = get_the_id(); ?> <div class="grid-box grid-block mod-box width33"> <div class="view view-first"> <?php if ( get_the_post_thumbnail($post_id) != '' ) { echo '<a href="'; the_permalink(); echo '" class="thumbnail-wrapper">'; the_post_thumbnail(array(300,600)); echo '</a>'; } else { echo '<a href="'; the_permalink(); echo '" class="thumbnail-wrapper">'; echo '<img src="'; echo catch_that_image(array(300,600)); echo '" alt="" />'; echo '</a>'; } wp_reset_query(); ?> <div class="mask"> <h2><a href="<?php the_permalink(); ?>"><?php echo get_the_title($currentid); ?></a></h2> <p><?php echo get_the_excerpt( $post->parent ); ?></p> <a href="<?php the_permalink(); ?>" class="info">Read More...</a> </div> </div> </div> <?php endwhile; endif; wp_reset_query(); ?> <div class="clearfix"> </div> <?php get_footer(); ?>
[ -0.017970800399780273, 0.0014060265384614468, 0.003927018027752638, 0.0072845760732889175, -0.0397573858499527, -0.008837377652525902, 0.005873582325875759, -0.004372606053948402, -0.019374877214431763, 0.019145525991916656, -0.03348049893975258, 0.013684243895113468, -0.0015877520199865103,...
[ 0.23756179213523865, 0.15446260571479797, 0.8438012003898621, -0.04991951584815979, -0.5447590947151184, -0.24787254631519318, 0.6023579239845276, -0.3793541193008423, -0.2907622158527374, -0.39617928862571716, 0.06306745857000351, 0.5314244627952576, -0.30765265226364136, 0.15053956210613...
I'm having trouble getting a PostGIS raster to load and display in MapServer. I'm not sure, but suspect it's a version problem. My mapfile layer and the error message are shown below. LAYER NAME bio12_2_5m TYPE raster STATUS ON DATA "PG:host=localhost port=31600 dbname='wdata' user='postgres' password='*****' schema='worldclim' table='bio12_2_5m'" PROCESSING "NODATA=-9999" PROCESSING "SCALE=AUTO" END msDrawMap(): Image handling error. Failed to draw layer named 'bio12_2_5m'. msDrawRasterLayerLow(): Unable to access file. Corrupt, empty or missing file 'PG:host=localhost port=31600 dbname='wdata' user='postgres' password=' _*_ *' schema='worldclim' table='bio12_2_5m'' for layer 'bio12_2_5m'. Sorry, but table bio12_2_5m doesn't seem to have regular blocking arrangement. Only tables with regular blocking arrangement can be read from now This is WorldClim data loaded into PostGIS with raster2pgsql like: raster2pgsql -s 4326 bio12.bil worldclim.bio12_2_5m > bio12.sql To get around what seems like a blocking problem I've also tried: raster2pgsql -s 4326 -C -r bio12.bil worldclim.test > test.sql raster2pgsql -s 4326 -t 10x10 bio12.bil worldclim.test2 > test2.sql And with the MapServer layer I've tried setting mode=1 or mode=2, still no success. Versions ... Ubuntu 12.04 LTS, PostgreSQL 9.1.11, GDAL 1.10.0, PostGIS 2.0.4 with raster support, and MapServer 6.0.1. But (always a but) to be fair I started with GDAL 1.7.3 via the default ubuntu repository, and at the end updated GDAL to 1.10.0 via the Ubuntu GIS "unstable" repository since it seems like the newer versions of GDAL are required to address/support the blocking items. Thanks, very close I suspect and not sure if installing everything from scratch would help starting with the 1.10.0 GDAL. Also tried building GDAL from source and no success there.
[ -0.01859242096543312, -0.004041898529976606, 0.00971953198313713, 0.022073622792959213, 0.018123392015695572, 0.02332020178437233, 0.009215058758854866, 0.009238652884960175, -0.01339007169008255, -0.009169659577310085, 0.0009610714623704553, 0.009587856940925121, 0.009118693880736828, 0.0...
[ -0.24116764962673187, 0.0490824393928051, 0.5436416268348694, -0.04666123911738396, -0.0786559209227562, -0.10147085040807724, 0.23660814762115479, 0.04202987626194954, -0.27670547366142273, -1.0181511640548706, -0.00673891743645072, 0.35199984908103943, -0.081673264503479, 0.4571661651134...
I have a WordPress option that stores some data, like: <h1>Header</h1> <p>Paragraph</p> [shortcodesomething/] [shortcode]Contents[/shortcode] I'm displaying this option's value using `echo get_option('my_option'));`. Of course the shortcodes doesn't work, and I'm wondering how to force them to do what they are supposed to do? `echo do_shortcode(get_option('my_option'));` doesn't work either (I know this is an absolutely wrong approach, but I don't know any other way of displaying them).
[ 0.01446928083896637, 0.007620992138981819, -0.007930805906653404, 0.013074127957224846, -0.005812019109725952, -0.004534277133643627, 0.006737383082509041, 0.009459205903112888, -0.01856347732245922, -0.0029062109533697367, -0.019992191344499588, 0.0042658038437366486, 0.0055695343762636185,...
[ 0.18351493775844574, -0.09839753806591034, 0.39484408497810364, -0.12925419211387634, 0.23081539571285248, 0.20168359577655792, 0.4421146512031555, -0.16181261837482452, -0.04391472041606903, -0.6368229389190674, -0.18437977135181427, 0.5344310998916626, -0.3422432243824005, -0.11328424513...
Does anyone know how to get back/ what mission to do to get Sanctuary 'back' , so I can place the AI in Moxxie's radio?
[ 0.02659749984741211, 0.013196837157011032, 0.001671827514655888, -0.030149349942803383, 0.013587779365479946, -0.005270824301987886, 0.013808228075504303, -0.0025125278625637293, -0.026475924998521805, -0.01212908886373043, -0.004353792406618595, 0.019004669040441513, -0.019913721829652786, ...
[ 0.7179590463638306, 0.15197913348674774, 0.21096207201480865, 0.1607818901538849, -0.01745547540485859, -0.3185311555862427, 0.4135493040084839, 0.017416931688785553, 0.035897426307201385, -0.026574380695819855, -0.14082196354866028, 0.7652873396873474, -0.1008950024843216, 0.0996617898344...
A javascript application release won't technically be closed source - at best it can be obfuscated. Given that, my question has two parts: > * Is it at all possible to release a technically open source application > under a proprietary license? > * Are there any examples of applications which are technically open > source, have been kept under a proprietary license, and are being profitably > sold? >
[ 0.0067622363567352295, 0.0063711898401379585, 0.008263394236564636, 0.010859299451112747, -0.029213760048151016, 0.010950884781777859, 0.009438376873731613, 0.028970271348953247, -0.026331577450037003, -0.050322026014328, -0.01865242049098015, 0.0290937889367342, 0.02348412573337555, 0.016...
[ 0.38463491201400757, 0.36608150601387024, 0.051712002605199814, 0.06523028761148453, -0.20703105628490448, -0.5176824927330017, 0.2186656892299652, -0.02469305694103241, -0.15096740424633026, -0.3840503990650177, -0.5040057301521301, 0.23331551253795624, -0.38254275918006897, -0.1091683954...
in Yang-Mills-Theory with matter fields a dirac spinor $\psi$ transforms under BRST as $$\psi \to \delta_\Omega\psi=i\eta\psi $$ with $\eta$ being a ghost field. If I want to get the transformation of the adjoint spinor $\bar \psi$ I get by using the invariance of $\bar \psi \psi$ $$ 0=\delta_\Omega (\bar \psi\psi)=(\delta_\Omega\bar\psi)\psi - \bar\psi (\delta_\Omega\psi) \quad \Rightarrow \quad \delta_\Omega \bar\psi=i\bar\psi \eta$$ If I now want to get the transform directly, I get $$ \delta_\Omega \psi^\dagger \gamma_0=[\psi^\dagger,\Omega]_+\gamma_0=([\psi,\Omega]_+)^\dagger \gamma_0=(i\eta\psi)^\dagger\gamma_0=-i\bar\psi\eta $$ So I get different results. Where is my error? What I am not sure about, is whether if I have $(\eta\psi)^\dagger$ if this is equal to $\psi^\dagger\eta$ or $-\psi^\dagger \eta$ as the transpose part should be purely in the dirac space. Thanks in advance.
[ -0.006453622132539749, 0.0031647835858166218, -0.009111771360039711, 0.02100548706948757, 0.015052424743771553, -0.01298561878502369, 0.010685894638299942, 0.004509482067078352, -0.010064372792840004, -0.005052250809967518, -0.012790335342288017, 0.01251192670315504, -0.013377415016293526, ...
[ -0.09678016602993011, 0.022708669304847717, 0.6490642428398132, -0.2397405207157135, -0.27496638894081116, 0.35360926389694214, -0.19023682177066803, -0.43066656589508057, -0.15450520813465118, 0.2083885222673416, -0.05465526133775711, -0.07879441976547241, -0.627129852771759, 0.8432969450...
I have the kill-screen up right now. I just finished a game that I really had fun with so I was wondering if it was possible to get the full game script somehow?
[ 0.010559738613665104, 0.0014026773860678077, -0.006708495318889618, 0.0060233501717448235, -0.01586015895009041, -0.029652703553438187, 0.008107994683086872, 0.0017819240456447005, -0.04789650812745094, 0.012163090519607067, -0.00653361901640892, 0.04183528572320938, 0.010628839023411274, ...
[ 0.9188976287841797, 0.21072296798229218, 0.41839325428009033, -0.05127442628145218, -0.28772082924842834, -0.21328891813755035, 0.24604667723178864, 0.008537339977920055, -0.1345788687467575, -0.6730607748031616, 0.5074384808540344, 0.8564391136169434, -0.07328979671001434, -0.105480268597...
I'm studying basic physics. I'm using the text available at http://www.anselm.edu/internet/physics/cbphysics/downloadsI.html. It develops the universal law of gravitation by postulating the existence of a vector at each point of the form $$ g_P = \sum G\frac{m_i}{|r|_i^3}r_i, $$ Where $m_i$ and $r_i$ are the mass of and separation vector from $P$ for all particles that aren't at the point $P$. It examines the effect of one particle on another. If the separation vector is $r$ then from the above equation, we see that $$ g = G\frac{m_1}{|r|^3}r $$ and that when a particle of mass $m_2$ is placed at the given point, the force will be $$ F = G\frac{m_1m_2}{|r|^3}r $$ The authors then go on to claim that we can either repeat the development to see the effect of the second particle on the first one or apply Newton's third law. How does Newton's third law apply through a field? If it's the field exerting the force, then Newton's third law would require a force on the field and not the object "generating" the field, correct?
[ -0.01591995730996132, 0.014777598902583122, 0.0031107342801988125, 0.007700128015130758, -0.007125892210751772, -0.010346522554755211, 0.005065836012363434, -0.013215569779276848, -0.009749732911586761, -0.014857803471386433, -0.003541609738022089, 0.006511644460260868, -0.005962695926427841...
[ 0.13418515026569366, 0.3343340754508972, 0.6941930055618286, -0.1661793738603592, -0.009623270481824875, 0.5175532102584839, -0.3234400749206543, -0.41704660654067993, -0.3227353096008301, -0.2167988270521164, 0.10350514203310013, 0.014833052642643452, -0.29574334621429443, 0.5411808490753...
I am new to probability course and find concepts very confusing. I am learning maximum likelihood estimation and as a starting point for that do we find the density or the distribution? In general, do we find the distribution from density or is it vice versa and how does one find it? For example, for Gaussian distribution do we find the density of a variable or the distribution so as to estimate unknown parameter A in the case **x[n] = A + w[n]** where w[n] are uncorrelated gaussian noise. This may sound too trivial but I am really confused. In many contexts I have come across where pdf = prob. density function and in some areas I find pdf = prob distribution function !! Please help.
[ -0.004952127113938332, 0.005563968792557716, 0.000918630335945636, 0.004663477651774883, -0.024939894676208496, -0.00911729596555233, 0.006772956810891628, 0.0010776352137327194, -0.02116328664124012, -0.022054489701986313, 0.0023732464760541916, 0.014776984229683876, -0.01630726084113121, ...
[ 0.3036077618598938, -0.222293421626091, -0.034511029720306396, 0.1268739253282547, -0.31886112689971924, -0.09622558951377869, -0.3147757649421692, -0.13499663770198822, -0.1138116791844368, -0.6967877745628357, 0.18489472568035126, 0.4806588590145111, -0.11096963286399841, 0.4467224478721...
I was an idiot and said, "No thanks, I don't want that godly weapon you have yet. Let me see what else you have." and the blacksmith responded, "Oh, you don't want it? Let me melt it down _immediately_ then so no one can ever have it again." Now he doesn't have anything for sale. What will cause the blacksmith to have more weapons for me to look through?
[ 0.027143161743879318, 0.020172759890556335, -0.006300922948867083, -0.0023931325413286686, -0.028057962656021118, 0.005564131774008274, 0.006552548613399267, -0.002980200806632638, -0.025682872161269188, -0.002457440597936511, -0.007885460741817951, 0.021371707320213318, -0.02220418304204940...
[ 0.4576060473918915, 0.6621832847595215, -0.13044621050357819, 0.25863176584243774, -0.2698877155780792, 0.09923869371414185, 0.3096723258495331, 0.115789033472538, -0.21358972787857056, -0.012247022241353989, 0.13991990685462952, 0.31101515889167786, -0.11883771419525146, 0.475705921649932...
I'm considering buying an android phone (probably a Galaxy S variant). But I only run Linux (Gentoo if it matters). So, I'm curious what features work with Linux, and whether there are any features that I'll miss out on by not syncing my phone to a Windows or OS X PC. (I'm being optimistic in the title and assuming that most things will Just Work -- please correct me if I'm wrong.)
[ -0.0033969059586524963, -0.0033608637750148773, -0.008945737034082413, -0.006521681789308786, -0.014763363637030125, -0.011640271171927452, 0.006447531748563051, 0.020161021500825882, -0.018520908430218697, -0.03244528919458389, -0.01495598815381527, 0.009892662987112999, 0.01504539046436548...
[ 0.5380985140800476, 0.22654448449611664, 0.2128361463546753, 0.18359455466270447, 0.524629533290863, 0.2211223542690277, 0.17444048821926117, 0.4919792413711548, -0.01315572950989008, -0.46599847078323364, 0.22832584381103516, 0.6433073282241821, 0.022695055231451988, -0.041520021855831146...
Im currently evaluating a front-end architecture for a big app that will be developed by a team of about 6 developers + 1 web designer so that said, having a robust SVN friendly architecture is a must. So far I'm evaluating this two options: * A mixture of JQuery/UI/Mobile + Backbone/Underscore + Require JS + Twitter-Bootstrap * ExtJS 4 This are the requirements the architecture needs to meet: * SVN Fiendly (or Friendly depending on you previous experiences) * Active community and extensive documentation * Extensible * Easy to learn and to maintain (nice to have) * Cross-browser support (most of the time at least) * Support for Mobile devices * Easy to maintain (MVC, MVVM) * OpenSource (nice to have) The previously mentioned ones meet most of the requirements in different levels and so far they are the only ones I've found about, am I missing any others?
[ -0.0013363463804125786, 0.018459325656294823, -0.0018518281867727637, 0.01341758482158184, 0.006800467614084482, -0.006055518984794617, 0.008663114160299301, 0.03431525453925133, -0.012430701404809952, -0.027473364025354385, -0.019995085895061493, 0.02238740585744381, -0.010210748761892319, ...
[ 0.211930051445961, 0.1696430891752243, 0.6048141717910767, -0.08519899845123291, -0.26708751916885376, -0.0942700132727623, 0.17980797588825226, -0.4103899598121643, -0.09636326134204865, -0.6935756206512451, -0.24931170046329498, 0.43524497747421265, -0.10055913776159286, -0.1431370675563...
When talking about electronic drawings is it correct to use scheme or should I only use schematic?
[ 0.01361437700688839, 0.008018874563276768, -0.06183819845318794, 0.004620036110281944, -0.04461565241217613, 0.039707452058792114, 0.024572616443037987, -0.0006037958082742989, -0.012322596274316311, -0.14094586670398712, -0.03515121340751648, 0.017750844359397888, 0.0003219274221919477, 0...
[ 0.7913691401481628, 0.12316739559173584, 0.2208549678325653, 0.1454300731420517, -0.1220751404762268, -0.08714690059423447, -0.21055114269256592, -0.08368252962827682, -0.2690604627132416, -0.6790533661842346, 0.44362151622772217, 0.3449828326702118, -0.26811742782592773, -0.27033486962318...
I tried to backup my partition using dd on another disk and I created a partition of the same size. When I finished the copy I had problems mounting the partition since I got `bad geometry: block count` . I noticed that in order to create a correct backup (avoid bad geometry error) not only the two partitions must have the same size in MB but also the same number of blocks (The number in blocks you can check using fdisk -l). I also noticed that the number of blocks varies if I create the partition starting from the beginning or starting from the end . Can someone explain me the behaviour and how to create two partition with the exact the same block size?
[ -0.008387193083763123, 0.013830428943037987, -0.01899290457367897, 0.019634906202554703, 0.007835413329303265, 0.030290141701698303, 0.010925740003585815, 0.00249122129753232, -0.0179403405636549, -0.03950094059109688, -0.01941862888634205, 0.018080325797200203, -0.006835984066128731, 0.00...
[ 0.2098359912633896, 0.1011953055858612, 0.033514898270368576, 0.02910706214606762, 0.06313595920801163, 0.20839112997055054, 0.2523576319217682, -0.2579384744167328, -0.4885910749435425, -0.6644969582557678, -0.03700504079461098, 0.5961734652519226, -0.25784483551979065, 0.4149596095085144...
I have an expression as below: Equations = 2.0799361919940695` x[1] + 3.3534325557330327` x[1]^2 - 4.335179297091139` x[1] x[2] + 1.1989715511881491` x[2]^2 - 3.766597877399148` x[1] x[3] - 0.33254815073371535` x[2] x[3] + 1.9050048836042945` x[3]^2 + 1.1386715715291826` x[1] x[4] + 2.802846492104668` x[2] x[4] - 0.6210244597295915` x[3] x[4] + 4.943369095158792` x[4]^2 I want to write it in an output file. So I use the below code: removebracketvar[x_] := StringReplace[ StringReplace[ ToString[x], {"[" -> "", "]" -> "", "," -> "", "*^" -> "e", ".*" -> ".0*"}], Whitespace -> ""]; SetDirectory["C:\\folder"]; WriteString["eqfile.txt", removebracketvar[ ToString[Equations , InputForm, NumberMarks -> False]] ]; Close["eqfile.txt"] The slight problem with the code for me is that it inserts the floating point numbers up to 16 digits of precision. I just want them to around up to 10 digits of precision. When I use `SetPrecision[Equations,10]`, it weirdly changes `x[1]` etc. to `x[1.0000000]`, etc.! I want to leave the variables as they are but want to change the floating points to less number of digits after the decimal point. What would be the best way of doing this?
[ -0.00660147937014699, 0.0013872929848730564, -0.008833244442939758, 0.0034806858748197556, -0.025533437728881836, -0.005229446105659008, 0.0006326834554784, -0.01203338522464037, -0.008074484765529633, -0.018856214359402657, 0.0016771923983469605, 0.004305321257561445, -0.008413680829107761,...
[ -0.32021981477737427, 0.3883098363876343, 0.4908929169178009, -0.16897225379943848, -0.05140317603945732, 0.2036243975162506, 0.4927844703197479, -0.3602643609046936, 0.18067221343517303, -0.6978958249092102, 0.21327941119670868, 0.5350609421730042, 0.0018851960776373744, 0.313300997018814...
Sometime back in an interview, I was asked to write following program: * * * _In a keypad of a mobile phone, there is a mapping between number and characters. e.g. 0 & 1 corresponds to nothing; 2 corresponds to 'a','b','c'; 3 corresponds to 'd','e','f'; ...; 9 corresponds to 'w','x','y','z'_. _User should input any number (e.g. 23, 389423, 927348923747293) and I should store all the combinations of these character mapping into some data structure. For example, if user enters "23" then possible character combinations are_ : > ad, ae, af, bd, be, bf, cd, ce, cf _or if user enters, "4676972" then it can be_ , > gmpmwpa, gmpmwpb, ..., hnroxrc, ..., iosozrc * * * Interviewer told that people have written code for this within 20-30 mins!! Also he insisted I have to **write on paper**. If I am writing a code then my tendency is as of I am writing production code, even though it may not be expected from me. So, I always try to think all the aspects like, _optimization, readability, maintainability, extensible_ and so on. Considering all these, I felt that I should be writing on PC and it needs decent 2 hours. Finally after 25 mins, I was able to come up with just the concept and some shattered pieces of code (not to mention of my rejection). My question is not the answer for the above program. I want to know that **is this a right way to judge the caliber of a person** ? Am I wrong / too slow in the estimates ? Am I too idealistic ?
[ -0.005417305510491133, 0.0162473376840353, -0.008829744532704353, 0.00825412292033434, -0.01011939812451601, 0.007034346926957369, 0.0069514075294137, -0.009518658742308617, -0.013125483877956867, -0.004655026830732822, -0.009893253445625305, -0.000506581855006516, 0.00473475456237793, 0.0...
[ -0.2016722708940506, 0.34435704350471497, 0.2772715389728546, -0.04251744598150253, 0.1449613869190216, 0.41118210554122925, 0.18067747354507446, 0.011049186810851097, -0.009222645312547684, -0.7131803631782532, -0.21052700281143188, 0.026614613831043243, 0.09047842770814896, 0.11681431531...
> This chapter is the final of four chapters covering the case study design of > the relational database model for the online auction house company. > **Business rules application to design encompasses** stored procedures, as > well as specialized and very detailed field formatting and restrictions. I can't make sense of the part that's in bold.
[ 0.014333758503198624, 0.01895791105926037, 0.0024530841037631035, 0.02333267405629158, -0.004011486191302538, 0.01519491896033287, 0.011774118058383465, 0.031285420060157776, -0.02219325490295887, -0.006341390311717987, -0.03498472273349762, 0.018142633140087128, 0.018089042976498604, 0.01...
[ 0.31706175208091736, 0.23382961750030518, 0.32115137577056885, 0.013994656503200531, -0.2860458195209503, -0.10722095519304276, 0.1997750997543335, -0.35352522134780884, 0.07235467433929443, -0.5213426947593689, 0.10140860080718994, 0.45420554280281067, 0.04733243212103844, 0.1905132830142...
I am working on a site that will be listing a few hundred products and we wanted to be able to quickly/easily show specific products on a page based on taxonomies. I have been doing some research and thought I was on the right track. First, I found this article on digwp.com which I thought was going to work great. I can not get it to pull the posts based on taxonomy though - just post_type. The Post Type is "Used Items" and I created a taxonomy named Color where one of the slugs was Red. [loop the_query="post_type=used-items&color=red&ord=ASC"] I also found this plugin for doing a query by shortcode, but same situation with it not pulling by taxonomy. **My question** Does anyone know what I can do to pull posts based on post type/taxonomies (would be using multiple each query) directly from a WordPress page. We want to be able to quickly query products without having to program the query into a page template for each possible option.
[ -0.0031466712243855, 0.00264097866602242, 0.00013170577585697174, 0.015022605657577515, 0.012199142947793007, 0.006136642768979073, 0.007006203290075064, 0.03166374936699867, -0.018234431743621826, 0.0033393269404768944, -0.010384831577539444, 0.005505596287548542, -0.006207369267940521, 0...
[ 0.6706686019897461, -0.031135547906160355, -0.1162957027554512, 0.050383929163217545, -0.09962169826030731, 0.17625488340854645, -0.2011539340019226, 0.23636741936206818, -0.4090628921985626, -0.5590834021568298, 0.6914425492286682, 0.40121567249298096, -0.28090986609458923, 0.602182686328...
I'm using the Turnkey Linux virtual machine. The version of Debian in there is apparently on the "Wheezy" upgrade path (if that's what you call it). I'm not completely comfortable with Linux, as you can see. I need some new features that MySQL 5.6 offers, while my current installation is MySQL 5.5.35. I'd like to get the latest version (5.6.17 as of this posting) or at least any 5.6.x. I've tried `apt-get update`, `apt-get upgrade`, `apt-get dist-upgrade`. These updated a bunch of things successfully, but none were MySQL. I tried `apt-get upgrade mysql-server`, which says I already have the latest version. A Debian Wheezy package repository list on the web shows MySQL 5.5 and nothing else, when I need 5.6. I then read something about "backports", which sounds like it might be able to get me things that my particular Debian isn't supposed to have yet -- but I can't seem to get it working. It requires adding lines to `/etc/apt/sources.list.d/sources.list`. When I try doing that, then running `apt-get update`, I get 404 errors. I've tried several other routes as well, one being compiling the vanilla MySQL download, which I failed at miserably and would rather not try again. I'm positively baffled at how this is done. Most of the URLs listed on forums/blogs for `apt-get`, and even for `wget`, seem to die soon after they're posted, so I can't understand how anyone even goes about finding the right information. I've been trying to do this for two days straight now and can't get anywhere. So assuming you've read this far, I would love to know if anyone has any way of upgrading MySQL in Debian in some relatively painless automated way (one where I wouldn't have to use `make` or export/import my current databases and configs manually).
[ 0.005918595008552074, -0.004008300136774778, -0.014580970630049706, -0.004611543845385313, -0.03134983777999878, -0.00667975191026926, 0.006359864491969347, -0.0000293022021651268, -0.015268937684595585, -0.02266460284590721, -0.010483498685061932, 0.0046024867333471775, -0.01302551664412021...
[ 0.5201148986816406, 0.18676531314849854, 0.2773328721523285, -0.03631379082798958, 0.09469543397426605, -0.22171156108379364, 0.2726826071739197, 0.3919300436973572, 0.006660933140665293, -0.8964865803718567, -0.06766579300165176, 0.8340709805488586, -0.028859818354249, 0.1825709193944931,...
Is there a way to make something like tabular such that in a loop we get each column in a line and render them in a custom format. Like this \begin{custom_tabular} l1c1 & l1c2 & l1c3 \\ l2c1 & l2c2 & l2c3 \end{custom_tabular} and render the the elements in a custom format, e.g. \textbf{l1c1}: l1c2$^{l1c3}$ for all the lines. Expected output: ![enter image description here](http://i.stack.imgur.com/UBtyH.png)
[ 0.007651114836335182, 0.015051251277327538, -0.0001745468471199274, 0.006007935386151075, -0.0021162049379199743, 0.004032779484987259, 0.004828443750739098, 0.009646875783801079, -0.01856910064816475, -0.014603637158870697, -0.002425146522000432, 0.006762555800378323, -0.006373882759362459,...
[ -0.11028770357370377, -0.41443169116973877, 0.3197413682937622, 0.12427356094121933, -0.054865121841430664, 0.5243871212005615, -0.21503636240959167, -0.17848899960517883, -0.2779514193534851, -0.33687636256217957, 0.2732083201408386, 0.2503001391887665, -0.4815957248210907, -0.43809905648...
![enter image description here](http://i.stack.imgur.com/KSVSn.png) Hooray! But I had to trade the shovel for it. That's two hundred rupees gone, if the store by the starting village even has one more to sell that is. What else can I give in exchange?
[ 0.0029718871228396893, 0.015796592459082603, -0.001621511997655034, 0.025655852630734444, -0.03908029943704605, 0.01271133404225111, 0.007070217747241259, 0.003781264880672097, -0.03240494802594185, -0.0046727159060537815, -0.013743149116635323, 0.008770463056862354, 0.0033983183093369007, ...
[ 0.188105046749115, 0.005006738472729921, -0.14142264425754547, 0.018625428900122643, -0.37317293882369995, 0.7301149368286133, 0.5757841467857361, -0.05942470207810402, -0.2997397184371948, 0.06999946385622025, 0.21784359216690063, 0.4780641496181488, 0.29861029982566833, 0.235627532005310...
If I make the height of the image more than 3cm, the table will automatically go to newpage which I don't want to. How can I make the table stay while having a nice image at the top-right of the paper? \documentclass{article} \usepackage{tikz} \usepackage{graphicx} \begin{document} \thispagestyle{empty} \noindent \noindent\begin{minipage}{.6\textwidth} \textbf{Standard Normal Cumulative Probability Table} \end{minipage} \begin{minipage}{\textwidth} \centering \includegraphics[height=3cm]{ztable.png} \end{minipage} \begin{center} \begin{tabular}{c|c c c c c c c c c c c} \multicolumn{11}{c}{\textbf{Cumulative probabilities for POSITIVE z-values are shown in the following table}} \\ \hline $\textbf{z}$&\textbf{0.00}&\textbf{0.01}&\textbf{0.02}&\textbf{0.03}&\textbf{0.04}&\textbf{0.05}&\textbf{0.06}&\textbf{0.07}&\textbf{0.08}&\textbf{0.09}\\ \hline \hline \textbf{0.0}&0.5000&0.5040&0.5080&0.5120&0.5160&0.5199&0.5239&0.5279&0.5319&0.5359\\ \textbf{0.1}&0.5398&0.5438&0.5478&0.5517&0.5557&0.5596&0.5636&0.5675&0.5714&0.5753\\ \textbf{0.2}&0.5793&0.5832&0.5871&0.5910&0.5948&0.5987&0.6026&0.6064&0.6103&0.6141\\ \textbf{0.3}&0.6179&0.6217&0.6255&0.6293&0.6331&0.6368&0.6406&0.6443&0.6480&0.6517\\ \textbf{0.4}&0.6554&0.6591&0.6628&0.6664&0.6700&0.6736&0.6772&0.6808&0.6844&0.6879\\ \textbf{0.5}&0.6915&0.6950&0.6985&0.7019&0.7054&0.7088&0.7123&0.7157&0.7190&0.7224\\ \textbf{0.6}&0.7257&0.7291&0.7324&0.7357&0.7389&0.7422&0.7454&0.7486&0.7517&0.7549\\ \textbf{0.7}&0.7580&0.7611&0.7642&0.7673&0.7703&0.7734&0.7764&0.7794&0.7823&0.7852\\ \textbf{0.8}&0.7881&0.7910&0.7939&0.7967&0.7995&0.8023&0.8051&0.8078&0.8106&0.8133\\ \textbf{0.9}&0.8159&0.8186&0.8212&0.8238&0.8264&0.8289&0.8315&0.8340&0.8365&0.8389\\ \textbf{1.0}&0.8413&0.8438&0.8461&0.8485&0.8508&0.8531&0.8554&0.8577&0.8599&0.8621\\ \end{tabular} \end{center} \end{document} ![enter image description here](http://i.stack.imgur.com/8j0DF.png)
[ 0.004724330734461546, 0.004523489158600569, 0.0053472924046218395, 0.017388805747032166, 0.00520577747374773, -0.0029372023418545723, 0.008567548356950283, 0.0049230488948524, -0.016517307609319687, 0.007254092022776604, -0.02504388988018036, -0.0038237920962274075, -0.0015512859681621194, ...
[ 0.2528974413871765, 0.07119685411453247, 0.9554333090782166, -0.1254889816045761, -0.06090794876217842, -0.034844089299440384, 0.4667329788208008, -0.22888247668743134, -0.240775004029274, -0.8128455877304077, 0.11326182633638382, 0.49434521794319153, 0.1537478119134903, 0.2341085374355316...
Assume I have a VM running on OSX with private IP 10.0.0.1 which can be accessed from the host machine. I was wondering how can I map a pseudo domain, `*.app.dev`, to the private IP on my host machine so that on my host machine I can resolve the domain `*.app.dev` to 10.0.0.1. The purpose of this setup is to have virtual environment for development and not pollute my host machine with unnecessary packages and services. edit: I realize that `/etc/hosts` can accomplish non wildcard domain names, I should have been more clear and mention wildcard domain names.
[ -0.015154603868722916, 0.0044040195643901825, 0.009337006136775017, 0.002084835199639201, -0.039681967347860336, 0.011091656982898712, 0.008545691147446632, 0.009711253456771374, -0.014487084001302719, -0.0011214589467272162, -0.01346834097057581, 0.01495507825165987, -0.00004745880141854286...
[ 0.30687880516052246, 0.1195598617196083, 0.26292479038238525, 0.15610477328300476, 0.21622304618358612, -0.14982885122299194, 0.1375838667154312, -0.12896865606307983, -0.054106712341308594, -0.7506693601608276, -0.1080671027302742, 0.5560598373413086, -0.5463299751281738, 0.10014024376869...
I recently experienced a logwatch issue that is fixed by running `export DATE_MANIP=DM5`. cron seems to forget this when I logout (presumably because it's not saved). Where can I add this custom environment variable? Since I don't manage the code for logwatch, I don't want to necessarily add an `export` command inside the script that launches. I tried adding `export DATE_MANIP=DM5` to `/root/.bash_profile` but that doesn't help when I'm not logged in. CentOS 6.x
[ 0.008553694933652878, 0.014007392339408398, -0.01580614410340786, 0.009416823275387287, -0.004572765901684761, 0.001252058194950223, 0.0089573934674263, 0.028689509257674217, -0.015656551346182823, -0.029046326875686646, -0.014766544103622437, 0.013994304463267326, -0.007326857186853886, 0...
[ 0.18674805760383606, 0.0786365270614624, 0.2588755190372467, -0.33447104692459106, -0.0005055064102634788, -0.03316449373960495, 0.41220054030418396, 0.08038459718227386, -0.6322119832038879, -0.6333293318748474, 0.11140652000904083, 0.8433523774147034, -0.035330284386873245, 0.16740398108...
I would like to know if there is a difference in the values taken from measurements of a given characteristic sampled at primary `forest=1` and secondary `forest=0` I'm using two vectors, one with the data itself (`DATA$Measurements`), and the other one with the category of sampling point (`DATA$Point`). How should I perform the `wilcox.test` fucntion? I'm doing this way: wilcox.test(DATA$Measurements~DATA$Point) But I notice that are other forms of performing this test: wilcox.test(DATA$Measurements,DATA$Point) wilcox.test(DATA$Measurements,DATA$Point,paired=TRUE) Which one is more suited for testing my question of interest, and what is the difference between these options? (Is there a difference between points for this given characteristic)?
[ -0.006413462571799755, 0.0068248361349105835, -0.0030494045931845903, 0.008550139144062996, -0.008568203076720238, -0.0031395871192216873, 0.009977178648114204, 0.01444482896476984, -0.012202379293739796, -0.0021663913503289223, -0.0006388515466824174, 0.008977917954325676, -0.02246229909360...
[ 0.34935590624809265, -0.3660849928855896, 0.35703060030937195, 0.22413209080696106, -0.08541399240493774, 0.21742713451385498, 0.036475684493780136, -0.5044653415679932, 0.05934647098183632, -0.2333911657333374, 0.4545682668685913, 0.42210280895233154, -0.2443883717060089, 0.22794808447360...
I have a relatively new Debian installation -- when it was first installed, it was without a gui, but I ran `apt-get install xinit` and `apt-get install openbox`, so when I run `startx` I get an openbox session. My openbox menu is rather limited, and `.xsession-errors` shows Openbox-Message: Unable to find a valid menu file "debian-menu.xml" Openbox-Message: Unable t o find a valid menu file "/var/lib/openbox/debian-menu.xml" When I check `/var/libopenbox/debian-menu.xml` there's nothing there. Is it possible to fix it so that openbox can show debian menu items? Edit: The following is an excerpt from my `~/.config/openbox/rc.xml` file: <menu> <!-- You can specify more than one menu file in here and they are all loaded, just don't make menu ids clash or, well, it'll be kind of pointless --> <!-- default menu file (or custom one in $HOME/.config/openbox/) --> <!-- system menu files on Debian systems --> <file>/var/lib/openbox/debian-menu.xml</file> <file>debian-menu.xml</file> <file>menu.xml</file> <hideDelay>200</hideDelay> <!-- if a press-release lasts longer than this setting (in milliseconds), the menu is hidden again --> <middle>no</middle> <!-- center submenus vertically about the parent entry --> <submenuShowDelay>100</submenuShowDelay> <!-- time to delay before showing a submenu after hovering over the parent entry. if this is a negative value, then the delay is infinite and the submenu will not be shown until it is clicked on --> <submenuHideDelay>400</submenuHideDelay> <!-- time to delay before hiding a submenu when selecting another entry in parent menu --> if this is a negative value, then the delay is infinite and the submenu will not be hidden until a different submenu is opened --> <applicationIcons>yes</applicationIcons> <!-- controls if icons appear in the client-list-(combined-)menu --> <manageDesktops>yes</manageDesktops> <!-- show the manage desktops section in the client-list-(combined-)menu --> </menu>
[ -0.0132406335324049, 0.0013343621976673603, -0.006659256760030985, 0.013553900644183159, -0.01685520075261593, -0.00953940860927105, 0.0091754961758852, 0.016232086345553398, -0.02252030000090599, -0.017981737852096558, -0.02712092362344265, 0.008601449429988861, -0.01246621459722519, -0.0...
[ 0.1335349678993225, 0.34618082642555237, 0.21095708012580872, -0.24228699505329132, 0.05845295637845993, -0.43422138690948486, 0.4872692823410034, -0.09727850556373596, 0.0042829061858356, -0.8726859092712402, -0.09323841333389282, 0.5463608503341675, -0.3784342110157013, 0.143981233239173...
I came across a large body of literature which advocates using Fisher's Information metric as a natural local metric in the space of probability distributions and then integrating over it to define distances and volumes. But are these "integrated" quantities actually useful for anything? I found no theoretical justifications and very few practical applications. One is Guy Lebanon's work where he uses "Fisher's distance" to classify documents and another one is Rodriguez' ABC of Model Selection… where "Fisher's volume" is used for model selection. Apparently, using "information volume" gives "orders of magnitude" improvement over AIC and BIC for model selection, but I haven't seen any follow up on that work. A theoretical justification might be to have a generalization bound which uses this measure of distance or volume and is better than bounds derived from MDL or asymptotic arguments, or a method relying on one of these quantities that's provably better in some reasonably practical situation, are there any results of this kind?
[ 0.002720070304349065, -0.001899453462101519, -0.009635915979743004, 0.00948295183479786, -0.023740654811263084, -0.01656505838036537, 0.008236055262386799, -0.01637374795973301, -0.01292543113231659, -0.030764129012823105, -0.005638535134494305, 0.014230909757316113, 0.006238051690161228, ...
[ 0.15558046102523804, -0.05797597020864487, -0.13536258041858673, 0.46389660239219666, 0.06186942383646965, -0.0546196885406971, -0.48293018341064453, -0.22280482947826385, -0.3593567907810211, -0.4736461639404297, -0.00704224081709981, 0.4690982401371002, 0.0796019434928894, 0.306577026844...
I'm working on this diagram: \documentclass[border=2mm]{standalone} \usepackage{tikz} \begin{document} \begin{tikzpicture} \clip (-0.1,-0.1) rectangle (5,3); \draw[help lines,->] (0,0) -- (4.2,0); \draw[help lines,->] (0,0) -- (0,3);% draw axis lines \draw[gray,dashed] (0,2) -- (4.2,2); % draw asymptote \draw[domain=0.1:4.6,very thick,red,->,samples=400] plot ({\x - 0.4},{1/(-\x) + 2} );% draw plot \draw[help lines,->] (1,0) -- (1,1.2); \draw[help lines,->] (2,0) -- (2,1.5); \draw[help lines,->] (3,0) -- (3,1.65); \draw[help lines,->] (4,0) -- (4,1.7); % scale to fit marginfigure \end{tikzpicture} \end{document} ![Asymptote drawing](http://i.stack.imgur.com/CHqkV.png) It gets the job done, more or less, but I'm wondering whether there's a cleaner or more elegant way to do it. Specifically, I'm wondering: 1. Is there a way to draw the curved line as a ray with an endpoint at (0,0)? The problem with drawing a demihyperbola and then clipping is that I want to label several points on that path (which is what the vertical arrows are for). I can't do that if I have to clip the diagram so closely. 2. Is there a way to describe the height of the vertical lines as relative to the curved plot (i.e., can I tell `tikz`, "Draw a vertical line, with an arrowhead, from (1,0) until it intersects the curved line")?
[ -0.006776082329452038, 0.0036112871021032333, -0.003375224070623517, 0.013906922191381454, -0.01637410745024681, -0.01215580478310585, 0.005748193245381117, 0.004982331767678261, -0.009611018002033234, 0.0077908579260110855, -0.014286205172538757, -0.0036845537833869457, 0.000069891102612018...
[ 0.24313393235206604, -0.1248336061835289, 1.1353719234466553, -0.06214825436472893, 0.1405911147594452, 0.07230590283870697, 0.03999956324696541, -0.1228233128786087, -0.20280061662197113, -0.937099039554596, 0.3208961486816406, 0.3225480318069458, -0.09118226915597916, 0.2034691572189331,...
I have a slightly complex SELECT query which uses a number of JOINs and variables inside it and the query is run frequently throughout the application. The result varies when one column is updated. Eg: When the product quantity is updated only, the result set changes. So my question is, is it a good idea to run a trigger and update a temporary table in the database, with my SELECT queries results, whenever the product quantity is updated? Or to run the complex SELECT query every time? Right now I tested with ~10 rows of data and the query execution time is 0.001 seconds. I want to use the best possible method which won't hinder the performance as the DB grows. Thank you!
[ 0.037578556686639786, 0.021656986325979233, -0.009394942782819271, 0.011597598902881145, -0.02566470019519329, 0.009742585942149162, 0.006963355466723442, -0.0057942308485507965, -0.016701191663742065, -0.030194159597158432, -0.016433611512184143, 0.015772083774209023, -0.008189202286303043,...
[ 0.2911975085735321, 0.1825113296508789, 0.3694884777069092, -0.1266816407442093, -0.10292955487966537, 0.3693799078464508, 0.10457003861665726, -0.4290878474712372, -0.0892144963145256, -0.5300567150115967, 0.49878430366516113, 0.45446932315826416, -0.21935229003429413, 0.33496344089508057...
I have a custom post type which represents a 'race' and there are meta fields for 'distance' (5km,10Mile) and 'type' (track,road,crosscountry). We get the race results as a csv file which contains the positions and times etc of the runners, each runner is linked to a wordpress user. The intention is to process the csv data and insert it into a custom database table. The results are then displayed via a WP_Table_List. My question is how i should best save the csv data so that it can be processed, the two options as i see are 1 - Save the csv data into the post content section. I'd then add a custom action 'Process Content CSV' which reads the content and processes the rows inserting into the database. 2 - As a custom meta field 'csv_results' which is setup to accept a file_upload. Once the file is uploaded and the post save, the content of the file is then processes and inserted into the db. With option one i'd be worried that wordpress will filter out my delimiter char and option 2 seems too complicated. Which approach and what issues would you watch out for?
[ 0.002654273761436343, 0.011848482303321362, 0.005929469130933285, 0.020009761676192284, -0.0018061019945889711, 0.00651366775855422, 0.008915439248085022, 0.012519173324108124, -0.013675960712134838, 0.008080151863396168, 0.005522296763956547, 0.011882700026035309, 0.011665202677249908, 0....
[ 0.58612060546875, 0.07890836149454117, 0.2697924077510834, 0.209157332777977, -0.19357170164585114, 0.17141492664813995, 0.5283973217010498, 0.02486152946949005, 0.00601684395223856, -0.5550007224082947, 0.20734551548957825, 0.11022201925516129, -0.07792264968156815, 0.1112074926495552, ...
How can I omit the date from being displayed when exporting to LaTeX from emacs org-mode? This (within the .org file) works to omit the date when exporting to ascii: #+OPTIONS: timestamp nil But it doesn't take effect when exporting to LaTeX. Nor does this work: #+ATTR_LATEX: :date nil
[ 0.02402418851852417, -0.0015902061713859439, -0.01513613574206829, 0.014311958104372025, 0.006090237759053707, 0.0112209627404809, 0.011609097942709923, 0.006685870699584484, -0.030396681278944016, 0.012533298693597317, -0.008158406242728233, 0.0002266102674184367, 0.005567170213907957, 0....
[ 0.2545199394226074, 0.060650184750556946, 0.3784198462963104, -0.2085699737071991, 0.24199585616588593, 0.08980730921030045, 0.30096468329429626, 0.12321943044662476, -0.3424873948097229, -0.6317555904388428, -0.11097659170627594, 0.2406732141971588, -0.2829574644565582, 0.4345909059047699...
How can I create a beautiful letterhead from the `article` class using say header or footer . I do not want to use `letter` class due to limitations. The letterhead should look outstanding.
[ 0.014230317436158657, 0.02695000171661377, -0.0123565923422575, 0.023647140711545944, -0.021942399442195892, -0.0029040537774562836, 0.017424723133444786, 0.0012747315922752023, -0.022690175101161003, -0.015699051320552826, -0.0203078705817461, 0.026097692549228668, 0.009869055822491646, -...
[ 0.13120535016059875, 0.5282099843025208, 0.2804657518863678, 0.15198996663093567, -0.060505032539367676, -0.14173860847949982, 0.2234748899936676, -0.03014882281422615, 0.04369904100894928, -0.6645143032073975, 0.01554679125547409, 0.06625153869390488, 0.029362116008996964, 0.3508987426757...
Is it safe to use Archlinux's repository within Manjaro Linux installation? If it is, all I need to do is just add some line to /etc/pacman.d/mirrorlist, right?
[ 0.02610870636999607, 0.026135368272662163, -0.004003544338047504, 0.011519394814968109, -0.029586559161543846, -0.008325602859258652, 0.014356298372149467, -0.004705656785517931, -0.022776653990149498, -0.020617857575416565, -0.01275724545121193, 0.021829878911376, -0.024965258315205574, 0...
[ 0.4155329167842865, 0.07781939208507538, 0.17397189140319824, 0.23243071138858795, 0.1285279393196106, -0.2047748565673828, 0.45757681131362915, 0.17884384095668793, -0.07735635340213776, -0.3623320758342743, 0.29664263129234314, 0.45304495096206665, -0.23840324580669403, -0.04442700743675...
I'm not familiar with chinese characters and fonts. But I have to type some characters into a document. `XeLaTeX` may be an option, but I'm looking for a solution with `pdflatex`. Here is my MWE, which I created with the help of other questions and answers: \documentclass{article} \usepackage{CJKutf8} \begin{document} \begin{CJK*}{UTF8}{gbsn} 中華人民共和國 ABC-1234 \begin{verbatim} 中華人民共和國 ABC-1234 \end{verbatim} \end{CJK*} \end{document} I have two questions: 1. Is there a monospace font for chinese characters available for using in `verbatim` environments with same size than latin characters (I'm using MacTeX/TeXLive 2012)? 2. Why are not all 7 chinese symbols in the resulted pdf? ![result of pdflatex](http://i.stack.imgur.com/CylkR.png)
[ 0.001844238955527544, 0.013416202738881111, -0.0008071355987340212, 0.0072568245232105255, 0.003578098490834236, 0.013524133712053299, 0.0074510094709694386, 0.012514293193817139, -0.01247716136276722, -0.009640706703066826, -0.002840633038431406, -0.00029118964448571205, -0.0001828176318667...
[ 0.06658109277486801, 0.3293350636959076, 0.4398086965084076, 0.24560119211673737, -0.3746394217014313, 0.30219677090644836, -0.018367193639278412, -0.6921299695968628, -0.0022252071648836136, -0.33424487709999084, 0.13620324432849884, 0.0882604718208313, 0.06788775324821472, -0.00138186337...
Playing Skyrim, trying to get married to Borgahk the Steel Heart and, upon waiting for the wedding to happen in the Temple of Mara the guests start fighting. I've determined that it might be something to do with Mjoll the Lioness, due to her being the only hostile character on my compass. Everyone in the temple will attack her until she is downed then resume attacking when she stands back up. Is there some way to fix this? Edit: Not Mjoll, thought it was but it's actually a Bandit Chief with very similar model and the same face-paint. Killing the bandit does end the cycle of fighting and waiting but I am continually told to leave the temple, and find myself unable to.
[ 0.020975030958652496, 0.011892415583133698, -0.0015049570938572288, -0.005040341056883335, 0.0014422833919525146, -0.01815921999514103, 0.009549440816044807, 0.0010486599057912827, -0.01642617955803871, 0.008003699593245983, -0.02350754290819168, 0.02044638991355896, -0.013145174831151962, ...
[ -0.18307402729988098, -0.38643768429756165, 0.35450613498687744, 0.12851162254810333, -0.33975526690483093, 0.24647241830825806, 0.34663137793540955, -0.011950142681598663, -0.6158381700515747, -0.03611750155687332, -0.06922971457242966, 0.4095918834209442, -0.20853669941425323, 0.11003508...
I use the `acronym` package with page numbers. Now I would like to have the dots between the acronym and the page number in same style as in the ListofTables: same spacing between the dots and between the last dot and the page number. The line spacing should be the same as well, but the dots have the higher priority. (In addition the alignment among the acronyms does not seem to work well in the mini example, but fine in my document.) \documentclass{article} \usepackage[utf8]{inputenc} \usepackage[printonlyused,withpage]{acronym} \begin{document} \parskip 6pt% space between paragraphs \renewcommand{\baselinestretch}{1.2}\normalsize% leading -- space between two lines \renewcommand{\bflabel}[1]{\normalfont{\normalsize\bf{#1}}\hfill}% font adjustment \section*{List of Abbreviations} \begin{acronym}[MSC] \acro{IMO}{International Maritime Organization} \acro{MSC}{Maritime Safety Committee} \end{acronym} \listoftables \begin{table}\begin{tabular}{c} \ac{MSC}\end{tabular}\caption{a}\end{table} \begin{table}\begin{tabular}{c} \ac{IMO}\end{tabular}\caption{b}\end{table} \end{document}
[ 0.01647067442536354, 0.007724948227405548, -0.02025620825588703, 0.02786904200911522, 0.009895465336740017, 0.018186628818511963, 0.008466836996376514, -0.015076735988259315, -0.013433748856186867, -0.016108490526676178, -0.01937466859817505, -0.0022449041716754436, -0.006140056997537613, ...
[ 0.2522317171096802, 0.1723361760377884, 0.5071657299995422, -0.2463100701570511, -0.1141924187541008, -0.15423235297203064, 0.27911198139190674, -0.19370585680007935, -0.13557754456996918, -1.0654923915863037, -0.027286743745207787, 0.6307832598686218, -0.1970529407262802, -0.3335952758789...
This question may look like a duplicate, but only at first glance. Of course, I would no longer need help in how to code a one-liner that extracts a fixed number of **continuous** lines (e. g. 5 in this example) from a data source, e. g. `top`: $ top -b -n1 | awk 'BEGIN {printf "%23s %7s\n","cpu","mem"} NR==8,NR==12 {printf "%-16s %6s%% %6s%%\n",$12,$9,$10}' This is even a very handy one-liner that will show the processes in the system that take most CPU, with the memory usage being printed in an additional column. So far, so good ... however, it's not _that_ trivial. To get this list, `top` is necessary and may (on low system load) show up _itself_ as process in this list. I'd rather not want that, since these calls are done in **intervals** and would regularly spawn `top` (if only for a short while). It is known that we want to begin at **line 8** (NR==8). However, what if a second `top` in another virtual desktop was forgotten about in a terminal which messes up the list as well? In this case, two `top` processes must be omitted, so the last line to process will be 14. So to improve this output and to get rid of `top` altogether, a counter seems mandatory (perhaps a `for` loop that we exit with a `break`?). Unfortunately my attempts with a for loop and `i = <number>` have been fruitless so far, because it would rather print every line as many times as `i` indicates. I've come up with a rather hackish solution, which works but may be unsuitable for more complex cases: top -b -n1 | grep -v ' \btop\b$' | awk 'BEGIN {printf "%23s %7s\n","cpu","mem"} NR==8,NR==12 {printf "%-16s %6s%% %6s%%\n",$12,$9,$10}' _(Note: This may give unwanted results if the user name in the second column happens to be "top" as well)_ Anyways, could I get a clue how to do that in `awk` please (and get rid of the `grep`)? Thanks in advance.
[ 0.006376959383487701, 0.010557638481259346, -0.01264952216297388, 0.012114901095628738, -0.006899591535329819, 0.0007974761538207531, 0.004678033292293549, -0.008704429492354393, -0.013324812985956669, 0.01818428933620453, 0.003390748519450426, -0.0002929065958596766, 0.003974512219429016, ...
[ 0.14820998907089233, 0.03819257393479347, 0.5113312602043152, 0.1571657359600067, 0.004925603047013283, 0.3101988136768341, 0.15834996104240417, -0.11613734811544418, -0.23031795024871826, -0.37985411286354065, -0.15521463751792908, 0.08186696469783783, -0.1039363220334053, 0.2297097742557...
Because of the standards in my field, for many plots I need to use (or at least end up using) rather narrowly defined colour schemes, and in particular the Matlab "Jet" scheme, which is well described in this SO answer. I've typically done this by having a bit of boilerplate code which I include when necessary, and which creates the object `ColorData["Jet"]`: Unprotect[ColorData]; ColorData["Jet"] = Function[x, Blend[Transpose[{ {0, 0.1, 0.3, 0.5, 0.7, 0.9, 1.0}, {Darker[Blue], Blue, Cyan, Green, Yellow, Red, Darker[Red]} }], x]]; Protect[ColorData] This is convenient because the new colour scheme can be handled using the same syntax as the in-built ones. In version 10, however, the choice of colour schemes has a bunch of additional niceties which I'd like to take advantage of it that's possible. For example, the autocomplete menu for `ColorDat...` gives you instant access to a drop-down menu with images of the schemes' gradients, and which inserts `ColorData["gradient-of-your-choice"]` when you click OK. ![enter image description here](http://i.stack.imgur.com/xkIdz.png) ![enter image description here](http://i.stack.imgur.com/Av0lK.png) Similarly, when you evaluate such `ColorData` objects, they come out with a nice preview of the gradient. ![enter image description here](http://i.stack.imgur.com/VazBL.png) So, my question is: is it possible to take advantage of this structure by 'injecting' additional data into `ColorData`, so that custom colour schemes come up in the choices and behave in the same way as in-built ones?
[ 0.008998217061161995, 0.010804908350110054, -0.010156488046050072, -0.007554314099252224, -0.007861833088099957, -0.005022031720727682, 0.00593878049403429, 0.003232283517718315, -0.009209627285599709, -0.012412323616445065, -0.0040084486827254295, 0.006893741898238659, -0.004684172570705414...
[ 0.26793134212493896, -0.47425490617752075, 0.7763734459877014, -0.1258522868156433, 0.27735626697540283, -0.031926125288009644, -0.09253416955471039, -0.48995232582092285, -0.030817992985248566, -0.6772148013114929, -0.4774576723575592, 0.599048376083374, -0.47971585392951965, 0.0073084319...
I want to root my Galaxy Ace but don't want to install Custom ROM or CWM! Just only root so that, I can run some apps which requires root.
[ 0.02045106142759323, 0.015596941113471985, 0.01292532216757536, 0.013836639933288097, -0.04999645799398422, 0.035827282816171646, 0.01639898680150509, 0.0014304567594081163, -0.02222560904920101, -0.004681373946368694, -0.009646990336477757, 0.015504168346524239, -0.018980542197823524, 0.0...
[ 0.387403666973114, 0.3118816316127777, 0.31935036182403564, 0.0867001935839653, 0.05724753811955452, -0.1330011785030365, 0.5305408835411072, 0.4397321939468384, -0.1964864730834961, -0.808235228061676, 0.04326072707772255, 0.3894631266593933, -0.08677350729703903, 0.1126064583659172, 0....
I need to set some TikZ options that have been written to a file by a script. I thought I could use a helper macro defined with `\edef` and the expandable version of `\input` as described in Why is \input not expandable?, but the compilation of my minimal document fails with Runaway definition? ->\tikzset { every node/.style={ fill=orange!50,draw=black,thick ! File ended while scanning definition of \auxmacro. <inserted text> } l.12 \@@input inputfile.txt Here's the test document \documentclass{article} \usepackage{tikz} \usepackage{filecontents} \begin{filecontents}{inputfile.txt} fill=orange!50,draw=black,thick \end{filecontents} \makeatletter \edef\auxmacro{\noexpand\tikzset{ every node/.style={ \@@input inputfile.txt } }} \auxmacro \makeatother \begin{document} \begin{tikzpicture} \node {Testnode}; \end{tikzpicture} \end{document} How can I apply TikZ options stored in a file?
[ 0.02217322774231434, 0.015596933662891388, 0.0012408606708049774, 0.017622236162424088, -0.006415796000510454, -0.003864530473947525, 0.009138235822319984, 0.006336824968457222, -0.021048085764050484, 0.00037322589196264744, -0.01229312177747488, -0.004187070764601231, 0.006282552611082792, ...
[ 0.11824315041303635, -0.2725338935852051, 0.4949249029159546, -0.30556246638298035, 0.4823760390281677, 0.09472884982824326, 0.1732906550168991, -0.08065908402204514, -0.1321728676557541, -0.9218065142631531, -0.10637837648391724, 0.7598475813865662, -0.3825332820415497, -0.113703414797782...
Yes, I have checked that previous answers to "Ran out of iterations..." questions do not solve my problem. I have fault data on Firefox, 899 faults and 1395 (estimated) censored faults. The censoring all happens on one of half a dozen start days and half a dozen end days (the initial/final release of a version). library(survival) ff_usage=read.csv("http://www.coding-guidelines.com/R_code/ff_usage.csv", as.is=TRUE) f_sur=Surv(ff_usage$start, ff_usage$end, event=ff_usage$event) plot(survfit(f_sur ~ 1)) f_cox=coxph(f_sur ~ total_usage+cluster(fault_id), data=ff_usage) The Kaplan-Meier curve looks about right. `total_usage` is an estimate of the number of Firefox users up until the fault is reported. This is very time dependent and so each fault timeline is broken up into 7 day intervals clustered on `fault_id`; unsplit original. The dependency on `total_usage` (or its log) could be close to 1 (I am hoping for one or the other). I have tried setting `init` and increasing `iter.max`; also `strata(src_id)` and subsetting on `src_id`. Most of the start/end times are estimated and have a regular interval, I have tried adding some randomization, e.g., `runif(n, -3, 3)`. No change. All I ever see is: Warning message: In fitter(X, Y, strats, offset, init, control, weights = weights, : Ran out of iterations and did not converge Suggestions for things to try welcome.
[ -0.012883423827588558, 0.02059231698513031, -0.018477395176887512, 0.0151245491579175, -0.004006681963801384, 0.013912700116634369, 0.007791070267558098, 0.016698993742465973, -0.017200520262122154, 0.008072055876255035, 0.0032043899409472942, 0.01095248106867075, -0.0019091451540589333, 0...
[ 0.2651018798351288, 0.39904356002807617, 0.39315345883369446, 0.0024282592348754406, 0.3330816924571991, 0.18265096843242645, 0.5270002484321594, -0.08833613991737366, -0.37061774730682373, -0.40150898694992065, -0.19265125691890717, 0.4168989658355713, 0.0029457793571054935, 0.17962345480...
I have Ubuntu 10.04 installed as my primary operating system, and I installed PC BSD in a different partition: `/dev/sda4` without installing it's boot loader. I figured out that I need to edit `/etc/grub.d/40_custom` to add an entry for PC-BSD. So far, nothing seems to work, though. EDIT: this sort of works, but doesn't fully boot the OS, it then asks me for the MOUNTROOT partition. menuentry "PC-BSD 8.1" { insmod ufs2 set root=(hd0,4) kfreebsd /boot/kernel/kernel } The selected answer below is correct. If you are dual-booting with Linux I suggest _NOT_ installing the PC-BSD bootloader as the documentation suggests, unless you enjoy pain.
[ -0.001769116148352623, 0.0026242295280098915, -0.018177613615989685, 0.018319841474294662, -0.04608334228396416, -0.00004002801142632961, 0.007986616343259811, 0.012359734624624252, -0.016577009111642838, -0.008510472252964973, -0.0185878686606884, 0.012052093632519245, -0.01754138246178627,...
[ -0.19976972043514252, 0.28458675742149353, 0.5116740465164185, -0.3483637571334839, -0.07642470300197601, -0.11276889592409134, 0.08492951840162277, 0.2830370366573334, -0.17430956661701202, -0.7936913371086121, -0.12466476857662201, 0.7338230013847351, -0.3801553547382355, 0.3077060282230...
In our webapplication we have a list of questions that have to be answered by the user. These questions are served to the user one by one and will be saved once the last question has been answered. The problem we faced was saving all the 'help'-data that goes with this: storing the index of the last question, returning whether or not you're at the last question, returning the answered questions for the overview, etc. Initially we stored this data each into its own session. This worked, but it also meant we had about 5 different session variables for each type of question list and a bunch of casts. I've removed these session variables by creating a few extra fields in the viewmodel and storing the viewModel in its entirety inside a session. This made sure we had our temporary data persisted troughout requests (each question solved meant a new request), removed a great deal of sessions and made the code more readable. Another example of usage: our local user object gets overwritten every request because it's being obtained from a repository/databasecontext that's re- created every request (ninject). This also meant that we couldn't just keep a temporary list in our user that holds the already answered questions their answers, since it'd be emptied every request. Using this approach we can save this list in the session object, write it to the local user at the start of the action method, perform the action (save a new answer) and afterwards obtain this list and write it to the viewmodel. It's a bit of a workaround, but it made sure we could keep this data. I believed this to be a decent solution, but now one of the project members (it's a school project due tomorrow) expressed his doubt about this method and said it was a very dirty solution (no alternative provided though). We're using ASP.NET MVC 4. Have I approached this the right way? How should I have solved it differently?
[ 0.010131263174116611, 0.006803994067013264, -0.00974833033978939, -0.0016344126779586077, 0.005861629731953144, 0.01707947440445423, 0.007074680645018816, 0.023391520604491234, -0.010276059620082378, -0.010339438915252686, -0.011243918910622597, 0.01647878997027874, -0.00023616617545485497, ...
[ 0.23459522426128387, -0.05984432250261307, 0.204599529504776, 0.2615579664707184, -0.18096299469470978, 0.1308101862668991, 0.3086179196834564, -0.15179353952407837, -0.321020245552063, -0.4690694212913513, 0.12697762250900269, 0.4147224724292755, 0.06230892613530159, -0.08808949589729309,...
Most universities I checked have a Quantum Information group, and they usually mention the term "Quantum Computing" somewhere on their website, but it's not very clear how much research is actually done in Quantum Computing. I would like to study at an institution where there are at least 2-3 potential supervisors who perform active research in Quantum Computing and are willing to take new students. (I plan to begin my PhD in the 2015/2016 academic year.)
[ 0.028799427673220634, 0.0027295255567878485, -0.0033861827105283737, 0.01857197843492031, 0.002475462853908539, -0.022157184779644012, 0.008999026380479336, -0.010616040788590908, -0.02832251787185669, -0.020731743425130844, 0.016484571620821953, 0.028029076755046844, 0.0021379783283919096, ...
[ 0.607386589050293, 0.1615700125694275, 0.35250619053840637, 0.08762985467910767, 0.055584654211997986, -0.21168044209480286, 0.22721827030181885, 0.37657302618026733, -0.20025348663330078, -0.3617648482322693, -0.08546513319015503, 0.22234763205051422, 0.5793682336807251, 0.106841169297695...
Is it possible to dynamically get sidebar parameters from within a widget? That is, I am trying to access the `before_widget`/`after_widget`/`before_title`/`after_title`/`name` parameters of the containing sidebar. Suppose we have a sidebar registered like so: register_sidebar( array( 'name' => "Homepage Sidebar", 'id' => 'homepage-sidebar', 'before_widget' => '<div id="%1$s" class="widget-container %2$s">', 'after_widget' => '</div>', 'before_title' => '<h2 class="widget-title">', 'after_title' => '</h2>', ) ); How would I access those values from inside my widget's `widget()` function? Like, how would I do something like this: public function widget( $args, $instance ) { // outputs the content of the widget if ($someCondition) echo $sidebar->before_title . 'My widget title' . $sidebar->before_title; echo 'You are looking at ' . $sidebar->name; } Is this possible?
[ 0.00007377052679657936, 0.0082000233232975, 0.006357445381581783, 0.013331972993910313, -0.008270620368421078, 0.012806539423763752, 0.008315525017678738, 0.020686671137809753, -0.01822502538561821, 0.0007688477635383606, -0.008605614304542542, 0.007485278882086277, -0.005579439923167229, ...
[ 0.003118023509159684, -0.299894243478775, 0.5023847222328186, 0.008565975353121758, 0.2930589020252228, 0.009489607065916061, 0.12044603377580643, -0.19465641677379608, -0.17153310775756836, -0.41113507747650146, 0.057372961193323135, 0.3658015727996826, -0.2745282053947449, 0.291763365268...
This might be a bad idea. The more I think about it the more I come to the realization that I probably shouldn't do it... but I've been trying and failing so I REALLY want to know how to do it, even if it's a bad idea. What I want is for the bashrc file to be sourced every time I run the `clear` command. The reason for this is completely materialistic. I have system information echoed out when I source bashrc and it's cool to me and I'd like that to be at the top every time I clear. I've tried to set up some aliases for `clear` but I keep running into infinite loops. The obvious fix is to change the aliases to something else besides clear so that I can run the clear command in the alias without interfering but I type clear so often that it's kind of ingrained in my brain at this point. I'd like to be able to type clear and make it clear AND source the bashrc file.
[ 0.021257001906633377, 0.015822608023881912, -0.007116821128875017, 0.014109906740486622, 0.009138506837189198, -0.02144300565123558, 0.0030726976692676544, 0.005936206318438053, -0.015116719529032707, -0.005349906161427498, -0.007553522475063801, 0.005573965609073639, -0.012291617691516876, ...
[ 0.5857670307159424, 0.3162705898284912, 0.2059447169303894, 0.0007102207746356726, 0.1523943394422531, -0.42251619696617126, 0.1938200742006302, 0.16805264353752136, -0.18788129091262817, -0.32299181818962097, 0.36342304944992065, 0.3980877101421356, -0.1391618400812149, 0.3613763451576233...