text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
Does anyone know if adaptIntegrate accepts a vectorized integrand? for e.g. alpha<-c(1,2) f <- function(z){ (z[1]+z[2])*alpha } adaptIntegrate(f,lower=c(1, 3), upper=c(2, 4),tol=0.01) ##does not work Is what I want possible with adaptIntegrate? Thanks in advance. | [
-0.005850248038768768,
0.019094638526439667,
-0.0159088596701622,
0.007301829289644957,
-0.002323628170415759,
-0.0015825305599719286,
0.0073151662945747375,
-0.01934034749865532,
-0.019948389381170273,
-0.0001807419175747782,
-0.003888415638357401,
0.01319669745862484,
-0.014619201421737671... | [
0.022899113595485687,
-0.13257813453674316,
0.41430994868278503,
-0.02194896899163723,
0.3517012298107147,
-0.0508715957403183,
0.1988486349582672,
-0.34968820214271545,
0.2963508069515228,
-0.9085777997970581,
-0.06684642285108566,
0.8516625761985779,
-0.5438959002494812,
-0.2901170253753... |
In the previous Lego games there's always been the option to turn hints on or off. They're on by default in Marvel Superheros game, and I can't find any option to turn them off. It's also no longer just text popping up like in the previous games, it's Agent Colson telling me to do things, and it's getting old fast. Did they really remove the option to turn hints on/off? Is there a cheat code, difficulty setting, or hack I can use to remove the hints? How do I shut Agent Colson up? | [
0.024061407893896103,
-0.014497064985334873,
-0.013127045705914497,
-0.0010583889670670033,
-0.002022617496550083,
0.014196019619703293,
0.007600840181112289,
-0.015644706785678864,
-0.013324085623025894,
0.02297935262322426,
-0.009269798174500465,
0.006477655842900276,
0.019254565238952637,... | [
0.2045467495918274,
0.11924467235803604,
0.28436610102653503,
0.31746694445610046,
0.02175380475819111,
-0.19736304879188538,
0.194984570145607,
-0.021879417821764946,
-0.4209859073162079,
-0.21610820293426514,
-0.2823077142238617,
0.3276068866252899,
-0.27787530422210693,
0.08440048247575... |
If I am selecting 232 people from a pool of 363 people without replacement what is the probability of 2 of a list of 12 specific people being in that selection? This is a random draw for an ultra race where there were 363 entrants for 232 spots. There is an argument about whether the selection was biased against a certain group of 12 people. My initial attempt at calculating this was that there was 232 choose 363 possible selections. The number of combinations of any one person from the list of twelve is 1 choose 12 + 2 choose 12 + ... + 11 choose 12 + 12 choose 12. Thus 1 choose 12 + 2 choose 12 .... / 232 choose 363. Which ends up being a very low number which is clearly too low. How do I calculate this? | [
0.01071326993405819,
0.015675384551286697,
-0.012984195724129677,
0.013274330645799637,
-0.007540566846728325,
-0.023218590766191483,
0.009475715458393097,
-0.02217477187514305,
-0.01774095557630062,
-0.00614642258733511,
-0.004806037992238998,
0.003209835384041071,
0.00975003931671381,
0.... | [
0.7859625816345215,
-0.28281518816947937,
-0.2300339937210083,
0.2653301954269409,
-0.32876166701316833,
0.5790925025939941,
0.26991912722587585,
-0.6159911751747131,
-0.441831111907959,
-0.5728959441184998,
0.4306250214576721,
0.4536488354206085,
-0.48636817932128906,
-0.07571017742156982... |
I have a shapefile containing hundreds of polygons. I need to create a kml of each of the polygon. Is there any way that I may not have to deal each polygon individually and my polygon file may give me individual kml of each polygon. I'm using ArcGIS 9.3 or QGIS. | [
0.011260067112743855,
-0.0010704502929002047,
-0.008491873741149902,
0.03192904591560364,
-0.011149534955620766,
0.008669829927384853,
0.008659282699227333,
0.03288649022579193,
-0.026348454877734184,
-0.006343682762235403,
0.010970384813845158,
0.011732935905456543,
-0.012275001034140587,
... | [
0.25379228591918945,
0.019938556477427483,
0.1588221788406372,
-0.02297373302280903,
-0.1938781589269638,
0.17192015051841736,
0.15749187767505646,
0.028490375727415085,
-0.12966208159923553,
-1.1212056875228882,
0.19541041553020477,
0.31044596433639526,
-0.07126288115978241,
0.03690449148... |
How do I convert a .NET STAThread into .NET/Link Mathematica code? Alternatively what would be the easiest way to compile and run the program below from within mathemaitca? using System; using Wolfram.NETLink; using System.IO; using System.Drawing; using System.Windows.Forms; public class File{ [STAThread] public static void Main(String[] args) { MathKernel k = new MathKernel(); k.Compute("ExportString[Graphics[Circle[]],{\"Base64\",\"PNG\"}, Background -> None]"); string input = k.Result.ToString(); byte[] bytes = Convert.FromBase64String(input); Stream stream = new MemoryStream(bytes); IDataObject dataObject = new DataObject(); dataObject.SetData("PNG", stream); Clipboard.SetDataObject(dataObject, true); } } I am compiling it using the following code, in case anyone wants to try it out. csc /target:winexe /reference:Wolfram.NETLink.dll File.cs | [
0.006269104778766632,
-0.00033009113394655287,
-0.0035107177682220936,
0.01264871470630169,
0.03539220616221428,
0.010729756206274033,
0.009021366946399212,
-0.009596740826964378,
-0.018656039610505104,
0.011471633799374104,
0.00198066676966846,
0.011128679849207401,
-0.009661996737122536,
... | [
0.06868289411067963,
0.09280556440353394,
0.47932079434394836,
0.09291119873523712,
-0.17559725046157837,
0.1462962031364441,
0.29883137345314026,
-0.3834148347377777,
0.07916902005672455,
-0.7887209057807922,
-0.04825466498732567,
0.2445725053548813,
-0.20784278213977814,
0.16542564332485... |
Lets pretend that a very large company (revenue numbers with more than 8 figures) is looking to do a refresh on a software system, particularly the dashboard used by employees. This system was originally put together in the early 1990's to handle inventory tracking and storage across a variety of facilities (10+). Since this large company is now in the process of implementing some of these inventory processes with SAP they are in need of a major refresh. The existing system: * Microsoft Access project performs dashboard duties * Unique shipping/receiving configurations at different facilities require unique forms and queries within the Access project * Uses 3rd party libraries referenced by Access to directly interface with at control system (read: motors, conveyors, and counters) * Individual SQL Server 2000 instances (some traces of pre-update SQL Server 6.0 documents) at each facility The Issue: * This system started as a home brewed inventory tracking scheme with a single internal sponsor who is still in charge of the technical direction. The original sponsor prescribing the desired deliverables that are being called for in the current RFP. * The RFP describes a system based around a single Access project. * Any suggestion that Access is ill suited for a project of this scope are shot down under the reasoning that "it works for the scope now". Are there any case studies, notices, or statements that can be used to disuade this potential customer from repeating their mistake? Does Microsoft make any statements directly about when it is _highly_ recommended to ditch Access? **EDIT:** To answer some of the comments below, the system is getting a rewrite no matter what due to the need to integrate a even greater push towards the deployment of an ERP solution. Problems with the current solution involve additional maintenance of ODBC connections, Office deployments, deploying an Access file to hundreds of workstations, and about 15 years of someone who doesn't know how to program generating enormous amounts of technical debt. | [
0.00663767522200942,
0.0012437154073268175,
-0.0031284799333661795,
0.003560400800779462,
0.008941233158111572,
0.005299877841025591,
0.007738360203802586,
0.008532430045306683,
-0.009989047423005104,
-0.00266292504966259,
-0.022674962878227234,
0.011771176010370255,
0.0053282612934708595,
... | [
0.44007736444473267,
0.11754166334867477,
0.3704942762851715,
0.44413062930107117,
0.28869307041168213,
-0.07437071204185486,
0.09575977176427841,
0.24794578552246094,
-0.457324355840683,
-0.198024719953537,
0.2525040805339813,
0.6650447845458984,
-0.05340304598212242,
-0.03123885951936245... |
I'm designing a flexible Wizard system that presents a number of screens to complete a task. Some screens may need to be skipped based on answers to prompts on one or more previous screens. The conditions to skip a given screen need to be editable by a non-technical user via a UI. Multiple conditions need only be combined with `and`. I have an initial design in mind, but it feels inelegant. I wonder if there's a better way to approach this class of problem. **Initial Design** **UI**  where The first column allows the user to select a question from a previous screen. The second column allows the user to select an operator applicable to the type of question asked. The third column allows the user to enter one or more values depending on the selected operator. **Object Model** public enum Operations { ... } public class Condition { int QuestionId { get; set; } Operations Operation { get; set; } List<object> Parameters { get; private set; } } List<Condition> pageSkipConditions; **Controller Logic** bool allConditionsTrue = pageSkipConditions.Count > 0; foreach (Condition c in pageSkipConditions) { allConditionsTrue &= Evaluate(previousAnswers, c); } // ... private bool Evaluate(List<Answers> previousAnswers, Condition c) { switch (c.Operation) { case Operations.StartsWith: // logic for this operation // etc. } } | [
0.0039271460846066475,
0.018555065616965294,
-0.008304436691105366,
0.015623538754880428,
-0.009198964573442936,
0.006894449703395367,
0.0067633530125021935,
0.01397718396037817,
-0.013030655682086945,
0.022882791236042976,
-0.024247977882623672,
0.008930529467761517,
-0.006889326497912407,
... | [
0.621414840221405,
0.024569282308220863,
0.29324987530708313,
0.0813063308596611,
0.06974931061267853,
0.0003172120777890086,
0.2452968955039978,
-0.19118642807006836,
-0.506646454334259,
-0.8660410642623901,
0.0663411021232605,
0.2873878479003906,
-0.2231195718050003,
-0.1528194099664688,... |
I am trying to install ODBC driver for Debian according to those instructions: https://blog.afoolishmanifesto.com/posts/install-and-configure-the-ms-odbc- driver-on-debian/ However, when I type `sqlcmd -S localhost`, I get the error `libcrypto.so.10: cannot open shared object file: No such file or directory`. What to do? I tried 1. $ cd /usr/lib $ sudo ln -s libssl.so.0.9.8 libssl.so.10 $ sudo ln -slibcrypto.so.0.9.8 libcrypto.so.10 2. Add /usr/local/lib64 to the /etc/ld.so.conf.d/doubango.conf file 3. sudo apt-get update sudo apt-get install libssl1.0.0 libssl-dev cd /lib/x86_64-linux-gnu sudo ln -s libssl.so.1.0.0 libssl.so.10 sudo ln -s libcrypto.so.1.0.0 libcrypto.so.10 4. sudo apt-get install libssl0.9.8:i386 but none of those helped. | [
-0.0038957700598984957,
-0.006587161682546139,
-0.019417740404605865,
0.008536934852600098,
-0.008957618847489357,
0.003620233852416277,
0.01076516043394804,
0.01350120734423399,
-0.017299680039286613,
-0.04920507222414017,
-0.012374194338917732,
0.0034330575726926327,
-0.008726413361728191,... | [
0.22626081109046936,
0.3749076724052429,
0.23347122967243195,
-0.33484306931495667,
-0.14095406234264374,
-0.09917540848255157,
0.4498724341392517,
0.059176553040742874,
0.07315456867218018,
-0.7795075178146362,
0.06779643893241882,
0.7711941599845886,
-0.5108910799026489,
0.24781584739685... |
Theres a way to convert links like: http://www.localhost.lh/?attachment_id=41 into: http://www.localhost.lh/attachment/id or: http://www.localhost.lh/author/attachment/id | [
-0.009033916518092155,
-0.0019987556152045727,
-0.013355743139982224,
0.027283649891614914,
0.00132221810054034,
0.032687168568372726,
0.01422637328505516,
-0.0019360959995537996,
-0.02486535534262657,
-0.008832170628011227,
-0.009801166132092476,
0.002978494158014655,
0.00572353508323431,
... | [
-0.0674232766032219,
0.18862953782081604,
0.603999137878418,
0.07006924599409103,
-0.144798144698143,
0.22548808157444,
-0.14364005625247955,
-0.05120709165930748,
-0.35171157121658325,
-0.7367259860038757,
-0.17167426645755768,
0.16532562673091888,
-0.06723088771104813,
0.2924140095710754... |
I was reading this book "the linux command line" and in the introduction it states that linux is internet backbone starting from servers to router infrastructure. That got me thinking to what extent this would be true. Yes I do have dd-wrt installed on my home router. But what about stock firmware of my belkin router? Is it linux based? I saw a list of distributions for routers: http://en.wikipedia.org/wiki/List_of_router_and_firewall_distributions Incredibly long one! I know cisco develops IOS, and some of their low end routers are linux, but what about IOS? is it unix derivative? or written from scratch? | [
-0.005692847538739443,
0.0009888365166261792,
-0.007108853664249182,
0.00201963959261775,
-0.01557239145040512,
-0.012404819950461388,
0.007642299402505159,
0.007364691235125065,
-0.014595160260796547,
-0.012462083250284195,
-0.003072830382734537,
0.0059972903691232204,
0.010143804363906384,... | [
0.6048977375030518,
0.23469680547714233,
0.27176547050476074,
-0.07213374972343445,
-0.041081931442022324,
-0.06724435091018677,
-0.17510727047920227,
0.49624142050743103,
-0.193998783826828,
-0.5789957046508789,
0.004416335839778185,
0.4470902681350708,
-0.3219384253025055,
0.196952775120... |
I have a necromancer that does condition damage. What is the best armor for this type of necro? It seems that the best armor would be the Grenth Karma Exotic armor that I can get from the Temple of Grenth in the Cursed Shore, right? That armor seems to increase condition damage with every piece you wear. Is there better one? Thanks. EDIT: Thought I would add these guides... http://www.guildwars2guru.com/topic/80743-massive-guides-for-condition- necromancer-pve-wvw-fractals/ http://lopezirl.com/2012/12/19/a-condition-necromancers-guide-to-pve/ http://lopezirl.com/2012/12/30/a-condition-necromancers-guide-to-world-vs- world/ | [
0.005981870926916599,
-0.0011021399404853582,
-0.00892016850411892,
0.007914522662758827,
-0.0038843583315610886,
-0.0021302783861756325,
0.009084569290280342,
-0.02072443999350071,
-0.018396345898509026,
0.017025373876094818,
-0.004545782692730427,
0.01664518564939499,
-0.011214278638362885... | [
0.5278616547584534,
0.06109689176082611,
0.0637444481253624,
0.29204103350639343,
-0.27469074726104736,
0.28073832392692566,
0.48241889476776123,
-0.0853048712015152,
-0.16695739328861237,
-0.5901190042495728,
0.01023793127387762,
0.7751845121383667,
0.19169475138187408,
0.0640693753957748... |
It's told in Landau - Classical Mechanics, that in the Hamiltonian method, generalized coordinates $q_j$ and generalized momenta $p_j$ are independent variables of a mechanical system. Anyway, in the case of Lagrangian method only generalized coordinates $q_j$ are independent. In this case generalized velocities are not independent, as they are the derivatives of coordinates. So, as I understood, in the first method, there are twice more independent variables, than in the second. This fact is used during the variation of action and finding the equations of motion. My question is, can the number of independent variables of the same system be different in these cases? Besides that, how can the momenta be independent from coordinates, if we have this equation $$p=\frac{\partial L}{\partial \dot{q}}$$ Thank you very much! I hope that my question is clear. | [
-0.004719163756817579,
0.02313563972711563,
-0.005089107900857925,
0.008608168922364712,
0.014401989988982677,
-0.03132455050945282,
0.010774500668048859,
0.004719930700957775,
-0.015273351222276688,
-0.005082213785499334,
-0.021018698811531067,
0.022907372564077377,
-0.011298410594463348,
... | [
-0.27471885085105896,
-0.1893167942762375,
0.4579693377017975,
0.2263176143169403,
-0.30033552646636963,
0.3155948519706726,
-0.3267499506473541,
-0.5344662070274353,
-0.27148836851119995,
-0.551226019859314,
0.32832035422325134,
0.453133225440979,
-0.3133666217327118,
0.5002169609069824,
... |
I have been reading on here and Google results for a few hours now including this question and the links therein and many others. Including a URI building styleguide from the w3c and others. I have settled on a format, I understand about apache, redirects, file extensions, and SEO. I am pretty sure (and it seems to be confirmed in the link to the Google Webmasters in the first question) that I understand but it just seems wrong... mysite.com/directory/specific-page/ is OK for a file, right? (with trailing slash "/") It doesn't necessarily mean that it is pointing to the index or default file in the /specific-page/ folder, right? It is significant mainly because while I am using blogging software for my blog, I am hand-coding (to retain more control) the rest of the site. It is entirely plausible to me that WordPress would, rather than creating pages/files would actually create directories with index/default pages in them. Up until now, I always thought that the trailing slash pointed to a directory's index page but that appears not to always be the case, is this correct? Sorry, I feel like this topic has been well-discussed, but that is part of what is causing me problems. I should note, I was leaning towards omitting the trailing slash from all pages, CMS generated or not, until I found this article from 2008 about the lack of a slash causing problems with Pingbacks in WordPress. | [
-0.01348789967596531,
0.011693276464939117,
0.004327448084950447,
0.014721965417265892,
0.021186504513025284,
0.016635052859783173,
0.006443051155656576,
0.006557101849466562,
-0.019694101065397263,
-0.010813342407345772,
0.0015198036562651396,
0.010436591692268848,
0.007622012868523598,
0... | [
0.5184760689735413,
0.3311806619167328,
0.5453441143035889,
0.04597153514623642,
-0.02779017761349678,
-0.16795246303081512,
0.13711972534656525,
0.12023141980171204,
-0.2600637972354889,
-0.441375732421875,
-0.11761820316314697,
0.24829956889152527,
0.19397258758544922,
0.1476004123687744... |
What do you do when you have spent considerable effort finessing your resume in TeX and a recruiter asks you for your resume in MS-Word? Do you: 1. Spend the time to produce something that looks half as good as the TeXed result, 2. Ignore the openings advertised by that recruiter, or 3. Somehow convert the resume to a draft in Word that you then edit? If you take the third approach, please share what you do. I've used Word over the years when someone had been passing a form that needed filling, but I have yet to learn the actual basics, hence my question. _Let me comment here on the answers to benefit from the ability to format_ **Solution: Use TeX4ht's htlatex** A resume is likely to use either tabbing or tabular environments extensively. If you run htlatex on the file: \documentclass{article} \begin{document} \begin{tabbing} Job A {\centering Years A} \` Company A \\ Job B {\centering Years B} \` Company B \\ \end{tabbing} \begin{tabular}{lcr} Job 1 & Years 1 & Company 1 \\ Job 2 & Years 2 & Company 2 \\ \end{tabular} \end{document} You will find that `tabular` is handled correctly, but `tabbing` is not. **Solution: Use online conversion tools** One did indeed produce a decent output, but it would be nice to know that the web site is not run by a marketer, a spammer, or worse. Sites that provide a program to download to one's own computer reduce somewhat this worry. | [
0.021707508713006973,
0.005628584884107113,
-0.004531248472630978,
0.013868123292922974,
-0.011001674458384514,
0.0050652301870286465,
0.008684017695486546,
0.012579910457134247,
-0.0171702541410923,
-0.0017036376520991325,
-0.0042994050309062,
0.00969117134809494,
0.009110982529819012,
0.... | [
0.6612919569015503,
-0.13161858916282654,
-0.11448151618242264,
0.16422103345394135,
0.1925804764032364,
0.009386816062033176,
0.14542707800865173,
0.0007929436978884041,
0.23232786357402802,
-0.6114398837089539,
0.17607846856117249,
0.8145061731338501,
0.05533462017774582,
-0.355963885784... |
I saw the word ‘secondhand’ come after ‘things’ in the lead copy of July 17 Time magazine’s article, titled “10 Things You Should Be Buying Used”, as follows. > Buying things secondhand can save a few bucks and help keep junk out of our > landfills. Though I think ‘secondhand’ is used adverbially, and modifies ‘buying’ here, “Buying things secondhand can save a few bucks’ was confusing to me at the first because I took ‘secondhand’ for a post-position to ‘things.’ Although it’s essentially a matter of taste, isn’t it more straightforward and plain to say ‘Buy secondhand things,’ ‘I bought a secondhand book’, ‘I bought a second hand car at a used-car shop,’ ‘I got firsthand (secondhand) news from my colleague,’ rather than saying ‘Buy things secondhand,’ ‘I bought a book secondhand,’ ‘I bought a car secondhand at a used-car shop,’ and ‘I got the news firsthand (secondhand) from my colleague.’? | [
0.0019193080952391028,
0.01104656234383583,
0.0008417776552960277,
0.006408453918993473,
-0.0038207757752388716,
-0.01739559881389141,
0.011070264503359795,
0.031156329438090324,
-0.01743469014763832,
0.0018200250342488289,
-0.0133493822067976,
0.008576318621635437,
0.01722520962357521,
-0... | [
0.5659734010696411,
-0.03279510885477066,
-0.009095694869756699,
-0.04037413373589516,
0.019579190760850906,
0.09065492451190948,
0.40193691849708557,
0.30054032802581787,
-0.2196643352508545,
-0.5193339586257935,
0.407606303691864,
0.6215912699699402,
-0.07207414507865906,
0.2044258415699... |
On my computer (under Arch Linux), GDM is spawning on tty1 while I would like it to spawn on tty7. Is there any way to do that? | [
-0.017541704699397087,
0.023038743063807487,
-0.049234189093112946,
0.0026857643388211727,
0.006170244421809912,
-0.022052966058254242,
0.014942661859095097,
0.025594430044293404,
-0.020712517201900482,
-0.053306229412555695,
-0.026409970596432686,
0.003977624233812094,
-0.030196338891983032... | [
0.5232567191123962,
0.0722663626074791,
0.44834184646606445,
0.10789934545755386,
-0.2438303679227829,
-0.052982620894908905,
0.05571601912379265,
0.4434085786342621,
-0.3772816061973572,
-0.6819345951080322,
0.28176623582839966,
0.33013102412223816,
-0.39353466033935547,
0.334319561719894... |
The title speaks for itself - how do I choose a background color for table cells in LyX? | [
0.07452654093503952,
0.010042067617177963,
-0.027804125100374222,
-0.018969018012285233,
-0.054547227919101715,
-0.0010856931330636144,
0.021129431203007698,
0.05789259076118469,
-0.03177240490913391,
0.004392658360302448,
-0.03710123151540756,
-0.005296820309013128,
0.008633331395685673,
... | [
0.24447152018547058,
0.06845341622829437,
0.05524296686053276,
0.18558359146118164,
0.035885039716959,
0.2181205004453659,
-0.6339498162269592,
0.053394969552755356,
-0.2436636984348297,
-0.06808274239301682,
-0.028070557862520218,
-0.11053311824798584,
0.10285160690546036,
0.1161409988999... |
Sorry for the strange title, I could not think of a better one. I am working on a project, and basically what I need is a redstone torch that is one minute on, and the next minute off, and then a minute on again, and so one. Ideally I want to have a switch that turns the whole thing off. What would be the best way to go about this. Resources are no problem. | [
0.037643980234861374,
0.029273204505443573,
-0.014823829755187035,
0.013379301875829697,
-0.05011116340756416,
-0.01026885025203228,
0.007177312858402729,
0.0031100022606551647,
-0.018600285053253174,
0.003393803955987096,
-0.015132752247154713,
0.0009250935399904847,
0.014509476721286774,
... | [
0.5994068384170532,
0.058624267578125,
0.15637478232383728,
0.00645931763574481,
-0.06507702171802521,
-0.41806069016456604,
-0.039085350930690765,
0.18762341141700745,
0.15211221575737,
-0.4532257616519928,
0.33366674184799194,
0.5032212138175964,
0.07978536933660507,
0.44980695843696594,... |
I would like to make my text fit a certain size "box" at a specific location. The two problems I am encountering are as follows: This long string should wrap when it hits the 4.5 inch limit but it does not: \documentclass[landscape]{article} \usepackage[top=1.5in, bottom=1.125in, left=.25in, right=6.25in,textwidth=4.5in, textheight=5.875in]{geometry} \begin{document} ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd \end{document} And this one just remains constant but I would like to force it to fit the entire 4.5in x 5.875in box, as in enlarge the text to make it fit the box. \documentclass[landscape]{article} \usepackage[top=1.5in, bottom=1.125in, left=.25in, right=6.25in,textwidth=4.5in, textheight=5.875in]{geometry} \begin{document} {\Huge This should fit in a 4.5in x 5.875in box but that is close to the left edge of paper, but it is not conforming to the box size desired.... \end{document} Thanks for the help. | [
-0.004458380863070488,
0.004763823933899403,
-0.0018120750319212675,
0.00798980426043272,
0.031544312834739685,
0.007918440736830235,
0.005711079575121403,
0.013231049291789532,
-0.012374678626656532,
-0.015404779464006424,
-0.01026399340480566,
-0.0011654513655230403,
0.008068665862083435,
... | [
0.1866784542798996,
-0.12268142402172089,
0.6768053770065308,
-0.2661949694156647,
0.0783856213092804,
-0.043961867690086365,
0.4821259081363678,
-0.5566070675849915,
-0.038554076105356216,
-0.5702099800109863,
0.1650160700082779,
0.3598777949810028,
-0.14507706463336945,
-0.10881412774324... |
Recently, I changed carriers from Verizon to T-Mobile. However, when I inserted my T-Mobile SIM card into my phone and rebooted, the status is always "Searching for Service" (not just for data, NOTHING works). I've looked on Android forums and they all have instructions for turning the SCH-I535 into a world phone, but that's not what I want. I just want to switch to T-Mobile (do I need to have a "world phone" to do this?) I've installed APN Manager and tried to change APNs, but entering the details found here The phone is rooted and not on a stock TouchWiz ROM (I'm using Paranoid Android, a JB ROM) Why won't my phone recognize the T-Mobile SIM? | [
0.0017837485065683722,
0.0013165463460609317,
-0.003599591553211212,
0.004719194490462542,
-0.007239645346999168,
0.027107585221529007,
0.0063616372644901276,
0.01177644357085228,
-0.012551851570606232,
-0.012179160490632057,
-0.01981041580438614,
0.01505914144217968,
-0.014852987602353096,
... | [
0.2352372258901596,
-0.1616942584514618,
0.45242545008659363,
0.005521655082702637,
0.09364961087703705,
0.3494918644428253,
0.20781326293945312,
0.7049514055252075,
-0.4431818127632141,
-0.8559896945953369,
0.2032250016927719,
0.5413368940353394,
-0.5003696084022522,
0.21286803483963013,
... |
I am creating a template for a website. The example is at Framework Login Page The main CSS sheet is at: master.css I am trying to center the main parent div. I am using #body { width: 100%; background: url('pathtoimage.png'); } #inner_body{ width: 800px; margin: auto; } <body> <div id="body"> <div id="inner_body"></div> </div> </body> What could the issue be? | [
-0.015210598707199097,
0.00017407769337296486,
-0.013284564018249512,
0.015549490228295326,
-0.01217819843441248,
-0.00014024320989847183,
0.006891509518027306,
0.010465404018759727,
-0.011888924986124039,
-0.009666608646512032,
-0.009025810286402702,
-0.0025440920144319534,
0.01229120045900... | [
0.6052685379981995,
-0.018496327102184296,
0.317905068397522,
-0.13281309604644775,
-0.053398504853248596,
0.03042612597346306,
0.2595692276954651,
-0.34871843457221985,
0.06685923039913177,
-0.9947599768638611,
0.11770133674144745,
0.16201767325401306,
-0.13786663115024567,
0.129084840416... |
Being rather a theoretician than an experimental physicist, I have a question to the community: Is it experimentally possible to mode-lock a laser (fixed phase relationships between the modes of the laser's resonant cavity) in a way that the longitudinal modes of the cavity would be exclusively from a discrete set of frequency $\\{p^m\\}$, where $p$ a prime number and $m$ a positive integer? If yes, how? If not, why? For instance: $\\{2^m\\}$ or $\\{2^m,3^m\\}$ or ... Thanks | [
0.01698504574596882,
0.014772148802876472,
-0.0013245216105133295,
0.00473704282194376,
0.016162356361746788,
-0.02568986266851425,
0.006349323317408562,
-0.006199542433023453,
-0.011263234540820122,
-0.0033581051975488663,
-0.005308652296662331,
0.009906044229865074,
-0.010375439189374447,
... | [
0.42267683148384094,
0.07682902365922928,
0.4799546003341675,
-0.19528335332870483,
-0.037656910717487335,
-0.16586820781230927,
-0.0000880546067492105,
-0.478860467672348,
-0.05933443084359169,
0.026576077565550804,
-0.3385166525840759,
0.7395678758621216,
-0.35971832275390625,
0.17738725... |
I am a native speaker of American English, and I have only ever heard this usage of the word _revert_ from one person. This person is not a native English speaker (he is from India), so he may just be mistaken, but I'm curious if anyone else has seen/heard this usage. He will write an email, bringing up a point for discussion. He will explain the issue, and then end the paragraph with something like _Please do analyze and revert on the status._ The best I can tell, he is asking for a response, and not asking for the something to be undone, or changed back to the way it was before (which is the meaning that I associate with the word _revert_ ). Is _revert_ used with different meanings outside the US? | [
0.013942250050604343,
0.005674166139215231,
-0.0062424540519714355,
0.015131929889321327,
-0.0018779891543090343,
-0.008905370719730854,
0.005099319387227297,
-0.004647205583751202,
-0.012123774737119675,
-0.01206091046333313,
0.006946180947124958,
0.010045148432254791,
0.0006546278018504381... | [
0.48127177357673645,
0.5323067307472229,
-0.020920835435390472,
-0.01576964370906353,
0.05409105494618416,
-0.14314453303813934,
0.4496358335018158,
0.014574657194316387,
-0.14594095945358276,
-0.464133083820343,
-0.29543545842170715,
0.30382442474365234,
-0.03765183687210083,
0.3144735097... |
I'm trying to make a class file to help me draft bylaws for clubs, but I'm running into a problem with `\includegraphics`. I want to be able to define a club logo using a `\logo` command (see class file), with the option of passing options to `\includegraphics`, but it doesn't want to work. Class file: \ProvidesClass{bylaws_min}[2012/10/27 version 0.01 "alpha" Bylaws] \NeedsTeXFormat{LaTeX2e} \RequirePackage[margin=1in]{geometry}% Sets page margins \RequirePackage{graphicx}% Allows adding images to documents \newcommand{\logo}[2][\@empty]{\def\@logoopts{#1}\def\@logoimage{#2}} \renewcommand{\maketitle}{% \begin{center} \ifx\@empty\@logoopts\includegraphics{\@logoimage}\else\includegraphics[\@logoopts]{\@logoimage}\fi \end{center}% } Document file: \documentclass{bylaws_min} \logo[scale=2]{imagefile.jpg} \begin{document} \maketitle \end{document} I've tried both `\logo[scale=2]{imagefile.jpg}` (or some other option other than `scale=2`), or just `\logo{imagefile.jpg}`, and I get variations on this error: Package keyval Error: scale=2 undefined. See the keyval package documentation for explanation. Type H <return> for immediate help. ... 1.7 \maketitle I've played around with this for a while, and it seems that there are two problems. First, `\ifx` never sees `\@logoopts` as `\@empty` (in other words, the comparison always fails, and the `\else` statement always happens). Second, if I explicitly create an image like `\includegrahpics[scale=2]{imagefile.jpg}` in `\maketitle`, everything works, but if I try to pass it the options with `\@logoopts` it fails. But passing it the image file name with `\@logoimage` always works. | [
0.0015640859492123127,
0.004786505363881588,
-0.006768238265067339,
0.021494099870324135,
-0.019289720803499222,
-0.0027715866453945637,
0.006756923161447048,
0.00399060919880867,
-0.02297266758978367,
0.009730447083711624,
-0.011921552941203117,
-0.0022336626425385475,
0.008727503009140491,... | [
0.7028225064277649,
0.04199037700891495,
0.21795719861984253,
-0.28679636120796204,
0.05979619547724724,
-0.3547576069831848,
0.2862367630004883,
-0.24737998843193054,
-0.05301006883382797,
-0.5033659338951111,
0.1974543333053589,
0.42788165807724,
-0.4759686589241028,
0.07073535770177841,... |
How can I differentiate a function with respect to several variables and evaluate it at the same time ? I want to specify also the variable index that I want to differentiate and the number of times I do it for each one. | [
0.02773553691804409,
0.013594872318208218,
-0.037321221083402634,
0.0027820346876978874,
-0.03252698481082916,
-0.00003165704038110562,
0.014113223180174828,
0.06422790884971619,
-0.045410897582769394,
-0.03770587965846062,
0.015292869880795479,
0.015041965991258621,
-0.037733886390924454,
... | [
0.15863652527332306,
0.11053062975406647,
-0.08474496006965637,
-0.044894296675920486,
-0.12660519778728485,
0.3631324768066406,
-0.15704677999019623,
-0.3533461391925812,
0.1979089230298996,
-0.6526986956596375,
0.23821869492530823,
0.5223777294158936,
0.0020386395044624805,
0.31938287615... |
I was wondering, while putting a log in my fireplace, how much energy the piece of wood would give. The most famous formula poped into my head: $E=m \cdot c ^ 2$! Is this formula applicable to a burning object or is in only applicable to a nuclear reaction? | [
0.006835281848907471,
0.0059907990507781506,
-0.02162315510213375,
-0.006533886771649122,
-0.006466556340456009,
-0.0027279837522655725,
0.011066226288676262,
-0.02068370394408703,
-0.023239750415086746,
-0.04889601841568947,
-0.008600158616900444,
-0.0033419537357985973,
-0.0022438848391175... | [
0.6925304532051086,
0.4581969678401947,
0.31125912070274353,
0.09954256564378738,
0.022481771185994148,
-0.2733931541442871,
0.08935026824474335,
-0.22002480924129486,
-0.2965591847896576,
0.011462642811238766,
-0.2594456374645233,
0.2124629020690918,
-0.30116239190101624,
0.36597293615341... |
Does Google Analytics only count a Unique Visitor if he/she lands on the index (home) page? What if the UV makes his way to a side-page of your site but never visits the home page (index page)? Would he still count as a unique visitor under Google analytics? | [
-0.01761176437139511,
0.015995396301150322,
0.008181408047676086,
0.03313206881284714,
-0.01073808129876852,
0.009694436565041542,
0.013908888213336468,
-0.036917224526405334,
-0.03537890687584877,
-0.021332362666726112,
-0.018466120585799217,
0.01016524713486433,
0.0060814423486590385,
0.... | [
0.19683067500591278,
0.07788015156984329,
0.8232576251029968,
0.05557063966989517,
0.0016341038281098008,
-0.023318061605095863,
-0.059806402772665024,
-0.05811276659369469,
-0.43197986483573914,
-0.22618532180786133,
0.00795658491551876,
0.27767616510391235,
-0.2985260784626007,
0.1490117... |
We have data from random polling within a group of people, sampled over a long period of time. As the pool is relatively small and anonymity is crucial there is a high likelihood that we have duplicate samples (though we can't tell for certain because we'd expect responses to vary over time). Does this somehow invalidate the data, or mean that certain operations on it will not be meaningful? Or is it OK to proceed as normal so long as I state this likelihood? * * * I'm happy to provide more detail if needed - just say what would help in a comment. Also I'm making this CW. Please feel free to edit the question if there are other relevant implications of duplicate data that would be worth specifying. | [
0.011405713856220245,
0.004645040258765221,
0.005248693749308586,
0.015782222151756287,
-0.017464442178606987,
-0.0164569690823555,
0.0034219250082969666,
0.004237368702888489,
-0.008793240413069725,
0.004768003709614277,
-0.0021531861275434494,
0.011987466365098953,
-0.0027117731515318155,
... | [
0.9078264236450195,
0.2898750603199005,
0.20713059604167938,
0.13747799396514893,
0.13487280905246735,
-0.09629949182271957,
0.15800230205059052,
-0.07045890390872955,
-0.040258441120386124,
-0.26256856322288513,
0.1108773723244667,
0.29405373334884644,
0.30896899104118347,
0.1488605141639... |
_"You say tomato, I say tomato"_ and the song from the beginning. As an informal turn of speech, it can be used to show that two or more parties are talking about basically the same thing but not in same exact terms, or not quite agreeing on the specifics. Yet written down as _tomato-tomato_ or _potato-potato_ it looks just plain wrong and confusing (like in the video title above). Short of using the transcriptions _/təˈmeɪtəʊ/ - /təˈmɑːtəʊ/_ and _/pəˈteɪtəʊ/ - /pəˈtɑːtəʊ/_ , is there a way of writing it down that indicates the differences in pronunciation? | [
-0.003727530362084508,
0.0036307862028479576,
-0.007984375581145287,
0.019562577828764915,
-0.029150182381272316,
0.015076984651386738,
0.008496632799506187,
-0.013949778862297535,
-0.017005303874611855,
0.009101023897528648,
-0.015007161535322666,
0.0007119611836969852,
0.004009448923170566... | [
0.4088054299354553,
0.43718576431274414,
0.11160692572593689,
-0.419192373752594,
0.08773855119943619,
0.40967386960983276,
0.14738021790981293,
0.149163156747818,
-0.36830687522888184,
-0.4260323643684387,
-0.02560468390583992,
0.5575013160705566,
-0.16280698776245117,
-0.0715339183807373... |
I would like to typeset a book where the odd/even pages have the same format: each has a long text section on the left and a place to put small marginal notes on the right. If necessary, I want the environments for tables/code in the text to be able to expand into this right margin. But, I would like to get the standard odd/even page headers for chapters/sections. | [
0.015705572441220284,
0.0169112216681242,
-0.016788246110081673,
0.016711261123418808,
-0.012809226289391518,
0.026127124205231667,
0.0093422532081604,
0.018591511994600296,
-0.01979661174118519,
0.0024437063839286566,
-0.011575786396861076,
0.0013790576485916972,
0.025804445147514343,
-0.... | [
0.4499097168445587,
0.09960173070430756,
0.02440633252263069,
0.23714859783649445,
-0.1108408272266388,
0.005093676969408989,
-0.2507161796092987,
0.262023389339447,
-0.27805832028388977,
-0.6723031401634216,
0.10737490653991699,
0.20010198652744293,
-0.04429041966795921,
0.005406818352639... |
I hope I'm asking this at the right part of Stack Exchange. Please bear with me if I'm wrong. I'm developing some gps based applications. The demands I have for precision are not very high, but I need to know the possible errors. I have learned that the fourth decimal gives you ~10 meters precision. That should be enough for me. The real question is how fast will I get that precision realiably in different environments (indoors, outdoors free sky, forest, cloudy, city etc). The applications I'm developing is for handheld devices so I prefer to have the gps active in as short intervals as possible. As I do now, the intervals I use the gps are more governed by battery life than precision. Now I'm trying to balance the two. | [
-0.011387664824724197,
0.003755518002435565,
-0.018202777951955795,
-0.005116300657391548,
-0.01748594455420971,
-0.007147714961320162,
0.004385840613394976,
0.004290936980396509,
-0.006760809570550919,
-0.031730182468891144,
0.004353344906121492,
0.005107813514769077,
-0.0018868227489292622... | [
0.76905757188797,
0.20903944969177246,
0.37672555446624756,
0.3469671607017517,
0.17050868272781372,
-0.02910376898944378,
0.31442153453826904,
-0.11157025396823883,
-0.31864169239997864,
-0.6867523789405823,
0.35167011618614197,
0.19944630563259125,
0.282587468624115,
-0.07909499108791351... |
Does anyone know why B is better than A ? > A. Nowadays, public health is a topic that starts to get growing attentions. - > B. Nowadays, public health is a topic that is starting to get growing > attention. | [
-0.053992386907339096,
0.024950522929430008,
-0.002864896086975932,
0.032694485038518906,
0.0249771885573864,
-0.024175390601158142,
0.016628677025437355,
0.01999242790043354,
-0.019993264228105545,
-0.04017190635204315,
-0.03346249461174011,
0.01740250177681446,
-0.006696429569274187,
0.0... | [
0.7873867750167847,
0.45303764939308167,
0.09301017969846725,
0.12156307697296143,
-0.23668326437473297,
-0.09145323187112808,
0.23870113492012024,
0.5738556981086731,
-0.4160107970237732,
-0.6202890276908875,
0.17406108975410461,
0.5113348960876465,
-0.13686755299568176,
0.103642411530017... |
So I like to post a lot on forums. And often times, I'll link images. I usually use imgur as the image provider. But this thought just came into my head. Would it be a good idea to have the image on a page of site, thereby increasing my site ranking (right now, I'm the only person who's ever been on my site hah). So basically, instead of linking http://i.imgur.com/veCBW.p ng, I would link mysite.com/pagethatincludestheimage . And inside it would just contain img src of the image. It would basically appear exactly the same. Is this a decent idea? Is there any other way it may help my site? Btw, I also use amazon s3, so hotlinking will not be an issue. | [
-0.0038502339739352465,
0.010294437408447266,
0.0021255766041576862,
0.012022323906421661,
0.003887750208377838,
0.0022378519643098116,
0.0054650334641337395,
-0.01085231825709343,
-0.01185009628534317,
0.0015884567983448505,
-0.007966442964971066,
0.011463579721748829,
-0.001724457368254661... | [
0.5932173132896423,
0.05399179831147194,
0.514048159122467,
0.01804802566766739,
-0.34326088428497314,
0.0044775791466236115,
0.21946516633033752,
0.3528546392917633,
-0.4651648998260498,
-0.6772845387458801,
0.36646899580955505,
0.32914894819259644,
0.2358843982219696,
0.34391146898269653... |
When I used this code: $ {\bar{a}_x}_y $ MiKtex said: Double subscript $ {\bar{a}_x}_ How to fix? | [
0.02179090492427349,
-0.007911205291748047,
-0.024603266268968582,
0.029684191569685936,
0.0029756329022347927,
0.005956249311566353,
0.015446477569639683,
0.04444936662912369,
-0.022421924397349358,
-0.019176285713911057,
-0.041174791753292084,
0.0034055334981530905,
-0.0449589267373085,
... | [
0.09624285250902176,
0.20304173231124878,
0.024178501218557358,
-0.2048763632774353,
0.15888042747974396,
0.5343188047409058,
0.2587694823741913,
0.3890875279903412,
-0.24098481237888336,
-0.052681855857372284,
0.038309577852487564,
0.5919329524040222,
-0.6891068816184998,
-0.1210160702466... |
I want Indian Rupee Symbol (₹) font in my android, so I can type it in message or anywhere I would like to. But, I can't find it in my keyboard. Can anybody please suggest me where can I get it? I am using Samsung Galaxy Grand and Android Jelly Bean 4.2.2 | [
0.012424342334270477,
-0.007902099750936031,
-0.01046130433678627,
0.01922977901995182,
-0.022166956216096878,
-0.005864431615918875,
0.009173697791993618,
0.04185418039560318,
-0.023599563166499138,
-0.022552262991666794,
-0.013326511718332767,
0.004337728954851627,
-0.01709895394742489,
... | [
0.11289376765489578,
0.16926470398902893,
0.27541807293891907,
-0.004615866579115391,
0.01914074830710888,
0.4376121759414673,
0.12068181484937668,
0.08229734003543854,
0.013215035200119019,
-0.6905168890953064,
0.24507637321949005,
0.10583413392305374,
-0.05888565629720688,
0.186250984668... |
A sequel to this question. I have a dataset where: * $\frac{4}{5}$ of the points are drawn from: $(x, y) \sim \mathcal{U}_{2}(0,30)$, $(z) \sim \mathcal{U}_{1}(14.5, 15.5)$. * $\frac{1}{5}$ of the points are drawn from: $(x, y, z) \sim \mathcal{U}_{3}(0,30)$ Where $\mathcal{U}_{d}(x,y)$ is to be interpreted as a $d$ dimensional set of points which are in each dimension drawn from the range between $x$ and $y$. ## The implementation I have implemented this in matlab like this: General Init: dim = 3; uniP = 1/5; wallP = 4/5; uniformN = ceil(N * uniP); wallN = ceil(N * wallP); First distribution (wall): % parameters lowerWall = [0,0,14.5]; upperWall = [30,30,15.5]; % values [wallD] = blockUniformDist(lowerWall, upperWall, wallN, dim); Second distribution (noise): % parameters lower = 0; upper = 30; % values [uniformD] = uniformDist(lower, upper, uniformN, dim); Combine data and compute the density: % Data data = [wallD; uniformD] % Density uniDensity = 1 / ((upper - lower) ^ dim); wallDensity = 1; for i=1 : dim wallDensity = wallDensity/(upperWall(i)-lowerWall(i)); end wallSpace = (data(:,3) < upperWall(3)) & (data(:,3) > lowerWall(3)); trueValues = wallP * wallDensity .* wallSpace + ... uniP .* (ones(N, 1) * uniDensity); The wallSpace is a boolean array that indicates for each observation in data whether or not it lies within the wall. Since the range of the wall is equal to the range of the wall is equal to the range of the uniform data in dimension one and two I only consider the third dimension. If a point with index `i` isn't part of the wall its density is `uniDensity`, since `wallSpace(i)` is zero for such walls `trueValues(i)` equals uniDensity. A point with index `j` whose z is between 14.5 and 15.5 is in the wall, its density should thus be $\frac{4}{5} \cdot$ `wallDensity` \+ $\frac{1}{5} \cdot$ `uniDensity`. Since `wallSpace[i]` is one for these points, this is the density that is placed in `trueValues[j]`. `blockUniformDist(lowerWall, upperWall, wallN, dim)` is defined as: function [ data ] = blockUniformDist( lower, upper, N, dim ) %BLOCKUNIFORMDIST Samples N values from a uniform distribution with % dim dimensions. % INPUT % - lower: The lowest value allowed (per dimension) % - upper: The highest value allowed (per dimension) % - N: Number of samples to be taken % - dim: Dimension of the distribution % OUTPUT % - data: A vector of samples form the distribution % values data = rand(N, dim); for i= 1 : dim data(:,i) = lower(i) + data(:,i).*(upper(i) - lower(i)); end end And `uniformDist` as: function [ data ] = uniformDist( lower, upper, N, dim ) %UNIFORMDIST Samples N values from a normal distribution with mean mu %and standard deviation sd in dimension dim. % INPUT % - lower: The lowest value allowed % - upper: The highest value allowed % - N: Number of samples to be taken % - dim: Dimension of the distribution % OUTPUT % - data: A vector of samples form the distribution % values data = lower + rand(N, dim) .* (upper - lower); end ## The result The result of this is that each observation has one of two densities either $8.962962963e-04$ or $7.407407407e-06$. Plotting the data set with the density dictating the colour (points with density $7.407407407e-06$ in red and points with density $8.962962963e-04$ in blue) results in:  ## The actual question Shouldn't the points in the denser area of the plot all have the same density, and thus all the same colour? | [
-0.006151906680315733,
0.014091544784605503,
-0.013197657652199268,
0.004961282014846802,
0.013270320370793343,
-0.021876294165849686,
0.005299701821058989,
0.004189112223684788,
-0.007399127818644047,
-0.01365416869521141,
-0.00913148745894432,
-0.0027790688909590244,
-0.020757168531417847,... | [
-0.055096276104450226,
0.1773836314678192,
0.6277801394462585,
-0.0044078161008656025,
-0.10433895885944366,
0.25981131196022034,
-0.10524055361747742,
-0.34298405051231384,
-0.3216691315174103,
-0.5658503770828247,
-0.03203962370753288,
0.3580797016620636,
-0.11361205577850342,
0.35072079... |
Sometimes I read a sentence like the following one: > Objective-C does not provide a standard library, _per se_ , but in most > places.. I wonder how to interpret "per se." I'm non-native English speaker and in Swedish we have the expression "per se," but I don't understand it and maybe you can say that it means something like "in itself" (the strange Swedish expression is _i och för sig_ ) like Latin for _de se_ as distinct from latin _de facto_ , _de re_ , _de dicto_ , _de jure_ , etc. Do these expressions have a connection: "per se" and _de se_? Is it Latin and therefore I have difficulty to understand? What is the difference between these sentences? * Breaking a traffic rule does not, per se, make you a burglar. * Breaking a traffic rule does not, per definition, make you a burglar. * Breaking a traffic rule does not, in itself, make you a burglar. | [
-0.004237594548612833,
0.0026916905771940947,
-0.00860130600631237,
0.007831559516489506,
-0.006391276139765978,
-0.00189129076898098,
0.007647064048796892,
0.001978111220523715,
-0.011080758646130562,
0.024889905005693436,
0.000590296636801213,
0.007780879270285368,
0.009382893331348896,
... | [
0.31589606404304504,
0.1885775476694107,
0.12702928483486176,
0.06374657899141312,
-0.16892890632152557,
-0.15744179487228394,
0.5399608016014099,
0.2807757258415222,
-0.22135122120380402,
-0.14875349402427673,
-0.04312668368220329,
-0.1475430279970169,
-0.045438002794981,
0.24343264102935... |
I need a way to create a custom HTML template for the wp_nav_menu function. I've heard of custom walker classes, but these don't appear to be helpful enough to achieve what I'm trying to do; at least as far as I know because of the lack of documentation towards walker functions. What I need to do is be able to add a 'hoverable' class to all top level menu items. I only need the menu to go two levels; top level, then child menu items. I need to add a 'top- level' class to all menu item anchor elements who have a sub-menu. I need all sub-menu lists to have the class 'sub-nav'. And I need to have all of the last sub-menu list items (li) to have a class 'last'. Here's the code I have right now that generates my menu the exact way I need it to be generated using the get_pages function: <?php $pages = get_pages(array( 'parent' => 0, 'sort_order' => 'ASC', 'sort_column' => 'menu_order' )); $num_pages = count($pages); $p = 0; $exclude = '"pastor.php","service.php","gallery.php","audio.php","video.php"'; $exclude_list = $wpdb->get_results("SELECT GROUP_CONCAT(t1.ID) AS IDS FROM " . $wpdb->posts . " AS t1 INNER JOIN " . $wpdb->postmeta . " AS t2 ON (t1.ID = t2.post_id) WHERE t1.post_type = 'page' AND (t1.post_status = 'publish' OR t1.post_status = 'private') AND t2.meta_key = '_wp_page_template' AND t2.meta_value IN (" . $exclude . ") ORDER BY t1.post_date DESC"); foreach($pages as &$page) : $children = get_pages(array( 'sort_order' => 'ASC', 'sort_column' => 'menu_order', 'hierarchical' => 0, 'childof' => $page->ID, 'parent' => $page->ID, 'exclude' => $exclude_list[0]->IDS )); $num_children = count($children); $has_children = $num_children > 0; ?> <li class="nav-item<?php echo ($has_children ? ' hoverable' : '') . ($num_pages == ++$p ? ' last' : '') . ($page->post_name === $root_parent->post_name ? ' active' : '')?>"> <a href="<?php echo get_page_link($page->ID)?>" class="top-level"><?php echo $page->post_title?></a> <?php if($has_children) : ?> <ul class="sub-nav"> <?php $c = 0; foreach($children as &$child) : ?> <li class="nav-item<?php echo ($num_children == ++$c ? ' last' : '')?>"> <a href="<?php echo get_page_link($child->ID)?>"><?php echo $child->post_title?></a> </li> <?php endforeach;?> </ul> <?php endif;?> </li> <?php endforeach; ?> Is there a way to pull menu items in an order multi-dimensional array so that way I can just iterate through them and generate the above template manually, instead of all this wp_nav_menu and walker none-sense? | [
0.0027633612044155598,
0.01640026643872261,
-0.013838807120919228,
0.004680026322603226,
0.012519238516688347,
-0.00045866286382079124,
0.007839150726795197,
0.009245305322110653,
-0.01786644198000431,
0.019153673201799393,
-0.005294506903737783,
0.011981218121945858,
0.007278460077941418,
... | [
0.342998206615448,
0.020110266283154488,
0.1537618190050125,
0.22330822050571442,
0.2645249664783478,
0.4612749516963959,
0.18628083169460297,
-0.18148529529571533,
-0.23713374137878418,
-0.7950169444084167,
0.13344302773475647,
0.22732357680797577,
-0.02387813851237297,
0.223473459482193,... |
I have a custom post type (Packages) that I want to display a custom page for. The post type is set up like this: public function create_packages_type() { register_post_type('packages', array( 'labels' => array( 'name' => __('Packages'), 'singular_name' => __('Packages') ), 'public' => true, 'has_archive' => true, 'supports' => array( 'title', 'editor', 'thumbnail', 'revisions', ), 'show_ui' => true, 'show_in_menu' => true ) ); add_theme_support('post-thumbnails', array( 'packages' ) ); } And I set up a nice form for it using ACF. It generates a permalink like this: /packages/package-name which is perfect. Because it's a totally custom site, we are using the toolbox theme and building our own custom pages as we go along. I want this custom post type to render the page on a certain page template. How do I do this? At the moment it's calling the image template, and that is not right at all. Any assistance? | [
0.014403743669390678,
0.007078989874571562,
0.006818169727921486,
0.017797274515032768,
0.025295976549386978,
0.004367188084870577,
0.006411314010620117,
0.0001878822222352028,
-0.012063632719218731,
-0.006082519888877869,
-0.005454499274492264,
0.004879492335021496,
0.012715840712189674,
... | [
0.5280071496963501,
0.11813877522945404,
0.3603898882865906,
-0.18272826075553894,
0.060920070856809616,
0.038931071758270264,
0.18921416997909546,
-0.45002979040145874,
0.08506206423044205,
-0.7004197239875793,
-0.1385968029499054,
0.4132087230682373,
-0.2628535330295563,
0.13082104921340... |
I am trying to redirect a page via my .htaccess, but it does not seem to be working. Old page: /dyn/?q=customer%20reference&f=1,1,1,1,1,1&c=10,10,10,10,10,10&s=1,1,1,1,1,1,1&st=1 New page: /customer-references/ So it should be as simple as this: RewriteEngine On RewriteCond %{QUERY_STRING} ^q=customer(?:[\ +]|%20)reference&f=1$ [NC] RewriteRule ^dyn/$ /1? [R=301,NE,NC,L] But it does not seem to be working. Is it because the original page's dynamic URL? The new page is actually a different physical php page if that matters. BTW, I also tried a straight 301 Redirect in the .htaccess. That didn't seem to work either: redirect 301 /dyn/?q=customer%20reference&f=1,1,1,1,1,1&c=10,10,10,10,10,10&s=1,1,1,1,1,1,1&st=1 /customer-references/ And another failed attempt was this: RewriteCond %{QUERY_STRING} ^customer%20reference&f=1,1,1,1,1,1&c=10,10,10,10,10,10&s=1,1,1,1,1,1,1&st=1$ RewriteRule ^$ http://www.domain.com/customer-references/? [R=301,L] Am I making this more difficult than it needs to be? | [
-0.0054475488141179085,
0.01698264107108116,
-0.005626048892736435,
0.0062632933259010315,
0.001857380848377943,
0.0002983992453664541,
0.007324849255383015,
-0.009363329969346523,
-0.01201216597110033,
-0.021587900817394257,
-0.003393718507140875,
0.0015694651519879699,
-0.01143602468073368... | [
-0.1262594759464264,
0.30083468556404114,
0.8854822516441345,
-0.3475392758846283,
0.12140535563230515,
0.47830623388290405,
0.47726765275001526,
-0.05407434701919556,
-0.2143915444612503,
-0.7122218012809753,
-0.02638332173228264,
0.46033045649528503,
-0.20833049714565277,
0.3723352253437... |
So I heard this in a movie and I'm not sure if it's grammatically correct . . . Should it be: * 1.) _"Did you like what you_ **_saw_**?" or * 2.) _"Did you like what you_ **_see_**?" Which one is right, you guys? I'm getting a bit rusty I'm afraid. | [
0.01307628769427538,
0.020484916865825653,
0.005323813296854496,
-0.0056953392922878265,
0.025434117764234543,
0.008082756772637367,
0.006809566169977188,
0.0042077708058059216,
-0.012716170400381088,
0.02665458805859089,
0.0004435034061316401,
0.00863577425479889,
0.010874082334339619,
0.... | [
0.4403562545776367,
0.46711570024490356,
-0.033135607838630676,
-0.07617571949958801,
-0.3760870695114136,
-0.10372120141983032,
0.2506383955478668,
0.3574165999889374,
-0.32725101709365845,
-0.27972501516342163,
0.3322514593601227,
0.5896170735359192,
0.2129800021648407,
0.286829829216003... |
I have a custom content type "balloons", and I've assigned a parent category to a "balloon" item, "Water Balloon" I've made the Balloon category a primary menu item. I want a link to "Water Balloon" to show in the dropdown, as it would if it was a page. I don't know how to do this, or if it's possible. Any ideas? Thanks in advance. | [
-0.0010864821961149573,
0.003690193872898817,
-0.00462470343336463,
0.026565363630652428,
-0.016496604308485985,
-0.053711287677288055,
0.008090130053460598,
0.01743127591907978,
-0.02250869758427143,
0.019687488675117493,
-0.010956658981740475,
0.015123657882213593,
-0.006625858135521412,
... | [
0.4270569980144501,
-0.00008946366870077327,
0.4306507706642151,
0.43060198426246643,
0.10306213051080704,
-0.07102174311876297,
-0.23033997416496277,
0.4009966552257538,
-0.4443836212158203,
-0.5309731364250183,
0.09852638840675354,
0.13021373748779297,
0.07887028902769089,
0.302243143320... |
I have been told, and have found for myself, that lots of developers are not good at UI design (I don't know how true is this) but _it is true about me at least_. In web development good code development skills are not enough without great skills in UI design to go with them. So for me, and many developers like me, that only have half of the thing (good development skills) how should we complete our other half other than paying for a designer? Is using Open Source web templates with little modifications the best solution for this, or are there other options? | [
-0.005330931395292282,
0.013525409623980522,
-0.0044062030501663685,
-0.005122646689414978,
-0.02935352921485901,
0.008825237862765789,
0.007397711277008057,
-0.0005738338804803789,
-0.015713756904006004,
-0.016099683940410614,
-0.01081839483231306,
0.01878117211163044,
0.0067291054874658585... | [
0.5945939421653748,
0.266152560710907,
-0.2556014060974121,
0.23636066913604736,
-0.27002233266830444,
-0.11679067462682724,
0.3101641535758972,
0.052102427929639816,
-0.44153547286987305,
-0.6098065376281738,
0.26405683159828186,
0.6045972108840942,
0.2113693654537201,
-0.0375917442142963... |
I have a .net hosting account but they have provided a free asp.net enterprise manager, a software to manage MSSQL server. Is there any other Phpmyadmin like software. My server supports Php also. | [
0.008793474175035954,
0.01879776641726494,
0.007465484086424112,
0.022307194769382477,
0.005438599735498428,
0.03320582956075668,
0.01362893357872963,
0.032693538814783096,
-0.02912253513932228,
-0.06529103964567184,
-0.002213945146650076,
0.03860652074217796,
0.011954274959862232,
0.00152... | [
0.3233569860458374,
0.3683761656284332,
0.1657779961824417,
-0.07699340581893921,
-0.14794103801250458,
-0.13358739018440247,
0.24320945143699646,
0.35738787055015564,
-0.2815925180912018,
-0.5835393071174622,
0.33258432149887085,
0.0916832834482193,
0.021773237735033035,
0.323066651821136... |
Please note that I am very new on this website so have some difficulties in writings as required here but trying really hard to learn quickly. La-Tex is the main problem but please understand me that I am serious. How do Vectors transform from one inertial reference frame to another inertial reference frame in [special relativity]. A bound vector in an inertial reference frame ($x$,$ct$) has its line of action as one of the space axis in that frame and is described by $x$* _i_ *,then what would it be in form of new base vectors ( **a** ) and ( **b** ) in a different inertial system ($x`$,$ct`$) moving with respect to the former inertial system with $v$* _i_ * velocity.Let ( **i** ) and ( **j** ) be the two bounded unit vectors with the line of action as co-ordinate axis($x$) and($ct$) respectively and senses in the positive side of co-ordinates and similarly ( **a** ) and ( **b** ) are defined for co-ordinates ($x`$) and ($ct`$) respectively. | [
0.013842456042766571,
0.014916324988007545,
0.0010175895877182484,
0.01453966274857521,
0.002659756690263748,
-0.00967867486178875,
0.006703890394419432,
0.011678420007228851,
-0.01631106063723564,
0.0046164290979504585,
-0.004997379146516323,
0.014761457219719887,
0.0014564376324415207,
0... | [
0.36529701948165894,
-0.25389790534973145,
0.4579738676548004,
-0.004326601978391409,
-0.09768734127283096,
0.22087912261486053,
-0.03053872473537922,
-0.11874493956565857,
-0.08959686756134033,
-0.4976840913295746,
-0.10305686295032501,
0.35540837049484253,
-0.6101858615875244,
0.44840252... |
I use to have old latex docs that included pstricks pictures. I had to define objects that I could reuse several times in the pictures and did this with a newcommand/def (and everything was fine). I recently had to compile these files and it's not working anymore: objects defined in the newcommand line isn't displayed. I've tried to show a short, non-working, example: \documentclass{article} \usepackage{pdftricks} \begin{psinputs} \usepackage{pstricks} \end{psinputs} %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% \begin{document} \newcommand*{\example}{\qline(0,0)(1,1)} \begin{pdfdisplay} \begin{pspicture}(3,3) \qline(0,0)(1,1) %this line displays correctly \rput(1,0){\qline(0,0)(1,1)} %this line displays correctly \rput(2,0){\example} %this line isn't displayed! \end{pspicture} \end{pdfdisplay} \end{document} In the above code, the first two lines are correctly drawn, but the third isn't. I've been fighting for hours with it and can't figure out what's wrong. The only piece of information I managed to understand is that the figure is wrapped in a tex file, then compiled in postscript, then transformed in pdf AND that the \example code isn't passed with it: %%% This is the example-fig1.tex automatically generated file \documentclass{article} \input tmp.inputs \pagestyle{empty} \begin{document} \begin{pspicture}(3,3) \qline(0,0)(1,1) %this line displays correctly \rput(1,0){\qline(0,0)(1,1)} %this line displays correctly \rput(2,0){\example} %this line doesn't display! \end{pspicture} \end{document} Help would be really appreciated! Thanks. | [
-0.002776940818876028,
0.010552841238677502,
-0.009939240291714668,
0.017679382115602493,
-0.010657817125320435,
-0.006954949349164963,
0.007581275422126055,
-0.003759923158213496,
-0.015628835186362267,
-0.008861040696501732,
-0.010554613545536995,
0.00199075136333704,
0.012241780757904053,... | [
0.5036869049072266,
-0.0757279247045517,
0.4223279058933258,
-0.004061014391481876,
0.04774917662143707,
-0.05147780105471611,
0.234649658203125,
-0.050332408398389816,
-0.3802698850631714,
-0.5003636479377747,
0.30146360397338867,
0.5562534928321838,
-0.4010637104511261,
0.126642391085624... |
At the dawn of the modern era, Galileo discovered and described how composite bodies fall through the air (or at least the discovery is usually attributed to him). I'm interested in whether this had been discovered earlier and how, particularly since it seems to me that there are good grounds for this result to hold true purely on the basis of continuity and symmetry. Imagine three balls of the same size and weight, and at equal distances from each other, dropped from a tower at the same time. By symmetry, all three must hit the ground at the same time. Now repeat the experiment, but move the left-hand ball next to the middle one. This makes no difference to the result—the three balls still hit the ground at the same time. Now repeat the experiment once more after slightly increasing the contact area of the two adjacent balls. Again, I would expect them to hit the ground at the same time. By repeating this, the left-hand and middle balls eventually merge into a single larger ball, which will fall at the same time as the right-hand one. Did anyone make this argument in the pre-modern literature? I'd be interested to know whether any of the ancient atomists came up with similar arguments when they considered how atoms moved under gravity. It ought to have then been a simple step via the above argument to see that composite bodies fall at the same rate. | [
-0.014075545594096184,
0.025984615087509155,
-0.012092974036931992,
-0.0015234360471367836,
0.010458129458129406,
-0.01931399665772915,
0.008609578013420105,
-0.011416692286729813,
-0.017754320055246353,
-0.00835433416068554,
-0.004114711657166481,
0.017571745440363884,
-0.01209615170955658,... | [
0.05373865365982056,
0.05364714190363884,
0.07881193608045578,
0.42843401432037354,
-0.24780681729316711,
0.22151756286621094,
-0.23068709671497345,
-0.18775561451911926,
-0.9711146354675293,
-0.6014426350593567,
0.14339393377304077,
0.030418002977967262,
-0.29431506991386414,
0.3054270446... |
I'm trying to figure out the generic name for systems that allows users to contribute with the development of new features. Kickass Torrents has a very interesting app in it's site named IdeaBox and it's divided in stages such as: suggestions, planned, in progress, completed. It has a voting feature for everyone with more than 100 reputation and that's pretty much all. I wanna look at opensource alternatives but that's not possible if I don't know how it's generically named. | [
0.006888325791805983,
0.0020298429299145937,
0.010964649729430676,
0.014631133526563644,
-0.0007937013870105147,
-0.0035489655565470457,
0.005798534955829382,
0.012877696193754673,
-0.013311686925590038,
0.032324131578207016,
-0.020189540460705757,
0.007402331102639437,
0.025701012462377548,... | [
0.7311201095581055,
0.16168427467346191,
0.06295792758464813,
0.4681079685688019,
-0.29752951860427856,
-0.14861242473125458,
0.18168944120407104,
0.3551694452762604,
-0.3209916055202484,
-0.46444565057754517,
0.33448082208633423,
-0.07678920775651932,
-0.042729075998067856,
0.307121247053... |
I have this equation in latex: \{\Theta, \{\phi, \psi \}\} + \{\phi, \{\psi,\Theta \}\} + \{\psi, \{\Theta, \phi\}\} = 0 and codecogs.com/latex/eqneditor.php show me "Invalid Equation":  but stackedit.io/ is ok:  Because codecogs.com/latex/eqneditor.php doesn't work? | [
0.01045235525816679,
0.0032577868551015854,
-0.006622111424803734,
0.020232953131198883,
-0.010556323453783989,
-0.0003251758753322065,
0.006708228029310703,
-0.005739057436585426,
-0.011690998449921608,
0.003382660448551178,
-0.010892219841480255,
0.006234203465282917,
-0.02850256860256195,... | [
-0.13626737892627716,
-0.017898395657539368,
0.5920924544334412,
-0.17023302614688873,
0.20365260541439056,
-0.061274077743291855,
0.14475096762180328,
-0.3562709391117096,
-0.5885892510414124,
-0.4470294117927551,
0.10459832847118378,
0.6698639392852783,
-0.36052608489990234,
0.1533243954... |
I have some apps that I open on occasion. One in particular is a game that likes to push notifications up reminding me to play again. Is there an app I can use to automatically freeze an Android app when I exit it, and automatically thaw and open the app when I want to open it again? I hate how many apps have background services or receive intents that I don't care for. With my ROM, I can specifically deny permissions to apps based on certain intents, but that doesn't always help. I know you can freeze and thaw an app using Titanium Backup Pro, which I have. But that would require manually going in to TiBu and doing the freeze/thaw commands each time. In an ideal world, I would like a list of apps to be frozen as soon as I exit. And instead of the app listed in my app drawer, I would have a shortcut to first thaw the app and then open it (and have the same icon as the app). I don't care if this takes an extra few seconds; I simply want certain apps to only be running when I say so. Does this exist in any form on Android, whether in a kernel patch, an app, or a simple script? | [
0.0034970929846167564,
0.003463235916569829,
-0.006430958863347769,
0.005843105725944042,
-0.003119231667369604,
0.004332764074206352,
0.0064344098791480064,
0.011297321878373623,
-0.01899237558245659,
0.016876578330993652,
-0.010472061112523079,
0.004952657036483288,
0.01074807345867157,
... | [
0.5255534052848816,
0.10596885532140732,
0.05607222765684128,
0.20107077062129974,
0.2962658405303955,
-0.1966107040643692,
0.4106193482875824,
0.08020313829183578,
-0.49008113145828247,
-0.42199602723121643,
-0.009783945046365261,
0.5197674632072449,
-0.2534922659397125,
0.105580694973468... |
I have a _Samsung Galaxy Note GT N7000 India_. It rebooted automatically the day before yesterday and then it got stuck at the Logo screen. I have downloaded and flashed it with `N7000DDLSC_N7000ODDLSC_N7000DDLS6_HOME.tar.md5` using Odin and also using the recovery mode from external card. But on restarting the phone doesn't go further than the logo screen. I even wiped the cache. I dont have a backup of contacts, most of the contacts are in phone memory not in my Google account. Is there a way to read the files using the download mode which we use for uploading the firmware using Odin? | [
-0.012276200577616692,
-0.007925614714622498,
-0.004162374883890152,
0.021680649369955063,
-0.04253503680229187,
-0.007602615747600794,
0.0075710914097726345,
0.01615021750330925,
-0.012265803292393684,
0.006157089024782181,
-0.016901057213544846,
0.004982972517609596,
0.003065775614231825,
... | [
-0.01968313194811344,
0.4093644917011261,
0.6252196431159973,
-0.26178446412086487,
0.04407600313425064,
-0.1742389053106308,
0.6189094185829163,
0.18163171410560608,
-0.18132174015045166,
-0.7215589880943298,
-0.1177946925163269,
0.4040329158306122,
-0.3229394555091858,
0.2831390202045440... |
I want to get and improve the source of the `LaTeX2HTML` program, but the top Google results are outdated, with broken links. Does anyone know where the source resides? | [
0.03069206513464451,
0.015059774741530418,
-0.010881965979933739,
0.0013218725798651576,
0.04679053649306297,
-0.01986680179834366,
0.009654681198298931,
0.01250034011900425,
-0.02847401425242424,
-0.00596679886803031,
-0.018627198413014412,
0.02352229505777359,
-0.041993990540504456,
0.02... | [
0.6479524374008179,
0.1741599291563034,
0.17521117627620697,
0.271151602268219,
0.29182928800582886,
-0.5314961075782776,
0.41101548075675964,
0.45473408699035645,
-0.004543629474937916,
-0.20479795336723328,
-0.09012942761182785,
0.12801356613636017,
-0.08096204698085785,
0.34654536843299... |
As a software engineer, we are always eager to get effective tools to boost our productivity. And in our daily work, we are often unsatisfactory with the existing tools and would like to have better ways such as better GDB script config, Vim script and some Python script to make boring things automatic. However, it is actually a trade-off since making tools also needs time and energy. It does not give productivity boosting immediately. Therefore, how do you judge whether it is time to stop work and to make some tools to ease your future pain? | [
0.010489372536540031,
0.012921031564474106,
-0.006373549811542034,
0.0025155390612781048,
-0.03133920207619667,
-0.004082912113517523,
0.007785008288919926,
0.01976371742784977,
-0.015221895650029182,
-0.00014695800200570375,
-0.012353184632956982,
0.013847227208316326,
0.009923073463141918,... | [
0.5116655230522156,
0.30400946736335754,
-0.15842169523239136,
0.2881737947463989,
0.06449463218450546,
-0.3406825363636017,
0.4507220983505249,
0.2505718469619751,
-0.1640593558549881,
-0.5492422580718994,
0.15904660522937775,
0.9172959327697754,
0.028203006833791733,
0.2569446265697479,
... |
If you decide "Program to An Interface" trumps YAGNI and decide to create a supertype where you don't envision anything other than one obvious implementation - is there a convention for naming the obvious concrete type? E.g. this morning I wrote a a class called PlainOldConversionReport and I wonder if that name betrays ignorance of convention. Would there a more normal name, in Object Oriented development generally or .Net or C# specifically, to give to a non specialized subtype of IConversionReport? | [
-0.013154033571481705,
0.02469540759921074,
-0.0024216969031840563,
0.019455984234809875,
-0.01779920980334282,
-0.0020830302964895964,
0.008113210089504719,
0.0047781579196453094,
-0.008458254858851433,
-0.020912516862154007,
-0.0003379270201548934,
0.01272429246455431,
0.018107831478118896... | [
0.3547990620136261,
-0.11661120504140854,
0.0198398157954216,
0.009739690460264683,
-0.38186293840408325,
-0.26666736602783203,
0.14913690090179443,
0.1312035769224167,
-0.2603977620601654,
-0.4673839807510376,
0.02820403128862381,
0.5861920118331909,
-0.5497148633003235,
0.147068962454795... |
I am new to Bayesian Decision Theory and don't understand the following concept: So from what I understood, the Bayes error is used to report the performance of a Bayes classifier in terms of the probability of making and error. From the conditional error probabilities  we can obtain the total probability of making an error (the probability to mis-classify).  Now, if my Bayes classifier was designed to minimize the overall risk, I have a loss function that gives penalties to certain decisions.   So, if my classifier includes such a loss function when I optimize my classifier for minimum overall risk, shouldn't be the Bayes error also include the loss function term? Hope you can help me here, because I think I am missing something here ... **EDIT:** I'll try to express my problem using an 2D-classification problem: Let's assume I have two pdfs (e.g., p(x|c1) and p(x|c2) ) with slight overlap. And mis-classifying a pattern as c2 where it truly belongs to c1 is more costly than vice-versa. In this case I would assign a higher loss to **"classify pattern x as c1 when it is truly c2"** than **"classify pattern x as c2 when it is truly c1"** in order to calculate and minimize the overall risk. I would therefore increase the probability to classify a pattern x as c2 over c1 due to the minimum risk optimization. Isn't this something I have to also include in p(error)? | [
-0.0028331209905445576,
0.021025361493229866,
-0.000727752223610878,
0.023946233093738556,
-0.012255692854523659,
-0.0018641087226569653,
0.009353427216410637,
-0.02295975759625435,
-0.012349465861916542,
0.0018529873341321945,
-0.009667308069765568,
0.011138779111206532,
-0.0202562157064676... | [
-0.26816752552986145,
0.13096311688423157,
-0.20051336288452148,
-0.25734320282936096,
-0.2752523720264435,
0.4538252055644989,
0.3865250051021576,
-0.4445578157901764,
-0.27999529242515564,
-0.34107640385627747,
0.2298070639371872,
0.3407367467880249,
-0.03735216706991196,
0.4528144896030... |
Can anyone help me find a little bit more polite way to say "I am pissed off at you"? I know this is used to show anger or irritation, but this is bit harsh. So, do you have any other option? | [
0.00567378057166934,
0.008264397270977497,
-0.0012097187573090196,
0.03638548403978348,
-0.03430299088358879,
-0.005050874315202236,
0.006395101547241211,
-0.003813613671809435,
-0.027068162336945534,
-0.007000258658081293,
-0.00483004841953516,
0.00980434287339449,
0.01420595869421959,
-0... | [
0.40047454833984375,
0.12339463829994202,
-0.06643293797969818,
0.30384260416030884,
-0.5012794733047485,
-0.099888376891613,
0.4262036085128784,
0.5860743522644043,
-0.11931278556585312,
-0.27314651012420654,
0.3623224198818207,
0.5964467525482178,
0.013084206730127335,
0.0389167591929435... |
We have just received a large set of DEMs at work and I would like to generate contours from them. The DEMs have a resolution of 1m and a size of 1kmx1km. Output from gdalinfo: Driver: AAIGrid/Arc/Info ASCII Grid Files: 380000_6888000_1k_1m_DEM_ESRI.asc Size is 1000, 1000 Coordinate System is `' Origin = (380000.000000000000000,6888000.000000000000000) Pixel Size = (1.000000000000000,-1.000000000000000) Corner Coordinates: Upper Left ( 380000.000, 6888000.000) Lower Left ( 380000.000, 6887000.000) Upper Right ( 381000.000, 6888000.000) Lower Right ( 381000.000, 6887000.000) Center ( 380500.000, 6887500.000) Band 1 Block=1000x1 Type=Float32, ColorInterp=Undefined NoData Value=-9999 I know I can use gdal_contour to generate the contours (my blog post on the topic) but I'm wondering what some **best practices** for generating contours are. Are there any rules you should follow to get the most out of the contours but not make stuff up or loose too much information? Say I want to generate three sets of contours: * 250mm * 1m * 5m **Is there anything I should do to the DEM before each set?** **Is post smoothing of the lines a good way to go or is smoothing the raster a better option?** | [
-0.007441830355674028,
0.001799063989892602,
-0.0042660473845899105,
0.016035377979278564,
-0.018955156207084656,
-0.007993925362825394,
0.005818587262183428,
0.00026511267060413957,
-0.008159448392689228,
-0.005500513594597578,
-0.0009790798649191856,
0.007647490594536066,
0.006450470536947... | [
0.2655065655708313,
-0.22697857022285461,
0.5117746591567993,
0.022423749789595604,
-0.12855082750320435,
0.36791497468948364,
-0.13432598114013672,
-0.5510674715042114,
-0.20911335945129395,
-0.6760854125022888,
-0.16584844887256622,
0.07894893735647202,
-0.0978531688451767,
0.02834643423... |
_I'm a total n00b with Android, and theexisting questions tagged `boot-loop` didn't help me. Apologies if I missed something "obvious"!_ I've borrowed a used HTC Desire Z from my brother so I can decide whether to switch to Android. The phone was provided with CyanogenMod 7 and worked like a charm for a few days. I rebooted it today to see whether a reboot would clear up this other question and since then it keeps rebooting itself. By "rebooting" I mean this sequence: The device shows the `HTC` logo and then the `cyanogen(mod)7` logo with the robot on a skateboard. After that, it shows the SIM PIN entry screen while the status bar says that it's loading the SD card. Without me touching the device, the screen goes black and boots again from the `HTC` logo onwards. It's got the ClocworkMod Recovery v5.0.2.7 and I'm guessing that this is part of the CM7 package. It's (supposed to be) a helpful way to reset and recover the device. Here's what I've tried so far (with reboots in between), without success: 1. I made a backup and then cleared the `/cache`. 2. I restored the backup. 3. I then reset the phone with the `wipe data/factory reset` function. 4. I removed the SIM and the SD card, then did another factory reset. 5. I wiped the `Dalvik cache`, whatever that is. 6. With help, I downloaded a fresh CM7.1 zip file and installed that via the Recovery and SD card. Also wiped all caches again, just to be sure. 7. I downloaded a CM9 rom that was made for this device and installed that. It starts fine and seems to work, but the exact same problem still occurs. The phone now boots to the welcome screen that says "touch the Android to start" and then, as before, goes black and reboots without me touching it. By the way, the battery is about 70% charged so it's not due to an empty battery, and the problem is the same while plugged into the charger or not. **How can I get this device to run again?** I don't care (much) for the settings I had; I can begin all over if need be. But I need a working device first... **Update:** I added the steps 6&7 above. Didn't solve the problem but it seems to prove that it's not a ROM issue. **Is the device physically defective?** | [
-0.008671773597598076,
0.005918127950280905,
-0.008608624339103699,
0.015748871490359306,
-0.03916100412607193,
-0.013251146301627159,
0.0051681771874427795,
0.008854227140545845,
-0.012056587263941765,
-0.015180671587586403,
-0.006307630799710751,
0.007716168649494648,
0.0027022480498999357... | [
0.21128946542739868,
0.19895239174365997,
0.6034950613975525,
0.14445653557777405,
0.06399180740118027,
0.1317431479692459,
0.5169339776039124,
0.21983754634857178,
-0.19072186946868896,
-0.6879870295524597,
0.07175317406654358,
0.6027181148529053,
0.035356491804122925,
0.23432821035385132... |
I have a dataset that I want to fit a simple linear model to, but I want to include the lag of the dependent variable as one of the regressors. Then I want to predict future values of this time series using forecasts I already have for the independent variables. The catch is: how do I incorporate the lag into my forecast? Here's an example: #A function to calculate lags lagmatrix <- function(x,max.lag){embed(c(rep(NA,max.lag),x),max.lag)} lag <- function(x,lag) { out<-lagmatrix(x,lag+1)[,lag] return(out[1:length(out)-1]) } y<-arima.sim(model=list(ar=c(.9)),n=1000) #Create AR(1) dependant variable A<-rnorm(1000) #Create independant variables B<-rnorm(1000) C<-rnorm(1000) Error<-rnorm(1000) y<-y+.5*A+.2*B-.3*C+.1*Error #Add relationship to independant variables #Fit linear model lag1<-lag(y,1) model<-lm(y~A+B+C+lag1) summary(model) #Forecast linear model A<-rnorm(50) #Assume we know 50 future values of A, B, C B<-rnorm(50) C<-rnorm(50) lag1<- #################This is where I'm stuck################## newdata<-as.data.frame(cbind(A,B,C,lag1)) predict.lm(model,newdata=newdata) | [
0.02063640020787716,
0.013107871636748314,
-0.007095743902027607,
0.004386278800666332,
0.010269394144415855,
-0.021641269326210022,
0.006793783511966467,
0.00275418721139431,
-0.014072097837924957,
-0.01693057455122471,
-0.0070162443444132805,
0.008075668476521969,
-0.009268787689507008,
... | [
0.08147336542606354,
-0.2921323776245117,
0.5324527025222778,
-0.003549925284460187,
0.06783857941627502,
0.23389926552772522,
-0.049060575664043427,
-0.35760506987571716,
-0.37768009305000305,
-0.27846893668174744,
0.050616972148418427,
0.4990624487400055,
-0.3658992052078247,
0.271463900... |
1-In general, if a theory has action at a distance effects, where can that appear exactly in the theory? 2-Does it appear in the dynamical law of the theory? (does it appear in Newton's 2nd law? where can it be spotted?) 3-Does it appear in the force law of the interaction? (it is said that Newton's law of gravitation, $\displaystyle F\sim\frac{m_1m_2}{r^2}$, supports action at a distance effects. How can one see that from the form of the law?) 4-Before special relativity, causality roughly means cases always come before effects. Now if the force law allows action at a distance, as in Newtonian gravity in 3, then the interaction is instantaneous. It seems that the words "before" and "after" lose their meaning in this case, then how causality is defined then? | [
-0.008049755357205868,
0.017056904733181,
-0.00200299802236259,
0.015978001058101654,
-0.007792475633323193,
-0.024240754544734955,
0.00721175130456686,
0.008669106289744377,
-0.01315315067768097,
0.0011684049386531115,
-0.020143652334809303,
0.02488425001502037,
-0.009239865466952324,
0.0... | [
-0.045181289315223694,
-0.12211629748344421,
0.6564927697181702,
0.02131655067205429,
-0.1302410513162613,
0.0624798946082592,
-0.24357809126377106,
-0.4330957531929016,
-0.4649561941623688,
-0.36179685592651367,
-0.05477936193346977,
0.30466943979263306,
-0.4412860572338104,
0.13979537785... |
When `man declare` in bash, I get the manpage of `declare`in sql. But I would like to get the manage of `declare` in shell. How can I do that? | [
0.01739189773797989,
0.026776570826768875,
-0.0023584356531500816,
0.012685069814324379,
-0.020023252815008163,
-0.002762690419331193,
0.011598995886743069,
0.0008959653205238283,
-0.02427874319255352,
-0.03142406418919563,
-0.014809808693826199,
0.0157749205827713,
-0.04940496385097504,
0... | [
0.11561178416013718,
0.5417423248291016,
-0.3142203688621521,
-0.2599969208240509,
-0.11724307388067245,
-0.3627004027366638,
0.49349552392959595,
-0.28106600046157837,
-0.243960902094841,
-0.6015370488166809,
-0.04657465219497681,
0.8637969493865967,
-0.05130983144044876,
0.04937744885683... |
Wikipedia says > An installation program or installer is a computer program that installs > files, such as applications, drivers, or other software, onto a computer. > > Some installers are specifically made to install the files they contain; > other installers are general-purpose and work by reading the contents of the > software package to be installed. 1. Is an installer always used for binary installation and not doing any compilation work for source installation? 2. Does Linux have the concept "installer" for package installation? Wikipedia distinguishes between "installer" and "Package management system". In Ubuntu, it seems to me all binary installation is done by Package management systems `dpkg` or `apt` - so where is an "installer"? | [
-0.022554362192749977,
0.009858528152108192,
0.0025267070159316063,
0.0036938809789717197,
-0.004502438474446535,
-0.002031699987128377,
0.0066656977869570255,
-0.021755989640951157,
-0.009229227900505066,
-0.017391391098499298,
-0.008733374997973442,
0.0039810556918382645,
0.028233330696821... | [
0.5284225344657898,
0.06827685236930847,
0.07063174247741699,
-0.17853416502475739,
0.06220945343375206,
-0.537688672542572,
-0.10499683022499084,
0.10993707925081253,
-0.04779182747006416,
-0.6441518664360046,
-0.6255017518997192,
0.6617366671562195,
-0.5039775967597961,
0.037622001022100... |
I need to realize a database in this mode: TABLE restaurant with "classic" fields of a place: address, street number, phone, website, etc etc etc. I need to indicate also the days of opening // closing. E.g. From Monday to wednesday close, saturday only evening and sunday lunch + evening. So, I did think at another table with 14 rows: ID_RESTAURANT MONDAY_LUNCH MONDAY_EVENING [...] SUNDAY_EVENING 12 0 0 1 And "finally" every restaurant has his own cooking ("international", "chinese", "japan") Another table with a column for every cooking? Finally, several services. "Conditioned Air", "Credit Card", "Animals permitted" "children addicted" etc etc. Another table with a column for every "service"? ID_RESTAURANT CONDITIONED CREDIT_CARD ANIMALS 12 1 1 0 13 0 1 1 I hope that I'm clear. Thank you very much! | [
0.014789624139666557,
0.009498387575149536,
-0.013665618374943733,
0.004950486123561859,
-0.020526712760329247,
0.0303264781832695,
0.008473075926303864,
0.0037528052926063538,
-0.011040667071938515,
-0.0034452048130333424,
-0.02754366397857666,
0.005745402071624994,
0.011662888340651989,
... | [
0.3831278085708618,
0.4417734742164612,
0.32836249470710754,
-0.18567124009132385,
-0.017308520153164864,
-0.08793777972459793,
0.034614525735378265,
0.03087269701063633,
-0.4612158238887787,
-0.5157753825187683,
-0.10961498320102692,
0.2070220559835434,
0.16621053218841553,
0.251131445169... |
The Hamiltonian of an atom coupled to an EM field, both described quantum mechanically is: $$H = \frac{1}{2m}(\hat{p}-q\hat{A})^2 = \frac{\hat{p}^2}{2m} -\frac{q}{m}\hat{p}\hat{A}+\frac{q^2}{2m}\hat{A}^2$$ Under the condition of transversality ($\hat{p}$ and $\hat{A}$ are vectors). I have only seen it treated in the dipole approximation where the last term becomes a constant and so unphysical; but what is its general interpretation? Is it shifting the energy of the free photons, it doesn't look diagonal in number of occupation basis. | [
0.004256953485310078,
0.009291335940361023,
0.0015840384876355529,
0.004348722752183676,
0.0024253390729427338,
-0.01739022694528103,
0.005754526704549789,
-0.005232418887317181,
-0.006908295676112175,
0.011958390474319458,
-0.015316078439354897,
0.007287156768143177,
-0.016383882611989975,
... | [
0.02401077374815941,
-0.011811644770205021,
0.5468938946723938,
0.10902725160121918,
-0.039537250995635986,
0.09761972725391388,
-0.12130172550678253,
-0.7081939578056335,
-0.04954255744814873,
0.018671590834856033,
-0.07208989560604095,
0.22691385447978973,
-0.7359926104545593,
0.42451611... |
I am reading about Endianness and it seems both systems save their data differently. Sparc uses big endian and x86 uses small endian, meaning the byte order is different. That makes me wonder... is it possible to mount UFS partitions from Sparc on x86-systems? | [
0.0055532390251755714,
0.0023922494146972895,
-0.03070741519331932,
0.0311907809227705,
-0.01758149266242981,
-0.008929436095058918,
0.019449487328529358,
0.013849234208464622,
-0.030489936470985413,
-0.024760980159044266,
-0.02130943164229393,
0.014538666233420372,
-0.003993671853095293,
... | [
-0.10439104586839676,
0.050529200583696365,
0.30250412225723267,
0.10305779427289963,
-0.15568900108337402,
-0.1361844539642334,
-0.0828665941953659,
-0.015195927582681179,
-0.2903051972389221,
-0.7536597847938538,
0.06870057433843613,
0.1172424927353859,
-0.2532590329647064,
0.58986592292... |
Let me support my question with a quick scenario. We're writing an app for family meal planning. We'll produce daily plans with a target calorie goal and meals to achieve it for our nuclear family. Our calorie goal will be calculated for each person from their attributes (gender, age, weight, activity level). The weight attribute is the simplest example here. When Dad (the fascist nerd who is inflicting this on his family) first uses the application he throws approximate values into it for Daughter. He thinks she is 5'2" (157 cm) and 125 lbs (56kg). The next day Mom sits down to generate the menu and looks back over what the bumbling Dad did, quietly fumes that he can never recall anything about the family, and says the value is really 118 lbs! This is the first introduction of the discord. It seems, in this scenario, Mom is _probably_ more correct that Dad. Though both are only an approximation of the actual value. The next day the dear Daughter decides to use the program and sees her weight listed. With the vanity only a teenager could muster she changes the weight to 110 lbs. Later that day the Mom returns home from a doctor's visit the Daughter needed and decides that it would be a good idea to update her Daughter's weight in the program. Hooray, another value, this time 117 lbs. Now how do you reconcile these data points? Measurement error, confidence in parties, bias, and more all confound the data. In some idealized world we'd have a weight authority of some nature providing the one and only truth. How about in our world though? And the icing on the cake is that this single data point changes over time. How have you guys solved or managed this conflict? | [
-0.0033476499374955893,
0.006517409346997738,
-0.00141363765578717,
0.010491635650396347,
-0.00037313904613256454,
-0.00856659933924675,
0.006673811934888363,
0.00840736459940672,
-0.008409513160586357,
-0.02639404684305191,
-0.005607609637081623,
0.0017812995938584208,
0.008869203738868237,... | [
0.6348941922187805,
0.04001712054014206,
0.14731349050998688,
0.170071542263031,
0.34048280119895935,
0.483491986989975,
0.0953754186630249,
-0.11642815172672272,
-0.16721105575561523,
-0.4014718532562256,
0.4128613770008087,
-0.09001314640045166,
0.09050239622592926,
0.17827099561691284,
... |
I am trying to use latex to write my thesis, and I have modified a template to use, but am having trouble with some of my section headings. Specifically, with the weird spacing between words caused by the section heading being justified, rather than flush left. I am using the titlesec package and have tried: \titleformat{\section}{\raggedright\large\sffamily\bf}{\thesection}{1em}{} I also tried forcing titlesec to make everything raggedright by adding \usepackage[raggedright]{titlesec} to my preamble but neither seem to have any effect and the heading text is still strangely spaced. I think it is probably something else within my settings that is taking preference? \documentclass[11 pt]{report} %----------------------------------------- % Packages %----------------------------------------- \usepackage{titlesec} \usepackage[a4paper, scale=1.0, textwidth=145mm, textheight=237mm, layoutvoffset=0pt, layouthoffset=0pt, ignoremp, includehead, marginparsep=0pt, bottom=4cm, top=2cm, left=4cm, right=2.5cm, verbose=true, bindingoffset=0pt]{geometry} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage[english]{babel} \usepackage{setspace} \usepackage[toc, page, header]{appendix} \usepackage{fancyhdr} \usepackage[plain]{fancyref} \usepackage{textcomp} \usepackage{sectsty} \usepackage{balance} \usepackage{lastpage} \usepackage[format=plain,justification=centering,singlelinecheck=false,font=small,labelfont=bf,labelsep=space]{caption} %----------------------------------------------------------------------- % MAIN PAGE SETUP %----------------------------------------------------------------------- \renewcommand{\rmdefault}{ptm} % sets roman font to Times ('ptm') \renewcommand{\sfdefault}{phv} % sets serif font to Helvetica \renewcommand{\ttdefault}{lmtt} % sets text type font to Latin Modern Typewriter ('ptm') \widowpenalty=500 \clubpenalty=500 \pretolerance=10000 \tolerance=2000 % \emergencystretch=10pt \titleformat{\section}{\fillleft\large\sffamily\bf}{\thesection}{1em}{} %formats the section titles \begin{document} \onehalfspace \chapter[Introduction]{Introduction} \chaptermark{Introduction} \label{ch:introduction} % label for referring to chapter in other parts of the thesis \section[Importance]{Importance}\label{C1:Intro} Filling in later \section[Synthesis]{Synthesis of averylongcompoundnamethatis toobigtofitinthisbox} blahblahblah \end{document}  | [
0.005488666705787182,
0.009756041690707207,
-0.010350040160119534,
0.027484159916639328,
0.0059648482128977776,
0.017574261873960495,
0.009297346696257591,
0.008109652437269688,
-0.009834518656134605,
0.0000685681588947773,
-0.009358666837215424,
0.0055313375778496265,
0.003904469311237335,
... | [
0.33273303508758545,
0.1986187994480133,
0.48063939809799194,
-0.16526101529598236,
0.17346413433551788,
-0.3952612280845642,
0.022022029384970665,
-0.12116099894046783,
0.18513897061347961,
-0.2492227852344513,
-0.07716045528650284,
0.4683830142021179,
0.02591748721897602,
0.2108574211597... |
A physics journal I'm submitting to requests that image files not contain any non-standard fonts. Standard fonts, as far as they are concerned, are: Times- Roman, Times-Italic, Times-Bold, Times-BoldItalic, Helvetica , Helvetica- Oblique, Helvetica-Bold, Helvetica-BoldOblique, Courier, Courier-Oblique, Courier-Bold, Courier-BoldOblique, Symbol. I'm using the following code to generate a standalone pdf file and I can't seem to get rid of all the "non-standard" LaTeX fonts, specifically "CMMI10" and "CMSY7". Note that I am using XeLaTeX to compile the code. \documentclass{standalone} \usepackage{tikz} \usepackage{pgfplots} \usepackage{verbatim} \usepackage{mathspec} \setallmainfonts{Times New Roman} \setallsansfonts{Arial} \setallmonofonts{Courier New} \begin{document} \pgfplotsset{every axis plot post/.append style={line width=1.0pt}} \pgfplotsset{grid style=dotted} \begin{tikzpicture} \begin{loglogaxis} [clip marker paths=true, xlabel=background field amplitude $B_{\text{a}}$ / T, ylabel=losses $Q$ / J/cycle/m, xmin=2e-3, xmax=5e-2, ymin=1e-4, ymax=1e-1, xtick={2e-3,5e-3,1e-2,2e-2,5e-2}, xticklabel style={/pgf/number format/.cd,fixed,precision=3}, xticklabel={% \pgfmathfloatparsenumber{\tick}% \pgfmathfloatexp{\pgfmathresult}% \pgfmathprintnumber{\pgfmathresult}% }, minor xtick={3e-3,4e-3,6e-3,7e-3,8e-3,9e-3,1.1e-2,1.2e-2,1.3e-2,1.4e-2,1.5e-2,1.6e-2,1.7e-2,1.8e-2,1.9e-2,3e-2,4e-2}, grid=major ] \end{loglogaxis} \end{tikzpicture} \end{document} Any ideas how to strip the two remaining fonts from my file? | [
0.01539634633809328,
-0.0030812297482043505,
-0.017431356012821198,
0.010738003998994827,
0.02297053672373295,
-0.005946176126599312,
0.007686743978410959,
0.010465919971466064,
-0.009048632346093655,
0.00471968948841095,
-0.0010637401137501001,
-0.011937562376260757,
0.022661836817860603,
... | [
-0.1576051115989685,
0.3726448118686676,
0.6301788091659546,
0.24502965807914734,
-0.3100418150424957,
0.16401779651641846,
0.09415499120950699,
0.05857964977622032,
-0.2330348789691925,
-0.3651755750179291,
-0.21756188571453094,
-0.016151035204529762,
-0.30668920278549194,
0.1581452637910... |
We have many maps containing Feature Classes and Tables in a Personal Geodatabase that need their information updated. Is there a function to List Tables from the current map document like mxd = arcpy.mapping.MapDocument("Current") arcpy.mapping.ListLayers(mxd) in ArcGIS 10.1? | [
-0.005583019927144051,
0.010743199847638607,
-0.013010782189667225,
-0.0019203266128897667,
0.002200834220275283,
0.003726847469806671,
0.010488049127161503,
0.020427100360393524,
-0.02038097195327282,
-0.0029368698596954346,
0.002284491201862693,
0.012664012610912323,
-0.010477462783455849,... | [
0.43157249689102173,
-0.22576230764389038,
0.5873978734016418,
0.30008018016815186,
0.050286076962947845,
-0.2745993435382843,
-0.15553249418735504,
-0.14224544167518616,
-0.4991697669029236,
-0.716304361820221,
-0.1884092390537262,
0.3077608644962311,
-0.20319879055023193,
-0.093475192785... |
I came across this term when looking into PHP's "traits" (which are apparently bad, since, among things, they're not mockable), but I can't really find a definition for this "mockability". It might be related to polymorphism? There isn't even a Wikipedia entry I could find! | [
-0.008488836698234081,
0.009229094721376896,
0.0030759915243834257,
0.015564579516649246,
-0.002695702016353607,
-0.020016830414533615,
0.007285391911864281,
0.001591083244420588,
-0.02149640955030918,
0.021888958290219307,
-0.007025082595646381,
0.004969573114067316,
-0.013117889873683453,
... | [
0.608506977558136,
0.21182376146316528,
-0.2904886305332184,
0.07975071668624878,
-0.4504322409629822,
-0.18956899642944336,
0.847346842288971,
0.322699636220932,
-0.07558014988899231,
-0.31512534618377686,
0.19298191368579865,
0.14979572594165802,
-0.3166903853416443,
0.31440654397010803,... |
I'm trying to take a list of the form `{ {"65 + 3 months", 75}, {"65 + 4 months", 75.1} }`, and transform the string part to a number such as 65.25 or 65.333, respectively. What I really want to do is something like: list = { {"65 + 3 months", 75}, {"65 + 4 months", 75.1} }; list /. {y:NumberString~~" + "~~m:NumberString,p_}:>{y+m/12,p} That form doesn't work, I believe, because I need to use StringReplace to cause the `StringExpression` to do any matching. So since I can't make that work, I tried something like: list /.{ys_String, p_} :> {StringReplace[ys, y:NumberString~~" + "~~m:NumberString~~" months":>y+m/12], p} But in that case, I'm getting `{ { StringExpression[3/12 + 65], 75}, {StringExpression[4/12 + 65], 75.1} }`, and have no idea where the `StringExpression` heads came from in that case or what to do with them without making my expression uglier and even further from just describing the patterns that I'm trying to transform from/to. Is there a simple way to compose a pattern that does some string matching and some structure matching? If not, what's the best way to write the `StringReplace` version for the problem above. | [
-0.0022086366079747677,
0.0014119099359959364,
-0.01825757883489132,
0.003850384149700403,
0.006894310005009174,
0.017438266426324844,
0.005577470175921917,
0.014962790533900261,
-0.01516245398670435,
-0.008744446560740471,
0.00609943363815546,
-0.0006288720178417861,
-0.006148532498627901,
... | [
0.28457027673721313,
0.11245126277208328,
0.46925514936447144,
-0.21888791024684906,
0.396016389131546,
0.5396327376365662,
0.1179836094379425,
-0.2533119320869446,
-0.5353633761405945,
-0.22725871205329895,
-0.06111840531229973,
0.18250171840190887,
0.09967463463544846,
0.2308052182197570... |
Sometimes when you have long code you need to check some part of this code. the way I am using currently is to selected the part that I want and then copy it to new notebook and then evaluate it there. this process becomes annoying when repeated several times. is there a better way to automatically evaluate selection in new notebook without need to go through copy paste procedure? **Update :** Thanks to halirutan for his suggestions. But one of my concerns is to evaluate selection within the cell in to another window. For example:  how to evaluate the selection without manually copy and paste. If the cell itself contain long content, then the only way to debug the cell is be copying part by part and pasting into another notebook and evaluate as a whole new cell. | [
0.021793775260448456,
0.003750808071345091,
-0.015233084559440613,
0.014582273550331593,
-0.018726462498307228,
0.007564898580312729,
0.006339323241263628,
0.021749693900346756,
-0.019149864092469215,
0.025709301233291626,
-0.02080341801047325,
0.013886701315641403,
-0.0006848396733403206,
... | [
0.4837619662284851,
-0.001806278945878148,
0.12904953956604004,
0.3225302994251251,
-0.03052837774157524,
-0.024962129071354866,
0.046559643000364304,
-0.4817691147327423,
-0.5296648144721985,
-0.9752556085586548,
0.0779305100440979,
0.3638087213039398,
-0.15063290297985077,
0.092545010149... |
I want to add a facebook-fanbox on a wordpress-page. I can go about this in the following matter: I take the test from the facebook-developer site. Initialize the JavaScript SDK using this app: books-page... <div id="fb-root"></div> <script>(function(d, s, id) { var js, fjs = d.getElementsByTagName(s)[0]; if (d.getElementById(id)) return; js = d.createElement(s); js.id = id; js.src = "//connect.facebook.net/en_GB/sdk.js#xfbml=1&appId=473241986032774&version=v2.0"; fjs.parentNode.insertBefore(js, fjs); }(document, 'script', 'facebook-jssdk'));</script> Include the JavaScript SDK on your page once, ideally right after the opening `<body>` tag. Or do I need to include this? <div class="fb-like-box" data-href="https://www.facebook.com/FacebookDevelopers" data-colorscheme="light" data-show-faces="true" data-header="true" data-stream="false" data-show-border="true"></div> Which text has to be added into a text-block widget? Any suggestions | [
-0.004691106732934713,
-0.001997194951400161,
0.0025939857587218285,
0.012951415032148361,
0.0007298975251615047,
-0.00365156726911664,
0.004827828146517277,
0.01963028497993946,
-0.015344281680881977,
-0.0011268013622611761,
-0.004487678408622742,
0.01007063314318657,
-0.010829668492078781,... | [
0.33449244499206543,
-0.10935507714748383,
0.7348961234092712,
-0.185132697224617,
0.011717338114976883,
-0.1223302036523819,
0.3352142572402954,
-0.4652951657772064,
0.3278437554836273,
-0.8620070219039917,
0.10225716233253479,
0.3542826175689697,
-0.17153772711753845,
-0.1602369844913482... |
I'm guessing this is a really dumb question, but I'm far more familiar with PHP/WordPress and I'm just getting started with jQuery so I'm not even sure what to search on. Is there a way to pass the output of echo get_blogifo('siteurl') To jQuery? I'm trying to get this to work: $('a.getstarted').attr('href', '<?php echo get_bloginfo('siteurl'); ?>/payroll-deduction-authorization/?plan=basic'); Thanks in advance for your help! | [
-0.010733137838542461,
0.010982823558151722,
0.011012477800250053,
0.000455981120467186,
-0.004825483076274395,
-0.0020283376798033714,
0.006544193252921104,
0.029808634892106056,
-0.026034832000732422,
-0.02211456559598446,
-0.015754222869873047,
0.015018210746347904,
-0.0032643498852849007... | [
0.6581432223320007,
0.12947316467761993,
0.35888102650642395,
0.043848294764757156,
-0.05152959004044533,
-0.11773187667131424,
0.3847312927246094,
-0.12933555245399475,
-0.23470167815685272,
-0.312860906124115,
0.0557083785533905,
0.40591561794281006,
-0.1602436602115631,
-0.0345108918845... |
Galaxy S running 2.2. try to install SKYPE but get the error - "insufficient memory" when there is loads on the phone and i can install any other app quite normally. Any help appreciated. | [
-0.0035365743096917868,
-0.012066852301359177,
-0.017987873405218124,
0.03105362504720688,
-0.036903560161590576,
0.012546143494546413,
0.01102637778967619,
0.012146486900746822,
-0.025233933702111244,
-0.010018452070653439,
-0.02025091089308262,
0.0019412630936130881,
-0.01551361009478569,
... | [
0.08777663856744766,
-0.21600621938705444,
1.2104359865188599,
-0.14687249064445496,
-0.336368203163147,
-0.1831260621547699,
0.40922752022743225,
0.15658463537693024,
-0.21952767670154572,
-0.8664116859436035,
-0.411097913980484,
0.4179254472255707,
-0.3526080846786499,
-0.213270783424377... |
Given the result $V_{m\times n} \approx W_{m\times k} \cdot H_{k\times n}$, where columns of $V$ are data points and $m$ is data dimension, what is the function by which you assign the data points to the $k$ classes? Is the usual method to use the normalized columns of $W$ as centroids? It seems like that would be problematic since even a normalized version of $W$ is not unique. For example if $R$ is a rotation matrix $WR\cdot R^{-1}H = W \cdot H$ where $WR$ is still normalized. Thanks so much... | [
-0.005384721793234348,
0.005067752208560705,
-0.015032347291707993,
0.017510170117020607,
-0.007254980970174074,
-0.017449188977479935,
0.007778872735798359,
-0.003935228101909161,
-0.011642571538686752,
0.02254641242325306,
-0.0073026129975914955,
0.011035684496164322,
-0.001128195319324731... | [
-0.35800009965896606,
-0.11010627448558807,
0.4868049919605255,
0.06388966739177704,
0.11276112496852875,
0.30899178981781006,
-0.34191861748695374,
-0.44000744819641113,
-0.11270695179700851,
-0.43477120995521545,
-0.31165099143981934,
0.25589898228645325,
-0.14996156096458435,
0.53330576... |
This is my feeling. But more is different. If atoms form a solid, it is hard to say whether the solid will be ferromagnetic or not. | [
0.006086145527660847,
0.03865622729063034,
-0.017992397770285606,
0.047045283019542694,
-0.014411000534892082,
-0.04816366359591484,
0.014944481663405895,
-0.003526852698996663,
-0.039697401225566864,
0.04405470937490463,
-0.014960715547204018,
0.03236497566103935,
0.002055190037935972,
-0... | [
0.5473727583885193,
0.27970585227012634,
0.04262150824069977,
0.08114628493785858,
-0.2218056619167328,
-0.267299085855484,
0.12922325730323792,
-0.3048657476902008,
-0.23747357726097107,
-0.37362140417099,
-0.11077263951301575,
0.32967159152030945,
-0.017144540324807167,
0.606451749801635... |
I want to add logging to an application I'm currently working on. I've added logging before, that's not an issue here. But from a design perspective in an object-oriented language, what are the best practices for logging that follow OOP and patterns? _**Note:_** I'm currently doing this in C#, so examples in C# are obviously welcome. I would also like to see examples in Java and Ruby. * * * **Edit:** I'm using log4net. I just don't know what's the best way to plug it in. | [
-0.009142221882939339,
0.01129283756017685,
-0.019487878307700157,
0.012591399252414703,
0.003710901364684105,
0.032302454113960266,
0.00539239589124918,
0.021758638322353363,
-0.022267471998929977,
-0.017717700451612473,
-0.0010467316024005413,
0.009347202256321907,
-0.005231720861047506,
... | [
0.8094730973243713,
0.14778976142406464,
0.11788944154977798,
0.016289396211504936,
-0.05391141399741173,
-0.25301817059516907,
0.3468160033226013,
0.3228650987148285,
-0.5588058233261108,
-0.7159336805343628,
0.14214643836021423,
0.6859101057052612,
-0.2922917604446411,
-0.221724748611450... |
I would very much like to disable the drag and drop upload feature for tinyMCE (both the default one and the `wp_editor()` one) because it interacts with my custom pop up menus - (WP Trac). Right now I'm trying to just disable the one with the actual `WP_Editor()` function. For some reason it doesn't seem to actually disable any drag and drop functionality and was wondering if I was doing something wrong. My function call looks like this: wp_editor($meta_content, 'additionalTab', array( 'wpautop' => true, 'media_buttons' => false, 'textarea_name' => '_additional_content', 'textarea_rows' => 15, 'teeny' => true, 'drag_drop_upload' => false )); **The Expected Output** is that it should completely disable drag and drop uploads directly to the tinyMCE. **The Actual Output** is that it doesn't do anything, it still allows you to drag and drop upload into the tinymce. Is there something wrong with what I'm doing or is there something inherieitnly wrong with `wp_editor()`? | [
-0.00238616019487381,
0.006295633502304554,
0.0003282317193225026,
0.01582835428416729,
-0.0028479243628680706,
-0.013987425714731216,
0.009086746722459793,
0.01642252877354622,
-0.020165584981441498,
0.005204949993640184,
-0.025080226361751556,
0.01154596358537674,
-0.009670798666775227,
... | [
0.17990519106388092,
0.07152407616376877,
0.347032368183136,
0.03378560394048691,
0.03598558157682419,
-0.12606492638587952,
0.2516029179096222,
0.008840139023959637,
-0.40594881772994995,
-0.45450344681739807,
0.03260680288076401,
0.9456742405891418,
-0.704799473285675,
0.3114704191684723... |
I have often heard it claimed that there is a duality between hypothesis tests and confidence regions: If $\langle\Omega,\mathcal{F}\rangle$ is a measurable space and $\Theta$ is a parameter space for a family of probability measures $P$ on $\mathcal{F}$, then a confidence region procedure is a function $C$ from $\Omega$ to the power set of $\Theta$ (taking an observed $\omega\in\Omega$ to a set of parameter values "consistent" with $\omega$), and a hypothesis test procedure is a function $T:\Theta\to\mathcal{F}$ where $T(\theta_{0})$ is the acceptance region for the null hypothesis $\theta=\theta_{0}$. The duality is generally supposed to work like this: If $C$ is a confidence region procedure, then $C^{\ast}:\Theta\to\mathcal{F}$ defined by $$C^{\ast}(\theta_{0})=\\{\omega\in\Omega \,\vert \theta_{0}\in C(\omega)\\}$$ is the test procedure associated with $C$, and similarly for $T^{\ast}$: $$T^{\ast}(\omega)=\\{\theta_{0}\in\Theta\, \vert \omega\in T(\theta_{0})\\}.$$ It is immediately clear that, for instance, $(C^{\ast})^{\ast}=C$, and similarly for $T$. So far, so simple, except that I am having trouble with this definition of $C^{\ast}$. In the general case, there is no reason to assume that $C^{\ast}(\theta_{0})$ is actually an element of $\mathcal{F}$, which means that it may not be meaningful to assign a probability to rejection or acceptance of the null hypothesis. There are certain special cases where everything works fine. If $\Omega$ is countable, for example, it is reasonable to think that $\mathcal{F}$ is the entire power set of $\Omega$. Likewise, if $\Omega$ and $\Theta$ are intervals and we impose some convexity and continuity conditions, everything turns into an interval, and things work out just fine. So, what I'm asking is whether there is any agreed upon set of criteria beyond the definitions I wrote above to make this duality work properly? Or is the duality only ever understood to apply in well behaved one dimensional contexts? | [
-0.012214255519211292,
0.0107261436060071,
0.004610985051840544,
0.016935672610998154,
0.021400131285190582,
0.0005534531082957983,
0.009485075250267982,
-0.02717207930982113,
-0.007542449049651623,
-0.006100258324295282,
-0.02151651680469513,
0.02168414369225502,
-0.010216344147920609,
0.... | [
0.042206697165966034,
-0.08896790444850922,
-0.12854282557964325,
0.25141599774360657,
0.22173228859901428,
0.22855554521083832,
0.018006999045610428,
-0.15287810564041138,
0.033347103744745255,
-0.3780670762062073,
-0.02799268066883087,
0.6377672553062439,
-0.023934243246912956,
0.4346382... |
I have an experimental design problem and I'm not sure which would the best way to proceed. We have a micro-array experiment in which we compare gene expression profiles between 2 groups of patients. Each group has 12 independent biological replicates. Lets call these groups CT and EXP. The goal of this experiment is to identify differentially expressed genes between CT and EXP. Two sets of hybridizations where done: first and second. First hybridization had big issue: CT samples were processed at one time-point and EXP couple of months later, creating a perfectly confounded batch effect. Second hybridization did not have this problem - all of the samples were processed at the same time. I've analyzed both sets separately and came up with 2 lists of differentially expressed genes. My question is, should I discard the results of the first hybridization completely and stick with the second one since it didn't have an obvious design issue? Or should I use both sets of hybridizations and select only genes which show up as differentially expressed in every case (there are 2900 of such genes)? Appreciate your help! | [
0.013650452718138695,
0.007291455287486315,
-0.014724642969667912,
0.022387973964214325,
-0.034326501190662384,
0.0013885404914617538,
0.009286906570196152,
-0.008343774825334549,
-0.013580807484686375,
-0.026072490960359573,
-0.01432794239372015,
0.00723235122859478,
-0.009632017463445663,
... | [
0.4224449098110199,
-0.650563657283783,
-0.011861681938171387,
0.2567636966705322,
0.09536121785640717,
0.30999594926834106,
0.014906597323715687,
-0.14702557027339935,
0.0674346312880516,
-0.5834916830062866,
0.4403916299343109,
0.052344564348459244,
-0.4843412935733795,
0.413865536451339... |
Consider the formal definition: > f(n) = O(g(n)) Why is it not: > f(n) = O(f(n)) or > f(n) = O(c*f(n)) since for the Big O analysis, `f(n)=2n` and `g(n)=n` are identical? I am confused by the function `f(n)` using another function. * * * ### Update Why isn' t the definition as follows: > f(n) <= c*abs(g(n)) What does the formal `O(g(x))` add to the definition? It seems like it overcomplicates things. | [
-0.018610497936606407,
0.016258075833320618,
-0.026306210085749626,
0.014833971858024597,
-0.013717181980609894,
-0.011051863431930542,
0.008115828968584538,
-0.00026935606729239225,
-0.013730069622397423,
0.019024237990379333,
0.02051781490445137,
0.005420132540166378,
0.007544638589024544,... | [
0.047275204211473465,
-0.10552822053432465,
0.2494218647480011,
-0.20103341341018677,
0.05183783546090126,
-0.1822958141565323,
0.03185088932514191,
-0.25723394751548767,
-0.1485210657119751,
-0.412387877702713,
-0.22332502901554108,
0.7656387686729431,
-0.31541183590888977,
-0.01769181899... |
Hi I am working on a archive page which I can use for multiple custom post types. I need to make a variable in the `$args` array which can change to the `post_type` name on the basis of `<?php post_type_archive_title(); ?>` So something like this: <?php $post_type = post_type_archive_title(); $args = array( 'post_type' => $post_type, 'orderby' => 'title', 'order' => 'ASC', 'caller_get_posts' => 1, 'posts_per_page' => 20, ); query_posts($args); ?> But this doesn't work. Does anyone know how I can fix this? | [
0.00994872860610485,
0.019112911075353622,
-0.0038430793210864067,
0.017615552991628647,
0.027786262333393097,
0.01264180801808834,
0.0076971836388111115,
0.024012872949242592,
-0.015008309856057167,
-0.013460209593176842,
-0.009938424453139305,
0.010346313938498497,
-0.00821601040661335,
... | [
0.14861193299293518,
0.25572845339775085,
0.7290756702423096,
-0.35019582509994507,
-0.07044066488742828,
0.11110367625951767,
0.22426357865333557,
-0.30026111006736755,
0.018497036769986153,
-0.37942570447921753,
-0.03310460224747658,
0.1877441108226776,
-0.190886989235878,
0.398482590913... |
I have updated my MiKTeX to 2.9. But unfortunately it cannot compile pdf file by either PdfLaTeX and XeLaTeX. When use XeLaTeX it lag at the begining, and for PdfLaTeX it returns errors: Command Line: D:\CTEX\MiKTeX\miktex\bin\x64\latex.exe --src --interaction=errorstopmode --synctex=-1 "asd.tex" Startup Folder: E:\temp\exersice This is pdfTeX, Version 3.1415926-2.4-1.40.13 (MiKTeX 2.9 64-bit) entering extended mode (E:\temp\exersice\asd.tex LaTeX2e <2011/06/27> Babel <v3.8m> and hyphenation patterns for english, ... loaded. (D:\CTEX\MiKTeX\tex\latex\base\article.cls Document Class: article 2007/10/19 v1.4h Standard LaTeX document class (D:\CTEX\MiKTeX\tex\latex\base\size10.clo))latex.exe: GUI framework cannot be initialized. LaTeX failed to create a dvi file. For possible explanations start the command from the Command Prompt... My article is \documentclass{article} \usepackage{amsmath} \begin{document} $$\mathfrak{A}$$ \end{document} It is sound. So I think there must be some thing wrong with the config but I cannot find any. | [
0.012241535820066929,
-0.004382352344691753,
-0.014980473555624485,
0.013164565898478031,
0.01752108335494995,
0.018127676099538803,
0.009151346981525421,
-0.014676287770271301,
-0.01555970124900341,
-0.003964045085012913,
-0.02072267420589924,
-0.004987549036741257,
-0.00333490245975554,
... | [
-0.2298392653465271,
0.26667356491088867,
1.0253808498382568,
-0.14594577252864838,
0.14398568868637085,
-0.07299686968326569,
0.4540908932685852,
-0.2733176648616791,
0.1174931526184082,
-0.8815193176269531,
-0.1720929592847824,
0.7323534488677979,
-0.29095521569252014,
-0.025997346267104... |
I suspect this isn't valid code, but the docs are ambiguous. fontspec is loaded with `\fontspec[<font features>]{fontname}`. BoldFont is included in "font features". Then I should be able to recursively define: \setmainfont[SizeFeatures={ {Size={-8},BoldFont=Arial}, {Size=8-}}]{Hoefler Text} But this gives (fontspec) The following font options are not recognised: (fontspec) BoldFont=Arial Not allowed or there's a way around it? My hopes were to make the **whole** font definitions in this way (with `BoldFeatures={SizeFeatures=...}` again and so on). A minimal working file: \documentclass{article} \usepackage{fontspec} \setmainfont[SizeFeatures={ {Size={-8},BoldFont=Arial}, {Size=8-}}]{Hoefler Text} \begin{document} Normal text\\ {\small \textbf{Small bold}}\\ \end{document} Edit: Using Gentoo/Linux, Texlive-2010 | [
-0.004818301647901535,
-0.0047430433332920074,
-0.0005867135478183627,
0.00735912099480629,
0.0011696876026690006,
-0.0022984901443123817,
0.006688280962407589,
0.010282116942107677,
-0.009645681828260422,
-0.00511887576431036,
-0.0010093583259731531,
0.0005886834114789963,
0.010852732695639... | [
-0.000045776021579513326,
0.13678398728370667,
0.5489051938056946,
-0.3695497214794159,
-0.11286979913711548,
0.04243386536836624,
0.3101651668548584,
-0.1631045788526535,
0.052456967532634735,
-0.5737970471382141,
-0.04551563784480095,
0.7748202085494995,
-0.14470165967941284,
-0.14797860... |
My background is computer science. I am fairly new to monte carlo sampling methods and, although I understand the math, I have hard time coming up with intuitive examples for importance sampling. More precisely, could someone provide examples of: 1. an original distribution one cannot sample from but one can estimate 2. an importance distribution which can be sampled from and adequate for this original distribution. Thanks in advance! | [
0.00529079046100378,
0.0036816131323575974,
-0.030555814504623413,
0.03588838502764702,
-0.023289188742637634,
-0.00033262406941503286,
0.012668400071561337,
0.005154531914740801,
-0.02976929023861885,
-0.02954450435936451,
-0.01267360057681799,
0.0020211064256727695,
-0.021044600754976273,
... | [
0.545939028263092,
-0.0698566809296608,
-0.3487558364868164,
0.330620139837265,
0.22525854408740997,
-0.06030871346592903,
-0.05731673911213875,
0.272087424993515,
-0.06892582774162292,
-0.4008900821208954,
0.3520268499851227,
0.1618690937757492,
0.21539443731307983,
-0.012113045901060104,... |
I'm sure this is possible, I just don't think I am going the right way about it. Basically, I have two custom post types (Investigators & Centres), both of which have search forms on their respective archive pages (below) to filter the archive based on a users input and custom fields (which are being indexed by the Custom fields search plugin). http://dev.anklearthritis.co.uk/find-a-specialist/ (Investigators) http://dev.anklearthritis.co.uk/centres/ (Centres) At the moment however, no matter what the user searches, the first 6 entries are always shown. I'm using the following query (for investigators) in my search file $args = array( 'post_type' => 'investigator', 'paged' => $paged, 'orderby' => 'title', 'order' => 'ASC', 'posts_per_page' => 6, //Limits the amount of posts on each page 'post_title' => 'LIKE %'.$_POST['s'].'%' ); $loop = new WP_Query( $args ); I then use a while loop to show each result on the page. Is this even possible, and if so what am I doing wrong? Thanks | [
0.014620194211602211,
0.006303838919848204,
-0.0016385219059884548,
0.014820130541920662,
-0.010199988260865211,
0.010503904893994331,
0.005819423124194145,
0.014846177771687508,
-0.012433342635631561,
0.004925617948174477,
-0.0008850088343024254,
0.010945742018520832,
0.00128595856949687,
... | [
0.6474007964134216,
-0.15845198929309845,
0.4817016124725342,
-0.12368909269571304,
0.0117060961201787,
-0.045382898300886154,
0.4313049912452698,
0.14459119737148285,
-0.30567100644111633,
-0.35232189297676086,
-0.11878380179405212,
0.1584206074476242,
0.017332781106233597,
0.319520115852... |
I'm trying to plot a graph wit a logarithmic y-axis. Since I'm exporting the graph to pdf and later printing it, I want to manually set the frame and tick marks to a reasonable thickness. However the logarithmic tick marks do not change their thickness. (Note: I exaggerated the thickness of the tick marks on purpose to illustrate my point.) LogPlot[x^2, {x, 1, 3}, PlotStyle -> Red, Frame -> True, FrameStyle -> Directive[Black, AbsoluteThickness[2]], FrameTicksStyle -> Directive[Black, AbsoluteThickness[2]] ]  I'm working with Mathematica 10 on Mac OS X 10.9.4. In Mathematica Version 9 the logarithmic tick marks change their thickness as expected. Can anyone reproduce this behavior? Is this a bug or did the FrameTicksStyle change in Mathematica 10? | [
0.014440645463764668,
0.002108796499669552,
-0.010965881869196892,
0.011188004165887833,
0.016045652329921722,
-0.009497440420091152,
0.00923379696905613,
0.021867722272872925,
-0.015233234502375126,
-0.02513032779097557,
-0.009114431217312813,
0.006118446122854948,
-0.01818813942372799,
0... | [
0.23568178713321686,
-0.08707492053508759,
0.7551427483558655,
0.3372555077075958,
0.06226099655032158,
0.2927744388580322,
0.10147716850042343,
-0.27985215187072754,
-0.0944812148809433,
-0.7238907217979431,
0.029493406414985657,
0.49900582432746887,
-0.2924184799194336,
0.050047378987073... |
I have a few extra custom fields added into a user's profile and output them in several places within the frontend. One thing I want to do is add some of those custom fields into the WordPress Feed as RSS elements. As an example, I know that the Author Display Name is output by default but also want to add their Twitter handle into the feed too. Anyone know a solution to this? Thanks | [
0.016091393306851387,
-0.0013220291584730148,
-0.00121354463044554,
0.024555381387472153,
-0.006247277837246656,
0.017456427216529846,
0.009341548196971416,
0.0056498125195503235,
-0.021998321637511253,
-0.025024374946951866,
0.0028935279697179794,
0.013864962384104729,
0.0043268464505672455... | [
0.9135739207267761,
0.06852290779352188,
0.31485000252723694,
0.16744115948677063,
-0.030045488849282265,
-0.021316884085536003,
0.3563159704208374,
0.4955793619155884,
-0.25759923458099365,
-0.6073060035705566,
0.26963698863983154,
0.5408782958984375,
-0.20021012425422668,
-0.132721364498... |
Which one is correct? > I have previous work experience in IBM in PHP domain during the period > 2009-2010. OR > I had previous work experience in IBM in PHP domain during the period > 2009-2010. | [
0.027025140821933746,
0.010265085846185684,
-0.03965208679437637,
0.004452212247997522,
0.04685918986797333,
0.07910452783107758,
0.02272375486791134,
0.022308535873889923,
-0.02618219144642353,
0.06358283758163452,
-0.03945198655128479,
0.019844744354486465,
0.015798799693584442,
-0.00864... | [
-0.02931876853108406,
0.07565908879041672,
-0.08432342112064362,
0.37887516617774963,
-0.24999825656414032,
-0.10352801531553268,
-0.02030833810567856,
-0.16143277287483215,
0.13401968777179718,
-0.3778263032436371,
0.22644293308258057,
0.5940273404121399,
-0.015693126246333122,
-0.0349791... |
I've been using WPML to have multilanguage capabilities in my WordPress implementations. Now WPML has gone commercial, and I'm looking for a open source non-commercial replacement. My main concerns are: * It should be easy to use for the content administrator. * It should be fairly flexible. * It should let me decide the URL structure for each language (subdomain, folder, parameter, etc.) * It should perform relatively well (Specially the queries) * It should support all major WP features (eg: Custom post types, menus, widgets) I'm in the process of testing a few plugins, but I'd want to know if any of you have good advice. | [
-0.0011422683019191027,
0.015413708984851837,
-0.00046077766455709934,
0.027194295078516006,
-0.000396158080548048,
0.006809175480157137,
0.006617366336286068,
0.03649822622537613,
-0.014367004856467247,
-0.018383007496595383,
-0.011271842755377293,
0.018403712660074234,
0.004556143656373024... | [
0.22220678627490997,
0.07426241785287857,
0.7616751790046692,
-0.1899459958076477,
-0.08180558681488037,
-0.21022704243659973,
0.1609121710062027,
-0.09077919274568558,
-0.4180898666381836,
-0.385908842086792,
-0.19108229875564575,
0.41528844833374023,
-0.2667461633682251,
0.11943928152322... |
How to Add Font Awesome Icons to WordPress Menus | [
0.13030105829238892,
-0.0049363248981535435,
0.03309604898095131,
0.06449616700410843,
-0.06258035451173782,
-0.07956573367118835,
0.032757241278886795,
-0.014183927327394485,
-0.031516313552856445,
0.10893856734037399,
-0.09146197885274887,
0.026422277092933655,
0.014162496663630009,
0.03... | [
0.1387253999710083,
0.07080734521150589,
0.255990207195282,
0.25076407194137573,
0.2002524733543396,
0.23029404878616333,
0.2591555714607239,
-0.19917534291744232,
-0.005491006653755903,
-0.796553373336792,
0.1220262348651886,
0.4999328553676605,
0.015127131715416908,
-0.543228805065155,
... |
I am in the process of writing an open source State Space Analysis suite in C# (for fun). I have implemented a number of different Kalman-Based Filters (Kalman Filter, Information Filter and the Square Root Filter) and the State and Disturbance Smoothers that work with these implementations. My tests on these filters (using the Nile data from _Durban and Koopman's (DK)_ book _"Time Series Analysis by State Space Methods"_ and other more complex data) show that the filters and smoothers work and they produce very similar results (as you would expect for a local univariate model). The smoothed output for the basic Kalman Filter for the Nile data looks like (because everyone likes a graphic or two :])  The dotted line is the 90% confidence interval. Now, my problem; I am now attempting to write the first part of the parameter estimation code and for the parameter estimation of the covariances $\mathbf{H}_{1}$ and $\mathbf{Q}_{1}$ in the linear Gaussian State Space model $$y_t = \mathbf{Z}_{t}\alpha_{t} + \epsilon_{t},$$ $$\alpha_{t + 1} = \mathbf{T}_{t}\alpha_{t} + \mathbf{R}_{t}\eta_{t},$$ where $y_{t}$ is our observation vector, $\alpha_{t}$ our state vector at time step $t$ and $$\epsilon_{t} \sim NID(0, \mathbf{H}_{t}),$$ $$\eta_{t} \sim NID(0, \mathbf{Q}_{t}),$$ $$\alpha_{1} \sim NID(a_{1}, \mathbf{P}_{1}).$$ where $t = 1,\ldots, n$. We have an implementation of the Expectation Maximisation (EM) algorithm and this is estimating $\mathbf{H} = 19550.37$ and $\mathbf{Q} = 7411.44$ with the loglikelihood converging in 32 iterations to -683.1 with a convergence tolerence of $10^-6$. The DK book quotes $\mathbf{H} = 15099$ and $\mathbf{Q} = 1469.1$ and from Tusell's paper and this walkthrough using R he matches this prediction getting $\mathbf{H} = 15102$ and $\mathbf{Q} = 1468$. I have debugged my code and it really seems like my implementation of the EM algorithm is correct. So get to the bottom of what is going on I want to run the same data set using R and KFAS... Using R and the R Package KFAS, I have attempted to reproduce these estimates for $\mathbf{H}$ and $\mathbf{Q}$, but my R knowledge is weak. My current R script is as follows install.packages('KFAS') require(KFAS) # Example of local level model for Nile series modelNile<-SSModel(Nile~SSMtrend(1,Q=list(matrix(NA))),H=matrix(NA)) modelNile modelNile<-fitSSM(inits=c(log(var(Nile)),log(var(Nile))),model=modelNile, method='BFGS',control=list(REPORT=1,trace=1))$model # Can use different optimisation: # should be one of “Nelder-Mead”, “BFGS”, “CG”, “L-BFGS-B”, “SANN”, “Brent” modelNile<-SSModel(Nile~SSMtrend(1,Q=list(matrix(NA))),H=matrix(NA)) modelNile modelNile<-fitSSM(inits=c(log(var(Nile)),log(var(Nile))),model=modelNile, method='L-BFGS-B',control=list(REPORT=1,trace=1))$model # Filtering and state smoothing out<-KFS(modelNile,filtering='state',smoothing='state') out _How can I extend this R script using KFAS in order to estimate the parameters $\mathbf{H}$ and $\mathbf{Q}$?_ and _Should each method of parameter estimation be hitting the same values for $\mathbf{H}$ and $\mathbf{Q}$?_ Thanks for your time. | [
0.006657881196588278,
0.00797200296074152,
-0.0033389993477612734,
0.026180503889918327,
-0.019182980060577393,
-0.0008380420040339231,
0.008650753647089005,
-0.004999359138309956,
-0.014851071871817112,
-0.0005171387456357479,
-0.001659040106460452,
0.02030358463525772,
-0.02234479412436485... | [
0.0008602891466580331,
-0.0677172914147377,
-0.15064886212348938,
-0.0658295601606369,
-0.27947914600372314,
0.07084767520427704,
0.05887703225016594,
-0.4415978789329529,
0.047679536044597626,
-0.9995627403259277,
0.0565778948366642,
0.19110167026519775,
-0.2909036874771118,
0.32865318655... |
I have a dataset of some 100 samples, each with >10,000 features, some of which highly correlated. Here's what I am doing currently. 1. Split the data set into three folds. 2. For each fold, 2.1 Run elastic net for 100 values of lambda. (this returns a nfeatures x 100 matrix) 2.2 Take a union of all non-zero weights. (returning a nfeatures x 1 vector) 3. Select features corresponding to the non-zero weights returned from 2.2 4. Use these features for training and testing SVM. My problem is that in step 3, for each fold I get a different set of features. How do I get one final model out of this? One final list of relevant features? Can I take an intersection of the selected features in step 3 for all folds? Features that are selected in all three folds would appear to be the most stable/significant. Can I do this, or is it cheating? | [
0.02343885600566864,
0.004015165381133556,
-0.01642022468149662,
-0.00013958418276160955,
-0.010351349599659443,
0.008547157049179077,
0.005839366465806961,
0.0019488217076286674,
-0.010809401981532574,
-0.001537545584142208,
-0.002124573104083538,
0.014652678743004799,
-0.022231407463550568... | [
0.28036144375801086,
0.00962477084249258,
0.4177137315273285,
-0.15415939688682556,
-0.23441433906555176,
0.3224574327468872,
0.3737982213497162,
-0.6669468283653259,
-0.13545535504817963,
-0.7924150824546814,
0.19695934653282166,
0.5545310378074646,
-0.16767287254333496,
-0.04188865795731... |
I am trying to use PAM to stop password reuse. Adding 'remember=24' to pam_unix stores the old passwords hashed with MD5 which I find unnerving. I was recommended 'pam_pwhistory' instead. Here is the password section of my /etc/pam.d/system-auth file that /etc/pam.d/passwd includes password requisite pam_cracklib.so retry=3 try_first_pass password requisite pam_pwhistory.so remember=24 retry=3 use_authtok password sufficient pam_unix.so sha512 rounds=5000 shadow use_authtok password required pam_deny.so When I try to change my password, if I fail the password reuse test, I don't get to retry. Instead a "Password:" prompt appears $ passwd (current) Unix password: <current password> New password: <old password> Retype new password: <old_password> Password has already been used. Choose another. Password: <current password> passwd: Authentication token manipulation error I have a similar same issue if I fail the pam_cracklib test more than the retry #. Except that my current password works with the prompt $ passwd (current) Unix password: <correct password> New password: <too simple password> BAD PASSWORD: it too simple New password: <too simple password> BAD PASSWORD: it too simple New password: <too simple password> BAD PASSWORD: it too simple Password: <current password> passwd: Have exhausted maximum number of retries for service If I set retry=1 for pam_pwhistory, then the current password works for the prompt. $ passwd (current) Unix password: <current password> New password: <old password> Retype new password: <old_password> Password has already been used. Choose another. Password: <current password> passwd: Have exhausted maximum number of retries for service The inablility to retry if you use an old password and this strange 'Password: ' are the only problem. Old passwords are stored in /etc/security/opasswd using their original hash function. It looks like 1. pam_pwhistory can't retry because it's in the middle of the stack? 2. A failure in pam_pwhistory doesn't cause pam_unix to fail, it instead prompts user? I am on a CentOS 6.3 | [
0.0008789720013737679,
0.011683819815516472,
-0.010189628228545189,
0.0055661010555922985,
-0.010786124505102634,
0.004312689416110516,
0.008180240169167519,
-0.0031771832145750523,
-0.012695925310254097,
-0.001617650268599391,
-0.013553188182413578,
0.0017619726713746786,
0.0019575192127376... | [
0.30778300762176514,
0.5455219149589539,
0.3729614019393921,
-0.4495948255062103,
-0.10964996367692947,
0.04569125548005104,
0.712654173374176,
-0.4483855664730072,
-0.08143842220306396,
-0.988538920879364,
-0.0003354221407789737,
0.668692946434021,
-0.06487055867910385,
0.1171528697013855... |
I want to write a document in english and some words in chinese using pdfLaTeX. Here a MWE: \documentclass{article} \usepackage[UTF8]{ctex} \begin{document} nǔrén\v{a}amùtūou \end{document} This shows up as in the image below:  As you can see after the `ǔ` or the `ū` LaTeX puts automatically a free space even if it's not written in the code, but not after `\v{a}`. 1) Why happens this? 2) How can I avoid this problem? (As you can imagine its not easy to make the change every time) Thank you for any suggestions. | [
0.00039286864921450615,
0.011379982344806194,
-0.000611055875197053,
0.01936292089521885,
-0.009740743786096573,
0.003177206963300705,
0.008165593259036541,
0.01558005902916193,
-0.01226540096104145,
-0.004758274648338556,
-0.0024079098366200924,
-0.0017670406959950924,
0.001897106529213488,... | [
0.4728943705558777,
0.3365626335144043,
0.3429155647754669,
-0.2055492401123047,
0.032532140612602234,
0.05728199705481529,
0.3837609887123108,
-0.11927353590726852,
0.2895854115486145,
-0.5909102559089661,
-0.19612760841846466,
0.4325089752674103,
-0.030227208510041237,
-0.234796434640884... |
> I’m gonna open a can of whoop ass on you if you don’t come with me now! I know that "whoop someone's ass" means to beat someone's ass And "open a can of whoop ass" means to beat someone Is there any relationship between this "a can of whoop ass" and the really whoop ass can?  | [
-0.0066404640674591064,
0.005238478071987629,
0.00869931373745203,
0.014146378263831139,
-0.014650441706180573,
-0.013281522318720818,
0.00985715165734291,
0.005731950514018536,
-0.014402990229427814,
0.013634721748530865,
-0.007936598733067513,
0.00830474216490984,
0.01712282933294773,
0.... | [
0.6129379868507385,
0.140917107462883,
-0.34642621874809265,
0.1628253161907196,
-0.5956449508666992,
0.10339660197496414,
0.6604682803153992,
0.04279927909374237,
-0.3464743494987488,
-0.4316909909248352,
0.2338026762008667,
0.6502852439880371,
0.1737843006849289,
-0.3246806561946869,
-... |
By definition, average life of radioactive sample is the amount of time required for it to get decayed to 36.8% of its original amount. But what is the significance of 36.8% and why has that value been chosen? | [
0.01379457302391529,
0.0168171264231205,
-0.022634150460362434,
-0.000563538633286953,
0.0365862213075161,
-0.017025118693709373,
0.015525173395872116,
-0.024020420387387276,
-0.027218254283070564,
-0.0305355042219162,
0.0007967042620293796,
0.0015815763035789132,
0.00823277235031128,
-0.0... | [
0.7546396851539612,
0.10505125671625137,
-0.15936200320720673,
0.11587399989366531,
0.13405801355838776,
-0.04158968850970268,
0.5891031622886658,
-0.13718153536319733,
-0.02109159342944622,
-0.05177615210413933,
0.13426221907138824,
0.025815285742282867,
-0.1709539294242859,
0.75141513347... |
On the candy box statistics page there is a figure for > Number of saves which annoyed the candy merchant How do you annoy the candy merchant and can you tell if you have already annoyed the candy merchant? Is it something you have to do explicitly, or is it something you do randomly or accidentally? | [
0.017654942348599434,
0.014618891291320324,
-0.009978709742426872,
0.036982446908950806,
-0.029724903404712677,
0.008746996521949768,
0.01248043030500412,
-0.009840100072324276,
-0.026524698361754417,
0.01922040432691574,
-0.021973220631480217,
0.01532549224793911,
0.0023271343670785427,
0... | [
0.4622058570384979,
0.07880672812461853,
-0.008976798504590988,
0.20152096450328827,
-0.6219983696937561,
0.07195091247558594,
0.391737163066864,
-0.12002380937337875,
-0.49388280510902405,
-0.44375574588775635,
0.33640018105506897,
0.32375645637512207,
-0.36296361684799194,
0.213271394371... |
It is easy for me to add arrows to the axes of the a figure by taking advantage of `AxesStyle -> Arrowheads[]` when the differences between the horizontal and vertical coordinates is small. For instance, by using Plot[1/x, {x, -20, 20}, AxesStyle -> Arrowheads[{0.0, 0.03}]] the arrows appear at both the horizontal and vertical axis. However, I don't know how to add arrows to the ones whose differences are big. For example, when the following program is run Plot[1/x^5, {x, -20, 20}, AxesStyle -> Arrowheads[{0.0, 0.00003}]] the arrows cannot be seen obviously? I want to know how I can make the arrows to be found evidently just as the previous one? | [
-0.006588793825358152,
0.0059717679396271706,
-0.010012046433985233,
0.010174648836255074,
0.00951838493347168,
-0.029027745127677917,
0.005399693734943867,
0.009296173229813576,
-0.011461328715085983,
-0.011749635450541973,
-0.01756192371249199,
-0.0008942701388150454,
-0.013420118018984795... | [
0.11729328334331512,
-0.23429684340953827,
0.584825336933136,
0.007533286232501268,
-0.32388243079185486,
0.46365946531295776,
-0.04154200479388237,
-0.31565427780151367,
-0.37923961877822876,
-0.5273647904396057,
0.07224064320325851,
0.30993202328681946,
-0.2062770128250122,
-0.0556389242... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.