text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
After installing `Univers_ps` font I get the error `lunr8r` not found, probably exactly the same which is reported by a guy in a post here (german) The problem is: I don't understand the solution given there, since I am a Windows user and I have (almost) no clue about LaTeX. Can anybody give me a hint on how I might approach this problem? Windows 7 SP1 MikTex 2.9.4248 TexnicCenter 2.02 ME \documentclass[ngerman]{tudscrreprt} \usepackage{selinput} \SelectInputMappings{adieresis={ä},germandbls={ß}} \usepackage[T1]{fontenc} \usepackage{babel} \usepackage{blindtext} \begin{document} \pagestyle{empty} \end{document} The error msg is !pdfTeX error: pdflatex.exe (file lunr8r): Font lunr8r at 657 not found Probably it has to do with this class `tudscrreprt` Id like to use. I know it's difficult to state something from far, but maybe there is a rather general _to do_ for such a case or am I mistaken?
[ 0.007399165537208319, 0.0017273190896958113, -0.004389832727611065, 0.017559651285409927, 0.019049059599637985, -0.007350749336183071, 0.006087995134294033, 0.013859198428690434, -0.012159934267401695, -0.005146726965904236, -0.01302359625697136, -0.0006737524527125061, 0.0016956995241343975...
[ 0.1415850818157196, 0.045030999928712845, 0.814906656742096, -0.136474609375, -0.06118239834904671, 0.025536123663187027, 0.4910782277584076, 0.33622437715530396, -0.20070989429950714, -0.819502592086792, -0.21353209018707275, 0.8263956308364868, -0.2920942008495331, 0.1049404889345169, ...
I have a couple of measurements for an environmental variable on different locations (7 measurement stations) as time series with averages for every year. For the single locations I have calculated the deviations of single years from the long term average. The annual deviations are not normally distributed as there are trends in the data. I cannot describe these trends as I have not enough knowledge about the influences. Further more the data for the single locations show different trends because of local influences. Also the absolute hight of the measurement value is different for the single locations (by more than 10% in the long term average). To get an idea on how the locations compare to each other I calculated the relative root mean square deviation of the single years from the long term average and the 5th and the 95th percentile of the relative deviations to get an symmetric coverage interval for the uncertainty at the 90% level. My results look as follows: location RMSD[%] P5[%] P95[%] 1 3.8 -5.6 -1.3 2 3.1 -5.1 5.2 3 5.1 -0.6 8.6 4 3.3 -6.2 3.6 5 3.8 -6.7 1.8 6 2.8 0.2 4.0 7 3.7 -6.1 3.9 I want to get an estimate of the expected deviation of a single year from the long term average and the uncertainty of the deviation for an arbitrary location in the region the measurements are taken from. What would be a statistically valid way to do this? I am thinking about using the average over all measurement locations I have, but as the locations are a lot different from another, I am not sure if this is ok.
[ -0.006515152752399445, 0.023179732263088226, -0.020102068781852722, 0.01976456493139267, 0.01839439384639263, -0.01596037670969963, 0.009512858465313911, 0.002188638551160693, -0.012293117120862007, -0.054010942578315735, 0.010834207758307457, 0.020945359021425247, -0.014720825478434563, -...
[ 0.6555262804031372, 0.32402896881103516, 0.2958395183086395, -0.008058237843215466, -0.04101526364684105, 0.31136664748191833, 0.5756686329841614, -0.16835208237171173, -0.41651949286460876, -0.3856261074542999, 0.3031674921512604, 0.12926340103149414, -0.12807230651378632, 0.5760519504547...
Is there an option or possibility to show the exact number of max. and current health of enemies and allies?
[ 0.03199290111660957, 0.016917087137699127, -0.03591509908437729, -0.008391615934669971, 0.027080081403255463, 0.016510751098394394, 0.01380391139537096, -0.04135523736476898, -0.02660912461578846, 0.05263876914978027, 0.0019148624269291759, 0.04176284745335579, -0.02722565457224846, 0.0461...
[ 0.47509491443634033, -0.29431840777397156, 0.31015849113464355, 0.5139352679252625, 0.27801740169525146, 0.24418936669826508, 0.10534112900495529, 0.011872165836393833, -0.20136940479278564, -0.16640062630176544, 0.27016186714172363, 0.5570933222770691, 0.07813884317874908, -0.107078835368...
I have 100 minutes of free talk every week, 25 free SMS every week and 2GB of Internet data usage in a month. Nothing alerts me when one of these thresholds has been exceeded. Is there any way to keep track of my outgoing calls and messages, and then receive an alert when I'm going to exceed my limit? It would be preferable if the threshold would reset automatically every week and be configurable.
[ 0.007764726877212524, 0.0003823612059932202, -0.0232095830142498, 0.01526516955345869, 0.027447355911135674, -0.002195167588070035, 0.009747822768986225, 0.003217731136828661, -0.015676040202379227, 0.006817453540861607, -0.008410453796386719, 0.014622942544519901, 0.016800209879875183, 0....
[ 0.4413297772407532, 0.1712050586938858, 0.7430211901664734, -0.24026553332805634, -0.16065770387649536, -0.03577616438269615, 0.8678122162818909, 0.1313450187444687, -0.5002129077911377, -0.4282416105270386, -0.013798599131405354, 0.49378904700279236, 0.0011296768207103014, -0.015044338069...
I want to write a letter in Hindi. I want to use mikTeX to create the PDF document in Hindi. I don't want to have transliteration from English to Hindi. How can I do this?
[ -0.004423689097166061, 0.010873683728277683, -0.0038726371712982655, 0.029092226177453995, 0.0058430773206055164, -0.007550177630037069, 0.016180939972400665, 0.02056441828608513, -0.04137859493494034, -0.039505042135715485, 0.000511332880705595, 0.0048117004334926605, -0.007697352208197117,...
[ 0.43766698241233826, 0.44492119550704956, 0.06392255425453186, -0.21405968070030212, -0.31551867723464966, 0.11044397950172424, 0.5645272731781006, 0.22028331458568573, 0.24935461580753326, -0.5236636996269226, -0.02270905300974846, 0.1248735785484314, -0.0952610895037651, 0.50918501615524...
What are the most widely used measures of predictive power of attributes in scoring models? **Motivation** : I have a lot of attributes, more than I can study by myself and I want to select somehow the most promising ones. Is IV a good criterion for that? Are there any alternatives?
[ 0.020346984267234802, 0.013017232529819012, -0.011655840091407299, 0.02413724735379219, -0.003462432185187936, 0.003660219255834818, 0.007207866292446852, -0.021621663123369217, -0.03219442814588547, -0.022373169660568237, -0.001150932745076716, 0.028065260499715805, -0.02910071797668934, ...
[ 0.446546345949173, -0.015675632283091545, 0.30837321281433105, 0.44281473755836487, -0.24262204766273499, 0.10996736586093903, 0.2965942323207855, -0.04914506524801254, -0.11575169861316681, -0.4903087019920349, 0.44502562284469604, 0.6889532208442688, 0.09227912873029709, -0.1488942652940...
Ok, by my current understanding of the two words, the sentence: > _The preciseness of this precision is very definite_ Is grammatically correct. Correct me if I'm wrong, and if so; what is the distinction between the two words? Also, what is the word for the situation in which this can occur? (I want to say it's the opposite of an oxymoron?) Edit: feel free to edit the tags if I'm wrong.
[ -0.014172576367855072, 0.00881850067526102, -0.009851508773863316, 0.011223062872886658, -0.01924489624798298, -0.009510201402008533, 0.005806304048746824, -0.0018220626516267657, -0.012733238749206066, -0.014017060399055481, 0.0059336768463253975, 0.01325902994722128, -0.0020512037444859743...
[ 0.20106607675552368, 0.2903522551059723, 0.06389927119016647, -0.07080043852329254, -0.23490163683891296, -0.1782982349395752, 0.5981699824333191, 0.17830997705459595, 0.05647117644548416, -0.5056923031806946, -0.16675913333892822, 0.5624648332595825, -0.12037957459688187, -0.1626684367656...
. . I'm stuck on the if (practice) { perfect; } Section of level 7 of vim adventures. Any suggestions would be appreciated. I believe I have all of the keys that I should have, but I can't figure it out :(
[ 0.00014711050607729703, 0.013679488562047482, -0.006817685440182686, 0.02664969675242901, -0.009609173983335495, 0.0185640100389719, 0.00749829038977623, 0.025546202436089516, -0.028695905581116676, 0.007299968507140875, -0.01794263906776905, -0.00038463177043013275, -0.03603381663560867, ...
[ 0.19654276967048645, 0.1167527288198471, 0.42025044560432434, 0.1492471992969513, 0.3262999653816223, -0.3684009909629822, 0.5470325946807861, -0.1453615128993988, 0.01159878820180893, -0.3669162690639496, -0.012551817111670971, 0.3949716091156006, 0.20224633812904358, 0.1722669005393982, ...
As far as I understand it, it is a big system file that apparently controls a big part of the phone. I am interested because I wanted to do some themeing and so far I had edited SystemUI.apk, but if I ever ran into any problems (and being a beginner that was bound to happen) nothing seemed to happen other than the fact that I lost my notification bar. Problem is, I have heard of some people who touched framework-res.apk and they wound up not being able to even boot their phone. I decided before carelessly poking around with something that could potentially soft brick my phone, I would like to learn a bit more about it. So what exactly is framework-res.apk? What is it used for? What does it control? Where do we visually see its use? What could potentially happen to my phone should something go amiss with framework-res? Anything else I should know before poking my nose where it shouldn't be?
[ -0.004699442535638809, 0.0028594909235835075, 0.017888212576508522, 0.01596720516681671, 0.020707648247480392, -0.02204238623380661, 0.00443560816347599, -0.0020014008041471243, -0.016624953597784042, 0.0015338682569563389, -0.009672027081251144, 0.009176276624202728, 0.003874759655445814, ...
[ 0.47891709208488464, 0.06338204443454742, 0.3664058744907379, 0.22157318890094757, 0.026251820847392082, -0.12267205119132996, 0.4267141819000244, 0.14173321425914764, -0.21753567457199097, -0.4424438178539276, 0.2035568207502365, 0.4004136621952057, -0.02796703763306141, 0.037928823381662...
Ideally i m trying to use my laptop and a 3Gphone as a WiFi router to redirect FORWARD HTTP but not HTTPS Traffic to privoxy which then forwards the traffic via a SSH tunnel to a ziproxy VPS. for the sake of simplicity privoxy is currently set to defaults ie is not forwarding to another proxy. with exception to accept intersepts 1 also sysctl net.ipv4.ip_forward=1 the following iptable commands work locally but is ignored by FORWARD traffic ie users connected by wifi are not filtered by privoxy but the local user is, i want the opposite behaviour iptables -t nat -A POSTROUTING -o ${INTERNET_IFACE} -j MASQUERADE iptables -t nat -A OUTPUT -p tcp --dport 80 -m owner --uid-owner privoxy -j ACCEPT iptables -t nat -A OUTPUT -p tcp --dport 80 -j REDIRECT --to-ports 8118 iptables -A FORWARD -i ${WIFI_IFACE} -j ACCEPT How do I force FORWARD HTTP traffic to go through privoxy ?
[ -0.012917730957269669, 0.0003718094667419791, 0.008526244200766087, 0.02103470079600811, -0.02040875516831875, -0.01842786930501461, 0.009331552311778069, -0.003338986076414585, -0.00869540311396122, 0.000054255127906799316, -0.011778632178902626, 0.005811512470245361, -0.006688183639198542,...
[ 0.0699651911854744, 0.1348433792591095, 0.7673225998878479, -0.16229598224163055, -0.2844082713127136, -0.033747944980859756, 0.1459822654724121, -0.4603744447231293, -0.22956599295139313, -0.7810944318771362, -0.08871190994977951, 0.2686454951763153, -0.32601433992385864, 0.30287575721740...
When talking about distances (miles, kilometers, etc...), is there a word that describes the field that specializes in those sort of things. I remember someone told me about this last week. I Googled it but couldn't find anything. I would also like to know if that (possibly existing) word is popular or not. Edit: I'm sorry about changing the checked answer. Next time I'll make my question clearer.
[ -0.003506266977638006, 0.0038929185830056667, -0.0030388920567929745, 0.016384854912757874, -0.013420878909528255, 0.003625401994213462, 0.004442084114998579, 0.02818157710134983, -0.01699935272336006, 0.009596727788448334, 0.00013917677279096097, 0.01698807254433632, 0.03495379537343979, ...
[ 0.6663084030151367, 0.08124325424432755, 0.10930872708559036, 0.2666890621185303, -0.19878150522708893, -0.25333982706069946, 0.07717961817979813, 0.6209313273429871, -0.34489020705223083, -0.34273549914360046, 0.18955962359905243, -0.0968964546918869, 0.19155468046665192, 0.17869578301906...
Go and D provide garbage collection, and yet they claim to be system programming languages. What degree of low-level programming can be achieved with languages having garbage collection? For low-level programming, I mean close to the hardware or being able to: 1. Runs directly in limited memory, with no latency, and performs well. An example would be operating system kernels. 2. It runs on a software base, but still has to perform well. An example would be system utilities.
[ -0.0023709775414317846, 0.009625240229070187, -0.00708241481333971, -0.0016585118137300014, -0.0032536399085074663, -0.016327708959579468, 0.010970017872750759, 0.0077584185637533665, -0.016435934230685234, 0.009201358072459698, -0.024754052981734276, 0.015169598162174225, 0.0059971138834953...
[ 0.28079918026924133, -0.16706730425357819, -0.323114275932312, 0.40187007188796997, -0.25091952085494995, -0.4387044608592987, 0.06397693604230881, -0.011234338395297527, 0.04144144430756569, -0.5593151450157166, -0.35518679022789, 0.5153374671936035, -0.18140482902526855, -0.2679125666618...
> **Possible Duplicate:** > How to find web hosting that meets my requirements? We're looking for a good place to host our custom Django app (a fork of OSQA) and its postgresql backend. Requirements include: * Linux * Python 2.6 or (ideally) Python 2.7 * Django 1.2 * Postgres 8.4 or later * DB backup/restore handled by the hoster, not us * OS & dev-platform-stack patching/maintenance handled by the hoster, not us * SSH access (so we can pull source code from GitHub, so we can install python eggs, etc.) * ability to set up cron jobs (e.g. to send out dail email updates) * ability to send up to 10K emails/day * good performance (not ganged up with a zillion other sites on one CPU, not starved for RAM) * FTP or SCP access to web logs * dedicated public IP * SSL support * Costs under $1000/month for a relatively small site (<5M pageviews/month) * Good customer service We already have a prototype site running on EC2 on top of a Bitnami DjangoStack. The problem is that we have to patch the OS, patch postgres, etc. We'd really prefer a platform-as-a-service (PaaS) offering, like Heroku offers for Rails apps, where all we need to worry about is deploying our code instead of worrying about system software patching and maintenance. Google App Engine is closest to what we're looking for, but they don't offer relational DB access (not yet at least). Anyone have a recommendation?
[ -0.0014201016165316105, -0.0030507310293614864, 0.0002151448279619217, -0.0018966589123010635, -0.019834019243717194, 0.018475741147994995, 0.008460422977805138, 0.04420384764671326, -0.020343109965324402, -0.010660097934305668, -0.008778010495007038, 0.015298860147595406, 0.0336492583155632...
[ 0.5703413486480713, 0.5023960471153259, 0.4181883633136749, 0.01658615842461586, 0.01693425513803959, -0.25350698828697205, 0.20195721089839935, -0.19742119312286377, -0.09733270108699799, -0.6365206837654114, -0.22594769299030304, 0.45857974886894226, -0.124348945915699, 0.061868805438280...
I'm looking for new Wiimotes, for the legacy Wii. My goal is to buy the 3rd and 4th Wiimotes for my console. These devices will be primarily used to play some games with children, like Mario, not high-paced games. I can see a lot of "compatible" devices that are neither official nor built by Nintendo. What is the quality of these alternative devices compared to the official ones? Do these devices operate like an official one? Should I expect the same reactivity, precision, etc.? I assume the global finish of the product may give a cheaper feeling, but I can live with that.
[ -0.00314547261223197, -0.018611546605825424, -0.005842872895300388, 0.012168669141829014, -0.023200441151857376, 0.004986397922039032, 0.007921604439616203, 0.021929385140538216, -0.013241940177977085, -0.03425823152065277, -0.00566416559740901, 0.010781772434711456, 0.022614413872361183, ...
[ 0.41376522183418274, -0.03983807563781738, 0.6000580787658691, 0.4310849905014038, 0.03564106300473213, 0.05691782385110855, -0.1436925083398819, 0.09700670093297958, -0.2939819097518921, -0.42296648025512695, 0.053719744086265564, 0.30841636657714844, 0.17097477614879608, 0.48851886391639...
I just rooted by AT&T Galaxy S2 (SGH-i777) to troubleshoot battery usage. I also would like to tether my laptop (OS X Lion) to make use of my phone's data connection intermittently while traveling. I see both "USB tethering" as well as "portable WiFi hotspot" under `Settings --> Wireless and network --> Tethering and portable hotspot`. Should I use this feature or install an additional tethering application? What [if any] are the advantages of a separate tethering app over the built-in tethering controls?
[ -0.008643293753266335, 0.00464570801705122, -0.00007919617928564548, 0.00335650029592216, -0.030181333422660828, -0.008061015047132969, 0.007732695899903774, 0.019011225551366806, -0.017907924950122833, 0.0012631568824872375, -0.009645478799939156, 0.0031232016626745462, -0.01472288277000188...
[ 0.510842502117157, 0.19796109199523926, 0.5799609422683716, 0.27251631021499634, 0.14784792065620422, -0.109977126121521, 0.11316542327404022, -0.14964792132377625, -0.35445457696914673, -0.2246205061674118, 0.23591488599777222, 0.47151023149490356, -0.14036792516708374, 0.0783129930496215...
Is it possible to do model selection in this way? Suppose I need to select a good (logistic) model among three variables (var1, var2, var3). The deviance D* (-2*log-likelihood) of this full model would be the minimum among all possible models. Then I could try all 6 combination of sub- models(1,2,3,12,13,23) and compute their deviance D1~D6. Next I compute the difference: deltaD_i=D_i-D*, this should follow chi-square distribution with df=differences_in_variables_numbers. The models with deltaD within 95% confidence interval of D* would be within the confidence interval of the full model, that is, the variance explained by the reduced model is not significantly different than the full model. Then we could accept these model as good models. By doing this, we could end of with several "good" models. Is this somehow a possible way to do model selection? Thanks
[ 0.007936648093163967, 0.025779057294130325, -0.0070133693516254425, -0.004613159690052271, -0.011111280880868435, -0.0011559659615159035, 0.0068038091994822025, -0.0006383983418345451, -0.011810427531599998, -0.03466110676527023, -0.012581553310155869, 0.015261194668710232, -0.01617406494915...
[ -0.17960596084594727, -0.20215576887130737, 0.21856330335140228, -0.09977614134550095, -0.15429426729679108, 0.8468957543373108, 0.3055160641670227, -0.599839985370636, 0.04447269067168236, -0.557167112827301, 0.20011673867702484, 0.5124825835227966, -0.16426068544387817, 0.399573057889938...
> The ground was ice-cold, no hint of anyone having lay/laid/lain there at > all. Which one is the correct option?
[ -0.01686706207692623, 0.04335422441363335, -0.04677654802799225, 0.005979897920042276, 0.07393255829811096, 0.03130379691720009, 0.02036113850772381, 0.012317726388573647, -0.025452720001339912, 0.04858054220676422, -0.022527523338794708, -0.00047445212840102613, 0.03475680202245712, 0.020...
[ 0.2221541702747345, -0.0300439540296793, 0.006421325262635946, 0.2830558717250824, -0.1602085381746292, -0.06402695178985596, 0.25085416436195374, -0.14514194428920746, -0.22195924818515778, -0.5145187973976135, 0.00377988675609231, 0.33535444736480713, -0.2466173619031906, -0.187085956335...
I've built a pretty good/enormous first stage rocket that's pretty reliable, and I'd like to use it to put different payloads in orbit. The only problem is I can't seem to switch out my command module. Is there any way to do this, or to save my first stage in such a way that I can import it into different builds?
[ 0.02047128975391388, 0.008970005437731743, -0.010102258063852787, 0.013486551120877266, -0.02933293581008911, 0.0014407620765268803, 0.008033891208469868, -0.003657190827652812, -0.023719647899270058, -0.015545199625194073, 0.0015008611371740699, 0.01575854793190956, 0.006928201299160719, ...
[ 0.4506111145019531, 0.42464616894721985, 0.14356926083564758, 0.2710244655609131, 0.3546988070011139, 0.06232791393995285, -0.1086416244506836, 0.0764617770910263, -0.07475407421588898, -0.1013866439461708, 0.1391633152961731, 0.5363720655441284, 0.17974026501178741, -0.05062950775027275, ...
I have a formula which is producing a large (hundreds of elements) list of complex numbers of the simple form $x+iy$. However I'm only interested in the one value which satisfies the conditions: 1. The imaginary part is the _least negative_ value in the list 2. The real part is positive. What can write that will pick out this value and display it for me?
[ 0.015853675082325935, 0.0035971486940979958, -0.01889108493924141, -0.00124883942771703, -0.019185304641723633, 0.01445899810642004, 0.008631892502307892, 0.02365030348300934, -0.022053200751543045, 0.004254637751728296, -0.025741007179021835, -0.006255115382373333, -0.0017049215966835618, ...
[ 0.2703086733818054, 0.45314133167266846, 0.2717849910259247, -0.035762716084718704, -0.17019183933734894, 0.20276714861392975, 0.18253354728221893, 0.01159114483743906, -0.09557248651981354, -0.3404029905796051, 0.12415758520364761, 0.3632396459579468, -0.5097259283065796, 0.45077803730964...
I want to transmit sunlight over a distance of approx 4 m via optical fiber. Ideally I want to light up the whole house with sunlight. I was reading that the largest diameter for an optical fiber may not be the right way to solve this issue. I was wondering if there some sort of equation where I can insert the length that I would like to carry light over (length is a variable) and maybe the number of cores of optical fiber (also a variable) with the output of this equation being the efficiency or loss in transmission. This way I can integrate all variables in one equation and optimize against the cost to procure.
[ 0.004979315679520369, 0.008282494731247425, -0.017370086163282394, 0.000701748242136091, -0.0200646985322237, -0.02666076458990574, 0.008619126863777637, -0.0037448108196258545, -0.01550729013979435, -0.05053446441888809, -0.0013104971731081605, 0.014581388793885708, -0.014885952696204185, ...
[ 1.0188490152359009, 0.05461553856730461, 0.6736770272254944, -0.21134234964847565, -0.2024364322423935, 0.01808537170290947, 0.4335176348686218, -0.4498990774154663, -0.41287893056869507, -0.6886624097824097, 0.2800463140010834, 0.42719006538391113, -0.3774682283401489, 0.2927466630935669,...
I have a created a custom post type and have attached some custom fields to it. Now I would like for the search that authors can perform on the custom post list screen (in the admin backend) to also be performed on the meta fileds and not only look in the title and content as usual. Where can I hook in and what code I have to use? **Example image** ![enter image description here](http://i.stack.imgur.com/30leQ.png) Stefano
[ 0.00686107249930501, 0.0005649392260238528, 0.0035601495765149593, 0.009706079959869385, 0.017390871420502663, 0.012778266333043575, 0.006532459054142237, 0.006584052462130785, -0.01593805104494095, 0.0040031434036791325, -0.004498089663684368, 0.010812553577125072, 0.0038844202645123005, ...
[ 0.6570410132408142, 0.048754241317510605, 0.4780595600605011, 0.011947325430810452, -0.08826988935470581, -0.09819532185792923, 0.002233340870589018, -0.19108547270298004, -0.3253607153892517, -0.50835782289505, -0.2106412649154663, 0.48875391483306885, -0.17927895486354828, 0.018270995467...
I have a Fedora 14 x86_64 iso image and I want to install it using a USB stick. How do I get this stick to boot up, and use the image to start the installation? I'm running Debian.
[ -0.02632651850581169, 0.008549410849809647, -0.022434847429394722, -0.004287788178771734, -0.04369470849633217, -0.009794398210942745, 0.016092993319034576, 0.0013554535107687116, -0.019540617242455482, -0.003161344211548567, -0.016239330172538757, -0.01125116366893053, -0.03520089387893677,...
[ 0.8680613040924072, 0.28058508038520813, -0.014312191866338253, 0.02986197918653488, 0.06540369242429733, -0.3979265093803406, -0.032492745667696, 0.26587164402008057, -0.044907864183187485, -0.9174569845199585, 0.10429687052965164, 0.5739783644676208, -0.599634051322937, 0.215880721807479...
I'm running Skyrim on a PS3, fully updated (2.5). However, I can't seem to trigger the new kill cinematics I've heard of. My character level is 78, and archery skill is 100, but I still haven't seen any of the new animations. Is there something particular I need to do to trigger them? What might prevent the animations from being triggered?
[ 0.018229134380817413, 0.011954579502344131, -0.020463034510612488, -0.010659392923116684, 0.012668301351368427, -0.03645767271518707, 0.00818123109638691, -0.016724400222301483, -0.0226590633392334, 0.001478256075643003, -0.017519941553473473, 0.01825910434126854, -0.00011480839748401195, ...
[ 0.4171874523162842, -0.012539965100586414, 0.6080107688903809, -0.09612336754798889, -0.27351537346839905, 0.003776191733777523, 0.3572777211666107, -0.31257399916648865, -0.21304604411125183, -0.47578758001327515, 0.07928261905908585, 0.43078142404556274, 0.17753289639949799, -0.191196456...
i have table with latitude & longitude columns in sql. how can i select all records with lat-long 60 degree north> ( Arctic Region ) Thanks Pragnesh
[ 0.033844076097011566, 0.029451142996549606, -0.008086254820227623, 0.0013244602596387267, -0.01946094073355198, 0.046086523681879044, 0.016645895317196846, 0.025744931772351265, -0.020138995721936226, 0.003320413874462247, -0.0014566463651135564, 0.035064589232206345, -0.04675920680165291, ...
[ 0.15066027641296387, 0.36049821972846985, 0.39582008123397827, 0.46454837918281555, -0.056072816252708435, -0.24041879177093506, 0.4441080093383789, 0.3952599763870239, -0.2600235342979431, -0.4931030571460724, -0.0207807794213295, -0.3082161545753479, 0.31822827458381653, 0.42097097635269...
I have been trying a way round my latest WordPress problem all morning and I feel now is the right time to shout and ask for help. I have been trying to create a people directory, I'm there except for the overall listing page which I have decided to be rows of 5 images followed by the persons name below each image, followed by another a row etc... Each persons page has a set of custom fields one of which is `name` and another is `mainimage` the idea being that with the following code I can pull the image along with the name from the child pages of `people` and display them all on the one page with the image and name linked to the permalinkof the specific child page. <?php // Get the page's children $args = array( 'numberposts' => 100, 'child_of' => 35, 'post_type' => page, ); $children = get_pages( $args ); if (!empty($children)) { echo ''; foreach($children as $child) { // Get the 2 meta values from the child page $main_image = get_post_meta($child->ID, 'main_image', true); $name = get_post_meta($child->ID, 'name', true); // Display the meta values echo '<br />'; echo $main_image; echo '<br />'; echo $name; } echo ''; } ?> It works kind of, but I'm getting a number for the image as opposed something that outputs to an image. Upon checking the custom field I have found the image value is a number, looks to me like media ID. I need to either be able to interpret the ID within the people page as an image or a way to make sure actual urls are being put into the custom field values so I can use standard `img src` html to output the image. I am using Advanced Custom Fields to create the fieldsets for the people page as well as the individual person pages. Any help would be greatly appreciated as I have tried several variations of the code above but just cannot get the number to change in to the associated image.
[ -0.0033077728003263474, 0.00736432708799839, -0.010822131298482418, 0.0165487602353096, 0.0016389562515541911, 0.0013296930119395256, 0.005756507162004709, 0.01732528954744339, -0.012955092824995518, 0.005898267030715942, -0.011068987660109997, 0.00434235529974103, 0.005488811060786247, 0....
[ -0.02901824191212654, 0.087281733751297, 0.4039807915687561, 0.19563017785549164, -0.031086934730410576, 0.3076067566871643, 0.2922004759311676, 0.22461511194705963, -0.21508736908435822, -0.7271441221237183, 0.06550101190805435, 0.33201321959495544, 0.017021963372826576, 0.281372785568237...
Why is it that divergent series make sense? Specifically, by basic calculus a sum such as $1 - 1 + 1 ...$ describes a divergent series (where divergent := non-convergent sequence of partial sums) but, as described in these videos, one can use Euler, Borel or generic summation to arrive at a value of $\tfrac{1}{2}$ for this sum. The first apparent indication that this makes any sense is the claim that the summation machine 'really' works in the complex plane, so that, for a sum like $1 + 2 + 4 + 8 + ... = -1$ there is some process like this: ![enter image description here](http://i.stack.imgur.com/slTT9.png) going on on a unit circle in the complex plane, where those lines will go all the way back to $-1$ explaining why it gives that sum. The claim seems to be that when we have a divergent series it is a non- convergent way of representing a function, thus we just need to change the way we express it, e.g. in the way one analytically continues the Gamma function to a larger domain. This claim doesn't make any sense out of the picture above, & I don't see (or can't find or think of) any justification for believing it. Futhermore there is this notion of Cesaro summation one uses in Fourier theory. For some reason one can construct these Cesaro sums to get convergence when you have a divergent Fourier series & prove some form of convergence, where in the world does such a notion come from? It just seems as though you are defining something to work when it doesn't, obviously I'm missing something. I've really tried to find some answers to these questions but I can't. Typical of the explanations is this summary of Hardy's divergent series book, just plowing right ahead without explaining or justifying the concepts. I really need some general intuition for these things for beginning to work with perturbation series expansions in quantum mechanics & quantum field theory, finding 'the real' expanation for WKB theory etc. It would be so great if somebody could just say something that links all these threads together.
[ -0.01878402754664421, 0.013643170706927776, -0.009667232632637024, 0.007102474104613066, -0.0055261459201574326, -0.01629333570599556, 0.009496990591287613, -0.006564770825207233, -0.01808425784111023, -0.030268386006355286, 0.0031198631040751934, 0.010482588782906532, 0.0016220726538449526,...
[ 0.12360228598117828, -0.25170764327049255, 0.13110210001468658, 0.11220664530992508, 0.08511581271886826, 0.06235663965344429, -0.015242562629282475, -0.15816886723041534, -0.04910026863217354, -0.2084125429391861, -0.0034511385019868612, 0.3747965693473816, -0.353833943605423, 0.241528466...
I installed a Gameloft game Brother in Arm.. But the graphics of the game is shown as below. Use 2.3.3 on Samsung S2 Any idea why its not being displayed ? ![enter image description here](http://i.stack.imgur.com/ykzGz.jpg)
[ -0.022522302344441414, -0.01973939687013626, -0.010360250249505043, 0.01708308234810829, -0.005136226769536734, -0.02051137387752533, 0.00883509311825037, 0.04635888710618019, -0.024502959102392197, -0.010055100545287132, -0.03631757199764252, 0.007574963383376598, 0.00676538422703743, 0.0...
[ -0.30848875641822815, -0.14697769284248352, 0.760223925113678, 0.06617096066474915, 0.0032553570345044136, -0.381560355424881, 0.45136985182762146, -0.2688305675983429, -0.405030757188797, -0.3537186086177826, -0.05196571350097656, 0.42416030168533325, -0.28419652581214905, -0.099811941385...
I just wanted to know how to do **_Heteroscedasticity Test on a Univariate Model_**? * ex: an univariate autoregressive model * ex: an univariate ARCH/GARCH model If it is possible, how does one do that in `R`?
[ 0.01594146527349949, 0.021479545161128044, -0.00044786359649151564, 0.00998353399336338, -0.020549211651086807, -0.005125789437443018, 0.012082002125680447, 0.008144374936819077, -0.030612200498580933, -0.016045432537794113, -0.008408808149397373, 0.0023977821692824364, -0.00746855279430747,...
[ 0.09260797500610352, 0.13362064957618713, 0.10749919712543488, -0.020360121503472328, -0.20163264870643616, 0.26866358518600464, 0.2958372235298157, -0.007871744222939014, -0.00313773425295949, -0.20610950887203217, 0.09935776144266129, 0.2482217699289322, -0.36256545782089233, 0.165565848...
I can bring images from the web using the Import command like this: Import["http://blog.wolfram.com/wp-content/uploads/2008/06/se-30.jpg"] ![good image](http://i.stack.imgur.com/qlxsY.png) Sometimes the web request is more complicated so I must to use URLFetch. For that, this is what I have tried to do: ImportString[ URLFetch["http://blog.wolfram.com/wp-content/uploads/2008/06/se-30.jpg"], "JPG"] ![unreadable image](http://i.stack.imgur.com/jtH3X.png) Do you know how to make the image load correctly? I suspect that this is related to strings encodings, but I have not been able to solve it.
[ -0.005824430380016565, 0.004828820005059242, -0.011730487458407879, 0.0006117952289059758, -0.030161142349243164, -0.004462084732949734, 0.007831879891455173, -0.0015501156449317932, -0.015823673456907272, -0.011344125494360924, -0.005830076523125172, 0.0007646710146218538, 0.003883222816511...
[ 0.3830278515815735, 0.3092271089553833, 0.5344885587692261, -0.26284077763557434, -0.42181363701820374, -0.05809397250413895, 0.293894499540329, -0.029314184561371803, -0.49688392877578735, -0.6980172395706177, -0.05123389512300491, 0.5279196500778198, -0.19008642435073853, 0.1063311323523...
new Float type: Illustration How to make a list of illustrations, with captions in the list, but for the illustrations to appear in the text with zero captioning - image only.
[ 0.008090726099908352, 0.006798809394240379, -0.014683859422802925, 0.04016192629933357, -0.04028254747390747, 0.00699659064412117, 0.01441139169037342, 0.03677745535969734, -0.03863490745425224, 0.057357873767614365, -0.05105084925889969, -0.012597288005053997, -0.011808828450739384, 0.012...
[ 0.5540916919708252, 0.09008144587278366, 0.8175389766693115, 0.5273950695991516, -0.30248191952705383, -0.19983090460300446, -0.36579760909080505, -0.5935379266738892, -0.3506206274032593, -0.4299026131629944, 0.20616808533668518, 0.3778337836265564, 0.011771087534725666, -0.25292670726776...
I have 100 compressed pictures. Each of them is 2 MB. I would like to send via email everything at once as attachments (no hosting on Gdrive or somewhere else). Basically pictures will have to be sent in several emails. Is there any way to do it at once? To put it otherwise, I'm asking for a way to automatically send multiple email messages to send all the files. Ideally not one email per file, but each email should contain as many pictures as possible based on a maximum that I would specify. I use a Samsung Galaxy S3. I know the max size of the recipient's mailbox and pictures cannot be further compressed. I don't mind filling my recipient's mailbox.
[ 0.00717167416587472, -0.005003585945814848, -0.00718139111995697, 0.020681627094745636, 0.03286493569612503, -0.016498154029250145, 0.006375804077833891, 0.007716134190559387, -0.015691932290792465, 0.006664087530225515, -0.012733940966427326, 0.008643753826618195, 0.009205702692270279, 0....
[ 0.17278626561164856, 0.17191864550113678, 0.6824864745140076, 0.15570490062236786, -0.1264985352754593, 0.48862648010253906, -0.18454986810684204, -0.362363338470459, -0.5571919083595276, -0.7701069116592407, 0.15589621663093567, 0.3693561851978302, -0.33347323536872864, -0.018807325512170...
I have a general, presumably simple, question, but I couldn't find a conclusive answer so far. Assume I have a simple case of a General Linear Model with one categorical predictor variable, that has 3 levels. This corresponds to a one-way ANOVA with 3 groups. The Linear Model would then be: $Y_i = \beta_0 + \beta_1X_1 + \beta_2X_2$ I use the following standard Helmert contrast codes (I, II) for the three groups A, B, C: A B C I -2 1 1 II 0 -1 1 Based on these contrast codes, **will the intercept of the linear model represent the mean of all group means? More generally, does this always apply to Helmert contrast coded variables,** regardless of the number of predictor variables or variable levels? Obviously, the intercept represents the predicted value, when all predictor variables (in this case $X_1$ and $X_2$) equal zero, but I can't figure out if that represents the mean of the group means.
[ 0.009677538648247719, 0.02534465491771698, -0.010311741381883621, 0.01503516361117363, 0.016679270192980766, 0.0010294315870851278, 0.00837987381964922, 0.002081954851746559, -0.006600236985832453, 0.00008390448056161404, -0.008952871896326542, 0.007848644629120827, -0.0029406992252916098, ...
[ 0.01926039531826973, 0.07580914348363876, 0.17632684111595154, -0.18842510879039764, -0.20689162611961365, 0.5036930441856384, 0.4740956425666809, -0.6751850843429565, 0.0886160284280777, -0.5686464309692383, -0.014586836099624634, 0.2787277400493622, -0.5336108803749084, 0.272075295448303...
1. I've read in different textbooks that say Linux is light-weight(e.g. It could fit on a 1.4MB floppy). So why is the download from the Ubuntu or Fedora CD sized or larger? 2. Do the device drivers extend the kernel? For example: if I have new hardware and I have installed the device driver, will my kernel code get extended or is the driver installed as a service for the kernel to use? 3. When using a LiveCD such as Ubuntu, when system boots does all 700MB of the OS get loaded to RAM or just parts of it? I ask these questions because I feel they're common beginner questions and I think it would be good to have them all in one place.
[ -0.00898912362754345, -0.0052883862517774105, -0.009388871490955353, 0.016649454832077026, -0.0206168070435524, -0.007567834109067917, 0.008642788045108318, 0.010538224130868912, -0.010949214920401573, -0.03889816999435425, -0.019271980971097946, 0.011034408584237099, 0.0122238639742136, -...
[ 0.27904772758483887, -0.12091568857431412, 0.31599995493888855, 0.03536355495452881, -0.0009092476102523506, -0.2567083537578583, -0.3451037108898163, -0.2935875952243805, -0.37682726979255676, -0.64372718334198, -0.0809115320444107, 0.7415117025375366, -0.15605013072490692, -0.08985280245...
All In my application there is notification system, when user click on that icon I want to make ajax call. The problem id it works fine for admin user (Debug : 200 ok), but not for subscriber user (Debug : 301 moved permanently). Ajax call $("#notifications-button").click(function() { $.ajax({ type : 'POST', url: '<?php echo get_option('siteurl') . '/wp-admin/admin-ajax.php' ?>', data:'action=my_special_ajax_call&value=1', success : function(data){ $('.message-count').hide(); }, }); }); Function function implement_ajax() { if(isset($_POST['value'])) { global $wpdb, $user_ID; $wpdb->query( $wpdb->prepare("UPDATE wp_frm_items SET alerts_flag = 0 WHERE user_id = '". (int)$user_ID ."'" )); } } add_action('wp_ajax_my_special_ajax_call', 'implement_ajax'); add_action('wp_ajax_nopriv_my_special_ajax_call', 'implement_ajax');//for users that are not logged in. I do not able to get why this not working for subscriber? Is there anything wrong in code? Thanks In Advance !
[ -0.012324831448495388, 0.012228458188474178, 0.010187831707298756, 0.0009058685973286629, -0.00954638421535492, -0.009039850905537605, 0.008239846676588058, 0.024744439870119095, -0.01569872349500656, 0.002952336333692074, -0.01121340598911047, 0.009994527325034142, -0.003470170311629772, ...
[ -0.01768028363585472, 0.2358044981956482, 0.673025369644165, -0.2140965610742569, 0.03149958327412605, 0.007415951695293188, 0.41468721628189087, -0.08374043554067612, -0.17788414657115936, -0.7741819620132446, -0.03445469215512276, 0.6132107377052307, -0.30618539452552795, 0.0519365333020...
Is there a way to block indent whole paragraphs in `emacs` \+ `auctex`? I'm thinking in that functionality in such editors as `WinEdt`. I've search in different fora, and in the available documentation, but what I've found doesn't seem to work in `LaTeX` mode.
[ 0.006563328206539154, -0.00030391800100915134, -0.017074143514037132, 0.020698631182312965, 0.012070286087691784, 0.0018099644221365452, 0.007465189788490534, -0.005793234799057245, -0.025498313829302788, 0.007721708621829748, -0.0120621919631958, 0.0084911547601223, -0.004881696775555611, ...
[ 0.1189916804432869, 0.27681341767311096, 0.09831474721431732, 0.09930995851755142, -0.07616482675075531, -0.07879067212343216, 0.259878933429718, 0.38545453548431396, -0.13216497004032135, -0.4633627235889435, -0.18077906966209412, 0.28680703043937683, -0.19768217206001282, 0.1830333024263...
I have never registered for Quantcast and placed Quantcast code on my site. Still, on many website value checker sites I found some number as Quantcast for my site. Does this rank really matter for a SEO? If yes, then is there any powerful method get a better ranking?
[ 0.01766410656273365, 0.01062335167080164, 0.006934286095201969, 0.015818972140550613, 0.013617245480418205, 0.021794816479086876, 0.013173576444387436, 0.018674449995160103, -0.03206975385546684, -0.008714761584997177, -0.00020876839698757976, 0.0106410076841712, -0.027458397671580315, 0.0...
[ 0.30160996317863464, 0.029536889865994453, 0.61970454454422, 0.20359568297863007, -0.3355425000190735, -0.298836886882782, 0.16662456095218658, -0.11828113347291946, -0.23957081139087677, -0.2595860958099365, 0.29509851336479187, 0.2833380401134491, 0.13465218245983124, -0.1418089568614959...
I am trying to find the hook for adding a filter to the list under "Posts [Add New]": ![enter image description here](http://i.stack.imgur.com/2xDQ3.png)
[ -0.022995349019765854, 0.004446263890713453, 0.0033779607620090246, 0.0237656868994236, 0.02344091795384884, 0.001462537795305252, 0.009555038996040821, 0.022170716896653175, -0.026292694732546806, 0.02560548670589924, -0.017097655683755875, 0.007829675450921059, -0.025500427931547165, 0.0...
[ 0.5089803338050842, -0.006344877649098635, 0.24594305455684662, -0.18274104595184326, -0.3539433777332306, -0.2463660091161728, 0.2326255440711975, 0.24968621134757996, -0.39497265219688416, -0.7300146222114563, 0.18315233290195465, 0.10333938151597977, -0.30012211203575134, 0.178521037101...
I compile the file below with `lualatex --shell-escape` but I get the error : ! Undefined control sequence. l.14 \savedata {\data}[{0,0},{0.1,0.319802645},{0.2,0.6510256205},{0.3,0.993869... ? ! Emergency stop. l.14 \savedata {\data}[{0,0},{0.1,0.319802645},{0.2,0.6510256205},{0.3,0.993869... 298 words of node memory still in use: 2 hlist, 1 rule, 1 kern, 6 attribute, 41 glue_spec, 6 attribute_list, 1 write , 3 special, 1 dir, 1 pdf_colorstack nodes avail lists: 2:16,3:1,4:4,5:1,6:7,7:1,9:2 ! ==> Fatal error occurred, no output PDF file produced! * * * \RequirePackage{ifluatex} \documentclass{article} \ifluatex \usepackage{fontspec} \setmainfont{TeX Gyre Pagella} \else \usepackage{tgpagella} \usepackage{pst-plot} \definecolor{CyanTikz40}{cmyk}{.4,0,0,0} \fi \usepackage{auto-pst-pdf} \begin{document} \savedata{\data}[{0,0},{0.1,0.319802645},{0.2,0.6510256205}, {0.3,0.993869283},{0.4,1.3485339894},{0.5,1.7152200962},{0.6,2.0941279602},{0.7,2.4854579379},{0.8,2.8894103862},{0.9,3.3061856616},{1,3.7359841208},{1.1,4.1790061205},{1.2,4.6354520174},{1.3,5.1055221682},{1.4,5.5894169295},{1.5,6.087336658},{1.6,6.5994817104},{1.7,7.1260524433},{1.8,7.6672492135},{1.9,8.2232723776},{2,8.7943222922},{2.1,9.3805993141},{2.2,9.9823037999},{2.3,10.5996361063},{2.4,11.23279659},{2.5,11.8819856076},{2.6,12.5474035159},{2.7,13.2292506714},{2.8,13.9277274309}] \psset{yAxisLabel=Volume d'eau en L,xAxisLabel=Niveau d'eau en dm,linewidth=1pt,xAxisLabelPos={2.5,-2.5},xticksize=0 3.9375, yticksize=0 9,tickcolor=CyanTikz40,tickwidth=1pt,arrowscale=1.2} \begin{psgraph}[Ox=0,Dy=2,Dx=0.5]{->}(0,0)(3,16){9.cm}{4.5cm} \listplot[linewidth=1pt]{\data} \end{psgraph} \end{document}
[ 0.013369282707571983, 0.006846493110060692, -0.00036856913357041776, 0.0055893827229738235, 0.026665357872843742, 0.013652252033352852, 0.0045519499108195305, 0.0036207097582519054, -0.010561124421656132, -0.0056014284491539, -0.008731560781598091, 0.0014949878677725792, -0.00855585373938083...
[ -0.17687393724918365, 0.02714364044368267, 0.2758646309375763, -0.353372186422348, 0.5365372896194458, 0.2848529517650604, 0.42117419838905334, -0.30498337745666504, -0.13225194811820984, -0.39526641368865967, -0.06334934383630753, 0.8797476887702942, -0.4572940468788147, 0.278902918100357...
At home I'd like to configure one of my wifi routers to serve as a WI-FI signal extender. Since my main router is OpenWrt-based (Backfire 10.1.3), I'd like to use it for that purpose. I've looked into the problem and found that there is "Access Point (WDS)" option available in wifi interface setup. I thought this looks promising, especially since I've managed to configure my other router in that mode, too. However switching that option "on" doesn't do a thing. I've tried manually adding BSSID of the other side (WDS client) to /etc/config/wireless, but it didn't work. I still don't know where to configure security settings for the networks to pair (my network is WPA2 encrypted). The modem that would become a WDS client isn't running OpenWRT and I'm unable to get it installed there, so I'm trying with its current OS. But still, the main issue is with configuring OpenWRT side. Any ideas? EDITED: Found this two links: http://wiki.openwrt.org/doc/recipes/atheroswds https://forum.openwrt.org/viewtopic.php?id=11849 The first one gives not too many information, the second is rather outdated. Doesn't seem to work :(
[ -0.00008044613059610128, -0.009691845625638962, -0.012179505079984665, 0.011444099247455597, -0.003298424882814288, -0.0017171634826809168, 0.005488578230142593, 0.005809204652905464, -0.0074254488572478294, 0.0012908042408525944, -0.007936772890388966, 0.009671824052929878, -0.0093833338469...
[ 0.46727603673934937, 0.1147775799036026, 0.45438113808631897, 0.0748874694108963, 0.20349068939685822, -0.10837496072053909, 0.21715626120567322, -0.20938126742839813, -0.09442086517810822, -0.9277446269989014, -0.090584896504879, 0.43219101428985596, -0.12497982382774353, -0.0182241592556...
I have built a beta regression model with log link for predicting adherence. My dependent variable's range is 0 to 1.When I used a test set to calculate the predicted values with the parameter estimates from this model, the values were out of range (>1 and max was 2.3). Is it wrong if I scale them to 0 to 1 range? It does nt really alter the distribution graph much. What should I do if the predicted values are out of range in general? **Edit** : The Code I am using proc glmmix (sas). proc glimmix data=_exp0_.Beta_training; class gndr_cd medispan_drug_class_grp_desc ecom ; model days_adhered=norm_age gndr_cd drug_type norm_days_until_refill total_fills/ dist=beta solution link=log s; random residual; nloptions technique=quanew update=bfgs; run;
[ 0.014719652943313122, 0.02110023982822895, -0.012622809037566185, 0.0009686921257525682, -0.00856570154428482, -0.012385640293359756, 0.00829099677503109, -0.004514224361628294, -0.012366365641355515, -0.04051808640360832, -0.002889853436499834, 0.008998047560453415, -0.011828359216451645, ...
[ 0.3048955798149109, -0.06979372352361679, 0.6334664821624756, -0.1264323741197586, -0.19507473707199097, -0.15728095173835754, 0.6380017399787903, -0.5073826909065247, -0.1698196977376938, -0.40249133110046387, 0.2586940824985504, 0.6442089676856995, -0.3399587571620941, 0.1946527361869812...
I am running a post-hoc analysis on the data collected during an experiment in which 15 unique stimuli were presented to participants. Having run a least squares regression using the lm() function in R I have found significant results for a subset of the data including 90 observations from 6 participants with two continuous variables and their interaction. Taking advice from an article by Judd, Westfall & Kenny (2012) I attempted to use a combination of the lmer() function found in the lme4 package in combination with a Kenward-Roger approximation through the KRmodcomp() function in the pbkrtest package (see the appendix in the article) in order to control for random effects: lmer(Prediction_Difference_Scale~Diff_AWD_LRTI_End_Scale*Diff_AWD_BD_End_Scale + (1|Unique_ID) + (Diff_AWD_LRTI_End_Scale*Diff_AWD_BD_End_Scale|Block),data=Data) The first variable after the DV is the fixed effect, the second variable in parentheses indicates that the intercept is random with respect the unique stimuli (Unique_ID) and the third variable in parentheses indicates that both the intercept and the Condition slopes are random with respect to participant (Block) and that a covariance between the effects should be estimated. When running the lmer() function I get the following error message: Error in checkNlevels(reTrms$flist, n = n, control) : number of levels of each grouping factor must be < number of observations This is obviously because the number of observations equal the number of unique stimuli. The function works when excluding the (1|Unique_ID) random effect, which if I understand correctly is the same as carrying out a 'by stimulus' analysis. However, the authors warn against this by stating: "Conceptually, a significant by-participant result suggests that experimental results would be likely to replicate for a new set of participants, but only using the same sample of stimuli. A significant by-stimulus result, on the other hand, suggests that experimental results would be likely to replicate for a new set of stimuli, but only using the same sample of participants. However, it is a fallacy to assume that the conjunction of these two results implies that a result would be likely to replicate with simultaneously new samples of both participants and stimuli." I would like to control for the random effects of both stimuli and participants, but I am unsure how to proceed? The article can be accessed here: http://jakewestfall.org/publications/JWK.pdf
[ 0.029989691451191902, 0.009188011288642883, -0.018471047282218933, -0.0034086673986166716, -0.0025465316139161587, 0.010683185420930386, 0.005898483097553253, -0.003980457782745361, -0.01161965075880289, -0.005785736255347729, 0.0025415236596018076, 0.013571852818131447, 0.012905332259833813...
[ -0.005442388821393251, 0.018289806321263313, 0.4089895188808441, -0.4556668698787689, -0.15515722334384918, 0.3199290335178375, 0.46443161368370056, -0.7256876826286316, -0.06781060993671417, -0.4381410479545593, -0.0470675528049469, 0.25296902656555176, -0.05764620006084442, 0.01255278475...
In web development, I get used to place css/js files under separate folders (css folder, js folder ..etc). What I am experiencing in WordPress is that I cannot access these folders in WordPress Admin Editor Page, the editor only displays the files under the main theme folder, not the nested folders. Is there an option to enable showing nested folders? is there a plugin to accomplish that? If not, then what is the best practice to follow for placing files and folders under WordPress folder?
[ -0.0059777782298624516, -0.00008970108319772407, -0.003401128575205803, 0.039203353226184845, 0.0180449727922678, -0.003987596835941076, 0.008683430962264538, -0.0032836501486599445, -0.021094264462590218, 0.01721721701323986, -0.017377056181430817, 0.015690216794610023, -0.01263751182705164...
[ 0.4176255762577057, 0.16459093987941742, 0.21022385358810425, 0.007921235635876656, -0.07180256396532059, -0.21468238532543182, 0.25213128328323364, 0.10972529649734497, -0.32859745621681213, -0.9094409942626953, -0.09473782777786255, 0.2890562117099762, -0.07857795804738998, 0.44782117009...
I know this place maybe isn't the best place to ask for help with this specific question, but I guess you can help. I have a UEFI system with partitions arranged as follows: #  # /etc/fstab: static file system information # # <file system> <dir> <type> <options> <dump> <pass> # /dev/sda2 UUID=1839555a-70c8-4fec-8938-cfd7a16ecc6d /         ext4       rw,relatime,data=ordered,discard,noatime 0 1 # /dev/sdb1 UUID=6d26afd2-cc3e-44a4-b5bd-ea209cd4343d /home     ext4       rw,relatime,data=ordered,discard,noatime 0 2 # /dev/sda1 UUID=7628-D37A       /boot     vfat       rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=iso8859-1,shortname=mixed,errors=remount-ro 0 2 # /dev/sdb2 UUID=f37bdb49-1d0f-4455-8f2d-559b923bfb40 /var       reiserfs   rw,relatime 0 2 # /dev/sdb3 UUID=96c1dbed-cf83-4a2e-9fc2-dcd9fa01a6bb /tmp       reiserfs   rw,relatime 0 2 (note that this is my system's fstab) now, here is some info about how these partitions are arranged and their sizes: hardware: SSD - 60Gb HDD - 500Gb / - about 20Gb on the SSD /boot - about 1Gb on the SSD /home - about 100Gb on the HDD /var - 10Gb on the HDD /tmp - 1Gb on the HDD That means that I have about 389 Gb of empty space on the HDD, I want to use 100~200Gb from those to install Windows for games but I also want to keep some space for installing Gentoo. What would be the easiest way to install Windows? I don't want to put it on the SSD and won't use it for anything other than gaming, and I don't want to ruin my current Arch setup. Also can use the same /tmp partition in Arch or Gentoo? or it isn't possible? Same question but with a swap partition (that is if I eventually have to use one). BTW what about using a VM? is it worth it? I have 8Gb RAM and can easily get more... And do you have any other recommendations?
[ 0.003224676474928856, 0.002702336758375168, -0.002913185628131032, 0.009561166167259216, -0.008479097858071327, -0.003031140426173806, 0.005584515165537596, 0.009970657527446747, -0.015324455685913563, 0.013740855269134045, 0.0012118753511458635, -0.000060850172303617, 0.0010522432858124375,...
[ 0.32665160298347473, 0.06477489322423935, 0.2865190804004669, 0.31871095299720764, 0.07603257149457932, -0.09099402278661728, 0.0032051713205873966, -0.1248643770813942, -0.2729623317718506, -0.6707240343093872, -0.07738696038722992, -0.1603647619485855, 0.01067386195063591, 0.618266165256...
When I set `WP_DEBUG` to `true` in `wp-config.php`, I get to see all the strict standards and deprecated messages. I've set the `error_reporting` in my `php.ini`, `ini_set()` and `error_reporting()` to `E_ERROR | E_WARNING | E_PARSE`. But I still get to see the strict standards messages. I know the messages can be useful, but they appear in some of the plugins I am using and I am not interested in seeing them. How do I disable them?
[ -0.0035763003397732973, 0.011292606592178345, -0.005886421538889408, 0.00623581325635314, 0.012036532163619995, 0.038477398455142975, 0.00913206860423088, -0.011721085757017136, -0.01336668524891138, -0.019495263695716858, -0.010928342118859291, 0.01784679666161537, -0.006947864778339863, ...
[ 0.5084923505783081, 0.3149682879447937, 0.37411123514175415, -0.23101426661014557, -0.12036532163619995, -0.3428112268447876, 0.7382910251617432, -0.010141597129404545, 0.16471557319164276, -0.49475398659706116, -0.0008332091383635998, 0.8428632616996765, -0.258822500705719, 0.012424851767...
I'd like to create a new partition and move the contents of the `/var` directory to it for the security reason of having `/var/www` and other subdirectories "mounted" with `nosuid`, `noexec`, and `nodev` permissions. How do I do it for `/var` or any other root directory?
[ -0.003593920264393091, 0.013780603185296059, -0.0033383017871528864, 0.009820913895964622, 0.005562813952565193, 0.010112707503139973, 0.009055137634277344, 0.01999090053141117, -0.01634894125163555, 0.007929086685180664, -0.018356574699282646, 0.011436866596341133, -0.03330525383353233, 0...
[ 0.458617627620697, -0.07946747541427612, 0.2294498234987259, 0.07427149266004562, 0.30471768975257874, -0.141305074095726, 0.03952805697917938, 0.15840616822242737, 0.05148610100150108, -0.792740523815155, 0.14862224459648132, 0.34330591559410095, -0.06658587604761124, 0.462945818901062, ...
I am looking for a reference to an elementary (or at least fairly simple) treatment of the Weibull-distribution. I have a bright high school student doing a project on wind mills and it turns out that wind speeds follow a Weibull distribution. Ideally I am looking for a derivation of the mean and standard deviations of a Windmill. A reference a proof of the fact, that the distribution in fact is a distribution would also be nice to see.
[ -0.00870992336422205, 0.008765539154410362, -0.01273717638105154, 0.02312602289021015, -0.02805107645690441, -0.013289391063153744, 0.010285427793860435, 0.029835684224963188, -0.020209673792123795, -0.01917022280395031, 0.018780289217829704, 0.017041493207216263, -0.009595168754458427, 0....
[ 0.5175281167030334, 0.022195635363459587, 0.26042336225509644, 0.13937415182590485, -0.2575798034667969, 0.05838809534907341, 0.02393862046301365, -0.16780821979045868, -0.2401423305273056, -0.4911518394947052, 0.36421510577201843, 0.14331763982772827, 0.2293657660484314, 0.557462632656097...
I would like to start / stop bluetooth service automatically when I turn on / off rfswitch, is that possible ?
[ 0.047391701489686966, 0.032222144305706024, -0.01326584629714489, 0.009311930276453495, -0.006812791805714369, -0.03618035092949867, 0.018103698268532753, 0.04413996636867523, -0.024286804720759392, 0.06975100189447403, -0.05214953422546387, 0.04606463760137558, -0.05595207214355469, 0.036...
[ 0.42138975858688354, 0.10643220692873001, 0.5796518325805664, -0.1032101958990097, 0.23547807335853577, -0.2993934154510498, 0.4413279592990875, -0.3028702735900879, -0.1857369989156723, -0.17838747799396515, 0.0776287317276001, 0.7117507457733154, -0.2026279866695404, -0.10016418248414993...
I've imported shp directory files of OSM to geoserver and the city label layer look like red dots instead of showing labels of street names. Is there a fix for this ? (should I declare this layer kml instead of openlayers ?)
[ -0.028926905244588852, 0.0016848617233335972, -0.01795022189617157, 0.05559791624546051, -0.031598180532455444, 0.020960088819265366, 0.016377795487642288, 0.04916253685951233, -0.020572831854224205, 0.009228827431797981, -0.006958003621548414, 0.02663137950003147, 0.016270020976662636, 0....
[ 0.462740957736969, -0.07076160609722137, 0.6562855839729309, 0.07161884009838104, -0.19506961107254028, -0.2231016904115677, 0.3195585608482361, 0.09712083637714386, -0.2975502908229828, -0.9208087921142578, -0.010511666536331177, 0.4094482958316803, -0.4567619860172272, 0.3843853771686554...
I have a shared network drive full of files I want to access from my phone over wifi. Is there a easy way to mount it on my phone? I have a rooted Motorola Milestone.
[ 0.016372380778193474, 0.0017590353963896632, -0.013256250880658627, 0.012321061454713345, 0.02210342139005661, 0.016888659447431564, 0.013259935192763805, 0.043828949332237244, -0.02063731662929058, -0.013379674404859543, -0.003951664082705975, 0.007501159328967333, -0.005999820306897163, ...
[ 0.5287628173828125, 0.655276894569397, -0.021113164722919464, 0.20879675447940826, 0.25182658433914185, 0.08055340498685837, 0.1172976940870285, 0.05249977484345436, -0.4818991720676422, -0.6006238460540771, 0.4558993875980377, 0.13909287750720978, 0.3457506000995636, 0.2488292157649994, ...
I'm curious if there are any commonly used large enterprise frameworks for PHP that would be your whole or most of your environment if you were working in PHP. Something comparable to ASP.NET's WebForms or MVC in that when you're working in either one, most of your code and system is based off of working integratively within that framework. Edit: Specifically I'm looking for a framework that fit's the bill of basically being what you would use for 99% of your work in writing your website, this means: * data retrieval, some kind of ORM in the framework * data representation, some kind of data aware automatically configured UI controls that make reading and writing to a DB easy * data caching * data serialization * data communication, an easy way to generate and host web services of various sorts
[ 0.00802486203610897, 0.013000047765672207, -0.0063136667013168335, 0.0027207054663449526, 0.026079412549734116, -0.017166126519441605, 0.004500329494476318, -0.003657491412013769, -0.01025879755616188, -0.03877301514148712, -0.014049026183784008, 0.009730366989970207, 0.005535910837352276, ...
[ 0.6328105330467224, 0.21871674060821533, -0.053791698068380356, 0.3416465222835541, -0.2862493395805359, -0.1081150621175766, 0.18113085627555847, 0.21379637718200684, -0.3298894762992859, -0.42514660954475403, 0.26959460973739624, 0.4295642375946045, 0.035005033016204834, -0.1036342829465...
What are the potential pitfalls of combining related class like objects (interfaces, traits, custom exceptions) in the same source file? For code reuse and only loading what I need I always separate out full class's into separate files. But interfaces, traits, and extended exceptions don't seem to need that same level of granularity. I have been trained that all classes and class like objects (interfaces, extended exceptions, etc.) should be given their own source files. But now I am starting to work with traits and after getting advice about combining traits and interfaces, to have the ability to have something along the lines of multiple inheritance, I am combining the interface and the trait in the same source file. That way the only way to access traits is to implement the corresponding interface. Once the class calls the autoloader to get the source for the interface it also loads the trait. Not all classes that implement the interface will use the trait for functionality, but most will, and those that don't will just not _use_ the trait. I also will use extended Exceptions for an interface along the lines of PDOException. These are exceptions only thrown by classes that extend or implement a class. So now my **Stackable.php** interface file looks like: interface Stackable{ public function add($content); public function _registerParent($parent); public function _checkLoop($child); } trait StackableTrait{ private $containter_contents = array(); private $container_parent = NULL; public function add($content){ if(is_a($content, "Stackable")){ $content->_registerParent($this); //prevents adding twice and creating loops. }elseif(!is_string($content)){ $this->_checkLoop($content); } return $this->container_contents[] = $content; } public function _registerParent($parent){ if(empty($this->container_parent)){ $this->container_parent = $parent; return; } throw new StackableException("Object has already been added to a parent"); } public function _checkLoop($child){ if(in_array($child,$this->container_contents,TRUE){ throw new StackableException("Object has already been added to a parent"); }elseif(!empty($this->container_parent)){ $this->container_parent->_checkLoop($child); } } } class StackableException extends Exception{} As you can see all of the contents of the file are interrelated, any class that wants to use the StackableTrait will need to also implement the interface so the class checking in the trait functions will work. And only things using _StackableTrait_ or implementing _Stackable_ will have access to throw _StackableException_.
[ 0.027098335325717926, 0.028323454782366753, -0.010555148124694824, 0.01952289044857025, 0.012305633164942265, 0.019386015832424164, 0.010599059984087944, -0.00370557839050889, -0.019941508769989014, -0.004732468165457249, -0.0006737475632689893, 0.02613960951566696, 0.01757369562983513, 0....
[ 0.1853628009557724, -0.0530240423977375, -0.40473926067352295, 0.2293017953634262, -0.0526081807911396, 0.09610225260257721, 0.26567143201828003, 0.014843597076833248, -0.1672936975955963, -0.7712823152542114, 0.07062899321317673, 0.5687321424484253, -0.13732042908668518, 0.052386190742254...
I was wondering how to achieve the Golden Weapons and Clues in Max Payne 3. I tried it by choosing chapter 2, then I found all the weapon parts and clues but when chapter 3 was initiated and I quit, I seem to have made no progress in collecting them. Do I have to play through the story the "normal" way or was it because I had the infinite bullettime cheat activated?
[ 0.02694796770811081, 0.006288595497608185, -0.007400136440992355, 0.0034193203318864107, -0.04004068672657013, -0.0013969686115160584, 0.009065826423466206, -0.007261664140969515, -0.0321982242166996, -0.0016147458227351308, -0.008115916512906551, 0.014049997553229332, -0.03600513935089111, ...
[ 0.3668319284915924, 0.007322391495108604, 0.1673416942358017, 0.2532583773136139, -0.3810581564903259, -0.06758196651935577, 0.4987492561340332, -0.31554460525512695, -0.11955824494361877, -0.0645088404417038, 0.4119672477245331, 0.5979692935943604, -0.10555762052536011, -0.087861590087413...
I am creating a CV with the moderncv classic style. My header looks so far like this: ![enter image description here](http://i.stack.imgur.com/urpSm.png): I would like to achieve that 1. the name and title are not aligned at the lower, but at the upper end of the box. 2. the name is "overlapping" with the address box, because I have to make the name smaller at the moment to avoid a line break. I guess that I have to change _moderncvstyleclassic.sty_. Unfortunately I'm not that familiar with the syntax so I cannot figure out what to change. Here is my code: \documentclass[11pt,a4paper,sans]{moderncv} \moderncvstyle{classic} \moderncvcolor{grey} \renewcommand*{\namefont}{\fontsize{25}{36}\mdseries\upshape} \usepackage[ansinew]{inputenc} \usepackage[scale=0.75]{geometry} % personal data \firstname{Dr. Marcus} \familyname{Surname} \title{Biochemist} \address{Street abc}{D-80331 Munich} \mobile{+49 xxx / xx xx xx xx} \email{marcus.surname@dummy.com} \photo[90pt][0.4pt]{dummy} \begin{document} \makecvtitle \end{document} Many thanks!
[ 0.006327788811177015, 0.0079201590269804, -0.005627785809338093, 0.018527023494243622, -0.01664000004529953, 0.006378418765962124, 0.006244156509637833, 0.006895576603710651, -0.009073833003640175, 0.020100481808185577, -0.018143314868211746, 0.00891651026904583, 0.0005702257622033358, 0.0...
[ 0.5023007392883301, 0.04418962076306343, 0.8733007311820984, -0.277235209941864, 0.20115014910697937, -0.6160228848457336, 0.08591295778751373, -0.08723625540733337, -0.15766197443008423, -0.4871159791946411, 0.0033385739661753178, 0.9504613280296326, 0.06305033713579178, 0.422573804855346...
I want to have sign up/login functionality on my site with a "Login with Facebook" and "Login with Twitter" button. Considering the following scenario: * User A logs in with Twitter. * Account gets created. * User A logs out. * User A comes back next week and can't remember what they signed up with (or just feels like) using Facebook login this time. How can you combine these accounts? So far there's 2 options I can think of:- * Cookie (Not reliable) * Have a 'login' and 'signup' action. So if they 'login' with Facebook it will say "You never signed up for this site with Facebook. Did you maybe use Twitter?" (But maybe they will 'signup' with Facebook then and then I still have no connection with the previous twitter account)
[ 0.007994288578629494, 0.009648647159337997, 0.004648538306355476, 0.0007990115555003285, -0.006358457263559103, 0.004808387719094753, 0.005659294780343771, 0.012569688260555267, -0.011722888797521591, -0.006400679238140583, -0.01721954718232155, 0.010265419259667397, -0.0005923951976001263, ...
[ 0.6303581595420837, -0.006075267214328051, 0.35279610753059387, -0.06837110221385956, 0.05861634016036987, 0.17376138269901276, 0.40848883986473083, -0.07495371997356415, -0.4140905737876892, -0.7626346349716187, 0.1706995815038681, 0.49274957180023193, 0.020352527499198914, 0.124629497528...
Globular clusters are apparently very very old, and the density of these clusters appears to increase as one approaches the center of a cluster. Orbits are bound to be chaotic, since there is no particular orbital plane, unlike a spiral galaxy. From tidal effects alone it seems that over time many of the stars in the middle ought to have merged, forming new stars of greater and greater size. Eventually one should have seen supernovae occurring inside these clusters, or at least so it would seem, and there ought to be black holes in some of them. However, it appears that this does not happen, and the stars in these clusters do not merge. _What is keeping this from happening?_
[ -0.01090966910123825, 0.008021918125450611, -0.0011836757184937596, 0.008818738162517548, -0.0216625165194273, 0.001703497488051653, 0.005546571686863899, 0.004652657080441713, -0.01254350133240223, 0.004568669945001602, -0.0009107214864343405, 0.02440187707543373, -0.0052026379853487015, ...
[ -0.18952979147434235, 0.4047533869743347, 0.6267810463905334, -0.15441469848155975, 0.00407820101827383, -0.1971738487482071, -0.0005596545524895191, 0.783425509929657, -0.5133512020111084, -0.5383220314979553, -0.5178497433662415, 0.09283564239740372, -0.7050800919532776, 0.60185861587524...
I know that light does not interact with other light, but can interfere it, at least its amplitude. With that said, lights frequency can be changed via bouncing off matter, where matter might absorb some of that photons energy changing the frequency.. Does this imply that a high energy beam of light that intersects an ~~non-~~ visible beam of light, might be enough to bounce the light around at the intersection of the two laser beams to emit visible light? By changing its frequency? Could that also be controlled by the energy output of the other laser?
[ 0.023773416876792908, 0.0008845065603964031, -0.005572593305259943, 0.010317913256585598, -0.009079240262508392, -0.02650880068540573, 0.009927499108016491, -0.00029084409470669925, -0.014222093857824802, -0.024541549384593964, -0.01738595776259899, 0.023563386872410774, -0.01090628281235694...
[ 0.47580233216285706, -0.016726689413189888, 0.506618857383728, 0.045557230710983276, -0.3835439682006836, -0.3440748453140259, -0.04101558029651642, -0.47196102142333984, -0.37715357542037964, -0.3152960240840912, 0.21069730818271637, 0.6116265654563904, -0.2962169647216797, 0.237251609563...
Murder (said old Quong)—oblige me by passing my pipe—murder is one of the simplest thing in the world to do. Killing a man is a much simpler matter than killing a duck. Not always so safe, perhaps, simpler. But to certain gifted people it is both simple and entirely safe. Many minds of finer complexion than my own have discolored themselves in seeking to name the identity of the author of those wholesale murders which took place last year. Question: What exactly 'discolored themselves' means in this novel of "The Hands of Mr. Ottermole" by Thomas Burke ? Many thanks in advance !
[ -0.0008305066730827093, 0.026205290108919144, -0.0014335999730974436, 0.0007141599780879915, -0.00971213635057211, -0.015685927122831345, 0.007231357041746378, 0.01308231521397829, -0.010992002673447132, -0.02073092758655548, -0.006769509520381689, 0.011832870543003082, 0.010900923982262611,...
[ 0.3818624019622803, 0.4929221570491791, 0.04037228226661682, -0.15796500444412231, -0.30228134989738464, 0.3600374460220337, 0.33150163292884827, -0.050259750336408615, 0.06936047226190567, -0.4347153604030609, 0.07245536148548126, 0.4778086245059967, -0.2972410023212433, 0.067592538893222...
How should the following sentence be understood? > "There isn't no happiness". > > a) meaning: There IS some happiness . (Because the two negations cancel each > other out) > > b) meaning: There is NO happiness. In daily-life conversations, which of the two would most likely be meant by the speaker? (Why would one want to use a double-negation anyway?)
[ -0.014482916332781315, 0.020515291020274162, 0.00041868616244755685, 0.026082780212163925, -0.022462178021669388, -0.030835269019007683, 0.008247680962085724, -0.0062776957638561726, -0.011895859614014626, -0.009821426123380661, -0.016138004139065742, 0.006746279541403055, -0.007478680927306...
[ -0.12054223567247391, 0.0033870674669742584, 0.003520234487950802, -0.13636457920074463, -0.29698342084884644, 0.2059880942106247, 0.8892136216163635, 0.03136976808309555, -0.2532106339931488, -0.4077221751213074, -0.4530607759952545, 0.516859233379364, -0.5584484338760376, -0.092525154352...
To be clear, I am not referring to `/etc/issue` (shown before the prompt) or `/etc/motd` (shown after login), but the prompt itself, which is currently hostname login: For good measure, I am running Arch Linux, but I believe the answer will likely be cross-distro.
[ 0.0048398422077298164, 0.0015040547586977482, -0.010081899352371693, 0.0052498565055429935, -0.04148518294095993, -0.007281686179339886, 0.010079682804644108, 0.016507716849446297, -0.019570576027035713, -0.0006965263746678829, -0.021582214161753654, 0.004610000643879175, 0.00525920791551470...
[ -0.07099902629852295, 0.29666194319725037, 0.27999040484428406, -0.3576792776584625, 0.01968393847346306, -0.28146228194236755, 0.3677632212638855, 0.5416774153709412, -0.1771918684244156, -0.6317752599716187, -0.291896790266037, 0.7778944969177246, -0.41916316747665405, 0.3342774510383606...
I was wondering how to calculate Tukey's HSD to follow up a three-way repeated measures ANOVA by hand. I tried to infer how to do it by modifying what I would do in a between subject ANOVA, but I'm not sure whether I'm on the right path. MSE = 18109 n = 59 (I have a total of 59 subjects, which participate in all groups) df Error = 55 I'd like to make 4 comparisons using 8 means. Using a `k` = 8 (should this be 4?) and the `df E` = 55 I looked up `q` = 4.45. I divided 18,109 by 59 = 306, took the square root of it = 17.49 and multiplied it (17.49) by 4.45 = 77 = HSD. Does this make sense?
[ 0.010419094935059547, 0.012541966512799263, -0.02112969011068344, 0.020804690197110176, -0.02143806964159012, 0.030418673530220985, 0.008062656037509441, -0.009224919602274895, -0.013633211143314838, -0.0034806362818926573, 0.003954395651817322, -0.00009534740820527077, -0.005708886776119471...
[ 0.2217971533536911, -0.03588402643799782, 0.41107699275016785, -0.06936687976121902, -0.207218736410141, 0.6623899340629578, 0.6059173345565796, -0.6040830612182617, -0.42614486813545227, -0.22572945058345795, -0.006106879096478224, 0.23362676799297333, -0.09916765987873077, 0.285279542207...
The question should be clear enough from the title. Also: What are we supposed to call one who ousts? [If this warrants another question, I will edit this out and open another question.]
[ 0.010363807901740074, 0.018411725759506226, -0.006949807051569223, 0.023012813180685043, -0.01649206131696701, -0.009557381272315979, 0.010704903863370419, 0.014497669413685799, -0.032039228826761246, 0.0096502136439085, 0.008020742796361446, 0.02275700867176056, 0.014245176687836647, 0.01...
[ -0.037230752408504486, 0.06518173217773438, -0.1055268868803978, -0.06340310722589493, -0.2265823483467102, -0.38093987107276917, 0.3497453033924103, -0.039551082998514175, -0.14479711651802063, -0.1622360199689865, -0.061094675213098526, 0.575729489326477, -0.19900928437709808, 0.05736894...
I have a job from a client that needs a iPhone/iPad app. I'm new to the smartphone development so where do I begin with? Do you have some nice in depth beginning tutorials of some kind? JS libraries I need to look into? Is there a startup boilerplate for iPhone/Android out there with a beneficial starting setup kit or something like that? In general, is there something I really need to know before I start? My coding background is: * HTML/CSS * jQuery/Javascript (rookie) * C# * AS 3.0
[ -0.008051043376326561, 0.01253516785800457, 0.009624199941754341, -0.0021452840883284807, 0.0004827834200114012, 0.008165652863681316, 0.006261360365897417, 0.02643173187971115, -0.016728730872273445, -0.011006142012774944, -0.009850701317191124, 0.00747574120759964, 0.011689092963933945, ...
[ 0.7080146670341492, 0.2560615539550781, 0.009956962428987026, 0.1081097200512886, 0.21436767280101776, 0.0007755421102046967, 0.14146313071250916, 0.13512976467609406, -0.06922100484371185, -0.7131579518318176, 0.12515626847743988, 0.3622436225414276, 0.2891426086425781, -0.413001269102096...
I am trying to adjust equation range numbering from `(1.10)-(1.15)` to `(1.10-15)`. By calling this in the preamble: \crefrangelabelformat{equation}{(#3#1#4-#5#2#6)} and then this in the document: \Crefrange{eq:1-10}{eq:1-15} I have gotten the range in side the same brackets, i.e. `(1.10-1.15)`, but I am having trouble stripping the section number from the second label in the range. Is there some way to adjust the `{eq:1-15}` label prior to calling `\Crefrange` (i.e. changing `1.15` to `15`), and then change it back after the call? I need to change it back as in other locations I reference 1.15 individually and so need to retain the section number. Alternately can I strip the section number from the formatted text (`#2`) passed to `\crefrangelabelformat`? Any help appreciated.
[ 0.006182976067066193, 0.01065103430300951, -0.011769631877541542, 0.014405188150703907, -0.01620667055249214, -0.000544926559086889, 0.006681227590888739, -0.00011240527965128422, -0.012035434134304523, -0.010434103198349476, -0.0021972195245325565, 0.007166679482907057, -0.02489546500146389...
[ 0.007700917776674032, 0.43815359473228455, 0.5947988033294678, -0.4579320549964905, -0.2563578486442566, 0.14793086051940918, 0.2421686053276062, -0.45131662487983704, 0.005629478022456169, -0.5457428693771362, -0.10813353210687637, 0.7095884680747986, -0.03015977516770363, 0.1462201774120...
I know that it could be possible since I've found Trevor shirtless while switching character. But when in the wardrobe there's no option to have a shirtless character or to remove t-shirt. How can I make my character shirtless?
[ 0.012798900716006756, 0.02475188486278057, -0.01829666830599308, 0.0033874502405524254, -0.03442069888114929, -0.003964695613831282, 0.01283317618072033, 0.03380566090345383, -0.031655654311180115, -0.01898881606757641, -0.018758976832032204, 0.021949712187051773, 0.0007646878366358578, 0....
[ 0.5863449573516846, -0.09449788928031921, -0.08921702206134796, 0.350094199180603, 0.2895704507827759, 0.3148405849933624, 0.404126912355423, -0.1734572947025299, -0.1915496289730072, -0.49098536372184753, -0.030849739909172058, 0.3393116295337677, -0.032100945711135864, 0.252828449010849,...
I have two machines with two applications that talk to each other on few network ports (TCP and UDP). I want to count traffic that they send and receive. I need not only overall count but stats per machine per port per day. I tried darkstat, but it doesn't provide stats per day, but only overall counters. Is there other way that I can count that traffic (I can put some proxy or gateway between that two machines).
[ 0.017768027260899544, 0.0018526393687352538, -0.01200029905885458, 0.010182801634073257, -0.004982751794159412, -0.011326885782182217, 0.00948329083621502, -0.012863469310104847, -0.020275123417377472, -0.006019109394401312, -0.006968085188418627, 0.012179610319435596, -0.011918838135898113,...
[ 0.1293506771326065, -0.10717854648828506, 0.35213035345077515, 0.22567838430404663, -0.18683235347270966, 0.12949083745479584, -0.2230292111635208, 0.23732586205005646, -0.4223245680332184, -1.0576132535934448, 0.40924757719039917, 0.5797000527381897, -0.31416335701942444, 0.34207278490066...
I am playing as a gnome conjurer in EverQuest II. This is my second character on this account, so I got the option of starting in Kelethin. I am trying to fast-track citizenship in Kelethin so that I can get a house and store stuff there. I have completed Learning Faerlie so that I can talk to the necessary NPCs, and I have a reputation of 40 000 with Kelethin. According to both Considering Kelethin and Kelethin: Citizenship I need to speak to Gibrien Marsden to start the respective quest (before going to the Fae Royal Hall to speak to one of two NPCs there - I cannot access them at the moment). However, when I speak to Gibrien, he asks me about _changing_ citizenship. And offers me the choice of changing to Qeynos or Freeport. He does not offer me dialog options to get on the Kelethin citizenship track. Is it possible that I am already a citizen? If so, how do I confirm this?
[ 0.0167379267513752, 0.013925444334745407, 0.0003961946349591017, -0.0035888387355953455, -0.03060903772711754, 0.009930949658155441, 0.009087080135941505, -0.003220038954168558, -0.012956177815794945, 0.022533955052495003, -0.005413665436208248, 0.018107786774635315, -0.015592928975820541, ...
[ -0.22470396757125854, -0.24234366416931152, 0.6138516664505005, 0.18734320998191833, -0.4126163125038147, 0.09957857429981232, 0.2452288568019867, 0.19509364664554596, -0.2253078818321228, -0.8934630155563354, -0.23440170288085938, 0.16106747090816498, 0.1831364780664444, 0.008570599369704...
I have a fitted GLM model: `m1=glm(y~x,family=poisson,data=data)`. I would like to use this fitted model to simulate new data but `simulate(m1,nsim=1)` results only in y's for the original x-values used to fit the model. Can the simulate function be used to generate y's from new x-values?
[ 0.008449779823422432, 0.017999907955527306, -0.01091480627655983, 0.006883903406560421, 0.0009954733541235328, -0.004557881969958544, 0.011344150640070438, 0.008826649747788906, -0.014977732673287392, -0.0093832453712821, -0.009537964127957821, 0.00836104154586792, -0.0005588632775470614, ...
[ -0.04912254959344864, 0.039836056530475616, 0.5167388916015625, -0.0012949112569913268, 0.061056844890117645, 0.09430161863565445, 0.10629571229219437, -0.39418065547943115, -0.20729762315750122, -0.4206630289554596, -0.06693339347839355, 0.35622382164001465, -0.08931564539670944, 0.374509...
I'm in the city of Stormwind in the Mage Quarter where I talked to the Master Fire Eater. The 2nd quest in his training is for you to light torches, throw them up in the air, and catch them four times in a row. However, I can't figure out how to do it. What I can figure out is that standing near the big bond fire is sufficient to light the torch, and to throw them I only need to click on the button that appears next to the objective on the right. However, after that I am at a loss what to do. How does the mechanics of the torch catching work? How do I know if I caught the torch? Is there a counter to tell you how many in a row you've caught?
[ 0.01652015931904316, 0.006387598346918821, -0.010954842902719975, -0.01578499749302864, -0.0192578062415123, -0.00765578169375658, 0.0068302336148917675, 0.020105207338929176, -0.016483578830957413, -0.007612289860844612, -0.022811096161603928, 0.01496492512524128, -0.01686968468129635, 0....
[ 0.49294063448905945, 0.15004362165927887, 0.3433729112148285, -0.09373897314071655, -0.582202136516571, -0.2654058635234833, 0.42239874601364136, -0.30927345156669617, -0.4630753993988037, -0.8419064879417419, 0.23529326915740967, 0.0539485365152359, 0.03321802243590355, -0.019317632541060...
I am building a website powered by twitter's bootstrap. I have created a landing page with menus and text. The next thing I want to do is create a About page, but what is the easiest way to create a new page with the same layout. I don't think copy and pasting the layout would be a great idea. I have used PHP includes and $_GET variables for this in the past but I wondered if there is a more elegant solution. I hope this is the place for the question and I hope somebody can help me.
[ -0.009997419081628323, 0.015367397107183933, 0.0023942007683217525, -0.0045473952777683735, -0.00038903573295101523, 0.0020980658009648323, 0.003890007734298706, -0.0042546335607767105, -0.016353769227862358, -0.018024934455752373, -0.013135208748281002, 0.0069822631776332855, 0.002830873709...
[ 0.892612874507904, 0.2094125896692276, 0.16749058663845062, 0.11649726331233978, 0.06844151020050049, -0.04417496547102928, 0.2241382747888565, 0.42910873889923096, -0.05653674900531769, -0.7988525032997131, 0.431139200925827, 0.339085191488266, 0.10446483641862869, 0.31482723355293274, ...
Good afternoon all, I was originally using a GSM iPhone and I've switched to an android Samsung Galaxy Ace S5830L. I've got some problems connecting to the internet and was wondering if someone could help me resolve the issue. These are the steps I took from **Settings** >> **Wireless and networks** >> **Mobile networks** >> **Network operators** : ![enter image description here](http://i.stack.imgur.com/WOOi9.png) ![enter image description here](http://i.stack.imgur.com/TZhnL.png) ![enter image description here](http://i.stack.imgur.com/g3L3r.png) ![enter image description here](http://i.stack.imgur.com/rHl8W.png) I've selected SGP-M1-GSM and it says "Registered on network" but I still can't seem to connect to the internet. ![enter image description here](http://i.stack.imgur.com/thIUV.png) What do I have to do to enable internet connection? I'm currently using v2.3.4 if that matters.
[ 0.00265819882042706, -0.008977456949651241, -0.0036073140799999237, -0.005845475941896439, -0.0060618105344474316, 0.0005838682409375906, 0.004376956261694431, 0.019678644835948944, -0.014104494825005531, -0.00939706526696682, -0.005526646971702576, 0.0015211210120469332, 0.00546738784760236...
[ 0.2574065029621124, 0.2598618268966675, 0.92093425989151, -0.015954935923218727, -0.1519227772951126, -0.08106305450201035, 0.4909377098083496, -0.0063536386005580425, -0.2879810333251953, -0.45862987637519836, -0.5320409536361694, 0.5589521527290344, -0.030531875789165497, -0.085965245962...
I would like to setup socks proxy on my Debian VPS. I try to use squid but its to hard for me. I saw a Dante software and think about it. I would like to setup dante to work with the web browser and plug FoxyProxy. So on server (vps-> dante server) and on the client (Firefox + FoxyProxy-> socks5 (user + password)) It should be public access to the socks but allowed to use it with atentication. If not Dante, what other program? I need complete instructions on how to set up a publicly accessible server socks5 with authentication to works fine with the browser with the addition of FoxyProxy Can anyone help?
[ -0.010492382571101189, -0.004423703067004681, -0.00764826312661171, 0.01197023130953312, -0.003643551841378212, -0.006163917016237974, 0.010583102703094482, 0.01739337109029293, -0.0181092768907547, -0.020564517006278038, -0.015384836122393608, 0.013875873759388924, -0.023921430110931396, ...
[ 0.3809323012828827, 0.07201685756444931, 0.15860725939273834, 0.1613001823425293, -0.4377581477165222, -0.36845624446868896, 0.2351689636707306, 0.01938587613403797, -0.15107761323451996, -0.8721684813499451, 0.35562512278556824, 0.3923104405403137, -0.36096808314323425, 0.0779323950409889...
The battery percentage on my tablet has reached 30%, which is still reporting as about 11 hours left before it goes flat. However, my Dropbox app has stopped uploading files reporting: > Camera Upload (x left): Low battery, charge to resume I'm not anywhere near anything I can use to start it but I want it to continue uploading screenshots I take in the mean time. How can I force Dropbox to upload even if the charge is low?
[ -0.013781639747321606, 0.004817935172468424, -0.002610438270494342, 0.011617976240813732, -0.008453329093754292, -0.025665471330285072, 0.007771780714392662, 0.013027559034526348, -0.01384671125560999, -0.007879712618887424, -0.0307153332978487, 0.013919699937105179, 0.01800587959587574, 0...
[ 0.6256142854690552, 0.19120384752750397, 0.6885550618171692, -0.11664589494466782, 0.11381187289953232, -0.08910917490720749, 0.6112607717514038, 0.07025817781686783, -0.11273863911628723, -0.27819764614105225, 0.2176368236541748, 0.7533709406852722, -0.04377378523349762, -0.14559462666511...
Does the Calendar Man say or do anything special on certain holidays?
[ -0.026753606274724007, 0.08051139861345291, -0.009282087907195091, 0.0004625426954589784, -0.008924688212573528, 0.07661160081624985, 0.02379600517451763, 0.029865946620702744, -0.03879799321293831, -0.0600164458155632, -0.06451640278100967, 0.04516874998807907, 0.07184114307165146, -0.053...
[ 0.7008602619171143, 0.47430458664894104, 0.20972342789173126, -0.1111060306429863, 0.037440840154886246, -0.38406407833099365, 0.1807122528553009, 0.6901687383651733, -0.6795482039451599, 0.046063296496868134, -0.11347438395023346, -0.24644125998020172, 0.04067689925432205, -0.123655915260...
I finished a couple of cups getting 1st place in every race and still got a 2-star rating, so the ranking isn't only based on your position (or total points). Is there a defined formula or is it yet another Nintendo secret?
[ -0.027378685772418976, 0.021768873557448387, -0.007912276312708855, -0.0038435226306319237, 0.0072088902816176414, 0.009410695172846317, 0.011898595839738846, 0.00246973498724401, -0.028605369850993156, 0.045522384345531464, -0.03446001186966896, 0.020007643848657608, 0.019322946667671204, ...
[ 0.36644500494003296, -0.14542755484580994, 0.5768213272094727, 0.3674422800540924, -0.3177916407585144, -0.1393592655658722, 0.2781999409198761, -0.2436719685792923, -0.2646390199661255, -0.0321158766746521, 0.6118707060813904, 0.19723041355609894, 0.3018857538700104, 0.19255338609218597, ...
I have defined several macros in my LaTeX report. For instance, following is a macro I defined for writing bold letter H. I have several macros like this I use throughout my work. \newcommand{\channel}{\ensuremath{\mathbf{H}}} Whenever I use `\channel`, TeXstudio highlights it as red saying unknown- command. But it is indeed supported by the auto-completion feature. I want to stop TeXstudio from highlighting it as red (as it is annoying and affects readability). Is there any way to do this? Update--- After seeing the comments, I realize I forgot to include a important thing. I always write these macros and package calls all inside a single file. Then I input this file in the LaTeX file I am writing. But now I guess I am asking a bit too much. I mean I require the software to recognize the macros which I have input from another file. Nevertheless I was wondering if there was any solution.
[ 0.008060768246650696, 0.017071545124053955, -0.019542565569281578, 0.02308981865644455, -0.002088964683935046, -0.00838766060769558, 0.008509250357747078, -0.009821167215704918, -0.0187322236597538, 0.0070727220736444, -0.01747000962495804, -0.0011470452882349491, 0.017240986227989197, 0.0...
[ 0.049627870321273804, 0.041170183569192886, 0.42306816577911377, -0.3938398063182831, -0.02046198770403862, 0.03744104132056236, 0.3798525929450989, 0.09952230751514435, -0.09246724843978882, -0.7373883128166199, 0.24808092415332794, 0.6618778705596924, -0.5319793820381165, 0.0730890557169...
Recently there was a great answer in How to draw a 3D hexagonal structure with TikZ? But I noticed a small "bug" which I can't seem to be able to fix. I stripped the code down that shows the problem: \documentclass{article} \usepackage{tikz} \usetikzlibrary{shapes} \begin{document} \begin{tikzpicture} \begin{scope}[% every node/.style={anchor=west,regular polygon, regular polygon sides=6,draw,inner sep=0.5cm}, transform shape] \node (A) {A}; \node (B) at (A.corner 1) {B}; \node (C) at (B.corner 5) {C}; \node (D) at (A.corner 5) {D}; \node (E) at (D.corner 5) {E}; \foreach \hex in {A,...,E} { \foreach \corn in {1,...,6} \draw[fill=white] (\hex.corner \corn) circle (2pt); } \end{scope} \end{tikzpicture} \end{document} This code draws five nodes with hexagonal shape, but the hexagons are a few, but always a different "number of few", pixels "off", as shown in this picture: ![enter image description here](http://i.stack.imgur.com/04ZyX.png) * Which value is responsible for this? * Why is the "shifting" happening with different strength? * How to fix it?
[ -0.0017254528356716037, -0.00019440276082605124, 0.00518797105178237, 0.021100446581840515, -0.006502559874206781, -0.003120990237221122, 0.006479955743998289, 0.013465140014886856, -0.015992213040590286, -0.005179082043468952, -0.009243900887668133, -0.0019481608178466558, -0.02368363551795...
[ 0.2571021318435669, 0.07901722937822342, 0.480421781539917, -0.17488360404968262, 0.05352722108364105, -0.04327709227800369, 0.020032523199915886, -0.03242594376206398, -0.11496925354003906, -0.7448376417160034, 0.22614137828350067, 0.07294835895299911, -0.2775740921497345, 0.3728123009204...
I am looking for a browser plugin which inputs textin of input boxes with the same keyboard interface of the Vim text editor. I have adopted Vimium which makes general navigation much more effective, however I often find my self wish I had the Vim interface while I have selected an input field (like while I'm creating this post). I would very much like to be able to navigate the text in the same way I edit file with the vim text editor and I am curious if such a plugin exists?
[ 0.012781510129570961, 0.007230933289974928, 0.003875254886224866, 0.009670683182775974, -0.004829688463360071, -0.012870910577476025, 0.007349824998527765, 0.020077666267752647, -0.022762160748243332, 0.025599513202905655, 0.0014949676115065813, 0.011865601874887943, 0.014319238252937794, ...
[ 0.5317268967628479, 0.19273705780506134, 0.36084261536598206, 0.26939496397972107, -0.4009449779987335, -0.27504420280456543, 0.0356711745262146, 0.23789988458156586, 0.08723361790180206, -1.0720409154891968, 0.3082803189754486, 0.560345470905304, -0.04285982996225357, 0.5082205533981323, ...
I'm starting to mess with Android SDK (on Ubuntu GNU/Linux), and am seeing stuff like running the `tools/android` and its GUI, `adb` or `gradlew` for the first time. So, I'm trying to set this up, and in reading stuff around, I see mention of `~/.android/adbkey` and `~/.android/adbkey.pub`. As far as I understand, these are the automatically generated debug keys - but I cannot tell: are these the same on every computer where the SDK gets installed, or are they different? So I look into `~/.android`, and they are indeed there; and I guess in a moment of madness, I decided to delete them (I guess to see what happens `:)`). Anyways, I soon regretted it, and I tried to look up online how to regenerate them. Unfortunately, I cannot find any explicit reference saying what command I should run, in order to get these file generated/recreated. So that is my question - what tools should I run, to re-generate these keys? * * * EDIT: I tried re-running `sdk/tools/android`, it doesn't reconstruct the keys. Then I tried deleting the `~/.android` folder: rm -rf ~/.android sdk/tools/android ... and after re-running `sdk/tools/android`, the `~/.android` folder _does_ get reconstructed, but _not_ the `adbkey*` files. * * * Answer from comments (better formatted): Ah, found it (via linux - Android Debug Bridge (adb) device - no permissions - Stack Overflow); what I did was: $ sdk/platform-tools/adb start-server * daemon not running. starting it now on port 5037 * * daemon started successfully * $ sdk/platform-tools/adb devices List of devices attached $ sdk/platform-tools/adb kill-server $ ls ~/.android/ adbkey adbkey.pub androidwin.cfg avd cache repositories.cfg sites-settings.cf
[ -0.02297377772629261, 0.012123744934797287, 0.008879820816218853, 0.005352100823074579, -0.0532592311501503, 0.012299107387661934, 0.007626580074429512, 0.023640144616365433, -0.012378876097500324, 0.005445948801934719, -0.014324979856610298, 0.010806562379002571, -0.010741202160716057, 0....
[ 0.2285127341747284, 0.5146462917327881, 0.33820417523384094, -0.07073377817869186, 0.0054443576373159885, -0.04869213327765465, 0.22657355666160583, 0.16241450607776642, -0.01667160540819168, -0.6696597337722778, -0.11961540579795837, 0.7367783188819885, -0.40145301818847656, 0.29605025053...
Is it possible to solve Killing equations in _Mathematica_ for a general vector? I am looking for a way to create Killing equations and then find what the vectors are, but I have a problem with this. * * * ### Introduction First of all what they are. Without going in to the all gory details of general relativity, in short, Killing vectors are vectors that satisfy Killing equations: $\nabla_\mu X_\nu+\nabla_\nu X_\mu=0$ Killing vector, according to the dimensions we are working in (3D, 4D etc.), and what coordinates, is a list with number of elements equating the number of dimension. So If I'm working in 2D sperical coordinate system, and I'm only interested in radial and polar coordinates, I'll have a Killing vector of the form X = { Xθ[θ,ϕ], Xϕ[θ,ϕ]} If I'm working in 4D spherical coordinate system with coordinates $\\{t,r,\theta,\phi\\}$, I'll have a Killing vector with components X = { Xt[t,r,θ,ϕ], Xr[t,r,θ,ϕ], Xθ[t,r,θ,ϕ], Xϕ[t,r,θ,ϕ]} The above equation is given in terms of covariant derivative, and for covariant vector (with indices down) is $\nabla_\mu X_\nu=\frac{\partial X_\nu}{\partial x^\mu}-\Gamma^\lambda_{\mu \nu}X_\lambda$ Now $x^\mu$ is just coordinate for $\mu=t,r,\theta,\phi$, so $x^t=t, x^r=r$ etc. And $\Gamma^\lambda_{\mu \nu}$ are Christofell symbols that I can easily find. Oh, and sometimes the partial derivative is noted as $\partial_\mu$. * * * ### Example I'm working on an easy example, a 2D sphere. It's metric is given by $\begin{pmatrix} 1 & 0\\\ 0 & \sin^2\theta \end{pmatrix}$ My code is this xIN = {θ, ϕ}; n = 2; met = {{1, 0}, {0, Sin[θ]^2}}; inversemetric := Inverse[met] // FullSimplify coord = xIN; (*Christoffel symbols*) affine := affine = Simplify[ Table[(1/2) Sum[ inversemetric[[μ, ρ]] (D[met[[ρ, ν]], coord[[λ]]] + D[met[[ρ, λ]], coord[[ν]]] - D[met[[ν, λ]], coord[[μ]]]), {ρ, 1, n}], {ν, 1, n}, {λ, 1, n}, {μ, 1, n}]] listaffine := Table[{Style[ Subsuperscript[Γ, Row[{coord[[ν]], coord[[λ]]}], coord[[μ]]], 18], Style[affine[[λ, ν, μ]], 14]}, {λ, 1, n}, {ν, 1, n}, {μ, 1, n}] // FullSimplify; data = {#[[1]], "=", #[[2]], #[[3]], "=", #[[4]]} & /@ Partition[DeleteCases[Flatten[listaffine], Null], 4]; data = Insert[data[[#]], #, 1] & /@ Range[Length[data]]; TableForm[data] (*Derivations*) der[f_, σ_] := D[f, xIN[[σ]]] derxU[xU_, μ_, ν_] := Module[{λ}, der[xU[[μ]], ν] + Sum[affine[[ν, μ, λ]] xU[[λ]], {λ, 1, 2}]] // FullSimplify derxd[xd_, μ_, ν_] := Module[{λ}, der[xd[[μ]], ν] - Sum[affine[[ν, λ, μ]] xd[[λ]], {λ, 1, 2}]] // FullSimplify derxUup[xU_, μ_, ν_] := Module[{λ, ρ}, Sum[inversemetric[[ν, ρ]] (der[xU[[μ]], ρ] + Sum[affine[[ρ, μ, λ]] xU[[λ]], {\ λ, 1, 4}]), {ρ, 1, 4}]] // FullSimplify derxdup[xd_, μ_, ν_] := Module[{λ, ρ}, Sum[inversemetric[[ν, ρ]] (der[xd[[μ]], ρ] - Sum[affine[[ρ, λ, μ]] xd[[λ]], {\ λ, 1, 4}]), {ρ, 1, 4}]] // FullSimplify Now, I have specified the general form of my Killing vector: ξ = { ξθ[θ, ϕ], ξϕ[θ, ϕ]}; And I've set up Killing equations: Killeq = Table[ derxd[ξ, ν, μ] + derxd[ξ, μ, ν] == 0, {μ, 1, 2}, {ν, 1, 2}] // Flatten And I get my equations, in Table form $$ \begin{array}{c} 2 \xi \theta ^{(1,0)}(\theta ,\phi )=0 \\\ \xi \theta ^{(0,1)}(\theta ,\phi )+\xi \phi ^{(1,0)}(\theta ,\phi )-2 \cot (\theta ) \xi \phi (\theta ,\phi )=0 \\\ \xi \theta ^{(0,1)}(\theta ,\phi )+\xi \phi ^{(1,0)}(\theta ,\phi )-2 \cot (\theta ) \xi \phi (\theta ,\phi )=0 \\\ 2 \sin (\theta ) \cos (\theta ) \xi \theta (\theta ,\phi )+2 \xi \phi ^{(0,1)}(\theta ,\phi )=0 \\\ \end{array} $$ And that's what I should get, so the code is working (yaaaay! :D) Now, even though I could just specify `Killeq[[1]] = 0` and so on, is there an automatic way for _Mathematica_ to see if there are same, and just give me the list of the ones left (Some kind of `If` statement)? The problem could be identifying which equation is which later on, but I could just look at the original form of Killeq and see it from there. This would be useful if I need to make `TeXForm` later on. And the second part that is bothering me is: How do I solve this? I tried with DSolve[{ Killeq[[1]], Killeq[[2]], Killeq[[4]]}, { ξθ[θ, ϕ], ξϕ[θ, ϕ]}, {θ, ϕ}] But I got the error: > > `DSolve::overdet: There are fewer dependent variables than equations, > so the system is overdetermined.` > Is there a way of finding these kind of things with _Mathematica_? :\ * * * ### Edit I tried by separately solving each equation Flatten[ Table[ DSolve[ Killeq[[i]], ξθ[θ, ϕ], {θ, ϕ}], {i, 1, 4}]] And I get this: $$ \begin{array}{c} \xi \theta (\theta ,\phi )\to c_1(\phi ) \\\ \xi \theta (\theta ,\phi )\to \int_1^{\phi } \left(2 \cot (\theta ) \xi \phi (\theta ,K[1])-\xi \phi ^{(1,0)}(\theta ,K[1])\right) \, dK[1]+c_1(\theta ) \\\ \xi \theta (\theta ,\phi )\to \int_1^{\phi } \left(2 \cot (\theta ) \xi \phi (\theta ,K[1])-\xi \phi ^{(1,0)}(\theta ,K[1])\right) \, dK[1]+c_1(\theta ) \\\ \xi \theta (\theta ,\phi )\to -\csc (\theta ) \sec (\theta ) \xi \phi ^{(0,1)}(\theta ,\phi ) \\\ \end{array} $$ Now, given that my 2nd and 3rd equations are repeating, is there any way of solving this with _Mathematica_? I am interested, because it would greatly help me find Killing equations in higher dimensions.
[ 0.011586185544729233, 0.01888112537562847, 0.0009496260317973793, 0.011445934884250164, -0.005552499555051327, -0.006487501785159111, 0.006029222160577774, -0.010230333544313908, -0.017709137871861458, -0.020385414361953735, 0.001940486952662468, 0.02574380300939083, -0.0035926070995628834, ...
[ 0.27865487337112427, 0.025801125913858414, 0.14701688289642334, 0.2702662944793701, -0.2781623303890228, 0.09985978901386261, -0.0984065979719162, 0.22725114226341248, -0.19651015102863312, -0.711965024471283, -0.0747261792421341, 0.5415489673614502, -0.23176929354667664, 0.284372240304946...
I've been wondering: Are there, still, some advantages, for current research, to study Newtonian gravity? I mean, not experimentally, where Newton gravity is a very good approximation to everyday phenomenon or, even, to the solar system (aside some perihelion problems, gravitational lensing, etc...) but to some study of General Relativity or some approximation to the Einstein field equations where knowing the Newton potential is needed.
[ 0.012622020207345486, 0.022241024300456047, -0.012061471119523048, 0.0036580204032361507, -0.03647775575518608, -0.007478539366275072, 0.00940120778977871, -0.038167309015989304, -0.0167979933321476, 0.010468341410160065, -0.014196682721376419, 0.024155395105481148, -0.023971280083060265, ...
[ 0.33359798789024353, -0.08061698824167252, 0.08583022654056549, 0.3310808539390564, 0.12453188002109528, 0.025598278269171715, -0.2956683039665222, 0.14638154208660126, -0.46705543994903564, -0.3723408877849579, -0.016941534355282784, 0.3027430772781372, 0.2201339155435562, 0.2100590318441...
I use multiple Konsole terminals. And I want all the commands I type in every terminal to be saved in command history, so that next konsole i open will have all of them. To prevent each terminal from over writing the other terminal's command history, I gave the following settings in my `.bashrc` # avoid duplicates and commands starting with space export HISTCONTROL=ignoredups:erasedups:ignorespace # append history entries.. shopt -s histappend #My machine reboots without warning sometimes.Hence to save commands instantaneously. export PROMPT_COMMAND="history -a" export HISTSIZE=1000 PS1="\[\e[1;34m\]\! \[\e[0m\]"$PS1 I gave the last line to see the command number in my prompt. The command no. has never gone above 600, but still some of my old commands are disappearing from the history. There are many commands which are given repeatedly, but as expected from `ignoredups`, it never increases the command no in prompt. Yet old commands are still disappearing.And the number of commands in history is always remaining slightly more than 500. The `.bash_history` file still contains a lot of duplicates in spite of ignoredups. PS: The output of echo `$HISTSIZE` and `$HISTFILESIZE` are both `=1000` * * * Update: I found the problem in the above entry to `.bashrc`. Just calling `history -a` in `PROMPT_COMMAND` simply appends the last new command to the .bash_history. So the `ignoredups` and `erasedups` have no effect. Is there any way, I can still write to the .bash_history without duplicates from every terminal? I don't want to load the entire history at each command prompt by `history -r` and write it again back with `history -w`, because the commands I issued in one terminal will appear in another parallel running terminal also. I want the combined commands to appear only in a new terminal. The puzzle, why my history was getting trimmed to 500 is solved. I noticed it happened each time I ssh into this machine. Creating a `.bash_profile` with the following entry solved this problem. if [ -f ~/.bashrc ]; then . ~/.bashrc fi Now my .bashrc is executed each time I ssh too. And the history file size is now monotonically increasing.
[ 0.03812535107135773, 0.01654805801808834, -0.0103803351521492, 0.00905382540076971, 0.0046813893131911755, -0.007560583762824535, 0.010094386525452137, 0.015917403623461723, -0.017029784619808197, 0.02271721139550209, -0.010011568665504456, 0.010563126765191555, 0.004452180117368698, 0.036...
[ 0.2264174520969391, 0.31678012013435364, 0.28491368889808655, -0.07225824147462845, 0.02467842400074005, 0.3129163384437561, 0.2972998321056366, -0.035209041088819504, -0.1986638307571411, -0.7218199372291565, -0.1293165683746338, 0.5041807293891907, -0.20792654156684875, 0.266449719667434...
I have caught a giant web fish. It allows me to permanently turn my pet into a pet spider. Do I lose anything by doing this except the cuddly looks of my companion? What are the benefits of having a permanent pet spider? (Does it then keep those benefits if i feed it other fish to turn it into other creatures temporarily?) Is there any advantage to turning your pet permanently into a spider if you have piles of web fish to keep turning it into a spider temporarily?
[ 0.012371801771223545, 0.017126206308603287, 0.0028459595050662756, 0.015351751819252968, 0.014091886579990387, 0.014168672263622284, 0.00875919684767723, 0.003748511429876089, -0.017645245417952538, -0.025767799466848373, -0.008155225776135921, 0.02198011614382267, 0.01186733040958643, -0....
[ 0.3245690166950226, -0.28142255544662476, 0.029154641553759575, 0.1890161633491516, 0.13124892115592957, 0.26951801776885986, 0.24016337096691132, 0.6399572491645813, -0.9727671146392822, -0.5422996282577515, 0.4386001527309418, 0.019215276464819908, -0.26300159096717834, 0.217281594872474...
What does `-t` option of `named` stands for? It looks like `-d` is already taken by `debug-level`, but where is the `t` come from? Is it from `chrooT`?
[ -0.04044286906719208, 0.02293100766837597, -0.0066729942336678505, 0.005259753670543432, -0.004972350317984819, -0.0014443452237173915, 0.012766336090862751, 0.023400140926241875, -0.02351747639477253, -0.0004996577627025545, -0.025019610300660133, 0.010234067216515541, 0.003078487468883395,...
[ 0.1096557080745697, 0.011052947491407394, -0.03173446282744408, -0.005443803500384092, 0.23220692574977875, 0.058048512786626816, -0.0105515718460083, 0.45528241991996765, -0.20604228973388672, 0.1657422035932541, -0.24556107819080353, 0.13199637830257416, -0.2752700746059418, 0.3328673839...
I am doing some work on a pipe network (blue lines) and I am trying to attribute the information about the street to the pipes. I have 2 polygon layers one of the blocks (green hatch) and one of intersections (yellow hatch). I am trying to create a set of buffers/polygons around the street segments (dark gray lines) that fill the space not already covered by either of the 2 other layers(white space). It would be nice to get the double wide streets to have their polygons meet at the mid point between the 2 segments. I have tried creating a buffer and erase what falls underneath the other two layers and with an over sized buffer but I end up with spill over into the cross streets. I need to do this for an entire city, so I can’t just make an undersized buffer and snaping to the other layers never gets far enough and then manually adjust the vertices takes too long. Plus with both methods I end up with overlapping polygons when it buffers around a double street segment which causes problems assigning the street information to the pipe segments. What would be the best way to accomplish this? Let me know if you would like more clarification. ![Example of two intersections](http://i.stack.imgur.com/qcmYg.jpg)
[ -0.011222532019019127, 0.01656424067914486, -0.012893199920654297, 0.012717540375888348, -0.00974469818174839, -0.004786524455994368, 0.006974434480071068, 0.005391957703977823, -0.013453353196382523, 0.003981528803706169, -0.004502668045461178, 0.011499905958771706, 0.005125253461301327, ...
[ 0.43969303369522095, 0.0859903022646904, 0.2850987911224365, 0.18672144412994385, 0.029160091653466225, 0.5267795920372009, 0.10046026855707169, 0.08565808087587357, -0.07956396043300629, -1.268370509147644, -0.004599654581397772, 0.09751628339290619, -0.08563303202390671, 0.00006245018448...
I live in a crappy area with no cows, sheep, villages, or anything else particularly useful. Problem is, I've been based there for a while. I have two houses (one with a mine shaft underneath it) and have chests filled with precious items. In addition, I have no idea where I'm planning on going. What's the best way for me to pack up and leave?
[ 0.025252170860767365, 0.032235898077487946, 0.004105371423065662, 0.005178345832973719, -0.04391024261713028, 0.024800922721624374, 0.006065213121473789, -0.0031314236111938953, -0.02163871005177498, 0.003930109553039074, -0.012842950411140919, 0.007280154153704643, 0.018207449465990067, -...
[ 0.6788105964660645, 0.5513826012611389, -0.3894648253917694, -0.1468164473772049, 0.22720640897750854, 0.07822325080633163, 0.7963564991950989, 0.182316392660141, -0.1418856382369995, -0.1653100848197937, -0.13982054591178894, -0.139757439494133, 0.14273793995380402, 0.3652391731739044, ...
### Background Reading this post, I realized that I had a tendency to use _"typical **for** "_ rather than _"typical **of** "_. After a quick research, reading through several sources on the web, I found that the more I read, the more I got confused. I felt like there is no established rules to state which one is more appropriate, in which occasion. I even found the two usages in the same BBC's article, > "The painting is a little a-typical **for** Van Gogh because of the many > people appearing on it but also very typical because of the prominent role > for the mill." > > "But he added that other elements of the the work, with its bright colours > lathered roughly on the canvas, was typical **of** Van Gogh's style at the > time he was living in Paris." The searches for "typical of" and "typical for" here (EL&U) both returned substantial results, though "typical of" appears to be in favor. ### Confusion Trying to make sense of it, I was about to conclude that I should use "of" when the typicality is something _intrinsic_ (a property or a character of what was talked about), and use "for" when such typicality should be viewed _extrinsically_. As these two examples from Oxford Advanced Learner's Dictionary, > _This meal is typical of local cookery_. (intrinsic) > _A typical working day for me begins at 7.30._ (extrinsic) But soon I felt abash when I found someone mentioned question 7 in this English Grammar Test, Elementary Level # 71, > Such bad behavior is typical .......... the spoiled child. > > > a) for > b) with > c) about > d) of The question was listed under _"british vs. american english"_ , implying that each dialect prefers one usage over another. But I have a feeling that it might turn out to be untrue. Is there any good rule of thumb for the usage of **typical for** vs. **typical of**?
[ 0.006694710813462734, 0.017320983111858368, -0.004282322246581316, 0.010961558669805527, 0.010703181847929955, -0.017887340858578682, 0.005522672086954117, 0.009093375876545906, -0.01172393187880516, 0.006288934964686632, -0.004967464599758387, 0.007062471471726894, 0.0018362000118941069, ...
[ 0.6725077033042908, 0.0007636703085154295, 0.3651936948299408, -0.14902451634407043, -0.37750890851020813, 0.0038835606537759304, 0.3290468454360962, 0.3612107038497925, -0.33736690878868103, -0.3878920376300812, -0.1848565638065338, 0.3014073371887207, -0.411816269159317, 0.21096003055572...
I have an interpolated raster image with EPSG:27700 (British National Grid), after I warp it using gdalwarp into EPSG:3857 (WGS84 Web Mercator (Auxiliary Sphere)): gdalwarp -s_srs EPSG:27700 -t_srs EPSG:3857 input.tif output.tif which works fine (GDAL 1.10.1), (including a little bit of twisting (black cells) in the output one which is understandable as I guess there must be a little bit of distortion). Then, if I create a Web Map Service (wms) in Geoserver using the output file as datastore, when I display the wms in a web page, by using OpenLayers and Google Maps as background, the image is not perfectly located, as it is misplaced about 90-100 meters LEFT where it should be in the British National Grid projection. My question is whether gdalwarp does not the right transformation and/or OpenLayers does its best and does not adjust it, or it is definitely impossible to transform 27700 into 3857 accurately.
[ -0.009319091215729713, 0.002266691066324711, -0.00990256853401661, 0.02559802681207657, -0.001676758169196546, 0.01815587468445301, 0.010426570661365986, -0.0023945725988596678, -0.009878111071884632, -0.025385241955518723, 0.007387480232864618, 0.010867584496736526, -0.025128334760665894, ...
[ -0.02419896423816681, 0.04666289687156677, 0.9441794157028198, -0.28439322113990784, -0.13243897259235382, 0.5681428909301758, 0.2471097856760025, -0.0337265208363533, -0.43210092186927795, -0.9432755708694458, 0.03357182815670967, 0.13183912634849548, -0.16374571621418, -0.078082740306854...
Would it be correct to use "My minor endeavors" in this sentence: "My minor endeavors through my high school kept up my curiosity in computers". If not please suggest an alternative and it would be great if I could know the reason why it is wrong. P.S.: My reasons for asking are purely interest-based as this does not sound good to my ears.
[ 0.012846682220697403, 0.011438939720392227, -0.006829194258898497, 0.00893618818372488, -0.00247448799200356, 0.002525338903069496, 0.008401389233767986, -0.015827756375074387, -0.020036056637763977, -0.0028519488405436277, 0.009799778461456299, 0.028546342626214027, 0.016141755506396294, ...
[ 0.6192929744720459, 0.1694505214691162, 0.10711578279733658, 0.29403984546661377, 0.2493160516023636, -0.013193470425903797, -0.0780220478773117, 0.22462604939937592, -0.039954349398612976, -0.4096202850341797, 0.20562054216861725, 0.45626798272132874, 0.3695562779903412, 0.202486425638198...
Because of a change in brand, I want to redirect our `subdomain.domain.com` to `newdomain.com`. The content being exactly the same, I was thinking of using a 301 wild card redirect to `newdomain.com`. I noticed it is not possible to do a redirect in Google Webmaster Tools as you can with a root domain. Is there a way I can do this redirect without losing all the backlinks referenced with Google?
[ -0.0041832528077065945, 0.012503753416240215, -0.011621309444308281, 0.01409867312759161, -0.0001513835886726156, -0.003564884653314948, 0.008692246861755848, -0.004177544731646776, -0.017083613201975822, -0.017394421622157097, 0.001520557445473969, 0.004660194739699364, 0.005255970638245344...
[ 0.5916171073913574, -0.489934504032135, 0.8724156022071838, 0.05912408232688904, 0.1589130461215973, -0.4201784133911133, 0.4147167205810547, 0.19335901737213135, -0.4200763404369354, -0.5473975539207458, 0.21159929037094116, 0.20325911045074463, -0.030257532373070717, 0.8215674161911011, ...
I will work on climate data to predict if a disease it will happen or not on a crop. During search I found logistic regression is best choice for my research, but it has several problems led to reduce the prediction. I'm new to prediction analysis and I don't know how to increase my prediction accuracy. There are several selection methods used with it. Which one is best to select data to fit the model. Which type of regression help me in this research?
[ 0.01961592771112919, 0.01535199023783207, -0.018677322193980217, -0.0035682818852365017, -0.01786099374294281, -0.009923554956912994, 0.008156001567840576, -0.0017922681290656328, -0.012249106541275978, -0.020276904106140137, -0.0009968540398404002, 0.011343486607074738, -0.01161206886172294...
[ 0.483869343996048, 0.023506944999098778, -0.05563485994935036, 0.07379615306854248, 0.08780518919229507, 0.11383265256881714, 0.16520829498767853, 0.14151623845100403, -0.2025793045759201, -0.7527203559875488, 0.5696611404418945, 0.36007508635520935, -0.05766521021723747, 0.736123621463775...
I have a desktop running Windows 7, a laptop running OpenSUSE 13.1 and an Android 4.1 Tablet, all connected to the same network via a router (the desktop through a LAN cable and the others via WiFi). What would be the best way to share files over this network seamlessly among all three devices? I have somewhat managed to share files between my laptop and desktop via Samba, but even there it works sometimes if I'm lucky while at other times Dolphin simply crashes. Also I can only share one folder on Windows for some reason (the user folders).
[ -0.012516297399997711, -0.0001187990783364512, -0.015127546153962612, 0.007665019016712904, -0.004456883762031794, -0.026003899052739143, 0.008466578088700771, 0.023864274844527245, -0.016197318211197853, -0.03870858997106552, -0.007471869233995676, 0.01215357892215252, 0.0004513593448791653...
[ 0.10075068473815918, 0.18694597482681274, 0.35727381706237793, 0.14847223460674286, 0.1678372472524643, 0.36199647188186646, 0.007111303973942995, 0.22465896606445312, -0.3752509355545044, -1.1063543558120728, 0.16361626982688904, 0.618789553642273, 0.11509672552347183, 0.28606167435646057...
**When can you offer innovation, without being off-task?** I can't understand how creative innovation fits in Agile and other methodologies. **People just want what they asked for, right?** Say you're following an efficient methodology like Agile/SCRUM. When can you offer innovation, without being off-task? Say you're making an app and doing some creative stuff with the UX. The client gave you implicit instructions, and you're staying on track. Imagine you discover an innovation that's not a beta. If you introduce innovations the wrong way, you could appear off-task.
[ 0.009283171035349369, 0.016773849725723267, -0.015231091529130936, 0.004705942701548338, -0.020495403558015823, 0.002378336386755109, 0.01095878891646862, -0.0004194503417238593, -0.012529121711850166, -0.01556557696312666, -0.02409225143492222, 0.02054937556385994, 0.004891576245427132, 0...
[ 0.6019588708877563, 0.0924815908074379, 0.1295284479856491, 0.156198650598526, 0.1121000200510025, 0.1497914046049118, -0.12527909874916077, 0.24104957282543182, -0.4780474305152893, -0.5632408857345581, 0.0015504078473895788, 0.6448507308959961, 0.3726978600025177, -0.08971492946147919, ...
I realized there could be a lot of bad code in a system while writing Java. I wonder why schools don't teach what programming practices are bad. For example, in Java, using a big try/catch block that catches `Exception` is bad, but no one told me. And we see the typical "String + String" concatenation appear in every corner of a program. Isn't it the job of schools, assuming most programmers are taught at a school and not self-taught, to warn students away from the pitfalls in a programming?
[ -0.02432626113295555, 0.019938047975301743, -0.009691623970866203, 0.005492742173373699, 0.00030144862830638885, 0.01154312863945961, 0.0064790076576173306, 0.005195354111492634, -0.016487427055835724, -0.014407118782401085, -0.015325215645134449, 0.012484473176300526, 0.0009066014317795634,...
[ 0.40219035744667053, 0.3686586022377014, -0.4879717528820038, 0.07596743851900101, -0.23924201726913452, -0.32179853320121765, 0.698485255241394, 0.2763758599758148, -0.29539451003074646, -0.34831860661506653, -0.039923470467329025, 0.029309775680303574, -0.06622672080993652, 0.08975090831...
I'm building a web application that will be transferring polygons from server to client. Initially, the shapefile is coming from an external source that I have no control over. Often, I find polygons with way too many points to describe a straight line. The result is, the size of the payload increases and therefore the fetch takes longer. This obviously isn't an issue with only a few polygons, but it will add up as more shapes are added to my dataset and I'd like to optimize bandwidth. ![too many points - qgis](http://i.stack.imgur.com/DqYHJ.gif) Is there a plugin, or native QGIS function that will take some kind of tolerance/variance as input, and produce (approximately) the same polygon with much fewer actual points?
[ -0.017175965011119843, 0.012511838227510452, -0.003473640652373433, 0.01947365701198578, 0.010000274516642094, -0.012784688733518124, 0.006888349074870348, 0.004054780583828688, -0.011680561117827892, -0.008049894124269485, -0.007643831893801689, 0.0161835215985775, -0.010611029341816902, ...
[ 0.4895522892475128, 0.1128627136349678, 0.6168319582939148, 0.09829536080360413, -0.3502413034439087, 0.18071529269218445, -0.16663549840450287, 0.1266830861568451, -0.4916015863418579, -0.7334533333778381, 0.30347737669944763, 0.09581828862428665, -0.0479806587100029, 0.5315428376197815, ...
I want to create a bash alias for grep that adds line numbers: alias grep='grep -n' But that, of course, adds line numbers to pipelines as well. Most of the time (and no exceptions come to mind) I don't want line numbers within a pipeline (at least internally, probably OK if it's last), and I don't really want to add a sed/awk/cut to the pipeline just to take them out. Perhaps my requirements could be simplified to "only add line numbers if grep is the only command on the line." Is there any way to do this without a particularly ugly alias?
[ 0.0006246616831049323, 0.016647029668092728, -0.0100543312728405, 0.013996179215610027, 0.01532733254134655, 0.010708227753639221, 0.00886596366763115, 0.0022282274439930916, -0.020116999745368958, 0.005486730951815844, 0.0005900526884943247, 0.006601489149034023, -0.022773053497076035, 0....
[ 0.6439042687416077, 0.04925947263836861, 0.04878094792366028, -0.24045249819755554, -0.11736391484737396, 0.0856964960694313, 0.38946497440338135, 0.295958012342453, -0.18932729959487915, -0.4192003011703491, 0.06444327533245087, 0.5423582792282104, -0.07657252252101898, -0.073334820568561...
I'm trying to make sure my writing is correct for the following sentence: > The list of functionality has/have to be implemented But I'm confused about **has** or **have** there.
[ 0.005345796700567007, 0.01555218081921339, -0.020451296120882034, 0.010904614813625813, 0.022117692977190018, -0.013670606538653374, 0.011832705698907375, 0.013950112275779247, -0.02621431089937687, 0.024738704785704613, -0.011270729824900627, 0.019046710804104805, 0.03210631012916565, -0....
[ 0.3968139588832855, 0.3886091411113739, 0.33565759658813477, -0.14482663571834564, -0.041668299585580826, -0.17204074561595917, 0.6277300119400024, -0.013688222505152225, -0.1463862508535385, -0.734977662563324, 0.08705729991197586, 0.49656930565834045, 0.2109108865261078, -0.0279933977872...
Often I hear PMs (Project Managers) talk about feature and function. And I'm just so puzzled to differentiate them. Sometimes I think of a feature to be equivalent to a user story. Something like "As a user, Bob should be able to see a list of his payments", and they call it a feature. Sometimes it gets as big as a subsystem, something like "the ability to send SMS via web application". Function on the other hand sometimes gets as small as a task, "implementing digit grouping for number inputs", while there are cases when it gets as big as a whole CRUD operation. My question is, how can we differentiate feature from function?
[ -0.013901975937187672, 0.001756900455802679, 0.006417909171432257, 0.00027842423878610134, 0.004894453100860119, -0.0034986657556146383, 0.0075605325400829315, 0.011086661368608475, -0.010683275759220123, -0.010428700596094131, -0.01697956770658493, 0.00980578362941742, 0.018207332119345665,...
[ 0.5162416100502014, -0.07551207393407822, 0.218480184674263, 0.1608290672302246, 0.04731156677007675, 0.03932991251349449, 0.059952594339847565, -0.03604816272854805, -0.8997253179550171, -0.2761071026325226, -0.20055407285690308, 0.5726499557495117, 0.11094307899475098, -0.058131448924541...
If i ssh to my remote OpenWrt router ["via DynDNS"] 1. I get the message @ WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! @ 2. Ok, i delete the old line in `~/.ssh/known_hosts`, add the new one 3. But my password wasn't accepted (even though I copy+pasted the good password) 4. But: if someone reboots the remote router it says: @ WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! @ 5. Ok, again, I delete the line in `~/.ssh/known_hosts`, add the new one [the old one??], and presto! I can log in! Why? Is this because a dyndns update failed, and I'm trying to log in to the wrong IP? And if the router is rebooted, the dyndns IP will be updated, so that the IP is correct, and I can log in? Is that why it says “host identification failed”? It happened for the 3rd time... I don't know what's exactly is going on.
[ 0.01215837337076664, 0.008919725194573402, -0.002301886910572648, 0.04337324574589729, -0.03968064486980438, 0.013717453926801682, 0.008018734864890575, 0.0035786516964435577, -0.017288709059357643, 0.008664790540933609, -0.004686558619141579, 0.023882422596216202, 0.015753265470266342, 0....
[ 0.1407809853553772, 0.040004026144742966, 0.4233560562133789, -0.20102466642856598, 0.1627494841814041, -0.23839737474918365, 0.8264557719230652, 0.07047954946756363, -0.23694732785224915, -0.8015490174293518, -0.47172796726226807, 0.571354329586029, -0.33634740114212036, 0.552602291107177...
I am running a `GetFeature` request against a WFS server which does not support to download all data at once. Can I use `ogc:PropertyIsGreaterThanOrEqualTo` to split the dataset into chunks at download them step by step? I wonder how I can construct the actual URL to include the paging filter. Here is what I tried: http://example.com/wfs.aspx?request=GetFeature&service=WFS&version=1.1.0 \ &typeName=example:example&maxFeatures=50000 \ &FILTER=<ogc:Filter xmlns:ogc="http://www.opengis.net/ogc"> \ <ogc:PropertyIsGreaterThanOrEqualTo><ogc:PropertyName>OBJECTID</ogc:PropertyName> \ <ogc:Literal>50000</ogc:Literal></ogc:PropertyIsGreaterThanOrEqualTo></ogc:Filter> This is not a valid URI though... That's why uri-encoded the filter part: http://example.com/wfs.aspx?request=GetFeature&service=WFS&version=1.1.0 \ &typeName=example:example&maxFeatures=50000 \ &FILTER=%3Cogc%3AFilter%20xmlns%3Aogc%3D%22http%3A%2F%2Fwww.opengis.net \ %2Fogc%22%3E%3Cogc%3APropertyIsGreaterThanOrEqualTo%3E%3Cogc%3APropertyName%3EOBJECTID \ %3C%2Fogc%3APropertyName%3E%3Cogc%3ALiteral%3E50000%3C%2Fogc%3ALiteral \ %3E%3C%2Fogc%3APropertyIsGreaterThanOrEqualTo%3E%3C%2Fogc%3AFilter%3E This actually works. Is there a "nicer" way to do this?
[ -0.01586763747036457, 0.003945061005651951, -0.009579516015946865, 0.014649353921413422, -0.0013218219392001629, 0.013445230200886726, 0.0068232109770178795, 0.020704170688986778, -0.019845550879836082, 0.006814796477556229, -0.0040058293379843235, 0.018918288871645927, -0.011636799201369286...
[ 0.5166640877723694, 0.05316200107336044, 0.24986475706100464, -0.3085775673389435, -0.1849367916584015, -0.22830699384212494, 0.3989126682281494, -0.07732520997524261, -0.13899312913417816, -0.685661792755127, 0.02355683036148548, 0.2454679012298584, -0.24919404089450836, 0.357842981815338...
I want to plot the phase diagram of prey predator versus prey refuge to see how the prey refuge influences the population of prey and predator.And this is the system $x'=\alpha x(1-x/k)-\beta\frac{(1-m)xy}{1+a(1-m)x}$ $y'=-\gamma y+c\beta\frac{(1-m)xy}{1+a(1-m)x}$ The prey predator with Holling type II model is incorporating a prey refuge, $mx$ and $k$, $\alpha$, $\gamma$, $c$ and $\beta/\alpha$ are the carrying capacity, growth rate of prey, death rate of predator, conversion factor denoting the number of newly born predators for each captured prey and maximum number of prey that can be eaten by each predator in unit time respectively. I have the numerical value for $a=0.02$, $k=100$, $\alpha=10$, $\beta=0.6$, $\gamma=0.09$, $c=0.02$. Thanks so much!
[ -0.008948052302002907, 0.010775899514555931, 0.0004585632123053074, 0.009243618696928024, 0.013097196817398071, -0.006647960282862186, 0.006882213056087494, -0.006441999226808548, -0.01045231893658638, -0.028881827369332314, -0.008911214768886566, -0.0027136486023664474, -0.01741928234696388...
[ 0.1495123952627182, 0.12902846932411194, 0.4234564006328583, -0.15304946899414062, 0.13510431349277496, -0.059998054057359695, -0.046145565807819366, -0.6221323013305664, -0.3221833109855652, -0.4783039391040802, 0.17028029263019562, 0.3213977813720703, -0.4927237629890442, 0.2641692459583...