text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
Specifically I want different versions of 'Item1Numbered'. For example one with a frame and another with a background colour. How is that done?
[ 0.033246248960494995, 0.037952277809381485, -0.012750960886478424, 0.029591038823127747, -0.009493080899119377, 0.008461801335215569, 0.016799280419945717, 0.06289786845445633, -0.029491370543837547, 0.028348878026008606, -0.01400364376604557, -0.0038031567819416523, 0.00633851857855916, -...
[ 0.5685485005378723, -0.11203756183385849, 0.00457244785502553, 0.33658427000045776, -0.19534756243228912, -0.030105944722890854, 0.2170041799545288, -0.3458743095397949, -0.3929218649864197, -0.4099448025226593, -0.003974143881350756, 0.09489408880472183, -0.18075096607208252, 0.0403130762...
This problem appear in an exam put by Chris Sims (3): http://sims.princeton.edu/yftp/emet04/ConfidenceCredibilityEx.pdf * * * Suppose the following model: $y=\beta_1 +\beta_2 X_2+\beta_1 X_3 +\epsilon$ with $\epsilon \sim \text{ Normal}(0,\sigma^2I_n)$ With non informative prior, flat prior for $\beta$ and $log(\sigma)$ The sufficient statistics for our data are: $\mathbf{X}=\begin{bmatrix}40 & 360 &20\\360 & 4240 & 280\\ 23 & 80 & 23\end{bmatrix}$ ; $\mathbf{X'Y}=\begin{bmatrix}42& 514& 36\\ \end{bmatrix}$; $ Y'Y=80 ; n=40$ **Part a: Calculate the least-squares estimate of β and the posterior covariance matrix of β around this estimate under a dσ/σ prior.** **Part b :Plot a 95% joint posterior HPD region for the coefficients** $ (\beta_2 , \beta_3)$ I don´t know if this problem has analytical solution * * * May be this is useful: With a flat prior on β we obtain, after some algebra, the conditional posterior distribution: $p(\beta|\sigma^2, {\bf y}) \sim N(\beta| \ {\bf ( X}^{\text{T}}{\bf X})^{-1}{\bf X}^{\text{T}}{\bf y}, \sigma^2({\bf X}^{\text{T}}{\bf X})^{-1})$. The conditional posterior distribution of $\beta$ would have been the desired posterior distribution had $\sigma^2$ been known. Since that is not the case, we need to obtain the marginal posterior distribution by integrating out $\sigma^2$ as: $p(\beta|{\bf y})= \int p(\beta|\sigma^2, {\bf y})p(\sigma^2|{\bf y}) d\sigma^2$ The marginal posterior distribution of $\beta$ is a multivariate t density ![Multivariate t](http://i.stack.imgur.com/3ednA.png) Demonstration in:http://www.biostat.umn.edu/~sudiptob/ph8472/BayesianLinearModel.pdf
[ 0.0033664507791399956, 0.010732375085353851, -0.019524194300174713, 0.014463899657130241, -0.0055275894701480865, -0.012297989800572395, 0.00675445981323719, -0.016429685056209564, -0.008662878535687923, -0.018532644957304, -0.022951260209083557, 0.004118426702916622, -0.018556613475084305, ...
[ -0.38031816482543945, -0.18945276737213135, 0.21079930663108826, 0.06611409038305283, 0.2582457661628723, 0.09935664385557175, -0.07799273729324341, -0.433701753616333, -0.1267814040184021, -0.6457712054252625, -0.027633704245090485, 0.742974579334259, -0.3369144797325134, 0.04302152618765...
Coming from basically no time series back ground, this is likely a simple question, but what is the relationship between "being able to" use an additive decomposition of a series into seasonal, trend and remainder and a Box Cox transformation? From Professor Hyndman's blog: > Because not all data could be decom­posed addi­tively, we first needed to > apply an auto­mated Box-​​Cox trans­for­ma­tion. I was wondering: 1) What makes an additive decomposition attractive relative to a multiplicative one (which I understand is basically the other choice). 2) What is the requirement for an additive decomp and what does Box Cox do to make this possible? I think of Box Cox for ANOVA and reducing heteroskedasticity. Is there a tie in with decomposition of a series?
[ -0.008037317544221878, 0.01095703337341547, -0.0026240013539791107, 0.025197576731443405, 0.011339210905134678, -0.01809464767575264, 0.009112250059843063, 0.006729796063154936, -0.01713477075099945, -0.0101846344769001, -0.012242919765412807, 0.023011673241853714, 0.00001872144639492035, ...
[ 0.32944756746292114, -0.3508036732673645, -0.09933789819478989, -0.08311356604099274, 0.17112864553928375, -0.09453392028808594, -0.00032366064260713756, -0.4798518717288971, -0.19738195836544037, -0.44473737478256226, -0.17789369821548462, -0.024587411433458328, -0.07203304767608643, 0.44...
I created a poll on a sports forum asking people how they think their favorite NBA team did this off-season. The poll options were: Top 10, Middle 10, Bottom 10. I expect dozens of people to respond, which isn't a huge sample. The hypothesis I'm testing (unbeknownst to the voters) is that there will be a "homerism" bias which makes the "Top 10" vote more likely. In terms of Bayesian methodology, I would like to compare two models, one has preference weighted uniformly across all three poll choices, and the other has the weight heavily biased towards the "Top 10" choice. So, my question is how to create the prior and the likelihood functions for this type of problem. I don't need full details, but if you could point me in the right direction, I'd appreciate it.
[ 0.016300685703754425, 0.023860186338424683, -0.006725231651216745, 0.023694366216659546, -0.01737794652581215, -0.012200004421174526, 0.0074244411662220955, -0.01973603293299675, -0.008671018294990063, -0.0000925031490623951, -0.013460242189466953, 0.006637427024543285, -0.007689585443586111...
[ 0.3563016653060913, 0.02647501602768898, -0.03710249438881874, 0.17560778558254242, 0.011150793172419071, 0.15198375284671783, 0.045816171914339066, 0.21161171793937683, 0.03720077499747276, -0.5936747789382935, 0.22245235741138458, 0.4722239375114441, 0.12967398762702942, -0.1367634683847...
I have a taxonomy in which term order matters, but varies from post to post. So I want to be able to set the order that terms are displayed in on a per- post basis: * **My first post** | Tagged: Apples, Oranges, Bananas * **Some other post** | Tagged: Oranges, Bananas, Apples Since this would vary from post to post, I need something other than the taxonomy-wide term ordering supported by plugins like Custom Taxonomy Sort and Taxonomy Terms Order. I am looking for advice on how to implement this. Here is an outline of my current thinking: ## Per-post term ordering using a custom field 1. Remove standard taxonomy metabox using `remove_meta_box()` 2. Add custom metabox containing term ordering interface, e.g., drag-and-drop (or simple text input next to each term for an integer describing term's order. (Or perhaps it's possible to add drag-and-drop to the existing meta_box without having to remove and add a new one?**) 3. Using javascript, save a comma-separated list of term IDs, in order, in a hidden custom field, _mytermorder_taxonomyname. (Separate custom field for each taxonomy, e.g., _mytermorder_category, __mytermorder_tags, etc.) 4. Create a custom function `my_get_the_terms()` that loops through the IDs in the _taxonomyname_termorder custom field. If a post has terms but no _taxonomyname_termorder field (as with old posts), just return get_the_terms(). 5. In relevant theme files, replace instances of `get_the_terms()` or other native term functions with `my_get_the_terms()`. Do you have advice, alternatives, warnings, or related snippets? I'd be grateful for any thoughts. Thank you! ### Concerns * What if the list in the custom field gets out of sync with the actual term list (as it would if terms were added through an interface other than the post edit screen, such as the quick-edit/bulk edit form or Front-End Editor). **Solution:** At the beginning of `my_get_the_terms()`, we can check whether the IDs in the custom field match the terms returned by get_the_terms. If not, default to get_the_terms. * New terms wouldn't be captured on first use, because they don't yet have an ID. You'd have to save a new term before it would work. (I could live with this, but it is a problem.) * The meta box has to list terms in the right order when the post edit page loads. * * * ** For drag-and-drop sorting in the default metabox, I discovered that I can use jQuery UI's .sortable, already available in the admin; for a non- hierarchical taxonomy, it's `$('#taxonomyname .tagchecklist').sortable()`.
[ -0.003533362876623869, 0.014846747741103172, -0.012339867651462555, 0.02607996203005314, 0.02681383118033409, 0.00005646469071507454, 0.008766671642661095, 0.02901925891637802, -0.019214507192373276, -0.0026647355407476425, -0.018451686948537827, 0.0037649511359632015, 0.0019081961363554, ...
[ 0.21713103353977203, -0.06590361893177032, -0.007974953390657902, -0.19125361740589142, -0.4780566692352295, 0.49472537636756897, 0.0488007590174675, 0.24103960394859314, -0.3072996437549591, -0.7590787410736084, 0.2013125717639923, 0.22883987426757812, -0.2990817129611969, 0.0709131285548...
After rooting my android tab 4.1,I noticed that my internal storage was corrupted. All my apps dissapear.and when I try to install it ,it said that I have insufficient storage.How come this happen when I still have 4 gb of memory on my internal storage?I haven't use memory card since I purchased this product.Please somebody help me.
[ -0.007238004822283983, 0.005513356998562813, 0.006521233357489109, 0.01721125654876232, -0.002649730769917369, -0.00786601286381483, 0.007447423413395882, 0.010536810383200645, -0.016392862424254417, -0.007784697227180004, -0.016927283257246017, 0.018833795562386513, 0.005529230926185846, ...
[ 0.35488054156303406, 0.4235552251338959, 0.180559441447258, 0.12733909487724304, 0.21128451824188232, 0.19306480884552002, 0.7139251232147217, 0.17859986424446106, -0.39896681904792786, -0.6237563490867615, 0.08189795166254044, 0.7479775547981262, -0.17307984828948975, 0.4541682302951813, ...
Is there any direct/indirect relationship with the number of search results in choosing a successful keyword phrase? I have about 22,500,000 results for my keyword phrase, which I intend to target. I think doing seo according to standards set by google, I will won half the battle. Now, if I give 1 year with other marketing techniques like adwords. **How much time will it take me to be on the first page for my targeted keyword phrase?**
[ 0.011566707864403725, 0.012835178524255753, -0.030643966048955917, 0.020454050973057747, -0.015608911402523518, 0.0016784608596935868, 0.009088984690606594, -0.01976395584642887, -0.021187715232372284, -0.018212132155895233, -0.0014592772349715233, 0.016930652782320976, -0.002821744419634342...
[ 0.3852904140949249, 0.2435009926557541, 0.7652050852775574, -0.14884468913078308, -0.24686060845851898, -0.014909068122506142, 0.5770472288131714, -0.168777734041214, -0.06298435479402542, -0.468258798122406, 0.036148492246866226, 0.5328288674354553, 0.1945880651473999, -0.1759552955627441...
Anyone knowns how to create (please sample code) a cover page like this?: http://context.aanhet.net/general/manuals/metafun-p.pdf The sources are available from: http://context.aanhet.net/svn/manuals/metafun/ BUT, i'm unable to compile (ConText) doing `context metafun.tex` ...
[ 0.007545696571469307, 0.007940873503684998, -0.008379561826586723, 0.017858432605862617, 0.01622154749929905, -0.0005818178178742528, 0.00911187194287777, 0.005358151625841856, -0.027081403881311417, -0.007421671412885189, 0.0037708524614572525, 0.005785275250673294, 0.0025862250477075577, ...
[ 0.464921772480011, 0.5177484154701233, 0.08367947489023209, -0.11912339925765991, -0.011579720303416252, 0.004837345331907272, 0.20482613146305084, -0.0732387974858284, 0.0805063545703888, -0.26356443762779236, 0.12326541543006897, 0.6176011562347412, -0.12662534415721893, -0.1344426721334...
My Droid Incredible did a system update & went to factory reset. Everything is gone. Can I go back and restore?
[ 0.043561678379774094, 0.016156505793333054, 0.001633900566957891, -0.011902480386197567, -0.04016232490539551, 0.013396204449236393, 0.0169328935444355, 0.03277423977851868, -0.028076762333512306, -0.06092742457985878, -0.03176550194621086, 0.05725884810090065, 0.06803403794765472, -0.0067...
[ 0.29310500621795654, 0.33998867869377136, 0.362858384847641, 0.1910223662853241, 0.007216301281005144, 0.21780605614185333, 0.6801583170890808, 0.22351181507110596, -0.21782730519771576, -0.2990090548992157, -0.15496575832366943, 0.44698482751846313, -0.17161612212657928, 0.722299695014953...
When I hold the home screen on a spot a set of options appears to add to Home. Two of these options say **program** and **widget**. To add my calendar app which syncs with my google calendar I had to use widget on the home screen. To add my task app which syncs with my google tasks, I added a program to the home screen. What is different from a Widget and a program. They both must be programs. Is a widget something which is dynamic all the time, and loads on boot so the viewer can see it on the home screen live and updated like a clock?
[ -0.0388677641749382, -0.007340300362557173, 0.0030669858679175377, 0.013715257868170738, 0.003932923078536987, -0.004355268552899361, 0.012367097660899162, 0.019629526883363724, -0.015975072979927063, -0.018941441550850868, -0.009573452174663544, 0.0036871046759188175, 0.009781183674931526, ...
[ 0.6840060949325562, -0.11691514402627945, 0.6276455521583557, 0.044067174196243286, -0.14799785614013672, 0.18344742059707642, 0.10188152641057968, 0.03372370824217796, -0.28987085819244385, -0.8235386610031128, 0.36364126205444336, 0.653917670249939, -0.03355798125267029, 0.38192209601402...
## Introduction I am a mathematically minded individual. I do not intuitively comprehend physics, and as a sophomore in high school who has only taken Intro to Physics in his freshman year, I may very well have a completely erroneous view of many concepts within physics. Recently, Einstein's model of gravity has been on my mind. I was watching the following lecture here wherein Professor Sera Cremonini explains that gravity is due to the curvature of 4 dimensional spacetime, and that this curvature is caused by the mass of objects within the universe. Furthermore, the higher the curvature (equivalently, the higher the mass of an object) the higher the force of gravity near that object. To illustrate this, she uses a stretched rubber disk and puts a large ball on it whilst smaller balls are on the disk. The large ball causes the rubber to stretch, and this curvature causes the smaller balls to be pulled toward the large ball just as gravity causes orbits of the solar system. This made me think. ## Question Theoretically, is it possible for the curvature of the universe at one particular point to be so strong that the gravitational force is infinite? If infinite is ridiculous in this context, consider the question reformulated: Is there, or can there be, a point in the universe where the curvature is greater than any other point in the universe? If so, why doesn't the strength of gravity at this point cause the entire universe to be contracted to this point? That is, isn't this just like the Big Crunch? ## P.S. Once again, I don't know if I've conveyed this using the proper terminology and I'm aware it's a very bizarre idea. I hope you all can entertain my thoughts and attempt to answer my question as best as possible, and correct my understanding and terminology as warranted. Thank you.
[ -0.03115694969892502, 0.010320322588086128, -0.01129124965518713, 0.00628203758969903, 0.02266925945878029, -0.004419016186147928, 0.007162113673985004, -0.0010872441343963146, -0.009167376905679703, -0.0074745058082044125, -0.000982489320449531, 0.006751890294253826, -0.011155286803841591, ...
[ -0.12338132411241531, 0.14491304755210876, 0.35456717014312744, 0.3105294704437256, -0.044189609587192535, -0.03476738929748535, 0.11545932292938232, 0.441667765378952, -0.18521732091903687, -0.5687345266342163, -0.26705965399742126, 0.3065331280231476, 0.2503020167350769, 0.14073230326175...
Please help me to further understand custom channels. As Google says it is a way to map your ads, but I still have a few questions: * Is it correct that a single custom channel per 1 ad is not very useful, since you can specify Ad blocks in the AdSense reports? * I have multiple Ads in multiple custom channels. After this I created 1 custom channel and added all the ads to it. I made this channel targetable, so people can target through this channel on all ads at once. Is this a good way to do it? In other words, is it possible to have ads in multiple custom channels (without targeting, just for analyzing) and then create 1 custom channel with targeting that embraces all the (desired) ads? * Why is it not possible for me to analyze custom channels (or ad blocks & formats) per site in the Adsense (reports). Or am I doing something wrong? If not, I have to create different custom channels per site to see how certain ads are doing on a site level?
[ -0.0020620685536414385, 0.025818761438131332, -0.006621159613132477, 0.010559196583926678, 0.00183470593765378, -0.008454802446067333, 0.007366127800196409, 0.003301019547507167, -0.020562784746289253, 0.02911507897078991, -0.009397249668836594, 0.013490529730916023, -0.004694705363363028, ...
[ 0.9521896839141846, 0.09317292273044586, 0.46734490990638733, 0.3005835711956024, -0.3653462827205658, 0.0658164918422699, 0.12706981599330902, -0.20595155656337738, -0.6592892408370972, -0.5999469757080078, 0.6109219789505005, 0.5228447914123535, -0.3366750180721283, 0.24788840115070343, ...
In programming contests like Google CodeJam, Facebook Hacker Cup, each question is provided with an input file for testing one's code for that program. I get struck at this point only. I don't understand whether that input file should be read by my program or it is just for testing the correctness of my program for various test cases given in that input file and I can copy past each test case given in the input file one by one manually and see whether the output provided by my program is right or wrong. It would be very helpful for me if anyone who has been actively participating in such competitive contests can share their views.
[ -0.0023660967126488686, 0.011532463133335114, -0.0011080089025199413, 0.009016647934913635, 0.010288795456290245, 0.03054739721119404, 0.0077663627453148365, 0.004656015895307064, -0.015830226242542267, -0.0033449181355535984, -0.00876929983496666, 0.009072085842490196, 0.012036226689815521,...
[ 0.3021824359893799, 0.43203678727149963, -0.26907360553741455, 0.5380736589431763, -0.48655447363853455, -0.17362353205680847, 0.016920529305934906, 0.20502597093582153, -0.057002559304237366, -0.3940153121948242, -0.08714528381824493, 0.28415900468826294, -0.013901972211897373, 0.03518872...
I need to create a guide or wiki that outlines different standardized techniques for marking up reusable elements in our software. When adding new features our developers are often able to reuse or mimic existing markup, and I'd like to make this as easy as possible for them. In order to do this now they have to dig through the interface to find something that's close enough. Are there any good examples online that show code samples for standardized markup? Is there a tool (wiki, etc.) that would make this somewhat easy to create?
[ 0.013464482501149178, 0.008392819203436375, -0.01230545062571764, 0.005789803806692362, 0.0111628333106637, 0.011301889084279537, 0.006463020108640194, 0.015282402746379375, -0.0198542270809412, -0.0052962396293878555, 0.004435205366462469, 0.006626937072724104, 0.0030717819463461637, 0.00...
[ 0.7356225252151489, 0.3071659207344055, -0.24896064400672913, 0.25893542170524597, 0.014286857098340988, 0.0701882392168045, 0.16833266615867615, -0.025674697011709213, -0.1706952005624771, -0.5438309907913208, 0.23473110795021057, 0.34067586064338684, -0.05743375048041344, 0.1448299139738...
In The Secret World, is it possible to cancel a quest? Without completing each quest that you pick up the only way that I have found is to select another quest that will 'replace' the quest you're currently on. Without replacing the quest you can move onto the next area but there will consistently be a quest mark on the Agartha entrance until such a time you replace the quest. So, is it possible to cancel a quest in The Secret World or are the only options to complete/replace the quest?
[ 0.014909432269632816, 0.01673658937215805, 0.0013789469376206398, 0.011614570394158363, 0.0017164520686492324, -0.012580713257193565, 0.008998773992061615, -0.004239706788212061, -0.018475469201803207, 0.024973006919026375, -0.01632639765739441, 0.02275831066071987, -0.01817210577428341, 0...
[ 0.30134227871894836, -0.1623198539018631, 0.08229248225688934, 0.2915506660938263, -0.10305564850568771, -0.21681290864944458, 0.2647280991077423, 0.24477195739746094, -0.22538697719573975, -0.2977040410041809, -0.04949144273996353, 0.13887105882167816, -0.3449432849884033, 0.6083077788352...
bbpress has a large number of callbacks attached to various hooks. I am familiar with this feature at the wordpress level but I am unable to locate where bbpress registers a large number of these callbacks. Specifically I would like to change the behaviour of the callback which is triggered via: do_action( 'bbp_template_before_forums_loop' ); If someone could help me locate this callback so that I can change the HTML it outputs that would be greatly appreciated.
[ 0.0008120047277770936, 0.006827315781265497, 0.0009120182367041707, 0.01509265135973692, 0.011458251625299454, 0.01101691834628582, 0.008368274196982384, 0.03766045719385147, -0.018881602212786674, -0.008212744258344173, -0.015261233784258366, 0.011539551429450512, -0.0058869956992566586, ...
[ 0.550361692905426, 0.07929504662752151, 0.25113236904144287, -0.1393343061208725, -0.3821251392364502, 0.025766195729374886, 0.1554262340068817, 0.10747536271810532, -0.41674506664276123, -0.4160730242729187, 0.06098276004195213, 0.32023555040359497, -0.5194622874259949, 0.0123686501756310...
Currently accepted scientific theory says that all elements heavier than hydrogen, helium, and a little lithium have been created in supernova explosions. My questions, specifically, is has anyone done the math to see if the observed amounts and the distribution of, the heavier elements in the universe agrees with the 13.7 billion year age of the universe? With the birth, life, and death of stars at the currently observed and accepted rates... does the math work out? Can you please refer me to a resource where I can review the material or get further clarification. (I freely admit that my maths aren't up to the task, but I'm hoping for a deeper understanding of the phenomenon.)
[ -0.004218744114041328, 0.008672854863107204, -0.014398027211427689, -0.0018166456138715148, 0.012663865461945534, -0.008416365832090378, 0.0069466400891542435, 0.013731695711612701, -0.013668611645698547, -0.03645419701933861, 0.002650336129590869, 0.008225416764616966, -0.004501680377870798...
[ 0.6197842359542847, 0.6805031299591064, -0.2974267899990082, 0.042408447712659836, 0.023652544245123863, -0.42515942454338074, 0.027533402666449547, 0.07252894341945648, -0.4047068953514099, 0.10029617697000504, -0.094831183552742, 0.08160147815942764, -0.15170244872570038, 0.5291112661361...
Among the literature, some line charts use mean/SE but others use mean/SD. Which is correct or better?
[ -0.0047804866917431355, 0.04293918237090111, -0.06655982881784439, 0.0095903892070055, -0.03756542503833771, 0.08376181125640869, 0.025578709319233894, 0.0393017902970314, -0.008604558184742928, -0.12441910058259964, -0.027418354526162148, 0.014826744794845581, -0.006653229705989361, 0.027...
[ 0.11572717130184174, -0.27542996406555176, 0.3202544152736664, 0.19309504330158234, -0.4009195566177368, -0.022744519636034966, 0.32143861055374146, -0.06545713543891907, 0.02833379991352558, -0.7145156264305115, -0.163199320435524, -0.21636059880256653, 0.4531301259994507, 0.0584143288433...
I'm running some latency measurements using Web Page Speed Test on a static javascript file. I compared hosting the file at four locations: Host Monster, Amazon S3, Amazon Cloudfront, and Heroku Static Sites. The results I'm getting are around 240ms-390ms, the worst being Host Monster. The tests were run from the default Web Page Speed Test data center at Dulles, VA, USA. The S3 zone I used was "US standard". My question is - isn't 240ms+ latency a bit too much for just serving a javascript file (33kb minified before gzip - Host Monster and Heroku have out- of-the-box gzip, it's a bit less seamless for Amazon Web Services do not so I didn't enable it for this test). What top latency should I expect when serving a static 30kb file from the US to the US? How can we to the sub 100ms latency zone? **Edit** Here is a sample latency breakdown from Web Page Speed Test on Amazon CloudFront: DNS Lookup: 49 ms Initial Connection: 59 ms Time to First Byte: 123 ms Content Download: 28 ms
[ -0.03576083108782768, -0.002354085212573409, -0.009514310397207737, 0.0026014382019639015, 0.0017956197261810303, 0.01921047642827034, 0.014141455292701721, -0.011290399357676506, -0.014803340658545494, 0.01589697040617466, -0.000998436938971281, 0.008951286785304546, 0.002997057978063822, ...
[ 0.368613600730896, 0.13889944553375244, 0.600967288017273, 0.005743886344134808, 0.09874008595943451, 0.028460199013352394, 0.41502660512924194, -0.06675656884908676, -0.08100437372922897, -0.5151479840278625, 0.008653378114104271, 0.32848218083381653, 0.14128006994724274, 0.12520951032638...
I know this has been covered everywhere, but I still don't get it. I `apt-get install git-core`. I'm trying to install websocketcpp here https://github.com/zaphoyd/websocketpp. I `wget https://github.com/zaphoyd/websocketpp.git`. I have no idea what to do next. Please help. (I've read http://stackoverflow.com/questions/315911/git-for-beginners-the- definitive-practical-guide and http://www.thegeekstuff.com/2011/08/git- install-configure/, but I'm obviously missing something)
[ -0.00489081721752882, -0.011910833418369293, 0.009807969443500042, 0.017155593261122704, -0.008652586489915848, -0.01389545202255249, 0.00701941829174757, -0.012336906045675278, -0.02043028734624386, -0.005571359768509865, 0.001358355046249926, 0.006539627909660339, -0.02443460375070572, 0...
[ 0.35091206431388855, 0.36334073543548584, 0.36317208409309387, 0.11464974284172058, 0.17820152640342712, -0.3216635584831238, 0.17424620687961578, 0.08677166700363159, -0.3060378134250641, -0.836941659450531, -0.04138821363449097, 0.669809877872467, -0.014798098243772984, 0.369272649288177...
# Problem I am using biblatex together with the `authortitle-dw` style. As it is desired in most cases, citations show up in the footnotes in the form "`name`, `shorttitle`, `pages` (if indicated)". However, as a historian, I work with a great amount of unpublished documents I've collected in various archives. To deal with these, I have modified the entrytype `@unpublished` in a rather dilettantish way (I guess). The problem is that in the footnotes, as it is expectable, only the title of my unpublished documents appears, while in this special case a full citation in the form "`title`, in: `archive`, `shelf`, Bd. `box`" would be desirable. The `\fullcite` command doesn't work for me here, because it doesn't allow the content of certain fields to be replaced by "ibid.", which is absolutely necessary. # Example Here's my MWE: \documentclass{article} \usepackage[ngerman]{babel} \usepackage[ style=authortitle-dw ]{biblatex} \DefineBibliographyStrings{ngerman}{ chapter = {Bd.} } \DeclareBibliographyDriver{unpublished}{% \usebibmacro{bibindex}% \usebibmacro{begentry}% \usebibmacro{title}% \newunit \usebibmacro{in:}% \printlist{institution}% \newunit\newblock \printlist{location}% \newunit\newblock \printfield{chapter}% \newunit\newblock \setunit{\bibpagerefpunct}\newblock \usebibmacro{pageref}% \usebibmacro{finentry}} \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @unpublished {test1, title = {Source1}, institution = {Archive1}, location = {Shelf1}, chapter = {Box1} } @unpublished {test2, title = {Source2}, institution = {Archive1}, location = {Shelf1}, chapter = {Box2} } @unpublished {test3, title = {Source3}, institution = {Archive1}, location = {Shelf1}, chapter = {Box2} } \end{filecontents} \addbibresource{\jobname.bib} \begin{document} \footcite{test1} \footcite{test2} \footcite{test3} \end{document} This example produces the following output: 1 Source1 2 Source2 3 Source3 However, the way I'd like to have it is the following: 1 Source1, in: Archive1, Shelf1, Bd. Box1 2 Source2, in: ibid., Bd. Box2 3 Source3, in: ibid. Is it possible to achieve this in any way?
[ 0.014346491545438766, 0.012863455340266228, -0.0028337596450001, 0.022191502153873444, 0.002192297950387001, -0.005072476342320442, 0.007218247279524803, 0.015268961898982525, -0.013312647119164467, -0.020003465935587883, -0.010104445740580559, 0.007160637062042952, -0.017514297738671303, ...
[ -0.12709259986877441, 0.5419346690177917, 0.4880518615245819, 0.0042630815878510475, -0.19261647760868073, -0.08258267492055893, 0.2924925684928894, 0.19717343151569366, -0.4046318829059601, -0.4780372977256775, -0.2663820683956146, 0.44828635454177856, -0.43340349197387695, 0.135283306241...
When including an Inverse Mill's ratio to account for selection in a mixed model using LMER in R (following Heckman's [1979] two-stage procedure), do I still need to estimate robust SEs or does the mixed model result in heteroscedasticity consistent SE?
[ 0.04410116374492645, 0.019607774913311005, -0.023601491004228592, 0.016017364338040352, -0.019643718376755714, 0.029575329273939133, 0.016378162428736687, -0.002909950213506818, -0.017664408311247826, -0.027723731473088264, -0.0006026180926710367, 0.01587897352874279, -0.02935105375945568, ...
[ -0.07565116137266159, -0.1981504112482071, -0.060989800840616226, -0.12779712677001953, -0.37923866510391235, 0.2510507106781006, 0.03429926931858063, -0.6109796166419983, -0.01536510605365038, -0.32868877053260803, 0.40202176570892334, 0.33749422430992126, -0.15195365250110626, 0.28928339...
I know the diametral pitches must match for spur gears in order for them to run together. How to check worm gear and worm? Thanks
[ -0.01686893403530121, 0.01316372025758028, -0.012450643815100193, 0.04045933112502098, -0.016390668228268623, 0.09663084149360657, 0.02088790200650692, 0.0466892383992672, -0.033240530639886856, -0.027499724179506302, 0.014222180470824242, 0.02014087699353695, -0.016625547781586647, 0.0173...
[ 0.5725015997886658, -0.040023066103458405, -0.06927742063999176, 0.40245285630226135, 0.06247317045927048, 0.26257920265197754, 0.28560736775398254, 0.4470730423927307, 0.0189208984375, -0.5820642709732056, 0.49155476689338684, 0.6199486255645752, 0.4212818741798401, 0.2425737828016281, ...
After adding a WMS weather layer to my map I could see that it added many layers including several radar sub layers each one representing 5 minutes previously in time. So my current plan is to use the TimeExtent "slider" and if the user picks "weather" from the list of time layers it will change the visibility of each of these layers on each "tick" of the slider to make it animate. 1.) Is this the best way to do this? Are there any drawbacks to this hack. 2.) Is there a better way? Have you added animated weather radar to your silverlight map?
[ 0.00695703225210309, 0.010144289582967758, -0.014760136604309082, 0.009861142374575138, 0.0069178929552435875, -0.009488707408308983, 0.007396417669951916, 0.0012296151835471392, -0.013776984065771103, -0.0043653459288179874, -0.003795467084273696, 0.01763976737856865, -0.007244824431836605,...
[ 0.7891299724578857, -0.21712373197078705, 0.5653578042984009, 0.40328601002693176, -0.06262758374214172, -0.12063474208116531, 0.3421376943588257, -0.16222643852233887, -0.2545579671859741, -0.7869170904159546, 0.3827604651451111, 0.08605337142944336, -0.12787893414497375, 0.09355625510215...
I will be starting a postdoc in Astrophysics, particularly VHE gamma-rays with the VERITAS and HAWK observatories. I am coming from particle physics and am looking for some good books to get me up to speed. I have taken 1 astrophysics course, but it didn't even have a book, so I would like some options on both general overviews, and some more detailed to VHE gamma-ray in specific. Thanks so much!
[ 0.014661988243460655, -0.003113129176199436, 0.008948429487645626, 0.014879124239087105, 0.007646039593964815, -0.011033294722437859, 0.008101345039904118, 0.007551244925707579, -0.02014143392443657, -0.00368508487008512, 0.0009542138432152569, 0.009143593721091747, -0.014783614315092564, ...
[ 0.80052250623703, 0.31738919019699097, 0.4678596556186676, 0.18121051788330078, 0.0373646579682827, -0.6124762296676636, 0.3346908688545227, 0.8267495036125183, -0.23229451477527618, -0.6024121642112732, -0.17750878632068634, 0.46235188841819763, 0.7612020373344421, 0.08799721300601959, ...
EDIT: It was simply a matter of outdated package versions. On a mac there are those useful commands sudo tlmgr update –self and after that sudo tlmgr update –all. The following throws the error "undefined control sequence", but when I switch to \tcbmaketheorem instead it works. Whats the matter? \documentclass{article} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage{lmodern} \usepackage[ngerman]{babel} \usepackage{amsmath} \usepackage[theorems]{tcolorbox} %\tcbmaketheorem{mytheo}{My Theorem}{colback=green!5,colframe=green!35!black,fonttitle=\bfseries}{th} \newtcbtheorem{mytheo}{My Theorem}{colback=green!5,colframe=green!35!black,fonttitle=\bfseries}{th} \begin{document} \begin{mytheo}{}{} Great theorem with beautiful box. \end{mytheo} \end{document}
[ 0.013265659101307392, 0.0074775684624910355, -0.0029673806857317686, 0.003559303004294634, 0.008722382597625256, 0.0061091408133506775, 0.0065637980587780476, 0.0009225686080753803, -0.010801229625940323, -0.03131968155503273, -0.01811148226261139, 0.00019709323532879353, -0.0147864017635583...
[ -0.06531544029712677, 0.14395369589328766, 0.5849804282188416, -0.14732593297958374, -0.06225935369729996, -0.1052762046456337, 0.40821218490600586, -0.21032774448394775, -0.18187160789966583, -0.4980615973472595, 0.12380584329366684, 0.7210923433303833, -0.40950530767440796, 0.19070784747...
I had a outage incident a few days ago and the support team has asked for any available logs for the period, while thinking about how to do this, it occurred to me that it would be generally handy to be able to parse many log files at once for entries between 2 dates. I see many tools that would make this possible, had we had the foresight to configure them, not least; `splunk`, `syslog-ng`, `rsyslog`, and others that stash logs to some searchable backend such as mysql. (I am definitely planning to install a central log host to fulfill this requirement) However for the short term I am interested in a tool that would blaze through the `/var/log/*` folder and dump to stdout, preferably with a simple options syntax (find options are nice +mtime etc) and written in something that can do this fast (so probably C rather than perl or ruby) I see a bunch of tools that I am looking at which may or may not be fit for purpose, such as `mergelog`, `logpp`, `sec`, but none are really jumping out as obviously good for this, which is odd because I would expect this to be the sort of thing that is done to death on unix tools.
[ 0.0134676368907094, 0.02289051190018654, -0.016448799520730972, 0.00786909181624651, 0.006712319329380989, 0.014498026110231876, 0.006586638279259205, 0.03492913395166397, -0.01787419244647026, 0.0024783886037766933, 0.00009992835111916065, 0.009865489788353443, -0.008749379776418209, -0.0...
[ 0.22148801386356354, 0.07643894851207733, 0.27872270345687866, 0.023353327065706253, 0.2545326352119446, -0.38521549105644226, 0.37392130494117737, 0.22186370193958282, -0.62375807762146, -0.32954302430152893, 0.0007119159563444555, 0.3148873448371887, -0.11998119950294495, 0.1035257801413...
I need to display two single quotes without anything between them (I'm preparing computer science lecture slides and I want to display `'\''`), but of course LaTeX wants to translate `''` to closing double quotes. What can I do to bypass what is now a limitation for me?
[ 0.027114292606711388, 0.010572054423391819, -0.02410782314836979, 0.009068381041288376, -0.032664231956005096, -0.007336951792240143, 0.010334745049476624, 0.012207644060254097, -0.02461768127977848, -0.02377474494278431, -0.02066810056567192, -0.011890433728694916, -0.008699703961610794, ...
[ 0.41577425599098206, 0.5302454233169556, 0.3501131534576416, -0.023155350238084793, -0.02322581596672535, 0.00067837507231161, 0.1472426801919937, 0.2942127287387848, 0.07371632009744644, -0.513495922088623, 0.02175108529627323, 0.4471054673194885, -0.11960098892450333, 0.3873099982738495,...
I have defined the following \providecommand*{\lstnumberautorefname}{line} and within my text \begin{lstlisting} |\label{sdd_1}|<sqlCallStatistics>YES</sqlCallStatistics> \end{lstlisting} When I now use in my text xxx. \autoref{sdd_1} then I receive the following output : xxxx. line 1 instead of (what I had expected) xxxx. Line 1
[ 0.011507936753332615, 0.01820039004087448, -0.01134512946009636, 0.010287470184266567, -0.008912713266909122, -0.005203068722039461, 0.00919623114168644, 0.01865365542471409, -0.009620155207812786, -0.01475185714662075, -0.012688416987657547, 0.0006000683642923832, -0.005850722081959248, 0...
[ -0.07896270602941513, 0.06589686125516891, 0.6791859865188599, -0.46753570437431335, -0.12536725401878357, 0.307636559009552, 0.07502471655607224, -0.24857041239738464, 0.02041744440793991, -0.7437006831169128, -0.24993237853050232, 0.35253575444221497, -0.24540701508522034, 0.217274010181...
I necessarily deleted a folder called app (it contained the basic apps .apk files and other files for example: calculator.apk) on my samsung galaxy mini. After that everything seemed to froze, so I restarted my phone and now it only repeats the samsung intro forever. Can you help me? Should I reinstall the android os, if so how can I do it? I alreay tried installing android 4.2 but when I went to the screen where I selected update form sd card and chose the file of android 4.2 it returned me an error "status 0". Please help me asap!
[ -0.008704360574483871, 0.0019805829506367445, -0.0016317926347255707, 0.02125830017030239, -0.009450153447687626, 0.02566695399582386, 0.006098203826695681, 0.020646588876843452, -0.020628593862056732, -0.016944732517004013, -0.01998681202530861, 0.009096269495785236, 0.016717255115509033, ...
[ 0.2510053515434265, 0.22309333086013794, 0.7442885041236877, 0.018846675753593445, -0.1180298775434494, 0.21419920027256012, 0.8324560523033142, 0.114386186003685, -0.3755994141101837, -0.4032118022441864, 0.05010068789124489, 0.7243703007698059, -0.44441792368888855, 0.15200108289718628, ...
I'm working with a client who wants visitors to be able to post comments on categories instead of on individual posts, but I am unsure on how to go about it. Is it possible to display the comment form in a category? Just pasting in `<?php comments_template(); ?>` didn't do it, unfortunately.
[ -0.0027110481169074774, 0.006690665148198605, -0.007833214476704597, 0.030412863940000534, 0.003577243769541383, -0.006744819227606058, 0.006209843792021275, 0.007126313168555498, -0.029616687446832657, -0.006005192641168833, -0.014678641222417355, 0.0031076965387910604, 0.012687106616795063...
[ 0.6272569298744202, 0.36065566539764404, 0.2424619346857071, 0.012422872707247734, -0.326717734336853, -0.2574247717857361, 0.07344554364681244, 0.22544662654399872, -0.4518495202064514, -0.4578145444393158, 0.2558978796005249, 0.25294229388237, -0.044708002358675, 0.36361363530158997, -...
As discussed in user25658's answer to this question, when one wants to compute $$ \beta = \mathbb{E}(x^Tx)^{-1} \mathbb{E}(x^TY) $$ but $\mathbb{E}(x^Tx)$ is not invertible, $\beta$ is not uniquely identified. A characterization of all possible values of $\beta$ is given by $$\beta= \mathbb{E}(x^Tx)^g\mathbb{E}(x^Ty)$$ where $\mathbb{E}(x^Tx)^g$ is the generalized inverse of $\mathbb{E}(x^Tx)$. In Hansen, Econometrics, pg 34, section 2.18 (as of today's version on Hansen's webiste) it is written that $x^T\beta = x^T \mathbb{E}(x^Tx)^g\mathbb{E}(x^Ty) $ is however uniquely identified. This result is not completely counterintuitive to me. I understand that that this is matrix multiplication and that the fact that for every $x$ there may be multiple $\beta(x)$, say $\beta(x)_1 \neq \beta(x)_2$, does not imply $x^T\beta(x)_1 \neq x^T\beta(x)_2$. Nevertheless, I have been unsuccessful in trying to prove the claim. Could someone help me with that?
[ -0.004286690149456263, 0.01766267791390419, -0.007076991721987724, 0.003828579094260931, 0.02244270220398903, 0.004247643984854221, 0.005545199848711491, -0.011075478047132492, -0.009223589673638344, -0.018765471875667572, -0.01171728316694498, 0.006227003410458565, -0.018130827695131302, ...
[ -0.2928004860877991, -0.01317349448800087, 0.07016162574291229, -0.36630958318710327, 0.06251148879528046, 0.27215635776519775, 0.15584781765937805, -0.30376559495925903, 0.19184629619121552, -0.5551415681838989, -0.2179582118988037, 0.23278038203716278, -0.5292513966560364, 0.316440731287...
We had an old site that had html pages in it. One of the pages in that site was linked to from many other sites. Now we moved to a WordPress site, and we created a page with that name, but of course it doesn't have the ".html" extension, so links to that page lead to an empty page... I saw a plugin that adds html extesnsions to all pages, but I don't want that - I have only one page to which I want to add .html. What can I do to accomplish this?
[ -0.0014617688721045852, 0.0052548968233168125, 0.006325920112431049, 0.01720157079398632, 0.006301301531493664, -0.003177125472575426, 0.004919489845633507, 0.011624084785580635, -0.019462930038571358, -0.036202866584062576, -0.0017405207036063075, 0.004972349386662245, -0.00966101698577404,...
[ 0.44675958156585693, 0.1706790328025818, 0.49716582894325256, 0.17454248666763306, 0.005039735697209835, 0.018437767401337624, 0.20267988741397858, 0.14335878193378448, -0.20881858468055725, -0.4777907729148865, 0.12144934386014938, -0.015421660616993904, -0.31649351119995117, 0.4286161363...
Although we know that day which we are talking about, why don't we say "What the beautiful day!" instead of "What a beautiful day!" ? In addition, a friend of mine have already said the following, but in my mind it is not persuadable- It couldn't conceive me yet. I am confused. Generally, we use 'a' for a generic thing, when many of them exist, and use 'the' when there is only one of them. This also depends on the context. Here, we've experienced many other days, and will most likely experience many more, so it's 'a' beautiful day. For another example, you might go to purchase 'a car' (since the place you're going has many), but might take 'the car' to work (if you only own one car yourself). In politics, you would have 'a Senator' (since there are 100, and two to each state), but 'the President' (since there is only one at a time).
[ -0.01869046874344349, 0.016652854159474373, -0.004824588540941477, 0.011847851797938347, -0.011996607296168804, -0.012616749852895737, 0.00635701697319746, -0.008150476031005383, -0.005344369448721409, -0.025370610877871513, 0.0014726700028404593, 0.003335843561217189, -0.0013712088111788034...
[ 0.38950061798095703, 0.22449621558189392, 0.10469749569892883, -0.4112841784954071, -0.25823232531547546, -0.11408888548612595, 0.6398623585700989, 0.4471474289894104, -0.35117995738983154, -0.48656585812568665, -0.03659474477171898, 0.23593559861183167, 0.0774601399898529, 0.6622722744941...
Im trying to figure out how to create a elevation raster from a known location (3D polygon) that projects at a specified slope (1%) for a specified distance(1-2 km). Does anyone have any idea how to do this?
[ -0.036851666867733, 0.02360064722597599, -0.01739809848368168, 0.019145473837852478, -0.03559372201561928, -0.022422417998313904, 0.013953248970210552, -0.023820390924811363, -0.03371676430106163, -0.00016226971638388932, 0.008356952108442783, 0.007769383490085602, 0.030854519456624985, 0....
[ 0.3454313278198242, 0.42009657621383667, 0.3812575936317444, 0.2124757468700409, -0.022263679653406143, 0.13127775490283966, 0.42164450883865356, -0.09543224424123764, 0.25871098041534424, -0.5122990608215332, 0.06939271837472916, 0.42374739050865173, -0.02398601919412613, -0.3714313209056...
What's next in the web industry, as social media reaches a plateau, what are the next milestones in this "experiment"? I guess I'm looking for in depth opinions on what's next in our beloved industry? What's beyond facebook, wikis, collaboration, shiny buttons, and interactive instantaneous communications? What will drive the next dot com boom?
[ -0.0013569785514846444, -0.009758890606462955, 0.008220503106713295, 0.008536448702216148, -0.01601008139550686, -0.011652779765427113, 0.01032643299549818, 0.014801664277911186, -0.026141002774238586, -0.037027742713689804, -0.004424204584211111, 0.010037967935204506, 0.031759653240442276, ...
[ 0.719578742980957, -0.10614190250635147, 0.3710651099681854, 0.34190601110458374, 0.32440027594566345, -0.18599377572536469, -0.22942692041397095, 0.5739672780036926, -0.4716935157775879, -0.7570852637290955, 0.37244781851768494, 0.4365008771419525, 0.38283389806747437, 0.20741412043571472...
In the collapsed Gibbs sampling version of LDA, the posterior distribution of topic assignments for each word is sampled. From what I have read (e.g. http://people.cs.umass.edu/~wallach/courses/s11/cmpsci791ss/readings/griffiths02gibbs.pdf) it looks like people just implement the Gibbs sampling procedure by letting it run until we are sure that we are sampling from the posterior then take the last sample as a good topic assignment. But shouldn't we be a bit more rigorous than this? For example, what if the sample we take happens to be from a part of the posterior with low probability (this would be unlikely but possible)? Wouldn't it be better to do something like take a Bayes estimator (or average, or mode) of the samples to make sure that we are using a topic assignment that is actually likely?
[ 0.002688984153792262, -0.005864029284566641, -0.010905123315751553, 0.016983389854431152, 0.0009820755803957582, -0.011897227726876736, 0.008579462766647339, -0.00019849208183586597, -0.01207477692514658, 0.008530722931027412, -0.00573806744068861, 0.007375581655651331, 0.00025735585950315, ...
[ -0.09369997680187225, -0.4198780059814453, -0.33377522230148315, -0.015509593300521374, -0.2249913215637207, 0.005779010709375143, 0.3681944012641907, -0.22065027058124542, 0.06518768519163132, -0.7132332921028137, -0.1781228929758072, 0.2743992209434509, -0.41603413224220276, 0.2447271049...
From a DD-WRT router I want to create a cron-job that wget's an image from a remote location. This image is to be passed on to a remote FTP server for storage. The process is to be repeated every minute, naming the destination image on the ftp with the current time : wget: get file from -> http://37.***.9.95/fullsize.jpg ftp: put fullsize.jpg -> ftp://83.***.151.224/shares/camera/dateTimeNow.jpg Where dateTimeNow is the current time. The ftp server for storage is not smart enough to have either cron nor something else than just FTP. the DD-WRT device however, has cron!
[ -0.010145753622055054, 0.0041869706474244595, -0.005662618204951286, 0.004907434806227684, -0.006205455400049686, -0.017369046807289124, 0.006207223050296307, 0.019408024847507477, -0.014954173937439919, 0.011403992772102356, -0.016312487423419952, 0.006935412995517254, 0.009171620942652225,...
[ 0.2276657074689865, 0.05859456583857536, 0.7724394798278809, 0.04852888360619545, -0.07089320570230484, 0.08494922518730164, -0.00009061570017365739, -0.16475607454776764, -0.4067898690700531, -0.6145159602165222, -0.36143481731414795, 0.4457740783691406, -0.22782017290592194, 0.2117934823...
One of the final bee-related quests requires you to breed bees to produce certus combs. In order to get this bee, you need to have access to ender bees. How can I get ender bees in Agrarian Skies?
[ 0.018247714266180992, -0.000299385137623176, 0.0023064231500029564, 0.0015192247228696942, -0.029011594131588936, 0.02787010744214058, 0.013077111914753914, -0.0010051580611616373, -0.027844656258821487, 0.015990402549505234, -0.0037353006191551685, 0.02151370979845524, -0.00933136697858572,...
[ 0.2930094003677368, -0.06727016717195511, 0.2114192694425583, 0.5007308721542358, -0.05355583503842354, 0.16827626526355743, 0.6945000290870667, 0.15144918859004974, -0.0975412055850029, -0.3510095179080963, 0.39983227849006653, 0.03257814794778824, 0.07818718999624252, 0.23989003896713257...
What is the difference between the words _random want_ and _whim_? The definition of **whim** can be easily found in the dictionary: > a sudden wish to do or have something, especially when it is something > unusual or unnecessary **Random want** is not on the dictionary. It turns out to be just two words combined together. But it seems to me to have nearly the same meaning as _whim_. I came across this word chunk in the phrase made up by a native American: > They thought the problem to be settled was how on earth to get posters to > understand that CONTEXT is not just some **random WANT** on the part of > those who would like to answer, but an ABSOLUTE REQUIREMENT. But later in the discussion the native speaker writes: > random want fits better than whim , which suggest that there is some difference for him between the words _whim_ and _random want_. So, what is the difference?
[ -0.012698987498879433, 0.013784595765173435, -0.00271895295009017, 0.024860963225364685, -0.02795742265880108, -0.017889879643917084, 0.009053231216967106, 0.0009403221774846315, -0.01570449396967888, 0.036152299493551254, -0.008337482810020447, 0.0024527288042008877, 0.023651279509067535, ...
[ 0.9681875109672546, -0.21353404223918915, 0.3201734125614166, 0.03463004529476166, -0.1301022320985794, 0.2560809254646301, -0.09784231334924698, 0.028612393885850906, -0.4781397879123688, -0.6736974120140076, 0.5297513604164124, 0.13425758481025696, -0.32276856899261475, 0.166239961981773...
Note: My question is duplicate of Why doesn't water come out of tap/faucet at high pressure when I turn it on?. None of the answers given there explains how the continuity equation fits properly. That's why I am asking this question. * * * I do not understand the _Hydraulic analogy_ explained on wikipedea. > I asked my _Network analysis_ teacher for an explanation . > He told me that cross-section area of a pipe represents its resistance and > flow of water represents current passing through that pipe. Like in a tap > when we change the area of aperture the flow rate of water changes, this is > analogous to change in current in a wire due to the change cross-section > area of the wire. > > * * * > > There is a contradiction in my understanding. **Contradiction** : When we decrease the area of the mouth of the tap by our thumb the amount of water flowing out remains same but if we decrease the area of aperture of the tap by turning the knob the amount of water flowing out decreases. **why?** Is it due to the change in type of flow i.e the flow changes from _laminar_ to _turbulent_ or _choked_? ![image](http://upload.wikimedia.org/wikipedia/commons/thumb/1/14/Tap.png/220px- Tap.png) I found a flaw by applying equation of continuity as: Suppose two parallel resisters are connected to voltage source as shown: ![image 1](http://images.tutorvista.com/content/electricity/parallel- resistors-energised-battery.jpeg) Let's name the three wires as _pipe-0_ (having the battery) , _pipe-1_ (having resister $R_1$) and _pipe 2_ (having resister $R_2$). > All three pipes are of same length. Resistances of different wires are > equivalent to the respective aperture radius of different pipes. Our analogy is to interchange the terms current $i=dq/dt$ (amount of charge crossed in a unit time) with flow rate of water i.e $i \equiv dm/dt$ (amount of mass of water crossed in unit time). > Let's remove the wire having $R_2$. The current $I$ in _pipe 0_ will decrease as the resistance of circuit increase. removing the resistance is analogous to change the area of pipe-2 to $0$. On the other hand removing pipe 2 will not change water flow $dm/dt$ in _pipe-0_ because of _Equation of Continuity_. * * * I wrote my understanding about the situation to tell the community that my question is not _Home-work like_ **In essence my question is 1.How the tap works? And how can we apply the equation of continuity to the water flow when we turn the knob and when we cover the tap with thumb? 2\. Where I'm getting wrong with my understanding of the Hydraulic analogy.**
[ -0.026793574914336205, 0.005258324556052685, -0.0005101885180920362, 0.010132119990885258, -0.012093300931155682, -0.029946565628051758, 0.00766972079873085, 0.00965907797217369, -0.014510639011859894, 0.011255894787609577, -0.004110346548259258, 0.012057384476065636, 0.007560141384601593, ...
[ 0.5954968333244324, 0.23225238919258118, 0.2735802233219147, -0.04508127644658089, 0.0421934612095356, 0.16760694980621338, 0.3636340796947479, -0.30564942955970764, -0.23648904263973236, -0.38286465406417847, 0.2692222595214844, 0.6257734894752502, 0.12068070471286774, 0.28336071968078613...
I develop a C code on Debian. I am looking for a program to monitor memory consumption of my code in real-time. The only value I care is the maximum RAM consumption, it should hold the value even if the the program is killed (e.g. by SIGKILL) . Could anyone suggest a handy tool for it? Normally I use `top -p pidnumber`, but it outputs with too much delay and if my program gets killed, it does not show the RAM consumption just before it is ended.
[ 0.004910571966320276, -0.005265532061457634, -0.006583946757018566, -0.002563349436968565, 0.005183411296457052, -0.016059406101703644, 0.007051276508718729, 0.0031176141928881407, -0.016237225383520126, 0.006632679142057896, -0.005127455573529005, 0.004989855457097292, 0.013782422989606857,...
[ 0.5391797423362732, 0.3569709062576294, -0.24857506155967712, -0.040209244936704636, -0.022232310846447945, -0.2242925614118576, 0.010706641711294651, 0.05768927186727524, -0.14358018338680267, -0.2735542058944702, 0.07119684666395187, 0.6190202832221985, -0.498017817735672, -0.06067924201...
I just moved from Ubuntu to Debian Wheezy and I thought I'd be finally free of all the unnecessary clutter. I guess I was wrong. Here is the situation: In the `Downloads` folder, Nautilus displays the following message: > Personal File Sharing > > _You can receive files over Bluetooth into this folder_ ![enter image description here](http://i.stack.imgur.com/RHtf0.png) I'd really like that to go away. My PC is not bluetooth-enabled, and even if it were, I'd figure out a way to share my files without the constant reminder. But I digress. Here's what I've tried so far: * The launch preferences dialog doesn't seem to make the notification go away, even when I enable sharing over the non-existent bluetooth device * Nautilus' `Preferences` dialog doesn't present a relevant option * Removing the `bluez` package via `synaptic` threatens to take the entire Gnome desktop with it. Same with trying to remove `gedit` or, indeed, `nautilus`. I feel like having Jules Winnfield over my head saying "I dare you. I _double_ dare you". In my previous Ubuntu installation I managed to remove bluetooth-related packages by switching to Thunar (and getting rid of the entire Unity family). This time I'd like to keep Nautilus if possible. The presence of a bluetooth package is something I can deal with later, what I'm worried about is the notification. Is there any way to make it go away? * * * ### Update: I decided to manually remove dependencies from the .deb files. Using this for guidance, I removed pretty much all but the most fundamental packages (such as X11 or sound packages) from `gnome`, `gnome-core` and `task-gnome-desktop`. I must have missed something though because removing unwanted packages such as bluetooth still wanted to take away Gnome. * * * ### Update 2: It turns out neither `gnome` nor `gnome-core` are essential to the system. In fact, from what I can tell, their only job is to provide dependencies and recommendations. I went ahead and purged them. This cased aptitude to see a bunch of useful packages as "no longer needed" and required to set the "manual install" flag on the ones I wanted - fair enough. A happy side-effect, bluetooth-related packages are gone. I did get rid of `nautilus-share` though as @grawity suggested. ~~This did not affect the notification at all. The problem remains~~. After a re-login, the offending area had vanished.
[ 0.003698123386129737, -0.004161311313509941, 0.005639340728521347, 0.0034791994839906693, 0.0328703373670578, -0.03501998633146286, 0.0036690514534711838, 0.005641178227961063, -0.012384383007884026, 0.010188003070652485, -0.02284756861627102, 0.0070350281894207, -0.022405892610549927, 0.0...
[ 0.17822696268558502, 0.12411567568778992, 0.24075797200202942, 0.037797193974256516, -0.025056596845388412, -0.18091139197349548, -0.008103392086923122, 0.6475886106491089, -0.579552173614502, -0.27142879366874695, -0.2657318115234375, 0.34411299228668213, -0.1451331526041031, 0.4740411341...
I have a classification problem where getting true positives is much more important than true negatives. To be clear, I know that roughly 10% of my population are actual positives, but I can assign some proportion (say 30%) of the population to be classified as positives without much cost, above all else I need to make sure that the actual positives are covered by this set. Unfortunately the classification tools I am using in WEKA seem to be balancing precision and recall, so that it is a) not assigning as many positives as it is allowed to, and b) getting quite a bad recall value. Is there a standard way of approaching this problem? My first guess would be to weight the cost function towards recall rather than F-score, but I don't see an easy way to do this in WEKA.
[ 0.01333252340555191, 0.007023458369076252, -0.014705697074532509, 0.021940328180789948, -0.013435116037726402, 0.009130969643592834, 0.007530934177339077, -0.00901952013373375, -0.018839947879314423, -0.0011672251857817173, -0.003685245756059885, 0.012354890815913677, -0.01201421208679676, ...
[ 0.3925219178199768, 0.36029744148254395, -0.12231861054897308, 0.1873447597026825, -0.41347548365592957, 0.13192754983901978, 0.515924334526062, 0.02923126146197319, -0.329264372587204, -0.39670106768608093, 0.12163019925355911, 0.13932017982006073, -0.15426790714263916, 0.2397157698869705...
here's a problem I have: I want my gantt bars to have two types of labels -- one on the side of the chart, and one inline -- and I'm having a hard time making it work. The code I use to generate the chart is listed below: \begin{ganttchart}[ vgrid, x unit=1cm, title height=1]{12}{12} \gantttitle{Quarters}{12} \\ \gantttitlelist{1,...,12}{1} \\ \gantttitle{Year 1}{4} \gantttitle{Year 2}{4} \gantttitle{Year 3}{4}\\ \ganttbar{WP1 Theoretical part}{1}{5} \\ \ganttbar{WP1 Experimental part}{2}{6} \\ \ganttbar{WP2 Theoretical part}{7}{11}\\ \ganttbar{WP2 Experimental part}{8}{12}\\ \ganttbar[bar/.append style={fill=gray,pattern=north east lines}]{WP3 Theoretical part}{6}{12}\\ \ganttbar[bar/.append style={fill=gray,pattern=north east lines}]{WP3 Experimental part}{7}{12} \end{ganttchart} And the idea is to have WPx labels on the left of the chart, but "Theoretical/experimental part" labels inside the bars. Is it possible to make this with pgfgantt?
[ 0.0026688817888498306, 0.006504883989691734, -0.004466206766664982, 0.00938736367970705, -0.008700461126863956, -0.012783952057361603, 0.007420147303491831, 0.012359365820884705, -0.0183415487408638, -0.007926699705421925, -0.0038471948355436325, -0.0004059466300532222, -0.007484326139092445...
[ 0.4729313552379608, 0.43946436047554016, 0.4655189514160156, -0.06315139681100845, 0.0018141252221539617, 0.1552104502916336, 0.19996578991413116, -0.6856006979942322, -0.2786279320716858, -0.7508569359779358, -0.03092094697058201, 0.22569502890110016, -0.08014888316392899, 0.0832063332200...
When I start a program that runs for a bit, say `make`, I could just type the next command in bash even before I got a prompt back and when the first command finished, the second one was executed as if I had typed it later. But suddenly this doesn't work any more. The additional command will be shown at the prompt after the first one finished but is not executed. If I type more commands, they are all shown at the prompt (without any space or newline between them). Any idea what is at fault here or how I can figure it out myself?
[ -0.005104993470013142, 0.0020871690940111876, -0.005439534317702055, 0.012213499285280704, 0.000033663585782051086, -0.00929933413863182, 0.005147761665284634, -0.003745472524315119, -0.013427842408418655, 0.012566667050123215, -0.01873566396534443, 0.005123776383697987, -0.00093900901265442...
[ 0.7984082698822021, 0.16997505724430084, 0.1339658498764038, -0.04803721606731415, 0.0396324060857296, 0.028174294158816338, 0.4073466360569, -0.29023823142051697, -0.1887095868587494, -0.5172622203826904, 0.10616867244243622, 0.6510542035102844, -0.08878351747989655, 0.3960179388523102, ...
Excess charge placed on a conductor distributes itself on the surface of a conductor. But why isn't the force which confines the charge to remain there, modelled in some way?
[ -0.021085316315293312, 0.01538628339767456, 0.0000908517322386615, 0.03274097666144371, 0.015251057222485542, -0.07231149822473526, 0.012969553470611572, -0.03354887291789055, -0.028786730021238327, 0.01253957487642765, 0.0019754553213715553, 0.03703662380576134, -0.015007304027676582, 0.0...
[ 0.5661184787750244, -0.15376822650432587, 0.5006598234176636, 0.2568519711494446, -0.1349223107099533, -0.3421655297279358, -0.2189655750989914, -0.47851285338401794, -0.14229504764080048, -0.4962415099143982, 0.16055810451507568, 0.46431198716163635, -0.12372968345880508, 0.49050667881965...
_( **Moderators note:** Title was originally "How can I add the "Page Attributes" and/or "Page Attributes > Template" selector to POSTS editor")_ WP currently only allows the assignment of a "template" to Pages (i.e. `post_type=='page'`.) I'd like to extend this functionality to Posts as well (i.e. `post_type=='post'`.) How can I add the _"Page Attributes"_ meta box and more specifically, the template switcher to the posts editor? I'm assuming this is code I will place in my `functions.php` for my theme. UPDATE: I've managed to add the hardcoded templates pulldown menu to my post editor, by simply adding the select box html to my existing custom meta options box. Here's the code I'm using for that... add_meta_box('categorydiv2', __('Post Options'), 'post_categories_meta_box_modified', 'post', 'side', 'high'); And here's the function that writes out the options and the template select box... //adds the custom categories box function post_categories_meta_box_modified() { global $post; if( get_post_meta($post->ID, '_noindex', true) ) $noindexChecked = " checked='checked'"; if( get_post_meta($post->ID, '_nofollow', true) ) $nofollowChecked = " checked='checked'"; ?> <div id="categories-all" class="ui-tabs-panel"> <ul id="categorychecklist" class="list:category categorychecklist form-no-clear"> <li id='noIndex' class="popular-category"><label class="selectit"><input value="noIndex" type="checkbox" name="chk_noIndex" id="chk_noIndex"<?php echo $noindexChecked ?> /> noindex</label></li> <li id='noFollow' class="popular-category"><label class="selectit"><input value="noFollow" type="checkbox" name="chk_noFollow" id="chk_noFollow"<?php echo $nofollowChecked ?> /> nofollow</label></li> </ul> <p><strong>Template</strong></p> <label class="screen-reader-text" for="page_template">Post Template</label><select name="page_template" id="page_template"> <option value='default'>Default Template</option> <option value='template-wide.php' >No Sidebar</option> <option value='template-salespage.php' >Salespage</option> </select> </div> <?php } And finally, the code to capture the selected values on save... function save_post_categories_meta($post_id) { if ( defined('DOING_AUTOSAVE') && DOING_AUTOSAVE ) return $post_id; $noIndex = $_POST['chk_noIndex']; $noFollow = $_POST['chk_noFollow']; update_post_meta( $post_id, '_noindex', $noIndex ); update_post_meta( $post_id, '_nofollow', $noFollow ); return $post_id; } Now, I believe all that's left is (1) capturing the selected template and adding it to the post meta for this post and (2) modifying index.php and single.php so that it uses the chosen template.
[ 0.002701272489503026, 0.0023442304227501154, 0.009140893816947937, 0.0229409858584404, 0.02477555349469185, 0.019426869228482246, 0.006862744223326445, 0.012409066781401634, -0.017206963151693344, 0.01639891415834427, -0.010485283099114895, 0.012788227759301662, 0.0003153369762003422, 0.01...
[ 0.6672735214233398, 0.2091415375471115, 0.5686482191085815, 0.08998828381299973, -0.21592094004154205, -0.29351314902305603, 0.32568633556365967, -0.48705151677131653, -0.02743404731154442, -0.5890524387359619, 0.22856396436691284, 0.5393480658531189, -0.17957668006420135, 0.19697901606559...
I wonder, if possible, how to create a matrix that looks like this: ![http://imageshack.us/photo/my- images/36/matrixbw.jpg/](http://i.stack.imgur.com/XfnL9.jpg)
[ 0.00002028957533184439, 0.013717476278543472, -0.008611978031694889, 0.01679980754852295, -0.009033248759806156, -0.018048416823148727, 0.0041213142685592175, 0.0017227530479431152, -0.030310101807117462, 0.005477766506373882, -0.011840817518532276, 0.013940954580903053, 0.004387104883790016...
[ 0.48354607820510864, 0.18991586565971375, 0.47505247592926025, -0.1636926680803299, -0.11511994153261185, -0.05504751577973366, -0.038162197917699814, -0.12585504353046417, -0.5304682850837708, -0.5064356923103333, 0.050792843103408813, 0.19617745280265808, -0.24028071761131287, 0.10698346...
I'm talking hypothetically but I'll use an example of what I worked on in a programming class a while ago. There's a better way to actually perform the operation, but that's not the point. An example to show what I mean I want to take a BufferedImage and individually manipulate each pixel. So something like this: for(int x = 0; x < out.getWidth(); x++) { for(int y = 0; y < out.getHeight(); y++) { int color = in.getRGB(x, y); int r = (color & 0xff0000) >>> 16; int g = (color & 0x00ff00) >>> 8; int b = (color & 0x0000ff); //Do something with r //Do something with g //Do something with b color = (r << 16) | (g <<8) | b; out.setRGB(x, y, color); } } return out; Would this loop be more or less efficient if I declared color, r, g and b outside of the for loops?
[ -0.013941321521997452, 0.008208210580050945, -0.004349823109805584, 0.0044829463586211205, -0.013164883479475975, -0.014187127351760864, 0.005376837216317654, 0.0018247272819280624, -0.014961136505007744, 0.011974876746535301, -0.005174135323613882, 0.005898683797568083, -0.00591868162155151...
[ 0.46712303161621094, -0.49988216161727905, 0.24895301461219788, -0.10404683649539948, 0.09873533248901367, 0.2664603292942047, 0.19213342666625977, -0.24487824738025665, -0.046338170766830444, -0.701473593711853, 0.18891355395317078, 0.3791300058364868, -0.5865827202796936, 0.0541511029005...
I want to make a tree like this one: ![original tree](http://i.stack.imgur.com/uGoB5.png) It doesn't need to be exactly the same, but I want to translate the text, so I need to rebuild it, I'd also want to add a node, but that's easy. My first try: \documentclass{standalone} \usepackage{booktabs} \usepackage{pgfplotstable} \usepackage{graphicx} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage{tikz} \usepackage{tikz-qtree} \begin{document} \begin{tikzpicture} \Tree [.{Métodos de otimização} [.{Métodos exatos} [.{Branch-and-X} [{Branch-and-bound} ] [{Branch-and-cut} ] [{Branch-and-price} ] ] [{Programação de restrições} ] [{Programação dinâmica} ] [{A*, IDA*} ] ] [.{Métodos aproximados} [.{Algoritmos Heurísticos} [.{Meta-heurísticas} [{Meta-heurísticas baseadas em solução única} ] [{Meta-heurísticas baseadasem população} ] ] [{Heurísticas específicas do problema} ] ] [{Algoritmos de aproximação} ] ] ] \end{tikzpicture} } \end{document} Giving me this ugly WIDE tree: ![enter image description here](http://i.stack.imgur.com/D0Ts9.png) The first problem is that I don't know how to break lines in each node. I tried `\\\\\\\` and it does nothing; should I use a tabular for each text line? The second problem is the subnodes of each subtree won't be placed on the same column. Is this fixable or hackable? Like "Metaheuristics" on the fourth row and "A*, IDA*" on the third, they are from different subtrees, but are on the same column. Any other ideas?
[ 0.001537685631774366, 0.007483168970793486, 0.0021214578300714493, 0.015291819348931313, -0.0011642036261036992, -0.017787694931030273, 0.005264647305011749, 0.031343989074230194, -0.019855335354804993, 0.002556468825787306, -0.0076086437329649925, -0.003160849679261446, -0.00143052684143185...
[ 0.15329262614250183, -0.05105162039399147, 0.3683324456214905, -0.2898373603820801, 0.06326230615377426, -0.03500888869166374, 0.5777012705802917, -0.18118853867053986, -0.2877102196216583, -0.8498924374580383, -0.003564264392480254, 0.2182207703590393, -0.33011019229888916, -0.14255501329...
Are my contacts encrypted on my Android and where are they? How can I access them? Any public APIs? (From PC)
[ 0.035042133182287216, 0.010022109374403954, 0.021117260679602623, 0.0037941799964755774, -0.028191817924380302, -0.018061146140098572, 0.01017184741795063, -0.004984972532838583, -0.01917160116136074, -0.029410069808363914, 0.003954790998250246, 0.027115846052765846, -0.0034576018806546926, ...
[ 0.4030137062072754, 0.44923338294029236, 0.27327290177345276, 0.21431118249893188, 0.3533938229084015, 0.23670761287212372, 0.605758547782898, 0.4416707456111908, -0.2545117139816284, -0.8372259736061096, 0.02353512868285179, 0.6573054790496826, -0.2052547037601471, 0.09510575979948044, ...
I'm trying to locate minima / maxima in matrices on a heat map. So I have my 2D matrix and did a `MatrixPlot`, ![my matrix plot](http://i.stack.imgur.com/84jU7.png) and I can find the position of the smallest value. My question: how do I place a graphical marker at that position? Like a box to outline field (7,6)? Completely stumped.
[ 0.0028802354354411364, -0.002357007237151265, -0.009020069614052773, 0.0072356644086539745, -0.0038286850322037935, 0.011826393194496632, 0.00619270047172904, -0.0005385970580391586, -0.021678488701581955, 0.0063486043363809586, -0.009419729001820087, 0.011608533561229706, -0.016417948529124...
[ 0.19326502084732056, 0.2646651268005371, 0.32806748151779175, -0.23889268934726715, -0.34396687150001526, 0.009735040366649628, 0.1699741631746292, -0.5464238524436951, 0.0014543852303177118, -0.5182877779006958, 0.09254468232393265, 0.058793939650058746, 0.05229148641228676, 0.23408089578...
I have recently set up a virtual machine and I want to share files between the host and the guest. Samba seems to be a perfect solution, and indeed it does all that I expect it to. However, if the host machine is not connected to the internet, the Samba server appears offline too. Can somebody show me how to make it work? Note: `ps -ef | grep smb` shows that `smbd -F` is running, so I think the process is started properly. EDIT: I'm using VirtualBox for the virtualization. Below is my config in `/etc/samba/smb.conf` (I have never changed it, comments removed for simplicity). Sharing is enabled through the GUI in nautilus. [global] workgroup = WORKGROUP server string = %h server (Samba, Ubuntu) dns proxy = no log file = /var/log/samba/log.%m max log size = 1000 syslog = 0 panic action = /usr/share/samba/panic-action %d encrypt passwords = true passdb backend = tdbsam obey pam restrictions = yes unix password sync = yes passwd program = /usr/bin/passwd %u passwd chat = *Enter\snew\s*\spassword:* %n\n *Retype\snew\s*\spassword:* %n\n *password\supdated\ssuccessfully* . pam password change = yes map to guest = bad user usershare allow guests = yes
[ -0.002848378848284483, 0.0012956802966073155, 0.011543739587068558, 0.013472220860421658, -0.01156750787049532, -0.004317862913012505, 0.005892234388738871, 0.012449109926819801, -0.011516676284372807, -0.013791820034384727, -0.016557037830352783, 0.008218510076403618, 0.0006748728919774294,...
[ 0.08412622660398483, 0.17932802438735962, 0.48540160059928894, 0.1014309972524643, 0.2804979383945465, -0.14911231398582458, 0.23255173861980438, -0.027509231120347977, -0.5115565061569214, -0.8256014585494995, 0.11783181875944138, 0.35399574041366577, -0.2598286271095276, 0.12329390645027...
I want to add bibliography to a document mainly written with Arabic language but intervened with some English text. Therefore, I opted for `arabxetex` approach rather than `polyglossia`, becasue IMHO gives you more flexibility to switch between the two directions than when you have much text in English and few text in Arabic (I am open to other opinions as well). Anyway, the problem is with bibliography and how to produce an Arabic counterpart to that of English without any transliteration or outright compromises. I came from this question trying to find an answer but it was in lualatex with a suboptimal output, and hence inadequate. **Questions** 1. As you can see only the citation key did appear, how to fix that? How and to make the citation numbers appear in Hindi numbers (Waahid-red, Ithnaan-green, Thalaatha,etc) instead of Arabic (1,2,3,etc)? 2. No `Bibliography` section did appear at the end of the document? And if it appears how to make it appear in Arabic language like `المراجع`? 3. What is the best approach to enter the fields of information for references in Arabic text, what am I missing? MWE Code \documentclass[oneside,16pt]{scrartcl} %=====================Graphics =================================== \usepackage[demo]{graphicx} \usepackage{xcolor} \usepackage{tikz} %=====================Floats ===================================== \usepackage{floatrow} \usepackage{float} \usepackage{scrhack} \usepackage{caption} \captionsetup[figure]{labelfont=bf,labelsep=period,format=plain,skip=0pt,font=small} \usepackage[framemethod=tikz]{mdframed} \usepackage{subcaption} %=====================Hyphenation ================================ \usepackage[none]{hyphenat} \usepackage[protrusion = true,final]{microtype} \emergencystretch=2em %=====================Hyperref =================================== \usepackage[hyperindex=true, hyperfootnotes=true]{hyperref} \hypersetup{ colorlinks, linkcolor={blue}, citecolor={blue}, urlcolor={blue} } %=====================Font issues =================================== \usepackage{fontspec} \usepackage{csquotes} \usepackage{amsmath,amsfonts,amssymb} \usepackage{arabxetex}% default options are "novoc" and "fdf2alif"; % other options are "voc", "fullvoc", "utf", "trans" and "fdf2noalif" \newfontfamily\arabicfont[Script=Arabic]{Amiri} % or put whatever open type Arabic font you have on your machine inside the two braces %=====================Bibliography =================================== \usepackage[style=verbose,backend=biber]{biblatex} \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @book{ibnsina:shifa:physics:ar, author = {ابن سينا}, year = {١٤١٧}, title = {السماع الطبيعى من كتاب الشفاء}, publisher = {دار المناهل}, location = {بيروت}, } \end{filecontents} \addbibresource{\jobname.bib} %=====================End of Preamble =================================== \begin{document} \begin{arab}[voc] وَهَذِهِ فِقرَةٌ بِالعَرَبِيَة مَعَ كَلِمَة اِنكلِيزِيَة رقم 1 او 2 باللغة اللعربية. \noindent (\textLR{Words in \textbf{English} with \textsc{small caps} or \textit{italics} are also possible}). عند الحاجة الى المصدر باللغة العربية ‎‎‎‎\cite{ibnsina:shifa:physics:ar} ‎ تذهب الى باب المراجع فلا تجده. \end{arab} \printbibliography \end{document} MWE Output ![enter image description here](http://i.stack.imgur.com/1RSVG.png) **Update 1** Based on the informative answer from PLK, some bugs are now fixed, what remains is the comma issue (circled in red), it is not in Arabic style (should be directed upwards instead of downwards). \documentclass[oneside,16pt]{scrartcl} \usepackage{fontspec} \usepackage{csquotes} \usepackage{polyglossia} \setmainlanguage{arabic} \setotherlanguage{english} \setmainfont{Arial} \newfontfamily\arabicfont[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \newfontfamily\arabicfontsf[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \newfontfamily\arabicfonttt[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \usepackage[style=numeric,backend=biber]{biblatex} \DefineBibliographyStrings{english}{% references = {المراجع}, } \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @book{ibnsina:shifa:physics:ar, author = {{ابن سينا}}, year = {1417}, title = {السماع الطبيعى من كتاب الشفاء}, publisher = {دار المناهل}, location = {بيروت}, } \end{filecontents} \addbibresource{\jobname.bib} %=====================End of Preamble =================================== \begin{document} وَهَذِهِ فِقرَةٌ بِالعَرَبِيَة مَعَ كَلِمَة اِنكلِيزِيَة رقم 1 او 2 باللغة اللعربية. \textenglish{(Words in \textbf{English} with \textsc{small caps} or \textit{italics} are also possible).} عند الحاجة الى المصدر باللغة العربية ‎‎‎‎\parencite{ibnsina:shifa:physics:ar}‎ تذهب الى باب المراجع فتجده، لكن الفارزة ليست على الطراز العربي بل الافرنجي. \printbibliography \end{document} **Output** ![enter image description here](http://i.stack.imgur.com/hw7ld.png) **Question Updated:** How to fix the comma? **Update 2** Based on PLK's great help to fix the comma issue, the final code is below: \documentclass[oneside,16pt]{scrartcl} \usepackage{fontspec} \usepackage{csquotes} \usepackage{polyglossia} \setmainlanguage{arabic} \setotherlanguage{english} \setmainfont{Arial} \newfontfamily\arabicfont[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \newfontfamily\arabicfontsf[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \newfontfamily\arabicfonttt[Script=Arabic,Mapping=arabicdigits]{USAMA NASKH} \usepackage[style=numeric,backend=biber]{biblatex} \DefineBibliographyStrings{english}{% references = {المراجع}, } \makeatletter \def\abx@comma{\ifdim\lastkern>\z@\unkern\fi\abx@puncthook{،}}% \makeatother \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @book{ibnsina:shifa:physics:ar, author = {{ابن سينا}}, year = {1417}, title = {السماع الطبيعى من كتاب الشفاء}, publisher = {دار المناهل}, location = {بيروت}, } \end{filecontents} \addbibresource{\jobname.bib} %=====================End of Preamble =================================== \begin{document} وَهَذِهِ فِقرَةٌ بِالعَرَبِيَة مَعَ كَلِمَة اِنكلِيزِيَة رقم 1 او 2 باللغة اللعربية. \textenglish{(Words in \textbf{English} with \textsc{small caps}, or \textit{italics} are also possible).} عند الحاجة الى المصدر باللغة العربية ‎‎‎‎\parencite{ibnsina:shifa:physics:ar}‎ تذهب الى باب المراجع فأنك حتما ستجده و مع الفارزة هذه المرة على الطراز العربي الاصيل. \printbibliography \end{document} **Output** ![enter image description here](http://i.stack.imgur.com/4ZFAT.png)
[ -0.0007485728710889816, 0.030511144548654556, -0.01374015025794506, 0.02319447137415409, -0.020900480449199677, 0.012615159153938293, 0.008841196075081825, 0.0147112300619483, -0.016275541856884956, -0.014891410246491432, -0.010163274593651295, 0.00003693241160362959, -0.04009391367435455, ...
[ 0.45695579051971436, 0.5096909403800964, -0.09133431315422058, 0.06648614257574081, -0.6200650930404663, -0.14654570817947388, -0.21262578666210175, 0.24735824763774872, -0.16555407643318176, -0.6935803294181824, 0.04817631095647812, 0.3543415069580078, -0.48036518692970276, 0.085561461746...
I have some centered text and mathematical stuff: \documentclass{article} \usepackage{amsmath} \begin{document} \begin{center} blablabla $\rho: H\longrightarrow GL_n(K)$ bla blabla bla blablablabla $H$ blabla blablabla blabla bla $\varphi:G\longrightarrow H$ blabla bla $\rho\circ\varphi:G\longrightarrow GL_n(K)$ blabla blablablabla bla $G$. \end{center} \end{document} Now, I'd like to give this piece of content a referable number, just like equations, and let the lines break automatically. I tried the following, but that didn't work. \documentclass{article} \usepackage{amsmath} \begin{document} \begin{equation} \begin{split} \text{blablabla $\rho: H\longrightarrow GL_n(K)$ bla blabla bla blablablabla $H$ blabla blablabla blabla bla $\varphi:G\longrightarrow H$ blabla bla $\rho\circ\varphi:G\longrightarrow GL_n(K)$ blabla blablablabla bla $G$.} \end{split} \end{equation} \end{document} ![enter image description here](http://i.stack.imgur.com/BUOWY.png) How can I solve this?
[ 0.010121221654117107, 0.005074084270745516, -0.003894967958331108, 0.0078959409147501, -0.006729140877723694, -0.005197761580348015, 0.005766786634922028, 0.01165512204170227, -0.009846147149801254, -0.010500103235244751, -0.0070177712477743626, -0.005366547964513302, 0.0016610726015642285, ...
[ 0.11434556543827057, 0.1766110211610794, 0.6494683623313904, -0.042468633502721786, 0.0029791120905429125, 0.1661314219236374, 0.16234108805656433, 0.0966607928276062, 0.09409471601247787, -0.5566378235816956, 0.15577848255634308, 0.18841201066970825, -0.14038649201393127, 0.16965292394161...
Is it possible to the output of `find … -exec ls -ls ;` sorted alpabetically, by filename? This is my cron command: find /home/setefgge/public_html -type f -ctime -1 -exec ls -ls {} \; This command works okay, for the most part. But the results are not sorted in any meaningful sequence. It would be very helpful if they would be sorted by the file name field.
[ 0.009398479014635086, 0.004079324658960104, -0.020554568618535995, 0.0029832404106855392, -0.013543774373829365, -0.0037633569445461035, 0.007459028623998165, 0.008330008015036583, -0.019668515771627426, 0.021373214200139046, -0.011323422193527222, 0.002924152184277773, 0.01210157759487629, ...
[ -0.27283474802970886, 0.311482697725296, 0.27209949493408203, -0.3122060298919678, 0.3698831796646118, 0.0037512858398258686, 0.020181993022561073, 0.10078846663236618, -0.15646329522132874, -0.6014053225517273, -0.13694648444652557, 0.7219404578208923, -0.00033329110010527074, 0.078735001...
I would like to use italicized superscript letters for footnotes rather than numbers. How could I do this? Example: > This is text. _a_ And more text. _b_ > > * * * > > _a_ Footnote 1. > > _b_ Footnote 2.
[ 0.008501690812408924, 0.003361542010679841, -0.022697949782013893, 0.024057665839791298, -0.0004215784720145166, -0.0057547553442418575, 0.01227541547268629, -0.002990553854033351, -0.03378589451313019, -0.009455148130655289, -0.015554179437458515, -0.017344415187835693, -0.01915851980447769...
[ 0.25624778866767883, 0.1940172016620636, 0.3157253563404083, -0.18980486690998077, -0.5791118741035461, 0.0413070023059845, 0.21140629053115845, 0.04019298031926155, -0.06563010066747665, -0.8351102471351624, 0.2917037606239319, 0.05830962583422661, -0.5293418169021606, -0.0100813079625368...
Method 1 (failed): tlmgr install siunitx sudo texhash Method 2 (failed): 1. Download zip folder containing .dtx .ins files 2. Unzip into MacSSD/Users/Username/Library/texmf directory 3. Run sudo texhash What am I doing wrong? * Mac OS X 10.8.5. * tlmgr revision 31673 * tlmgr using installation: /usr/local/texlive/2013 * TeX Live (http://tug.org/texlive) version 2013
[ 0.016774503514170647, -0.0028771148063242435, -0.009011970832943916, -0.007145494222640991, 0.01313998643308878, -0.020489457994699478, 0.010019328445196152, 0.009998152032494545, -0.016128242015838623, -0.02422529272735119, -0.007022809237241745, -0.014114894904196262, -0.009500504471361637...
[ 0.15800625085830688, 0.12736521661281586, 0.46458154916763306, -0.2324584722518921, -0.22017788887023926, 0.02578972838819027, 1.088491678237915, -0.6078609228134155, 0.051297999918460846, -0.08911693841218948, -0.0074293361976742744, 0.616454541683197, -0.41841763257980347, 0.157246693968...
I'm developing a mobile site which runs on port 8888. I want to see how it looks on my Android phone. How can I use my phone's web browser to view pages on port 8888?
[ 0.00026608776533976197, 0.004383134189993143, 0.013307939283549786, -0.010414463467895985, -0.07296347618103027, 0.024874744936823845, 0.011966777965426445, 0.04737751558423042, -0.02782483398914337, -0.0330633819103241, -0.008458131924271584, 0.0264061838388443, 0.008531903848052025, 0.03...
[ 0.5380560755729675, 0.39141348004341125, 0.33483126759529114, 0.10907632857561111, 0.24527272582054138, -0.1022181361913681, 0.22192172706127167, 0.34118592739105225, -0.21461006999015808, -0.8465847969055176, 0.1826399266719818, 0.39037320017814636, 0.0135102029889822, 0.2666495442390442,...
I have a dataset containing ( _x_ , _y_ ) information from the years 1985–2013. The _x_ data is the pick number and the _y_ data is performance. Each year has 60 ( _x_ , _y_ ) data points. When looking at my data for individual years, I personally think the median of performance of each pick number is a better representation of performance than the mean. I want to run a regression to look at the relationship between pick number and performance. Is it 'wrong' statistically to first find the median of each pick's performance data and then run a regression in order to find a relationship between pick and performance? Any help would be greatly appreciated.
[ 0.014020555652678013, 0.011085576377809048, -0.025706259533762932, -0.0012075970880687237, 0.011435830034315586, -0.010777361690998077, 0.007379214745014906, 0.0024978802539408207, -0.006677163764834404, -0.009652139618992805, -0.00715814670547843, 0.01009304728358984, -0.0013336301781237125...
[ 0.15393832325935364, -0.10534024238586426, -0.29605135321617126, 0.11869857460260391, -0.2007998526096344, 0.849848747253418, -0.25174903869628906, -0.09123493731021881, -0.19086837768554688, -0.38160765171051025, 0.552582859992981, 0.06172855570912361, 0.4890325367450714, 0.58120745420455...
I've just aquired Cass and according to the wiki, she can be good at both melee and guns. If I give her both a melee weapon and a gun, will she swap weapons during combat accordingly? I.e. if the monsters start nibbling on her shins, will she switch to melee? Or are the choices you select in the panel set in stone? Or even better, can I give her a sniper rifle and a shotgun and let her decide when it's best to use either? As it is now, she has `All American` and `Riot shotgun` and tries to snipe enemies from the other side of the desert with the shotgun.
[ 0.004728063475340605, 0.008436381816864014, 0.0025011515244841576, -0.003961140755563974, -0.011649318039417267, -0.014254948124289513, 0.009181845933198929, -0.011717465706169605, -0.014415851794183254, -0.014488184824585915, -0.012960312888026237, 0.014308152720332146, -0.00360423931851983...
[ 0.644470751285553, -0.06365372240543365, -0.3638997972011566, 0.1969517171382904, 0.3591671288013458, 0.3362475037574768, 0.4925687909126282, -0.5221756100654602, -0.6642246842384338, -0.13014274835586548, 0.3959483802318573, 0.701999306678772, -0.026859905570745468, -0.11940508335828781, ...
I have data that looks like this: 011100111110100111 111111111111110010 111100001111000011 1D lanes of streams of data. Each row signifies the presence of a thing at that time with a 1 or a 0. Things tend to exist in consecutive chunks across the stream. The vertical order of the rows doesn't matter, patterns can occur between any set of rows. I am seeking patterns where dimensions coincide at two or more indices, both at the **start of '1' group's** , and also **where any '1' groups overlap** across all rows/dimensions. Pattern can be offset with nearby adjacent indices, if it's proportion is maintained. D = Dimension/Row N = Index in stream (( D1(N), D25(N+4), D900(N-1) ), ( 3, 67, 90, 3000 )) D1(3), D25(7), D900(2) D1(67), D25(71), D900(66) D1(90), D25(94), D900(89) D1(3000), D25(3004), D900(2999) An example of a pattern match that migh occur at multiple places. Dimension01 at N, Dimension25 at index N + 4, Dimension900 at index N -1 occur at indices 3, 67, 90 and 3000. ### The returned patterns: * Are ordered by the number of matching indices or by the number of dimensions in the pattern. * Match at least two dimensions at at least two points How can I go about this?
[ -0.013377425260841846, 0.01603938452899456, -0.00638943538069725, 0.0075491140596568584, 0.01659998670220375, 0.005818682722747326, 0.00629928195849061, 0.021904198452830315, -0.011408759281039238, 0.012727431952953339, -0.00516936369240284, 0.008365394547581673, 0.0043478296138346195, 0.0...
[ 0.532770037651062, -0.09114471077919006, 0.11290846765041351, 0.3292250633239746, -0.18527524173259735, 0.23057594895362854, -0.01707744225859642, 0.2855708599090576, -0.4357069134712219, -0.7965600490570068, 0.13218660652637482, -0.07748972624540329, -0.28915196657180786, 0.26446446776390...
On GitHub is a wonderful installation instructions, but most users GitHub, use Linux, it's not worked on Windows.... https://github.com/openlayers/ol3/wiki/Developer-Guide How can I build OpenLayers 3 on Windows? Can't install Pystache Traceback (most recent call last): File "C:\Openlayers\pystache-0.5.3\setup.py", line 111, in <module> import setuptools as dist ImportError: No module named setuptools And Closure Linter Warning (from warnings module): File "C:\Openlayers\lib\distutils\dist.py", line 267 warnings.warn(msg) UserWarning: Unknown distribution option: 'install_requires'
[ 0.0015989854000508785, 0.0006293595070019364, -0.009600641205906868, 0.011692099273204803, -0.00569413136690855, 0.0045899455435574055, 0.00943642109632492, 0.009296937845647335, -0.015646865591406822, -0.029810700565576553, -0.0030767463613301516, 0.004423183389008045, -0.005098672583699226...
[ 0.3060497045516968, 0.29427510499954224, 0.5885077714920044, -0.07590794563293457, -0.047675445675849915, -0.31217461824417114, 0.2916281521320343, -0.10490317642688751, -0.3562420904636383, -0.8495031595230103, -0.27620258927345276, 0.387696772813797, -0.46176666021347046, 0.0370050519704...
I cut this out so that there is no confusion. I have this picture: ![enter image description here](http://i.stack.imgur.com/ON81O.png) And I've managed to draw this: ![enter image description here](http://i.stack.imgur.com/dzlas.png) The code is as follows: \begin{figure} \begin{circuitikz}[european] \draw[o-] (0,2.2) to (0,4); \draw (0,4) -- (1,4); \draw (1,4) to[R,l_=\SI{1}{\kilo\ohm},-*] (1,2) to[R,l_=\SI{500}{\ohm},-*] (1,0) -- (0,0); \draw[-o] (0,0) -- (0,1.5); \draw (1,2) -- (3,2); \draw (3,2) -- (4,2) node[component]{A} -- (5,2); \draw (3,0) -- (3,1) node[component]{V} -- (3,2); \draw (6,2.27) node[njfet]{} node[anchor=east] {G} node[anchor=north] {D} node[anchor=south] {S}; \draw (6,0) -- (6,2); \draw (6,3) -- (6,4); \draw (1,0) -- (8,0); \draw (2,0) to[C,l^=\SI{4.7}{\nano\farad},*-*] (2,2); \draw (7,0) -- (7,2) node[component]{V} -- (7,4); \draw (6,4) -- (6.5,4) node[component]{A} -- (8,4); \draw[o-] (8,2.2) to (8,4); \draw[-o] (8,0) to (8,1.5); \end{circuitikz} \end{figure} I guess you can see where is the problem: 1. how to add + and - on the nodes? 2. how to make writings not overlap each other? 3. when I tried to make the code for the jfet like the one from the manual I got errors about (njfet. G) stuff... So any advice on how to solve this? EDIT: Oh and can I somehow modify the elements? Because, for instance, npn BJT has arrow pointing out of emmiter and I need it to point it in...
[ -0.008389085531234741, 0.002790065249428153, 0.0012497346615418792, 0.017082901671528816, 0.013364555314183235, -0.009223246946930885, 0.005428715609014034, -0.002833029953762889, -0.012873595580458641, 0.0054457406513392925, -0.011865769512951374, -0.008504313416779041, -0.01002711988985538...
[ 0.27563241124153137, -0.208339124917984, 0.7445470690727234, -0.23793365061283112, 0.13182015717029572, 0.1558784693479538, 0.3179769814014435, -0.37850505113601685, -0.2983964681625366, -0.3690045177936554, -0.05011007562279701, -0.06421159952878952, -0.274073988199234, -0.053649071604013...
I published a service to ArcGIS Server 10.1 from an .mxd with no halos around the labels and created a flex application using the flex viewer. When I browse to the map here and zoom in on an individual trail segment, the labels become very grainy and appear to have a white halo around them, making them very difficult to read. It becomes exceptionally noticeable when imagery is used as a basemap. Is there any thing I can do to stop the labels from getting all messy looking or is this just how they look when using flex?
[ -0.019384058192372322, -0.0006094839773140848, -0.014838110655546188, 0.018980832770466805, 0.0005653202533721924, -0.014309262856841087, 0.009133890271186829, 0.005954528693109751, -0.015676846727728844, 0.01104708295315504, 0.001719171879813075, 0.009800818748772144, -0.0038556319195777178...
[ 0.810042142868042, -0.2059938609600067, 0.11005578935146332, 0.1593295782804489, -0.2073153704404831, -0.29875919222831726, 0.1499931961297989, -0.2711668908596039, -0.27195650339126587, -0.6586885452270508, 0.344333291053772, 0.2886638641357422, -0.24401098489761353, -0.2329665720462799, ...
When running openSUSE 12.3, the updater pops up with mysql/mariadb updates available. Being a trusting person, I always update whenever any update is offered. I have spent the day regretting this decision. I have found that **when the mysql daemon is started it promptly changes the file permissions of some very important files and directories.** For instance it changes the owner and group to `mysql:mysql` for the directory `/usr`, `/usr/bin` and all the files in `/usr/bin`. This **impacts the system dramatically** , like **not being able to use`su` or `sudo`**. Also found that it did similar things in `/etc`. I have disabled the system starting the mysql daemon and am now **I'm trying to repair the damage**. Can anyone tell me how to stop this behaviour as I am now reluctant to start the daemon again as I suspect it will change all the file permissions back again.
[ -0.007247529923915863, 0.005080036353319883, -0.0022650728933513165, 0.009240198880434036, -0.01182003878057003, 0.004340930841863155, 0.008853260427713394, 0.015919048339128494, -0.013206461444497108, -0.025827636942267418, -0.018450219184160233, 0.011890175752341747, -0.0038801771588623524...
[ 0.21749471127986908, 0.23664814233779907, 0.4242137670516968, -0.334588885307312, -0.23490846157073975, -0.43711057305336, 0.5879694223403931, 0.05836724117398262, 0.03167029470205307, -0.8716578483581543, -0.3783263862133026, 0.6701980233192444, -0.38330352306365967, 0.2003897726535797, ...
I am working with Modis Data (MOD02, MOD11, MOD03, MOD35). In particular I am looking for a solution to reproject the subdatasets below, into albers equal. Because I am using NLCD dataset which uses albers.(PROJ4: +proj=aea +lat_1=29.5 +lat_2=45.5 +lat_0=23 +lon_0=-96 +x_0=0 +y_0=0 +ellps=GRS80 +datum=NAD83 +units=m +no_defs) area in 1000 m resolution. I will run a model for whole USA so I thats why I chose Albers. Can somebody help me? By the way I am aware of the MRT Swath but I want to do this in Python environment. So far I tried this but there was shifts in the pixels. (Alignment problems). gdalwarp -tr 1000 1000 -of GTIFF -tps -t_srs "+proj=aea +lat_1=29.5 +lat_2=45.5 +lat_0=23 +lon_0=-96 +x_0=0 +y_0=0 +ellps=GRS80 +datum=NAD83 +units=m +no_defs” -r cubic 'HDF4_EOS:EOS_SWATH: "MOD021KM.A2007181.1750.005.2010199133141.hdf”:MODIS_SWATH_Type_L1B:EV_250_Aggr1km_RefSB’ x.tif This one works for the MOD021KM but I couldn't work it out for the rest. I am confused. These are the subdatasets that I want to extract as geotiffs. HDF4_EOS:EOS_SWATH:"MOD021KM.A2007157.1700.005.2010197153031.hdf":MODIS_SWATH_Type_L1B:EV_250_Aggr1km_RefSB HDF4_EOS:EOS_SWATH:"MOD021KM.A2007157.1700.005.2010197153031.hdf":MODIS_SWATH_Type_L1B:EV_500_Aggr1km_RefSB HDF4_EOS:EOS_SWATH:"MOD03.A2007157.1700.005.2010196183307.hdf":MODIS_Swath_Type_GEO:SensorZenith HDF4_EOS:EOS_SWATH:"MOD03.A2007157.1700.005.2010196183307.hdf":MODIS_Swath_Type_GEO:SolarZenith HDF4_EOS:EOS_SWATH:"MOD11_L2.A2007157.1700.005.2007158222249.hdf":MOD_Swath_LST:LST HDF4_EOS:EOS_SWATH:"MOD11_L2.A2007157.1700.005.2007158222249.hdf":MOD_Swath_LST:View_angle HDF4_EOS:EOS_SWATH:"MOD35_L2.A2007157.1700.005.2010197153939.hdf":mod35:Cloud_Mask Sorry for the simple question. I am a newbie in the gdal and python.
[ 0.009160186164081097, 0.013854332268238068, -0.017464663833379745, 0.009354829788208008, -0.005083512980490923, 0.009530121460556984, 0.005795128643512726, 0.0034570866264402866, -0.010556264780461788, 0.014105771668255329, -0.0012086232891306281, 0.011298207566142082, -0.019209731370210648,...
[ -0.2593003213405609, -0.1517895609140396, 0.3577463924884796, -0.21171148121356964, -0.1522856205701828, 0.1917579174041748, 0.05763961747288704, -0.641880452632904, -0.14014706015586853, -0.9278833270072937, 0.038363415747880936, 0.5445430874824524, -0.19420012831687927, 0.010092233307659...
I have tried different commands to see sizes of partitions and space used/left and it seems that they always differ in their results. Here is my space used/left with from the command `df -h`: Size Used Avail Use% Mounted on 395G 355G 21G 95% /home Notice that used and availabe only add up to 376G but size is 395G
[ -0.02683912217617035, 0.015101492404937744, -0.028077634051442146, 0.013627531938254833, -0.0022681851405650377, -0.015002476051449776, 0.012570383958518505, -0.029888682067394257, -0.0159805566072464, 0.007197302300482988, 0.008002661168575287, -0.005033851135522127, 0.004648968111723661, ...
[ 0.05146007239818573, 0.41756686568260193, 0.6879528164863586, -0.23119810223579407, 0.0034463172778487206, 0.3910658657550812, 0.07333686202764511, -0.33200886845588684, -0.14373323321342468, -0.7860572338104248, 0.05698287487030029, 0.22061605751514435, 0.23272648453712463, 0.092701122164...
Will a door or a trapdoor diminish or block out a light source on the other side of it? If so, by how much, and does the direction in which the door is placed change this effect?
[ 0.021554235368967056, 0.018501650542020798, -0.0033198355231434107, 0.00873823557049036, -0.04047522693872452, -0.02268114872276783, 0.01196127850562334, -0.007520647253841162, -0.02068440243601799, -0.007020903751254082, -0.03233882784843445, 0.022656992077827454, -0.004926653113216162, 0...
[ 0.4464012384414673, 0.016910728067159653, 0.1524091511964798, -0.04160492494702339, 0.35230812430381775, 0.04146275296807289, 0.5969988703727722, -0.6176815629005432, -0.181792214512825, -0.707308828830719, -0.08899927139282227, 0.21726088225841522, -0.0987100824713707, 0.11322900652885437...
I've got a shapefile with about 30 fields that I'll need to update from time to time by importing a CSV with new data, joining it to the shapefile, and using the field calculator to carry the new values into the old fields one at a time: ![enter image description here](http://i.stack.imgur.com/RPfz3.jpg) However, this is a tedious process, and it seems like something ripe for automation with pyQGIS. The target field names and corresponding join field names will not change over time. I have a sense of what this would look like in SQL, but the python version escapes me, and this documentation is thin on field calculation. Is it even possible to automate a batch of field calculations like this?
[ 0.00765202334150672, 0.01007002592086792, -0.009896054863929749, 0.017933625727891922, -0.021836720407009125, 0.02653297409415245, 0.007124339696019888, 0.01918691024184227, -0.01925927773118019, 0.00032896525226533413, 0.00365145574323833, 0.012365465983748436, -0.005868135020136833, 0.00...
[ 0.7074631452560425, 0.13900206983089447, 0.7019408345222473, -0.2308412790298462, 0.017459886148571968, 0.2500802278518677, -0.06618862599134445, -0.44494783878326416, -0.0003017180133610964, -0.8240736722946167, 0.028544792905449867, 0.16575683653354645, -0.1710958331823349, -0.2405178099...
I use yii framework that implements Active Record pattern as ORM base. It has CActiveRecord class that is a table wrapper class with attributes reflecting table columns. So each object of this class represents a database row. Wiki says about Active Record pattern: > Active record is an approach to accessing data in a database and > A database table or view is wrapped into a class. Thus, an object instance > is tied to a single row in the table. So far so good. But where should I put complex raw sql query that retrieves statistics data for example? And, more generally, where should I put methods that retrieve some data that can not be an active record object (like data retrieved with aggregation queries) or if I knowing do not want to retrieve an object but an array instead for example?
[ -0.0033200017642229795, 0.019457977265119553, -0.012643937021493912, 0.012662782333791256, -0.0008627772331237793, 0.00445591751486063, 0.01021056529134512, 0.0024824347347021103, -0.011884200386703014, -0.004674170631915331, -0.011369653046131134, 0.02069680579006672, 0.009617699310183525, ...
[ 0.364698588848114, -0.005560468416661024, 0.08947006613016129, 0.2059808224439621, -0.5189715623855591, -0.19178947806358337, 0.0018169264076277614, -0.2596569359302521, -0.3068383038043976, -0.8063125014305115, 0.26432228088378906, 0.3897832930088043, -0.2710244953632355, 0.11150042712688...
Long story short, i just accidentally deleted my entire home folder. Thankfully it seems like the hidden files are still there. I'm not sure, but aren't all of the folders within the home folder (Desktop, Downloads and whatever else is in there) empty by default? If that is the case, could some super nice person just name all of the files located in the home folder so that i can rebuild it? Thanks a bunch in advance
[ 0.01924603432416916, 0.01026294194161892, 0.006575006991624832, 0.014223850332200527, 0.0030529838986694813, -0.019846001639962196, 0.004781011492013931, -0.005583083722740412, -0.023522760719060898, -0.01812831498682499, -0.0032858573831617832, 0.005112781189382076, 0.02329007349908352, -...
[ 0.4063410758972168, 0.17628386616706848, -0.05107065662741661, 0.5772165656089783, 0.1453818529844284, -0.36050960421562195, 0.4378872811794281, 0.7315950989723206, -0.3411467671394348, -0.11432207375764847, 0.07238417863845825, 0.5650786757469177, 0.20500372350215912, 0.26536068320274353,...
I'm trying to create a presentation with LaTeX `beamer` using `pstool`, and since updating my system to Ubuntu 13.10 (pstool 1.5a, TeX Live 2013) the compilation of the pstool auxiliary TeX files (*-pstool.tex) fails, with the following message in the log: ! Improper \spacefactor. \@->\spacefactor \@m l.19 \immediate \write \@ mainaux {\@percentchar <*PSTOOLLABELS>} I could trace the problem to the following section in the created files: \makeatletter \def \thepage {\csname @arabic\endcsname \c@page } \setcounter {page}{15} \@input {presentation.oldaux} \begin {document} \immediate \write \@mainaux {\@percentchar <*PSTOOLLABELS>} \makeatother \centering \null \vfill If I either move the `\begin{document}` below the `\makeatother` or add another `\makeatletter` right after it, I can manually compile the auxiliary file (latex, dvipdf), but of course this is not a solution. Apparently, `\begin{document}` cancels the `\makeatletter`. The problem is that the problematic section is automatically generated by pstool, and manually changing it for each and every EPS graphic is hardly feasible. Any ideas how I can work around this problem? **Edit:** Here is a complete example of the TeX file using `pstool`. The errors occur during the compilation of the auxiliary file created by `pstool`, `test-pstool.tex` in this case. The EPS file `test.eps` contains only a box with the label "test". \documentclass[a4paper]{beamer} \usepackage[crop=pdfcrop]{pstool} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \EndPreamble \begin{document} \begin{frame}{Test} \begin{figure}[p] \begin{center} \pstool[width=0.5\textwidth]{test}{ \psfrag{test}[][]{Test} } \caption{Test} \end{center} \end{figure} \end{frame} \end{document}
[ 0.009272417053580284, -0.00685306079685688, 0.0016545464750379324, 0.03700696676969528, -0.01574481651186943, 0.001423802226781845, 0.008868943899869919, 0.010437246412038803, -0.011850766837596893, -0.012002304196357727, -0.013095937669277191, -0.00670311925932765, 0.01596200466156006, 0....
[ -0.1500890851020813, 0.2706328332424164, 0.7094548940658569, -0.38703665137290955, 0.183492973446846, 0.03531711921095848, 0.6436787843704224, 0.1824788600206375, 0.0354030467569828, -0.728946328163147, -0.11923318356275558, 0.5010977387428284, -0.4495541453361511, -0.04074914753437042, ...
I have a query var `show=othertemplate` that I can add to any URL to show any post, page or archive in Wordpress in a different template, like this: **mysite.com/[any URL]/?show=othertemplate**. I'd like for it to be accessible via clean URL, like **[any URL]/othertemplate** or **[any URL]/show/othertemplate**. I have tried achieving this by registering the query var with a query_vars filter and adding code like below to .htaccess, but Wordpress intervene and throws a 404. WP Debug Bar says that the query is looking for attachment=print. RewriteEngine On RewriteBase / RewriteCond %{REQUEST_URI} !^/index\.php RewriteRule ^(.*)/othertemplate/?$ /index.php/$1?view=othertemplate [L] I have also tried using the Wordpress rewrite functions, but the same thing happens. function st_myqueryvars($vars) { $vars[] = 'show'; return $vars; } add_filter('query_vars', 'st_myqueryvars'); function st_myrewriterules( $wp_rewrite ) { $newrules = array(); $new_rules['^(.*)/othertemplate/?$'] = 'index.php/$matches[1]?show=othertemplate'; $wp_rewrite->rules = $new_rules + $wp_rewrite->rules; } add_filter('generate_rewrite_rules','st_myrewriterules'); Is there something I'm missing here? > **Update:** After flushing permalinks I see that this WP function rewrites > **/ _[any URL]_ /template** to **index.php?show=template** instead of > **index.php? _[the URL's corresponding query string]_ &show=template**. I > can't 'hard code' a query string like > post_type=$matches[1]&name=$matches[2]... into the rewrite rule because it > needs to work for any page/post/archive, which have different permalink > structures. How do I get around this?
[ 0.02168804407119751, 0.006929852068424225, -0.00164166244212538, 0.017474427819252014, 0.001082731643691659, 0.01670260913670063, 0.008663810789585114, 0.016796791926026344, -0.015487534925341606, -0.009384453296661377, -0.004260512068867683, 0.012526728212833405, -0.00939021073281765, 0.0...
[ 0.4889433979988098, 0.23402488231658936, 0.4695059657096863, -0.07935034483671188, -0.2404152899980545, -0.21546034514904022, 0.13232652842998505, 0.08453146368265152, -0.14055369794368744, -0.6457488536834717, 0.1251201182603836, 0.21097874641418457, -0.02678520604968071, 0.33758935332298...
I would really like to get out of the 9-5 "cubicle" existence and do something where I can work from home, take off when I like, etc. Don't get me wrong... I don't mind working hard, but I just am not enjoying the whole corporate thing where my hours, vacation days, dress code, place of work, etc. are set. I really think this is possible because other developers are somehow pulling this off. But so far it is just in the "dream" stage and I have no idea how to move it forward. I don't know if I want to be freelancer, start my own company or what. In case it matters, I have been working in .Net / MVC for the past few years (both desktop and web) but have experience with other languages as well (Java, C++, Perl, etc.). I don't have much experience with "web design" side of things since in all my jobs other people have handled that while I have worked on the "programming" work. Just to make it more concrete what I want... I wake up on a random Tuesday in June and it's a gorgeous day. I am ahead of schedule on my projects and I don't feel like working ... I want to go hiking. I want the freedom to do that without dealing with some PHB. Another scenario... I do feel like working but I want to go visit an old friend or relative for a week in Bismarck / Budapest / Boston. Said friend has reliable Internet. Why can't I do that? I'm not talking about being irresponsible just a bit more free.
[ -0.005814306437969208, 0.007850788533687592, -0.000066372100263834, 0.01007629930973053, -0.009920205920934677, -0.0024282962549477816, 0.004180251155048609, 0.0039150407537817955, -0.012766438536345959, -0.00032084877602756023, -0.013221470639109612, 0.013917968608438969, 0.0183516778051853...
[ 1.1321429014205933, 0.0917971208691597, -0.09065715223550797, 0.16564765572547913, 0.22637198865413666, -0.102574422955513, 0.3111211955547333, 0.679139256477356, -0.08737452328205109, -0.6013748645782471, -0.08869744837284088, 0.22433987259864807, 0.05470110476016998, 0.31046849489212036,...
I saw on the Internet a lecture with the title 8 QUEENS PROBLEM USING BACK TRACKING. There is a drawing of the Board following. How to make this using the Latex? (I tried using the package Chessboard, but the result is very different) \documentclass{article} \usepackage{chessboard} \storechessboardstyle{8x8}{maxfield=d8} \begin{document} \chessboard[style=4x4,setwhite={Qd4}, pgfstyle=straightmove, arrow=to,linewidth=0.2ex, color=red, pgfstyle=straightmove, markmoves={d4-h8,d4-a7,d4-a1,d4-g1,d4-d8,d4-d1,d4-a4,d4-h4}, shortenstart=1ex,showmover=false] \end{document} ![enter image description here](http://i.stack.imgur.com/yooWe.jpg)
[ 0.002020439598709345, -0.002415746683254838, -0.010055694729089737, 0.015471201390028, 0.03665514290332794, -0.004976246505975723, 0.0076029859483242035, -0.004468151368200779, -0.014660081826150417, -0.01966097205877304, -0.006814639084041119, 0.001790927373804152, -0.0038387393578886986, ...
[ 0.261859267950058, -0.32245132327079773, 0.37342768907546997, 0.07868686318397522, 0.021720685064792633, -0.04333271086215973, -0.022506490349769592, -0.3305550813674927, -0.5881911516189575, -0.472690224647522, 0.3049701750278473, 0.11121151596307755, -0.3139009177684784, 0.03605582937598...
Is there a printable guide(pdf) or book for HearthStone - Heroes of Warcraft? I want to learn how to play this game properly, so I don't want to browse the net for scraping guides. Book is always a best solution to most problems :) For example I am interested in stuff like, which cards are good for paladin or shaman or whatever? Or which set of cards are good for which hero, for example a set of murloc cards? Something like that. In more detail: I am not searching for some advanced advice how to play the game, I just want to know basic tips, much like the tips that are shown when you start a game. Am I clear? What goes with what.
[ -0.005639822222292423, 0.003741868771612644, -0.0017891719471663237, 0.008808470331132412, -0.0038819722831249237, 0.021625220775604248, 0.005251608323305845, 0.0007629208266735077, -0.02442827820777893, -0.031292080879211426, 0.007246020715683699, 0.007528452202677727, 0.0009033595561049879...
[ 0.36723482608795166, 0.24624747037887573, 0.025624442845582962, 0.40454065799713135, -0.2766430675983429, -0.18679462373256683, -0.03632759302854538, -0.00007917140464996919, -0.36058998107910156, -0.27719733119010925, 0.37780463695526123, 0.5206835865974426, 0.2713792622089386, -0.1107237...
Being fairly new to TF2 (about 100 hours play time) I am becoming steadily proficient in most classes, and am having a great time with them. However, one class, the scout, is eluding me. When I play as scout, I am killed very quickly and am lucky to get a single kill. However, when other people play as scout, they run circles around me and kill me quickly. Clearly they are doing something that I'm missing. How do I play well as a scout?
[ -0.014161291532218456, 0.004342761822044849, -0.01837390474975109, 0.002810328733175993, -0.0021056742407381535, -0.035098567605018616, 0.005897181574255228, -0.008140728808939457, -0.015339385718107224, 0.004089009948074818, -0.000042134452087339014, 0.015932880342006683, -0.003693548496812...
[ 0.3730151355266571, -0.1529935896396637, -0.2468586266040802, 0.018716003745794296, -0.3671746253967285, -0.6732047200202942, 0.7365942597389221, -0.040485039353370667, -0.26074567437171936, -0.7775065898895264, 0.13661828637123108, 0.5515876412391663, 0.3518880605697632, -0.48848739266395...
I have a lyx file in which there are several figures and tables. The preamble has been amended to have: \usepackage{multirow} \usepackage{rotating} \usepackage{subfig} \usepackage{longtable,lscape} \usepackage[nolists,tablesfirst]{endfloat} \usepackage{caption} \captionsetup[table]{skip=20pt} When I added a new figure after all my other figures, in the lyx screen it numbers it Figure 5. However, when I actually export to PDF LaTeX seems to number it Figure 1. Any ideas on this? This latest figure is an eps graphic. The other figures are all pdf documents. EDIT: Interesting: I figured out the issue (or atleast what fixes it). This figure was specified to be sideways. Once I removed that the ordering became correct. Any ideas what's going on?
[ 0.03153327852487564, 0.01593708246946335, -0.0023192400112748146, 0.020512837916612625, 0.007503345608711243, 0.02027304098010063, 0.007981546223163605, 0.01430498342961073, -0.010208673775196075, -0.0014301877235993743, -0.019545143470168114, 0.004300796892493963, -0.01338307373225689, 0....
[ 0.3399622142314911, -0.4116744101047516, 0.9135486483573914, -0.25264987349510193, 0.09580342471599579, 0.4389024078845978, -0.5029524564743042, -0.19833220541477203, -0.38616064190864563, -0.5102617144584656, 0.18490929901599884, 0.06752888113260269, -0.09981557726860046, -0.1421049088239...
I am curious on how to install my Xbox360 controller with my Wireless Receiver. For some reason, I've done research and my light on my receiver is not green, and I cannot install my driver because of an error message, "Access is Denied". I am the Administrator too. I need to know also where to find the files that installed with my Disc. Thanks! OS: Windows 7 Home Premium Imgur Image of the Error: ![Driver Error](http://i.imgur.com/a0OVgMF.png)
[ -0.001958402106538415, -0.018517224118113518, -0.011075462214648724, 0.022418376058340073, 0.002558366395533085, -0.0027441810816526413, 0.006820170674473047, -0.016239499673247337, -0.015595337375998497, -0.019162844866514206, -0.007723018527030945, 0.011064215563237667, -0.0012215333990752...
[ 0.2692950665950775, 0.1246354728937149, 0.4972240924835205, 0.05861737206578255, 0.01965373568236828, -0.20043408870697021, 0.3298304080963135, 0.1387513279914856, 0.06924908608198166, -0.5397706031799316, -0.03675151243805885, 1.0634057521820068, -0.032095085829496384, 0.2614899277687073,...
I thought it would be nice to have an alias (in ~/.zshrc) to have 'python' aliased to 'ipython' ONLY when 'python' is run without args (otherwise, programs won't launch). First of all, how to express "without args" in an alias? Secondly, what do you think about it?
[ 0.033247265964746475, 0.018257861956954002, -0.00464245118200779, 0.024588918313384056, -0.03893830627202988, 0.0007575165946036577, 0.007515003904700279, -0.003353557549417019, -0.019727183505892754, -0.016573213040828705, -0.0198203157633543, -0.0011027618311345577, -0.01010196190327406, ...
[ 0.3771805763244629, 0.17082177102565765, -0.0876486673951149, 0.12415163964033127, -0.04778222367167473, -0.34053894877433777, 0.15868361294269562, 0.28784430027008057, -0.3786618113517761, -0.22502939403057098, 0.22421030700206757, 0.48587730526924133, -0.270869642496109, -0.2104083150625...
WolframAlpha can give step by step solution for indefinite integral. There seems to be similar question but for derivatives. Is there a way that I can generate my own step by step solution for indefinite integrals (for simple integrals having closed form)? Sorry, if the question is inappropriate.
[ 0.03199196979403496, 0.013947764411568642, -0.013073241338133812, 0.024647507816553116, -0.02680567093193531, 0.023095102980732918, 0.012556027621030807, 0.029651712626218796, -0.035578444600105286, 0.00399747071787715, -0.003972804173827171, 0.01486823707818985, -0.008057978935539722, -0....
[ -0.4122549593448639, 0.03442224860191345, -0.08000844717025757, -0.2934589684009552, -0.4909166991710663, 0.14375697076320648, 0.45102760195732117, -0.35219407081604004, -0.1934119164943695, 0.23962879180908203, 0.037430230528116226, -0.04951515421271324, -0.4279211461544037, -0.2180133759...
I have been using Symfony2 with Doctrine2 for some years. I have recently started using Microsoft's Entity Framework with MVC5. From my Symfony2 experience I understand that a repository's job is only to retrieve and return objects, no additional operations like Saving. Now every examples I have seen for EF has a method Save/Update as part of the repository. For symfony I have been creating manager classes as follows: interface IManager { function getClassName() ; IRepository getRepository() ; function Save(object); function Update(); } So I pass around the manager, if I need to retrieve objects I call the repository directly. If I need to save I call the manager's save method. Is a repository supposed to support save/update? What do you think of my IManager class, should I also use it for EF?
[ -0.010357221588492393, 0.03365052491426468, -0.00898002739995718, 0.013845507055521011, 0.004621980246156454, -0.009478477761149406, 0.009413042105734348, 0.017723683267831802, -0.012047355994582176, -0.014364118687808514, -0.025808265432715416, 0.02367858588695526, -0.009264427237212658, ...
[ -0.2797143757343292, -0.1912064105272293, 0.30496516823768616, -0.2126844823360443, -0.2987762689590454, 0.10467520356178284, 0.16040705144405365, -0.28795257210731506, 0.0007078036433085799, -0.5940883755683899, -0.043492767959833145, 0.7402960062026978, -0.3124321401119232, 0.11899170279...
Logistic regression can help to predict a value whether it would happen or no. I'd like to know how can I do that using sklearn. I'd like to know the probability if this event would happen or no. I have a huge dataset (20K lines and 20 columns). My data has 19 columns as predictors and last column as target (values between 0-10). To simplify work, I am using random data to understand how can I interpret data. A,B,C : Predictors target: as a target from sklearn import linear_model import pandas as pd dataset = pd.DataFrame({'A':np.random.rand(100)*1000, 'B':np.random.rand(100)*100, 'C':np.random.rand(100)*10, 'target':np.random.rand(100)}) predictors= dataset.ix[:,['A','B','C']].values target = dataset.ix[:,['target']].values lr = linear_model.LogisticRegression() lr.fit(predictors, target) linear_model.LogisticRegression(C=1.0, class_weight=None, dual=False, fit_intercept=True, intercept_scaling=1, penalty='l2', random_state=None, tol=0.0001) Now, should I plot (lr.predict_proba) to get probability of every element ? what should I do in order to have probability of every line.
[ 0.0027945172041654587, 0.018134357407689095, -0.010682351887226105, 0.0025948926340788603, -0.009302105754613876, 0.004963623359799385, 0.006065379362553358, 0.003914343658834696, -0.007946614176034927, 0.0021888515911996365, 0.0004875378217548132, 0.004724913276731968, 0.0008769998094066978...
[ -0.14913871884346008, -0.25217992067337036, 0.2536998689174652, 0.3616425395011902, -0.31491610407829285, 0.34262415766716003, -0.11565682291984558, -0.2132323831319809, -0.24929441511631012, -0.3516591787338257, 0.4913743734359741, 0.44442713260650635, -0.3855457603931427, 0.1885892301797...
I have a web-cam which I want to use to stream h264 video. Now I'm streaming mjpg video with mjpg-streamer. And now I'd like to stream h264. I've found that it could be done with gstreamer, but I'm totally new to it. Also after reading some blogs/articles/etc it seems to me that there were some issues with h264 and gstreamer. Is it still true? May be there are some other methods to do this besides gstreamer? Any info would be helpful. Thanks in advance.
[ -0.01595965214073658, 0.0020735375583171844, -0.00361384404823184, 0.014808961190283298, 0.009646549820899963, 0.0009002528968267143, 0.00681061577051878, 0.010027822107076645, -0.018422605469822884, -0.0244697704911232, 0.0025684735737740993, 0.014763972721993923, 0.018052946776151657, 0....
[ 0.7344784140586853, 0.24562981724739075, 0.39003053307533264, 0.36679762601852417, -0.1586974710226059, -0.3125564754009247, 0.13589875400066376, 0.28144571185112, -0.46636003255844116, -0.7361235022544861, 0.16463658213615417, 0.5151551961898804, -0.01367078348994255, 0.393004834651947, ...
In quantum mechanics we start with a Hamiltonian $H_0$ for which we know the exact eigenstates and energy eigenvalues. We perturb it by a known term $H$, and then attempt to compute (approximately) the new eigenstates and eigenvalues. In general relativity, my understanding is we start with a metric $g_{\mu \nu}$, and perturb it by a known $h_{\mu \nu}.$ But in my lecture notes (http://arxiv.org/pdf/0804.2595.pdf), the lecturer shows how to compute $h_{\mu \nu}$. I thought we perturbed a system by a known quantity; can someone clarify the regular procedure of perturbation theory in general relativity, and what typical 'goals' are? The only alternative I see is that we perturb a known solution $g_{\mu \nu}$ by an unknown perturbation $h_{\mu \nu}$, state how we would like the stress- energy $T_{\mu \nu}$ to change, and then try and compute $h_{\mu \nu}$ such that it does. Could this be the correct interpretation?
[ 0.004380425438284874, 0.015232188627123833, 0.0028794286772608757, 0.006736668758094311, -0.015687789767980576, 0.005835022311657667, 0.010270694270730019, -0.03672659397125244, -0.012295762076973915, 0.005765264388173819, -0.0010446356609463692, 0.011223557405173779, -0.01481795497238636, ...
[ 0.07258445769548416, 0.00873214565217495, 0.11456387490034103, -0.13651227951049805, -0.13573125004768372, -0.04633713513612747, -0.2714989483356476, -0.2794693410396576, -0.014790302142500877, -0.5161349177360535, -0.3132774233818054, 0.1730984002351761, -0.48754945397377014, 0.7205917835...
I am using quotchap as well a pre-existing PhD format cls file. On the first page of every chapter, the page number is placed at the bottom rather than at the top and the last line of the page overwrites the page number. Is there a way to move the page number lower, up to the top, or remove one line from the text? (I had tried using fancyhdr to do the trick but it seems to have no effect - possibly because I am using the pre-existing cls file...?) Thanks for any help!
[ 0.010570913553237915, 0.0144608523696661, -0.0066070095635950565, 0.02548895962536335, 0.00814707949757576, 0.007240308914333582, 0.008425140753388405, -0.004870899021625519, -0.01384657435119152, 0.008198178373277187, -0.011107038706541061, 0.0017266651848331094, -0.00666397949680686, 0.0...
[ 0.1680428832769394, 0.3549094796180725, 0.3175655007362366, 0.24327030777931213, -0.11673091351985931, -0.22994084656238556, 0.3745930790901184, 0.3851376473903656, -0.20439952611923218, -0.48459118604660034, 0.034935563802719116, 0.13716727495193481, -0.07857957482337952, 0.11884886026382...
I'm interested in transitioning my development environment onto the cloud. I currently use VirtualBox with Vagrant to manage separate virtual machines for each individual project. What would be the optimal setup for hosting these virtual machines in the cloud? Given than I have nearly a dozen VMs, setting up separate EC2 instance for each one is not economical. A cheaper option might be to snag a low-end VPS from one of the many cheap VPS providers out there. The downside, aside from reliability, is that I don't need all my VMs available to me at all times. For example, VMs that I use as my staging environments only need to be on a few times a week at best. As an alternative, I've thought about running VirtualBox and Vagrant on a powerful box with enough memory to support multiple machines running simultaneously. Is this a smart setup, or are there caveats I have not considered? As a side note to the aforementioned option, is there an alternative to VirtualBox that would be a wiser choice?
[ 0.02481902949512005, 0.009171582758426666, -0.0035685410257428885, 0.005380409769713879, -0.013133862987160683, 0.0035907903220504522, 0.006475750356912613, 0.017036736011505127, -0.01544223167002201, 0.004936262033879757, -0.007505288813263178, 0.02052566036581993, -0.009044296108186245, ...
[ 0.544023871421814, 0.2565605044364929, 0.29975131154060364, 0.3402806520462036, 0.27711841464042664, 0.015479043126106262, -0.000027642045097309165, 0.15571069717407227, -0.38247251510620117, -0.5984147787094116, 0.16459105908870697, 0.5178961753845215, -0.0996418371796608, 0.2942250370979...
I am coding a program which has persistent data (the MELT monitor, related to GCC MELT). The data is persistent because it is expected to be read and overwritten by most executions. (That data is in particular made of abstract syntax trees) Currently, this persistent data is an Sqlite database. Of course, I am backing up it in textual format, obtained by sqlite dump, and I want to manage this textual dump with a version control system (probably git, but perhaps also subversion ....). Unless I take special precautions, these SQL dumps will probably have quite long lines (e.g. several dozens of kilobytes). Wide SQL columns would probably contain JSON text of many kilobytes. Would `git` (and `svn`) be more happy with shorter lines (in particular would they perform slower with long lines, or have repositories using much more disk space)? I'm probably not mostly interested in the `diff` commands (e.g. `git diff` or `svn diff`), because I expect that using them for such textual dump is not very interesting for the human developers. I am coding the persistency routines, so I am able to change slightly the format (e.g. to add newlines in some JSON text, which `sqlite3 dump` is dumping verbatim). If you are curious, my current code snapshot is on http://starynkevitch.net/Basile/monimelt-bgc-21apr2014.tar.bz2 and contains a `state-monimelt.sql` dump file (which will of course become much larger, perhaps near or above a megabyte)
[ -0.009604120627045631, 0.002550081815570593, -0.0060669006779789925, 0.003837675554677844, 0.01577112078666687, 0.017681526020169258, 0.00708826445043087, 0.010163486935198307, -0.010531164705753326, -0.01829233393073082, -0.0015199343906715512, 0.00597608694806695, 0.011223304085433483, 0...
[ 0.5594752430915833, 0.3292202651500702, 0.4294777512550354, 0.08978025615215302, -0.005346981342881918, -0.15619970858097076, 0.0038180388510227203, 0.184501513838768, -0.08359265327453613, -0.4510171413421631, -0.10896012932062149, 0.35965871810913086, -0.44939371943473816, 0.040731787681...
I want to use the command `shutdown -r now` in Lazarus, but I don't know how to do it. Any tips or examples that you can give me I would be most grateful.
[ 0.04970790818333626, -0.0061472258530557156, -0.041077788919210434, -0.006452709436416626, 0.003001539036631584, 0.023292958736419678, 0.011416323482990265, -0.01739386096596718, -0.03709467127919197, -0.005401040893048048, -0.008680026046931744, -0.00519209960475564, -0.0078945467248559, ...
[ 0.6286683678627014, 0.4063946604728699, -0.14288455247879028, -0.11305700987577438, -0.14249074459075928, -0.2733781039714813, 0.41431549191474915, 0.39916765689849854, 0.29409584403038025, -0.4385657012462616, -0.30329054594039917, 0.03465421125292778, 0.1279953569173813, 0.24706158041954...
I have a dataset of educators and the courses that they designed. My original thought was to do a multilevel model where courses are nested within educators, and the outcome is whether the course ever was offered publicly to students. I was fitting this model using `melogit` in Stata. However, I realized that courses could have been designed by multiple teachers. So courses could potentially be nested within multiple teachers (or vice versa). Does anyone know how to fit this type of "multiple membership" model in Stata?
[ 0.004206713754683733, 0.018851125612854958, -0.0009344866848550737, 0.024106673896312714, 0.027995208278298378, 0.004388853907585144, 0.009410438127815723, 0.0274182241410017, -0.021596485748887062, 0.016983332112431526, 0.0010703367879614234, 0.021099787205457687, 0.0012135969009250402, 0...
[ 0.6391743421554565, -0.051670536398887634, -0.1986432820558548, 0.31931012868881226, 0.011105760000646114, -0.02703511342406273, 0.026147659868001938, 0.12214906513690948, -0.1703747808933258, -0.5076227784156799, 0.026238741353154182, 0.13018974661827087, 0.0617782361805439, 0.13452035188...
I'm currently analyzing a variety of signals. The problem I have is that I have several "missing" values. These "missing" values represent the absence of signal, they are not errors in sampling or something else. So it seems important to keep them when performing the data analysis. However, I'm unsure about the best way to include them in the analysis. I have devised 2 main options, both using standardized data: 1. To transform continuous data into categorical data using the standard deviations of the normalized data, and adding a category for "no-signal". 2. Replacing all missing values by a z-score of -4 or -6. I would like to ask about your opinion (pros and cons) of using one of these methods. PS: Maybe is worth to say that my data are right skewed ranging from ~ (-2) to (+5) SD.
[ 0.027884725481271744, 0.02068815752863884, -0.017612172290682793, 0.004831991158425808, 0.013641128316521645, 0.0005085030570626259, 0.008644520305097103, -0.008530953899025917, -0.009551200084388256, 0.000341540202498436, -0.0037232472095638514, 0.0069619594141840935, -0.018029745668172836,...
[ 0.13468989729881287, -0.05533086881041527, -0.11380408704280853, 0.1322077214717865, -0.15632766485214233, 0.13370542228221893, 0.42346271872520447, -0.06075393408536911, -0.2652316987514496, -0.16402097046375275, -0.10195555537939072, 0.5333006978034973, -0.41714245080947876, 0.2378575205...
I work with many different federal lands all of which can be associated with iconic logos (think National Park Service, National Wildlife Refuge System, etc.). I'd like to create a 'Style Reference' that has all of the different logos to denote point data. This will be a working style reference as I encounter new features from time to time. Any reference to maintaining or editing a style reference would be helpful. I'm assuming the preferred format would be .emf (vector) to maintain transparency, but I'm not at all familiar with the style format structure or file type. ![National Park Service](http://i.stack.imgur.com/5I5WP.jpg) ![National Wildlife Refuge System](http://i.stack.imgur.com/hPjlc.jpg)
[ 0.010458914563059807, 0.01293661817908287, -0.0007768361829221249, 0.01088340301066637, 0.014294530265033245, -0.013796464540064335, 0.006182155106216669, 0.027191467583179474, -0.012586712837219238, 0.008917925879359245, 0.009921560063958168, 0.013663779012858868, 0.004886907525360584, -0...
[ 1.091421127319336, 0.33166974782943726, 0.30420300364494324, 0.27005958557128906, 0.14837811887264252, -0.4305073320865631, 0.04372246563434601, -0.01477772556245327, -0.34404847025871277, -0.6003128886222839, 0.05359199270606041, 0.3152623772621155, -0.10852070897817612, 0.151137471199035...
I have recently noticed that when I send out e-mail campaigns that Gmail is now including 'Unsubscribe' in the subject along with my actual subject. Is this now a common practice among e-mail providers or is it just Gmail? _**UPDATE_** Nick was 100% correct and I have made a few changes I decided to post: I did not want to change my layout any and keep it the same yet Google pulls the first text it sees which was ALT text for an image. To correct this I inserted a 1x1 Transparent Picture with the ALT text that said what I want the Gmail Preview to say. Works like a charm. Here is the code for it: <p align="center"><img style="color:#ffffff" src="http://mysite.com/Images/skin/spacer1x1.png" width="1" height="1" alt="ALT Text" /></p> I put the color there to ensure the text blends in to the background because I do not want it visible in the e-mail for users who have images disabled.
[ -0.017290771007537842, -0.0007092201267369092, -0.0026692955289036036, 0.020493511110544205, 0.024465106427669525, -0.006017571315169334, 0.005895167123526335, 0.015522008761763573, -0.011071691289544106, 0.003031965345144272, -0.012167492881417274, 0.009125004522502422, 0.000419420655816793...
[ 0.589674711227417, 0.28049132227897644, 0.3812066614627838, 0.01408393308520317, -0.26062536239624023, 0.14541834592819214, 0.5218403935432434, -0.06935276091098785, -0.11747445911169052, -0.7130681276321411, 0.13873744010925293, 0.2926252782344818, -0.2198965698480606, 0.1520538479089737,...
Consider the minimal code: \documentclass{report} \usepackage{mathtools} \usepackage{lipsum} \begin{document} \lipsum*[2]% \begin{align*} m_{1}(x) &=x(x-1)=x^{2}-x, & m_{2}(x)&=x^{2}(x-1)=x^{3}-x^{2}, \end{align*} \begin{align*} m_{3}(x) &=x(x-1)^{2}=x^{3}-2x^{2}+x. \end{align*}% \lipsum*[3] \end{document} producing ![enter image description here](http://i.stack.imgur.com/OSFrb.png) What is the best way to have the same result but with a smaller vertical space between the two lines, but using only one environment? Of course I can try to use `\vspace{negative space}` between those two `align*` but I'm afraid it could be dangerous.
[ 0.02504800260066986, 0.004778278525918722, -0.006091705989092588, 0.010645845904946327, -0.006302332505583763, 0.00942493136972189, 0.00701881293207407, 0.0033641590271145105, -0.006765201687812805, 0.007082549389451742, -0.005602437071502209, -0.005343981087207794, -0.00929686613380909, 0...
[ -0.5857560038566589, -0.06106071546673775, 0.6536027193069458, 0.07032826542854309, 0.138662189245224, -0.012588715180754662, 0.19070230424404144, -0.3040499687194824, -0.001193420379422605, -0.9240681529045105, 0.021942030638456345, 0.7729030251502991, -0.07440879940986633, -0.21328103542...
I have a macro that inserts itemize environment in a longtable. The issue is that I have big spaces between top and bottom of the itemize environment. How can I control the spaces? ![enter image description here](http://i.stack.imgur.com/pqlkk.png) \documentclass{article} \usepackage{longtable} \usepackage{booktabs} \newcommand{\heu}[1]{% \textit{BB Heuristic:} #1% }% \newcommand{\whypri}[2]{% \begin{itemize} % \item \textit{Why}: #1% \item \textit{Priority Justification}: #2% \end{itemize}% } \begin{document} \begin{longtable}{c|p{0.4\textwidth}|p{0.5\textwidth}} \caption[Derivation Plan]{Derivation Plan} \label{d_plan} \\ \toprule \textbf{1} & \multicolumn{2}{l}{\textbf{Goal: Extensibility}} \\\midrule 1.1 & \heu{Reduce Data/Event Dependency} & \whypri{Extensible components should be working as a standalone that has minimum dependency on other components. The more dependencies, the more complexity Occurs To prevent extensible structure.}{Less complexity in extension introduced by reducing dependency.} \\\midrule %\st{BB Heuristic: } \bottomrule \end{longtable} \end{document}
[ 0.021141745150089264, 0.005175960250198841, -0.0020220777951180935, 0.023471787571907043, 0.015476053580641747, 0.013321274891495705, 0.008154578506946564, 0.009365810081362724, -0.014588904567062855, -0.0011252295225858688, -0.024694444611668587, 0.0006819871487095952, -0.000550160184502601...
[ -0.11370401084423065, 0.09558030962944031, 0.33750855922698975, -0.11579745262861252, 0.24959762394428253, 0.19678069651126862, -0.1224648654460907, -0.24885448813438416, -0.31211358308792114, -0.5672022700309753, 0.04432960972189903, 0.35465773940086365, -0.41387656331062317, 0.0423476099...
i have installed the theme-my-login plugin and i want change the password_fields function: this is the code of custom-password.php of the plugin: /** * Loads the module * * @since 6.0 * @access protected */ protected function load() { add_action( 'register_form', array( &$this, 'password_fields' ) ); add_filter( 'registration_errors', array( &$this, 'password_errors' ) ); add_filter( 'random_password', array( &$this, 'set_password' ) ); add_action( 'signup_extra_fields', array( &$this, 'ms_password_fields' ) ); add_action( 'signup_hidden_fields', array( &$this, 'ms_hidden_password_field' ) ); add_filter( 'wpmu_validate_user_signup', array( &$this, 'ms_password_errors' ) ); add_filter( 'add_signup_meta', array( &$this, 'ms_save_password' ) ); add_action( 'tml_new_user_registered', array( &$this, 'remove_default_password_nag' ) ); add_action( 'approve_user', array( &$this, 'remove_default_password_nag' ) ); add_filter( 'tml_register_passmail_template_message', array( &$this, 'register_passmail_template_message' ) ); add_action( 'tml_request', array( &$this, 'action_messages' ) ); add_filter( 'registration_redirect', array( &$this, 'registration_redirect' ) ); } /** * Outputs password fields to registration form * * Callback for "register_form" hook in file "register-form.php", included by Theme_My_Login_Template::display() * * @see Theme_My_Login::display() * @since 6.0 * @access public */ public function password_fields() { $template =& Theme_My_Login::get_object()->get_active_instance(); ?> <p><label for="pass1<?php $template->the_instance(); ?>"><?php _e( 'Password' ); ?></label> <input autocomplete="off" name="pass1" id="pass1<?php $template->the_instance(); ?>" class="input" size="20" value="" type="password" /></p> <p><label for="pass2<?php $template->the_instance(); ?>"><?php _e( 'Confirm Password', 'theme-my-login' ); ?></label> <input autocomplete="off" name="pass2" id="pass2<?php $template->the_instance(); ?>" class="input" size="20" value="" type="password" /></p> <?php } in functions.php i have added this but not working: function init_wp() { // remove parent theme's header content action remove_action('register_form', 'password_fields'); // add child theme's header content action add_action('register_form', 'cust_password_fields'); } add_action('init', 'init_wp'); function cust_password_fields(){ } any ideas ?
[ 0.01773129031062126, 0.0023536330554634333, -0.0017869460862129927, 0.014895174652338028, -0.021014807745814323, -0.013096416369080544, 0.007099333219230175, 0.0018716426566243172, -0.012652186676859856, 0.018723467364907265, -0.008896559476852417, 0.007669683545827866, 0.010792717337608337,...
[ 0.2283456027507782, 0.01877743937075138, 1.0024261474609375, -0.2674453854560852, 0.2260064333677292, -0.24525322020053864, 0.17238740622997284, -0.5065734386444092, -0.20699414610862732, -0.5542470216751099, -0.1401849091053009, 0.6259780526161194, -0.24997299909591675, 0.2602381408214569...
This is a follow up to Display Post by taxonomy and taxonomy child if exist I need to display a list of Accomodation, first by Type of Accomodation, then by Regions. I have a post-type, called `accomodation`, a taxonomy `Country`, a taxonomy `accomodation-type`, a taxonomy `Region` I am in the Country page, where I retrieve the value of the country via $taxonomy = 'country'; $terms=get_the_terms($post->ID,$taxonomy); if($terms) { foreach( $terms as $termcountry ) { So I am successfully getting `$termcountry->name;` as the value of current country. Next I need to do my query for the Accomodation based on that country, filtered by Accomodation Type, and by region in this type. So it should be * Acc-Type 1 * Region1 * Acco 1 * Acco 2 * Acco 3 * Region 2 * Acco 1 * Acco 2 * Acc-Type 2 * etc.... I am trying to do it this way $taxonomy2 = 'accomodation-type'; $termsacc = get_terms("accomodation-type",array('orderby' => 'slug', 'order' => 'ASC')); //Loop through Acc-Type foreach ($termsacc as $termaccomodation) { //Loop through Region in this type $taxonomyregion = 'region'; $termsreg = get_terms("region",array('orderby' => 'slug', 'order' => 'ASC')); foreach ($termsreg as $termregion) { $args = array( 'post_type' => 'accomodation', 'country' => $termcountry->name, 'tax_query' => array( 'relation' => 'AND', array( 'taxonomy' => 'accomodation-type', 'field' => 'slug', 'terms' => array($termaccomodation->name), ), array( 'taxonomy' => 'region', 'field' => 'slug', 'terms' => array($termregion->name), ) ) ); $query = new WP_Query($arg); //Do the loop but this is not doing it.... Scratching my head proper on this one and will appreciate some help/solution thx in advance
[ -0.006687586195766926, 0.018855243921279907, 0.004446698352694511, 0.016042182222008705, 0.009044022299349308, 0.0072311596013605595, 0.007255972363054752, 0.02357351779937744, -0.011371990665793419, 0.0012451717630028725, -0.004714117851108313, 0.004100551828742027, -0.01127350889146328, ...
[ 0.014823930338025093, -0.039449192583560944, 0.8150202631950378, 0.22869420051574707, 0.28556108474731445, 0.11038696765899658, 0.2369355708360672, -0.3263932466506958, -0.15883636474609375, -0.7440628409385681, -0.1062847152352333, 0.37714841961860657, 0.018795844167470932, 0.379777669906...
The settings: we have a sequence of numbers generated from a normal distribution with $\mu=1;\sigma^2 = 1 $. And with every new observation collected, we'd like to test for: $H_0:\mu_0=0$ (two-sided). From what I've learned in school, there is the Wald's group sequential for when comparing two simple hypothesis, and I would like to know what is available for complex hypothesis. What I am looking for is: 1. Some reference text (article/book) that describes the solution for this (or a similar) case. 2. If there is some implementation in R for performing such a test. Thanks.
[ 0.0012018660781905055, 0.014972508884966373, -0.011561150662600994, 0.012927509844303131, 0.00361511018127203, 0.0036454719956964254, 0.005869985558092594, 0.0020719249732792377, -0.008977554738521576, 0.013554869219660759, -0.008590202778577805, -0.0025441613979637623, -0.003906810190528631...
[ 0.2069825381040573, -0.012192553840577602, 0.40957364439964294, 0.1478615552186966, 0.05351492762565613, 0.25457024574279785, -0.033777885138988495, -0.3414483070373535, -0.17417962849140167, -0.6712960004806519, -0.0369577519595623, 0.6443156599998474, -0.11225631833076477, -0.00914006680...