text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
I'm a little new to data mining and would definitely appreciate some tips. I'm using clustering algorithms looking for possible grouping in some variables described below. I've been using the Excel data mining add-in which connects to SSAS and uses the EM algorithm by default. I'm also using R, so far with the Kmeans algorithm. I have two independent variables X, and G which are integers, and three dependent variables A, B, C which are related to X and G by the equations below: $$A=100\left(\frac{2G\cdot.06+K}{2G\cdot.4+2\cdot.06+K+X}\right)$$ $$B=100\left(\frac X{2G\cdot.4+2G\cdot0.06+K+X}\right)$$ C=100-((((2xG)x(0.06)+K)/((2xG)x(.4)+(2xG)x(0.06)+K+X)))x100 + (X/((2xG)x(.4)+(2xG)x(0.06)+K+X))X100) where K is an arbitrary constant which varies. Most of the data follows the formulas above with some variation. I'm using clustering algorithms to look for groupings in the A,B, C variables. Any advice on strategies for looking for clusters, and how to tell when I've been successful would be greatly appreciated.
[ 0.009096917696297169, 0.015280216000974178, -0.011386638507246971, 0.007936427369713783, -0.0018845831509679556, -0.0069434754550457, 0.007075001951307058, -0.011236922815442085, -0.013369400054216385, -0.006179140880703926, 0.006861301138997078, 0.003713466227054596, -0.0012871287763118744,...
[ 0.19543591141700745, 0.49593842029571533, 0.38445883989334106, -0.35866546630859375, 0.15612632036209106, 0.3303343653678894, 0.09469126164913177, -0.12377376854419708, -0.1900061070919037, -0.4282165467739105, 0.24684329330921173, 0.44321349263191223, -0.09043814241886139, 0.3395232260227...
It is very common in biology to draw diagrams like the one below to show regulatory interactions between genes: ![enter image description here](http://i.stack.imgur.com/oT81k.jpg) You can see that this is essentially a graph, with 3 key elements: * Nodes, which can be gene names (eg. `lin-1`) or events (`vulval fates`). * Positive (activating) edges, shown by arrows (often green), which indicate that a gene induces the other. * Negative (inhibitory) edges, shown by stubbly lines (often red), which indicate that a gene represses the other. Sometimes, it may be necessary to make the edges very long or curved (when the two genes are far away). Occasionally, dotted or dashed edges can be used to show putative or proposed relationships. Arrows may sometimes point to _other arrows_ , eg. when a gene A enhances the ability of gene B to activate gene C. A slightly less trivial example: ![enter image description here](http://i.stack.imgur.com/QaN0u.jpg) The above is the bare minimum of what I want to do. Obviously, the layout of the graph is a whole different can of worms, but let's ignore that for now and assume that the user will manually layout the nodes. ## Extra credit It is possible to make such diagrams much more complicated, to emphasize nuances in manner of regulation (transcriptional, post-translational, localization, direct vs. indirect, etc). Here is a showcase of a proposed symbolic vocabulary from Kohn 2006. Here is another proposed vocabulary from Pirson 2000 and an example usage
[ 0.0028395606204867363, 0.00884329155087471, -0.0026443605311214924, 0.0156567320227623, -0.020590953528881073, -0.014593495987355709, 0.0068991221487522125, 0.0021978463046252728, -0.011435079388320446, 0.004352258518338203, -0.012182692065834999, 0.0041868360713124275, -0.00859510712325573,...
[ 0.24665424227714539, -0.22321030497550964, 0.16214372217655182, 0.1856762170791626, -0.11749369651079178, 0.04090062901377678, -0.04561843350529671, -0.24365973472595215, -0.3422533869743347, -0.6051036715507507, 0.23856766521930695, -0.3549240231513977, -0.7347750663757324, 0.475791215896...
I'm trying to write a new "related posts" widget (for the single post view) for my custom post type, filtering by custom taxonomy terms that the related posts should have in common with the current post. I found this solution. But what I'd like to add is that the posts are ordered by relevance. I determine the relevancy by the number of taxonomy terms that they have in common: The more taxonomy terms in common, the more relevant. Example: Current post has tags (1), (2), (3) and (4). A related post with tags (1), (3) and (4) is more relevant than a related post with only the tags (1) and (2). Also, if possible, I'd like to add special severity/emphasis to one of the custom taxonomy terms (which I assume to be the "main" term of a post). Example: Current post has tags (1), (2), (3) and (4). (1) is the main term of the post. Posts should only be displayed as related posts, if they have at least this main term in common. A related post with tags (1) and (3) is more relevant than a related post with the tag (1) only. (Optional: A post with the tags (2), (3) and (4) is not considered a related post, or: this post has a lower relevance than any post that has the main term (1) in common.) Can you please help me with this? Thanks in advance! :) Regards, René
[ 0.004934198223054409, 0.01628510095179081, -0.009996859356760979, 0.018233705312013626, 0.024816077202558517, -0.005489058326929808, 0.008389536291360855, 0.022638602182269096, -0.013937386684119701, -0.006071953102946281, -0.005433207843452692, 0.004279337823390961, 0.0016122078523039818, ...
[ 0.5128080248832703, 0.034110307693481445, 0.6290704011917114, 0.2654861509799957, -0.3039769232273102, 0.007939138449728489, 0.0076229325495660305, -0.10659286379814148, -0.2667262852191925, -0.37544557452201843, 0.03427085280418396, 0.011659442447125912, -0.620591938495636, 0.214368313550...
I have a number of friends on my battle.net account. However, sometimes I don't want to appear online to some (or all) of them. Is there a way to be hidden from your friends on battle.net?
[ -0.003347561927512288, 0.014395897276699543, 0.008875691331923008, 0.030350688844919205, 0.036980029195547104, 0.01901899464428425, 0.007289911154657602, -0.027802325785160065, -0.030289243906736374, -0.014635569415986538, 0.01604609377682209, 0.024706650525331497, -0.009935320354998112, 0...
[ 0.7807347178459167, 0.04252270981669426, -0.19844931364059448, -0.14963199198246002, -0.0184155460447073, 0.09933122247457504, 0.598209798336029, 0.45802801847457886, -0.3112597167491913, -0.596062421798706, 0.09940989315509796, 0.3290065824985504, 0.35252153873443604, 0.07070621103048325,...
Some years back I have developed a little C# class library, which is still frequently downloaded. Since I do not have the time to support the project any more, I would like to donate it. What is the best place to do so? Some remarks: 1. I will not be the owner of that project any more, I just "want to hand it over" 2. It has to be "free of charge", non commercial, ... 3. Best would be if others could make changes and also re-distribute them 4. It must not be a legal hazard for me, so no obligations etc. Have you done something before, what would you recommend...? Update: I got excellent feedback in virtually no time, **so I have to choose a license**. I am not so much in that legal stuff, also it is not my main interest. What do you recommend for this? In plain words, I only have little demands (clarifying my intentions): 1. People can use it for commercial and non commercial software 2. I do not want to be in any kind of liabilities 3. All improvements (fixes etc.) should be also available for me and the public under the same license as in 4. 4. I do not want force people to publish their whole source code when they deploy the library as part of a bigger project. But I want them to re-publish any direct improvements or bug fixes. There are too many licenses to read them all....
[ -0.004523167386651039, 0.00851029809564352, -0.009516969323158264, 0.015246625989675522, 0.008842863142490387, -0.001572508830577135, 0.006444721017032862, 0.017115090042352676, -0.014693526551127434, -0.016798827797174454, 0.006910483352839947, 0.012390118092298508, -0.0015867071924731135, ...
[ 0.7509810924530029, 0.18619726598262787, -0.07134824246168137, 0.3072469234466553, 0.15513843297958374, -0.3740134835243225, 0.05538789555430412, 0.2743573486804962, -0.0558529794216156, -0.2631344199180603, 0.09436400234699249, 0.07620355486869812, 0.10085158050060272, 0.34497880935668945...
If an observer starts moving at relativistic speeds will he observe the temperature of objects to change as compared to their rest temperatures? Suppose the rest temperature measured is $T$ and the observer starts moving with speed $v$. What will be the new temperature observed by him?
[ -0.011563779786229134, 0.04377362132072449, 0.008209798485040665, 0.01558710541576147, 0.03173522278666496, -0.0011543924920260906, 0.018192462623119354, -0.04321260005235672, -0.028826240450143814, -0.014253659173846245, 0.010479077696800232, 0.03009260818362236, -0.009662174619734287, 0....
[ 0.2886382043361664, -0.23940517008304596, 0.6531461477279663, 0.3116503655910492, 0.2579086422920227, 0.2636399567127228, 0.16905619204044342, -0.3888123631477356, -0.3224041163921356, -0.487813800573349, 0.08371617645025253, 0.37569695711135864, -0.3948039412498474, 0.5553857088088989, ...
In 1985, Harmuth wrote that Maxwell's equations are incompatible with causality, and overcame the problem by adding a term for magnetic dipole currents, and as a consequence the problem of infinite zero-point energy and renormalization disappears. At least according to Harmuth's book: Calculus of finite differences in quantum electrodynamics By Henning F. Harmuth, Beate Meffert The foreword is readable at http://books.google.com/books?id=EQFENquY6fwC Are Harmuth's modifications generally accepted by the physics community as a more accurate description of reality than the unmodified equations ?
[ 0.021525880321860313, 0.010066568851470947, -0.01255507580935955, 0.011085809208452702, -0.022142989560961723, -0.015252301469445229, 0.009633276611566544, -0.014178469777107239, -0.01025199331343174, -0.015304102562367916, -0.0012499834410846233, 0.017298247665166855, -0.007514116354286671,...
[ -0.26876023411750793, 0.3227075934410095, 0.21006061136722565, -0.18288034200668335, -0.40375813841819763, 0.17801325023174286, 0.39638978242874146, -0.6552339196205139, -0.17844398319721222, 0.1531299501657486, 0.06265740096569061, 0.060882896184921265, -0.3629845976829529, 0.658949077129...
> **Possible Duplicate:** > How should I test randomness? One of the features to be developed in our application was allowing user to click a button which will choose an element at random out of (normally 5-10) values registered in the database. The randomness doesn't need to be secure or scientifically uniform, but we do need it to be somehow unpredictable and not noticeably biased to one of the value. Any suggestion how to write a good unit test and functional test for this feature?
[ 0.011381980963051319, 0.01644967496395111, -0.01138328481465578, -0.003311177482828498, -0.03649229556322098, -0.01304452121257782, 0.008405535481870174, 0.018878577277064323, -0.021429048851132393, -0.005317412316799164, -0.010813974775373936, 0.008364446461200714, 0.0001977197825908661, ...
[ 0.28465813398361206, -0.10780074447393417, -0.014418043196201324, 0.33784565329551697, -0.12680453062057495, -0.016152488067746162, -0.013116441667079926, -0.008397993631660938, -0.2965872883796692, -0.5828551650047302, 0.2971779406070709, 0.18820984661579132, -0.1334717720746994, 0.032625...
Master file: \documentclass[dvipsnames,cmyk]{book} % File Containing All the Control Settings \usepackage{../../Style/mystyle} \makeindex \begin{document} % Front Cover \include {Front_Cover/frontcover} % Chapters \include {Chapters/CH_Analysis_of_Algorithms/Analysis_of_Algorithms} \include {Chapters/CH_Abstract_Data_Types/Abstract_Data_Types} \include {Chapters/CH_Arrays/Arrays} \include {Chapters/CH_Linked_Lists/Linked_Lists} \include {Chapters/CH_Stacks/Stacks} \include {Bibliography/biblio} % Index \printindex \end{document} I compiled it as follows: $pdflatex master.tex => Empty `idx` file generated $makeindex master.idx **This is makeindex, version 2.15 [TeX Live 2012] (kpathsea + Thai support). Scanning input file master.idx...done (0 entries accepted, 0 rejected). Nothing written in master.ind. Transcript written in master.ilg.** $pdflatex master.tex Output: PDF with NO Index. I am compiling from the command line using `pdflatex` (don't use `latex` to avoid boxing issues with graphics) Please guide me on this issue.
[ 0.005817533936351538, 0.006517605856060982, -0.0030468287877738476, 0.029292481020092964, 0.01650809496641159, 0.008605169132351875, 0.007750139571726322, -0.011430355720221996, -0.013211844488978386, 0.013377778232097626, -0.010708276182413101, 0.004745559301227331, 0.010620888322591782, ...
[ -0.3177921175956726, 0.20076093077659607, 0.44545215368270874, 0.44752663373947144, 0.1834408938884735, 0.12581229209899902, 0.057536590844392776, -0.6186767816543579, -0.25764477252960205, -0.5161190629005432, -0.49531590938568115, 0.36544910073280334, -0.42269790172576904, -0.02175002917...
Do you need to have a generator or an app of some sort that would convert the Textile markup you did to its equivalent HTML?
[ -0.014608322642743587, 0.029170675203204155, 0.01678762398660183, 0.015168966725468636, -0.07421508431434631, 0.026846660301089287, 0.014869333244860172, 0.04254722595214844, -0.03413122147321701, -0.029756031930446625, -0.0027995039708912373, 0.017271654680371284, 0.04112003743648529, 0.0...
[ 0.7853681445121765, -0.06986440718173981, 0.14221060276031494, 0.3279336988925934, 0.0274221021682024, -0.09546536207199097, -0.0616954080760479, 0.17190346121788025, -0.434824138879776, -0.3805379569530487, 0.1659497618675232, 0.17917393147945404, -0.4296123683452606, -0.13872379064559937...
A string with a fixed frequency vibrator at one end forms a standing wave with 4 antinodes when under tension T1. When the tension is slowly increased, the standing wave disappears until tension T2 is reached with no resonances occurring between the two tensions. How many antinodes are there in this new standing wave? There would be 8 antinodes, right? Because as you jump up to the next frequency, the number of antinodes doubles. Actually, no, it would be 5 antinodes, right?
[ -0.015606033615767956, 0.02217589132487774, -0.01229831948876381, 0.02052273601293564, -0.016445403918623924, -0.017005804926156998, 0.012018399313092232, -0.004907801281660795, -0.01962575688958168, 0.015923205763101578, -0.012488537468016148, 0.0230440404266119, -0.027742905542254448, -0...
[ 0.2582157850265503, -0.6093751192092896, 0.6009319424629211, -0.0860457792878151, -0.4846900999546051, -0.00933182891458273, 0.18356597423553467, -0.5920552015304565, -0.690091609954834, 0.2942748963832855, 0.06391290575265884, 0.28264567255973816, -0.43287962675094604, 0.3359479010105133,...
I have a problem with the Hamiltonian, I don't think anything to solve it!! So could you give me some hints! * Knowing that: $$\left[ {{{\hat p}_i},{{\hat q}_k}} \right] = \frac{\hbar }{i}{\delta _{ik}}.$$ Hamiltonian is: $$\begin{array}{l} \hat H = \hat q_i^{},\hat q_i^2,\hat q_i^3,....,\hat q_i^n, \\\ (n \in N) \\\ \end{array}$$ Proof that: $$\left[ {\hat H,{{\hat p}_i}} \right] = - \frac{\hbar }{i}\frac{{\partial \hat H}}{{\partial {{\hat q}_i}}}.$$
[ -0.003847491927444935, 0.009298151358962059, -0.008524062111973763, 0.002395590767264366, -0.015484336763620377, -0.008426649495959282, 0.007195138372480869, 0.0044242460280656815, -0.01535478513687849, 0.006273274775594473, -0.017786268144845963, 0.0002995976246893406, -0.03752565383911133,...
[ 0.023240212351083755, 0.1935461312532425, 0.28048813343048096, 0.11819121241569519, 0.15776827931404114, 0.06512679904699326, 0.1957460194826126, -0.21581794321537018, -0.2133997082710266, -0.3446168899536133, 0.1416367143392563, 0.33353543281555176, -0.48241016268730164, 0.036141071468591...
In the `listings` package, they note that "a global `\documentclass`-option `draft` is recognized, so you don’t need to repeat it as a package option" I tried looking into its source to see how it accomplishes this, but I wasn't able to find the magic bit. How does one detect the global `\documentclass` options from within an independent style/package?
[ 0.003977660089731216, 0.006506944075226784, -0.007890396751463413, 0.02060161717236042, 0.0043749636970460415, -0.018083596602082253, 0.006900842767208815, 0.01723957620561123, -0.01940562017261982, 0.02364945411682129, -0.013577739708125591, 0.010343768633902073, 0.0025583356618881226, 0....
[ 0.2370406538248062, -0.013561572879552841, 0.36171576380729675, 0.07231465727090836, 0.1358160376548767, -0.5161274671554565, 0.13201658427715302, -0.25557929277420044, -0.2513595223426819, -0.5684969425201416, -0.17676901817321777, 0.451938658952713, 0.0034694161731749773, 0.1774506717920...
I've searched a lot for this and I seem to get different answers. In my situation, my jQuery function has to use a PHP variable from a different file. How can I accomplish this? Clear and simple answer would be greatly appreciated.
[ 0.016049271449446678, 0.035013604909181595, 0.017883816733956337, 0.003241775557398796, 0.03516724333167076, -0.023166920989751816, 0.006849282421171665, -0.0006953009287826717, -0.033903565257787704, -0.04387550428509712, -0.008531615138053894, 0.013615080155432224, 0.0023645649198442698, ...
[ 0.14637811481952667, 0.04980533942580223, 0.27231070399284363, 0.4672875702381134, -0.1265518218278885, 0.1423002928495407, 0.37320220470428467, -0.12650181353092194, -0.007473536767065525, -0.5199382305145264, 0.345081090927124, 0.2916617691516876, 0.13878275454044342, 0.29468291997909546...
Imagine a simple set up where where entities (e.g. a restaurant) are rated every year with 5 rating categories $c_t$: _excellent_ , _good_ , _medium_ , _poor_ and _bad_. Now assume that I have some underlying data $X^i_t=(x^{1i}_t,x^{2i}_t, ..., x^{pi}_t)$ at entity level $i$ on top of the rating history $c^i_t$. I am interested in a modelling approach to predict $c^i_t$ from $X^i_t$. In the past, I have randomly sampled each entity only once from the entire rating history (to minimise the correlation effects) and used an ordered logit model to estimate the coefficients of $X^i_t$. However, I believe I can make more efficient use of my data if I could use the entire data. Is there such a model? Could you please point to some literature?
[ 0.014568135142326355, 0.01867208071053028, -0.00238037365488708, 0.007763988338410854, 0.021502947434782982, 0.010333377867937088, 0.006422686390578747, -0.012652495875954628, -0.009549543261528015, 0.011855781078338623, -0.022753272205591202, 0.006370872259140015, 0.004148267209529877, 0....
[ 0.3378313183784485, 0.4194931089878082, 0.3900202512741089, -0.15829499065876007, 0.215650737285614, 0.00952284038066864, -0.047403037548065186, 0.007044536527246237, 0.05912889167666435, -0.4013878405094147, 0.053508054465055466, 0.5873059630393982, -0.006204687990248203, 0.51532739400863...
Is there a nice (built-in) way to apply Replace-style rules to key-value pairs in an association? Something like SomeFunction[<| "id" -> 3, "freq" -> 4 |>, "freq" -> #freq+1 &] (* should produce <| "id" -> 3, "freq" -> 5 |> *) This could also be used to add computed keys, like SomeFunction[..., "density" -> #mass / #volume &] This would be really useful when combining with `Query` and `Dataset`.
[ 0.0054125310853123665, 0.014620537869632244, -0.0062920693308115005, 0.009531480260193348, 0.002884763991460204, -0.010395806282758713, 0.006152174435555935, 0.0006823018193244934, -0.013937413692474365, 0.01702386699616909, 0.00003892887616530061, 0.008433357812464237, -0.001203509047627449...
[ 0.1178838312625885, -0.028055788949131966, 0.24885734915733337, 0.3640882968902588, 0.02596040442585945, -0.20630770921707153, -0.13668419420719147, -0.146736741065979, -0.13017794489860535, -0.3340104818344116, -0.11802004277706146, 0.4901021122932434, -0.29488202929496765, -0.20230303704...
We're looking to beef up the security on a site, among other things we thought of trying to block access to the DB based on IP (in a similar way you can do a FTP lock dependent on IP), so that only the static IP of our VPS can access it as well as our local office IP (also static). Is this a standard approach, I tried searching for it to see any notes on implementation / best practices, but I could see anything written about it?
[ 0.019798552617430687, 0.010033202357590199, 0.002932511270046234, 0.00018093430844601244, -0.0037007289938628674, 0.013090449385344982, 0.007263776380568743, 0.014142726548016071, -0.01793580874800682, 0.02640555053949356, -0.02270282618701458, 0.01559162326157093, -0.01707475632429123, 0....
[ 0.7481943964958191, 0.1747783124446869, 0.1884242296218872, 0.09474944323301315, -0.11423441767692566, -0.4005882441997528, 0.3380100727081299, -0.08173972368240356, -0.23479199409484863, -0.12837111949920654, 0.23303012549877167, 0.13880425691604614, -0.17386363446712494, 0.15396779775619...
I'm creating an age select menu in the admin, populated from a taxonomy of `age`. The taxonomy is hierarchical as follows: * 18-25 (parent, ID 183) * 18 (child) * 19 * 20 * 21 * 22 * 23 * 24 * 25 * 26-30 (parent, ID 184) * 26 * 27 * 28 * 29 * 30 I would like to only list the children (18, 19 etc) and not the parents (18-25, 26-30) etc. Currently I am using `get_terms` with the `parent` argument, but it doesn't accept more than 1 parent ID. Here's what I have so far, which shows the children from 18-25. $ages = get_terms( 'age', array( 'hide_empty' => 0, 'parent' => '183', )); Here's what I want it to do, but isn't supported. I have also tried it with an array but it doesn't work either. $ages = get_terms( 'age', array( 'hide_empty' => 0, 'parent' => '183,184', )); I see there is a get_term_children function but I'm unsure of how to use this either as it looks like it only accepts one value also. Eg: In this example it would build an unordered list but I could modify for select menu. <?php $termID = 183; $taxonomyName = "age"; $termchildren = get_term_children( $termID, $taxonomyName ); echo '<ul>'; foreach ($termchildren as $child) { $term = get_term_by( 'id', $child, $taxonomyName ); echo '<li><a href="' . get_term_link( $term->name, $taxonomyName ) . '">' . $term->name . '</a></li>'; } echo '</ul>'; ?>
[ -0.0002668179804459214, 0.011003969237208366, -0.013116404414176941, 0.013935131952166557, -0.01092330738902092, -0.003716300241649151, 0.007029109168797731, 0.017275091260671616, -0.016658397391438484, -0.013520397245883942, -0.011890709400177002, 0.00032220061984844506, -0.0103053711354732...
[ 0.38021987676620483, 0.3397166132926941, 0.5462248921394348, -0.10778269916772842, 0.2651313841342926, 0.5175045132637024, 0.18028847873210907, -0.25597211718559265, -0.1255904734134674, -0.7657797932624817, -0.11145473271608353, 0.11398770660161972, 0.22981266677379608, 0.4132695496082306...
I am trying to make a function that will check the time associated with a future event (cpt) with the current time, to see if the event has past. When I first collect the time it is in seconds and I can then convert it easily with date_i18n() to whatever, but I am not sure how to compare that results with the wordpress function the_time() to retrieve the current time. Is there a way to have the_time() spit out the result in seconds, and I can then compare the two simply by ><= ? or is there a way to compare two dates formatted in the same way? So formatting my date_i18n() result the same way as the_time() result and then comparing somehow
[ -0.006090093869715929, 0.00985533744096756, -0.012777432799339294, 0.011041899211704731, -0.0032163173891603947, 0.006318553350865841, 0.007955620065331459, 0.008314091712236404, -0.015073016285896301, 0.03384242579340935, -0.012792903929948807, 0.0010434433352202177, 0.006048486568033695, ...
[ 0.1706237941980362, -0.051110848784446716, 0.45024728775024414, 0.08165520429611206, 0.2982196509838104, 0.22810018062591553, -0.019073160365223885, 0.17590729892253876, -0.20957551896572113, -0.6110751032829285, 0.2609076499938965, 0.20622222125530243, -0.06998752802610397, 0.459860265254...
When I need to find my recently used files, this command: `ls -lt` lists in ascending order (by time), when there are lots files, I need to scroll to the top to see needed files, because wont fit in terminal screen. After finding out that `tac` inverses the output, I use: `ls -lt dir/ | tac` Is there more fun way of doing it, without using external scripts/utils?
[ 0.018258171156048775, 0.0011436102213338017, -0.025679200887680054, -0.015686726197600365, 0.012381424196064472, -0.0025146275293082, 0.007647709921002388, -0.03147896006703377, -0.015902753919363022, 0.006055232137441635, 0.00028423042385838926, -0.018351474776864052, 0.00813485961407423, ...
[ 0.08643507957458496, -0.0714457556605339, 0.4076564908027649, -0.42708054184913635, 0.2148837298154831, -0.052136920392513275, -0.07520676404237747, 0.3111272156238556, -0.17034733295440674, -0.7342197895050049, -0.13999207317829132, 0.7480630874633789, -0.41157999634742737, -0.03756663948...
A while ago I made my own distro of Linux using `SuseStudio` and may have left out some important packages. * Apps like `gnome-software` and `gnome-music` don't work * The computer runs really high process rates, doing simple tasks like `web browsing` * `vmware` can't recognize hardware acceleration * Overall sluggishness about the machine is existent In the past I would update to the latest package version, but now I feel that I'm missing some packages. Is there a `zypper` command I can run or a way to install all the base SuSe rpms?
[ -0.014214186929166317, -0.004812462721019983, -0.0018405478913336992, 0.007046927697956562, -0.011137891560792923, 0.00577877601608634, 0.006937914527952671, 0.007157891057431698, -0.011237792670726776, -0.0037200013175606728, -0.018070286139845848, 0.002745129633694887, -0.01397354714572429...
[ 0.11994504183530807, 0.20784437656402588, 0.3026956021785736, 0.15839730203151703, 0.055206671357154846, 0.04327739402651787, 0.16324946284294128, 0.24064448475837708, -0.10755784809589386, -0.3891063630580902, 0.20704331994056702, 0.8197259306907654, -0.2548224925994873, 0.347558021545410...
When I enabled USB debugging in settings, an unrecognized device appeared in device manager: H9000+ Then I followed this: http://www.zopomobileshop.com/zopo-rom/how-to-install- driver-for-zopo-mobile-phone-step-by-step-instruction/ MT65XX Preloader replaced with "MediaTek PreLoader USB VCOM Port (COM3)", then it disappeared. I unplugged USB, turned on the phone and plugged it back. No new drivers were installed. H9000+ is still unrecognized. Eclipse says: No compatible targets were found
[ 0.002391892485320568, -0.003805612213909626, 0.01722729578614235, 0.009822836145758629, -0.01601215824484825, 0.004688770044595003, 0.010657898150384426, 0.0021618609316647053, -0.014943559654057026, -0.008005384355783463, -0.006309692747890949, 0.009032705798745155, 0.010416720993816853, ...
[ 0.020329970866441727, 0.20414040982723236, 0.35911157727241516, -0.03335193917155266, -0.168508380651474, 0.13672955334186554, 0.1615753173828125, 0.23104582726955414, -0.15096931159496307, -0.36675918102264404, 0.077304907143116, 0.771720826625824, -0.37303945422172546, 0.2392476946115493...
I host a weekly Minecraft night. All LAN server. Tonight, the Minecraft authentication server was down for nearly an hour. Blew game night out of the water. There used to be a "play offline" ability, but now, when I start the launcher, it immediately goes to username and password, and even left blank, I can't get past that screen. Any way to play LAN multiplayer games offline?
[ -0.005524401552975178, 0.010514388792216778, 0.0010423788335174322, -0.0016383419279009104, -0.039243608713150024, -0.018850911408662796, 0.006555150728672743, -0.0018952072132378817, -0.017940707504749298, 0.013213472440838814, -0.02457243949174881, 0.010445350781083107, 0.03217344358563423...
[ 0.28611722588539124, 0.17921139299869537, 0.6366645097732544, 0.07391701638698578, -0.14001935720443726, -0.24147270619869232, 0.18813617527484894, 0.25165897607803345, -0.2162182629108429, -0.40363818407058716, 0.23732247948646545, 0.4690729081630707, 0.22346696257591248, 0.20207011699676...
I am using a Samsung Galaxy S (running Android 2.1) mobile phone and am thinking (only thinking mind you) of bying a Samsung Galaxy tablet (this one running Android 2.2). I am wondering: is it somehow **possible to sync the data used by programs on these two devices** so that software (like the marvelous Book catalogue) can be installed twice with the possiblity to copy the databases used by these program to both mobiles/tablets ? Preferably by copying them over using WiFi or so.
[ -0.012820814736187458, 0.0005812999443151057, -0.006002792622894049, 0.013621818274259567, -0.01157341431826353, -0.0007741222507320344, 0.00812392309308052, 0.031153138726949692, -0.011944770812988281, -0.039457499980926514, -0.015511224046349525, 0.013709990307688713, -0.001811709953472018...
[ 0.3527812361717224, 0.32868263125419617, 0.43193086981773376, 0.08603404462337494, 0.18931041657924652, 0.07202169299125671, -0.1340136080980301, 0.22738510370254517, -0.38538631796836853, -0.6764868497848511, -0.11475366353988647, 0.2429007589817047, -0.1252647340297699, 0.222346812486648...
ArcMap wont display the attribute table of a dataset. I get an error 'Could not load data from the data source' and it says 'An invalid SQL statement was used'. It then opens a table with the correct number of fields, but no field headers and no data. It does however display the attribute table if I make a selection and ask it to show only the selection, and I can select every record in the table and it works. It also displays the values if I use the identify button. The datasource is a table in a Geodatabase and was created from FME. FME reports no errors in loading the data, and loads all of the records it is given. Using ArcGIS 10.1 SP1. The table has no special characters in the table name, or the field names. Edit 10/05/2013 The Geodatabase will open in QGIS. If I export to another features class in a file geodatabase using ArcCatalog it the works. I have tried using the FILEGDB and GEODATABASE_FILE writers in FME and the attribute table will not open in ArcMap. Both work in QGIS 1.8
[ 0.0027125575579702854, 0.006253235973417759, -0.011574538424611092, 0.010547621175646782, -0.0002476838417351246, -0.016681108623743057, 0.006299878004938364, 0.007982378825545311, -0.014529719948768616, -0.013528146781027317, -0.008593607693910599, 0.01264314353466034, -0.006065273657441139...
[ -0.09151310473680496, -0.3038088381290436, 0.2843991816043854, 0.22366467118263245, -0.15203678607940674, -0.0468614436686039, -0.15506047010421753, -0.3094911575317383, 0.1294812113046646, -0.5658130049705505, 0.09651001542806625, 0.66605144739151, -0.4849822223186493, -0.2962610125541687...
I've a CentOS 5 server at domain.fr. I am trying to set a subdomain so that I can use it with git : git.domain.fr My repositories are in /home/git (example /home/git/repos.git) I've installed `git-http-backend` and Nginx. I've set a repostory like this: cd /home/git/repos.git && git --bare init I've set my `git.conf` (included in `nginx.conf`) as below. However, on my client shell, I get the fatal error "repository not found": `git clone http://git.domain.fr/repos.git` Does anyone know what I should do? It seems so simple, I'm getting frustrated because I'm sure it's nothing. server { listen 80; server_name git.domain.fr; root /home/git; location ~ /(/.*) { include /etc/nginx/fastcgi_params; fastcgi_param SCRIPT_FILENAME /usr/bin/git-http-backend; fastcgi_param GIT_HTTP_EXPORT_ALL true; fastcgi_param GIT_PROJECT_ROOT /home/git; fastcgi_param PATH_INFO $1; fastcgi_pass unix:/var/run/fcgiwrap.socket; } }
[ -0.006077671889215708, 0.0010696679819375277, 0.0001458687474951148, 0.020066045224666595, 0.022656716406345367, -0.0045746974647045135, 0.00961393117904663, 0.004783944226801395, -0.016397986561059952, 0.005343432072550058, -0.0051686582155525684, 0.0018190972041338682, -0.00189026934094727...
[ 0.23078249394893646, 0.36748823523521423, 0.3063122630119324, -0.12464733421802521, 0.14557430148124695, 0.22292566299438477, 0.3189662992954254, 0.11695430427789688, -0.23930969834327698, -0.8940660357475281, -0.10674070566892624, 0.40158388018608093, 0.08219201862812042, 0.25983634591102...
This is a comment from a discussion on The effects of exercise on depression: > For severe depression, this actually is pretty much the treatment routine. > Pills to get just the tiniest amount of motivation, cognitive behavioral > therapy to **pence** that motivation towards positive activities like > regular exercise instead of suicide. > > Sounds severe, but that's how problematic deep depressions often are. What does _pence_ mean mentioned above? I've looked it up, but I couldn't find an entry in the dictionaries that defines _pence_ in verb form.
[ -0.002398671582341194, 0.006935426965355873, -0.00662160012871027, 0.01704263500869274, -0.015109175816178322, -0.027871517464518547, 0.0070890155620872974, 0.013881813734769821, -0.012737523764371872, -0.0006186917889863253, -0.010098898783326149, 0.008626772090792656, -0.004475616849958897...
[ 0.5519557595252991, 0.06871753185987473, -0.2376338690519333, -0.20276778936386108, -0.3556879162788391, 0.13633662462234497, 0.45325174927711487, -0.07772113382816315, -0.21636860072612762, -0.22037440538406372, 0.47772499918937683, 0.5383114814758301, -0.10569997876882553, -0.07404390722...
What's an intuitive derivation behind ODO's definition that helps to internalise its meaning: > **to leverage** = Use borrowed capital for (an investment), expecting the > profits made to be greater than the interest payable: Etymonline relates only to a physical/mechanical 'lever', not this meaning in finance.
[ 0.002621383173391223, 0.006159466225653887, -0.013485230505466461, 0.002052895724773407, -0.008616330102086067, 0.010536767542362213, 0.011686215177178383, -0.006752840708941221, -0.015323524363338947, 0.007194740232080221, -0.005697291810065508, 0.00880415178835392, 0.011033117771148682, ...
[ 0.17027738690376282, -0.6973237991333008, 0.21367144584655762, -0.11208513379096985, 0.16706955432891846, 0.2996901571750641, -0.5569931864738464, -0.26506856083869934, -0.24463462829589844, -0.1419265866279602, 0.02475346066057682, 0.7958238124847412, 0.29057905077934265, 0.30830925703048...
I have built a framework for my applications that has been working well, and I have always referred to it in my mind as an MVC. However, after reading some more on what an MVC is (not that I didn't to start with, but I gave it a very brief overview, and looked at other frameworks that used the mvc pattern), and it has given me pause for thought. The structure is as follows: The framework has a bootstrap that loads the required files for the framework and the initiation file for the application (`app_bootstrap.php`). The `app_bootstrap.php` file has three functions. First, it creates constants for the framework to use (i.e. Database connection setup, profiling setup, base url path, and internal file paths). Second, it requires all the files that would be pertinent to the application, primarily, what I have called "model" classes which contain information on the name and structure of tables in the database, validation rules, associations to other "models", etc. These "models" extend an abstract class that contains static functions for selecting and creating rows from the database, which will create instances of an entity class that deals with the row, such as `__get`, `__has`, `__set`, `__unset` for fetching columns and associations, and `save` and `delete` for saving the row changes to the database. Third, it calls functions within the `router.php` file that will setup bindings for environmental data (like the url, user login status, ssl, ajax, cookie existence, etc) to "controller" classes. The framework then runs an execution function within `router.php` that will go through the defined environmental data rules, and determine which "controller" to load, if it finds one, it will `require` the file (from the application/controllers directory, throwing an error if not found) and call the class with the appropriate entry point function (index is none defined), or throw a 404 error. Once the "controller" has been started, there are no more functions that the framework will execute, and once the "controller" has finished, the program will end (unless the profiling has been activated, where it will output the data, then end). But the framework provides functions to the "controller", and to name a few, an `input` class (for dealing with all input, such as `$_GET`, `$_POST` or `$_SERVER` variables, cookies, user agent, remote ip, etc, as well as providing a sanitized interface), an `output` class (for headers, cookies, status messages, etc), a `queryBuilder` class (for creating queries to the database in the event functionality that can't be done within the "model"), and a `view` class (for parsing variables into a view for rendering and returning to the client). But after reading some more information on MVC's, it appears that the `router.php` acts similarly to a Controller, the "controller" acts like the model, the "model" acts like an ORM, and the view is pretty much correct. So, would this framework still be classed as an MVC, or does it fit under another pattern (if there is such a pattern)?
[ -0.024243682622909546, 0.022989783436059952, 0.016106240451335907, 0.004477035254240036, 0.024662768468260765, -0.016880303621292114, 0.006333607714623213, -0.026239244267344475, -0.017077207565307617, -0.012774461880326271, -0.0211722981184721, -0.000033198739401996136, 0.016581006348133087...
[ 0.43082913756370544, 0.3764961063861847, 0.4470869302749634, -0.2727653682231903, -0.07493642717599869, -0.24945107102394104, 0.10421034693717957, -0.3976484537124634, -0.3732057809829712, -0.30204272270202637, -0.003830453846603632, 0.5068524479866028, -0.6621017456054688, -0.273983210325...
In the output of netstat -a | grep LISTEN there are usually a lot of processes. How (where) can I find out information about them, what is their purpose in the system and if I can kill them? Which of them can be insecure and which are safe?
[ 0.013867940753698349, 0.011055667884647846, 0.008964975364506245, 0.014480399899184704, -0.006966509856283665, 0.041815098375082016, 0.009690907783806324, 0.005330575630068779, -0.021090958267450333, 0.0444970540702343, -0.0060770208947360516, 0.017202481627464294, -0.009367324411869049, 0...
[ 0.252872109413147, -0.027963031083345413, 0.038722384721040726, 0.13990484178066254, 0.05679077282547951, 0.003992051817476749, 0.18721231818199158, -0.21059641242027283, -0.3484289050102234, -0.596246600151062, 0.07862214744091034, 0.8020920157432556, -0.4793751537799835, 0.14836804568767...
I was attempting to add this Flash Count Down plugin. The shortcode works fine in preview: ![enter image description here](http://i.stack.imgur.com/eOgwo.jpg) But when published the shortcode is rendered as text: ![enter image description here](http://i.stack.imgur.com/7JbEg.jpg) Is there a general solution to this problem or do I have to get someone from the publisher to answer this question? They have apparently abandoned supporting these as their link is a 404: http://premiumcoding.com/premiumcode-forum/wordpress-plugins-group1/wordpress- plugins-support-forum2/ Here is the template index file: <?php /* Template Name: Index Page */ ?> <?php get_header(); ?> <div id="content" class="content content-group content-index"> <div class="pad"> <div class="post-group"> <?php get_template_part('loop','index'); ?> </div> </div> </div> <?php get_sidebar(); ?> <?php get_footer(); ?> Here's the main index file: <?php /** * Front to the WordPress application. This file doesn't do anything, but loads * wp-blog-header.php which does and tells WordPress to load the theme. * * @package WordPress */ /** * Tells WordPress to load the WordPress theme and output it. * * @var bool */ define('WP_USE_THEMES', true); /** Loads the WordPress Environment and Template */ require('./wp-blog-header.php'); ?> loop.php: <?php while (have_posts()) : ?> <?php the_post(); ?> <div id="post-<?php the_ID(); ?>" <?php post_class('append-clear'); ?>> <div class="title"> <h1><a href="<?php the_permalink() ?>" rel="bookmark" title="Permanent Link to <?php the_title_attribute(); ?>"><?php the_title(); ?></a></h1> </div> <?php // padd_theme_share_button(); ?> <?php //the_excerpt();?> </div> <?php endwhile; ?>
[ -0.0013087630504742265, -0.001911817817017436, 0.004163671750575304, 0.022244900465011597, -0.020768877118825912, -0.017567675560712814, 0.004591562785208225, 0.004045139066874981, -0.017058026045560837, 0.002463557990267873, -0.02336878702044487, 0.007366318255662918, 0.0038072995375841856,...
[ 0.3326543867588043, 0.013276892714202404, 0.36755654215812683, 0.00648828549310565, -0.2876126170158386, -0.18053659796714783, 0.2669921815395355, 0.029958028346300125, -0.22442497313022614, -0.24763065576553345, -0.12203842401504517, 0.3849291503429413, -0.4403460919857025, 0.031076174229...
I'll have to come up with some examples to show you my question: I know if a sentence is inside either, the punctuation is inside (I also know I use a lot of comma splices. I think of the way the sentences sound in my head and break them up that way.). I'm stumped. Would when I say, "break them up that way.)." be correct? Or if I say, "break them up that way.).", be correct? Or if I say, "break them up that way.).," be correct? I'm also using the entire sentence as an example, so please take that into consideration. For the paraphrased items at the end of a sentence should it also just be separated into a completely new sentence? Like: "I know if a sentence is inside either, the punctuation is inside. (I also know I use a lot of comma splices. I think of the way the sentences sound in my head and break them up that way.)" I'd also like to know the British vs English(American) rules on this.
[ -0.004237246233969927, 0.009659918025135994, -0.009716557338833809, 0.02906545251607895, 0.01234051026403904, 0.006803279276937246, 0.007464933674782515, -0.00692791398614645, -0.017384804785251617, 0.010122952982783318, -0.0029368549585342407, -0.008402124047279358, 0.006099113263189793, ...
[ 0.13143067061901093, 0.48517921566963196, 0.10642970353364944, -0.08410099148750305, -0.3664529323577881, 0.07288319617509842, 0.6696361899375916, -0.11967821419239044, -0.40360939502716064, -0.5440379977226257, -0.21174703538417816, 0.30322787165641785, 0.1946992725133896, 0.1419061124324...
How can I give shape to some arrow in tikz, to draw the inclusion arrow? I read a similar question How to typeset inclusion arrow? with given answer, but the answer didn't really used tikz and I really need tikz in order to produce nice biggish diagrams. For instance, in this diagram (taken from here ), I want, say, A to be included in A' : \matrix (m) [matrix of math nodes, row sep=3em, column sep=3em] { 0 & A & B & C & 0 \\ 0 & A' & B' & C' & 0 \\ }; { [start chain] \chainin (m-1-1); \chainin (m-1-2); { [start branch=A] \chainin (m-2-2) [join={node[right,labeled] {\eta_1}}];} \chainin (m-1-3) [join={node[above,labeled] {\varphi}}]; { [start branch=B] \chainin (m-2-3) [join={node[right,labeled] {\eta_2}}];} \chainin (m-1-4) [join={node[above,labeled] {\psi}}]; { [start branch=C] \chainin (m-2-4) [join={node[right,labeled] {\eta_3}}];} \chainin (m-1-5); } { [start chain] \chainin (m-2-1); \chainin (m-2-2); \chainin (m-2-3) [join={node[above,labeled] {\varphi'}}]; \chainin (m-2-4) [join={node[above,labeled] {\psi'}}]; \chainin (m-2-5); }
[ 0.005359966307878494, 0.0010019047185778618, -0.012712597846984863, 0.011026894673705101, -0.00860155001282692, 0.011780163273215294, 0.007424432784318924, -0.0023014340549707413, -0.019796978682279587, -0.01854478009045124, -0.006576893385499716, -0.0016680439002811909, -0.00795772206038236...
[ 0.16955037415027618, 0.1777552217245102, 0.22418922185897827, -0.1620761901140213, -0.3523639738559723, 0.3304829001426697, 0.05125156790018082, -0.3961769640445709, -0.33326131105422974, -0.5242214202880859, 0.14609986543655396, 0.04774068668484688, -0.37286412715911865, 0.003623420838266...
I have a Motorola XOOM and it wont let me type on signs IP addresses or let me name maps at all please find a fix
[ 0.048150211572647095, -0.01574357971549034, -0.04569154977798462, -0.045282892882823944, -0.03350364416837692, 0.050932057201862335, 0.01658838614821434, 0.06819701939821243, -0.01674525812268257, 0.039975401014089584, -0.02183234691619873, 0.030591169372200966, -0.006818755064159632, 0.04...
[ -0.12585866451263428, 0.5874646306037903, 0.21778935194015503, 0.3541909158229828, -0.03982412815093994, 0.09376819431781769, 0.7817611694335938, 0.21379618346691132, -0.0006792528438381851, -0.5607947707176208, 0.08318386226892471, 0.16110718250274658, -0.37363117933273315, -0.18005645275...
I have this bash script: for opt in string1 string2 string3 ... string99 do somestuff It works, but I would like to replace the explicit listing of my strings with a file which actually contains all the strings; something like this: strings=loadFromFile for opt in $strings do somestuff How should I do this?
[ 0.01332928892225027, 0.03403078019618988, -0.0009539726888760924, -0.002629448426887393, 0.005345991346985102, -0.03293155878782272, 0.008770416490733624, -0.007058205548673868, -0.03134394437074661, 0.0022315254900604486, -0.018294675275683403, -0.00010785215272335336, 0.008925598114728928,...
[ 0.4420981705188751, 0.05874525383114815, 0.12278669327497482, -0.20910927653312683, -0.06591304391622543, -0.36844420433044434, 0.35159215331077576, -0.26009127497673035, -0.06877856701612473, -0.4558230936527252, 0.3397601544857025, 0.7724034190177917, -0.31218448281288147, -0.16125287115...
I have a classifier that I want to use on short documents and I need to improve quality of the corpus used for training. I have 250,000 documents and manual inspection is expensive. The collection of short documents has been pre-labelled by a collection of rules as good or bad. The accuracy of this pre-processing step has been assessed by sampling to be around 98% or higher. When I train the classifier on 80% of the corpus and test its accuracy on the remaining 20%, it has a 98% accuracy. What I am wondering is this: if I filter the corpus and remove the documents for which there is disagreement and retrain and repeat, can I expect to arrive at a corpus for which the classifier agrees with the original labeling process within some small margin of error (less than 1%)? EDIT: After finding a corpus that appears to be clean, I would manually inspect the documents that were rejected. I expect there would be around 5000 (2% of 250,000) Thanks!
[ 0.0031982616055756807, 0.0055292085744440556, -0.012046709656715393, 0.023968834429979324, -0.020925186574459076, 0.016547277569770813, 0.009374981746077538, -0.015542509965598583, -0.012909958139061928, 0.01664760895073414, -0.008900023065507412, 0.018252359703183174, -0.014492826536297798,...
[ 0.27466636896133423, 0.5191807150840759, -0.0031276275403797626, 0.0713164433836937, -0.03522530570626259, 0.2546769380569458, 0.6009770035743713, -0.28365036845207214, 0.04114847257733345, -0.6867583990097046, -0.09987035393714905, 0.31477221846580505, -0.12788702547550201, 0.182607173919...
I just read from this question that the OP was > using Mapscript and C# to generate and display a Mapfile I was wondering if it is really possible to use MapScript to **generate** a Mapfile? And if so, is there an example of doing so with PHP MapScript? P.S. I am new to MapServer, and I feel that manually writing every mapfile's is a pain even if I learn all its grammar. It would be nice if a bunch of mapfiles can be generated from a program by varying certain parameters. If MapScript isn't the solution, is there any other way to generate mapfiles programmatically? P.S.2. This post asks a similar question about doing it in python, but there seems to be no solution in the answers. Thanks in advance, --- EDIT --- Just to clarify, I am interested in writing/serialization of Mapscript objects into a mapfile (in PHP), not writing a mapfile verbatim via a programming languages such as PHP or C++. From the SO question quoted, this seems possible in C#. I was asking whether there are specific examples of actually doing that (preferably in PHP).
[ -0.014709076844155788, 0.012035288847982883, 0.0004393074195832014, 0.01833334006369114, 0.00674099288880825, -0.004430505447089672, 0.005745135713368654, 0.007518574129790068, -0.029612470418214798, -0.018107376992702484, 0.0031039465684443712, 0.020702913403511047, 0.01175733469426632, 0...
[ 0.3468570411205292, 0.25728851556777954, 0.3846065104007721, 0.0665677860379219, -0.3000691831111908, -0.13598212599754333, -0.03227870538830757, -0.0009496124112047255, -0.3144185245037079, -0.7233161330223083, 0.34990665316581726, 0.6872696280479431, -0.16417016088962555, -0.202703520655...
Hoping this is an easy one... I want to build a web/mobile app for a college project which allows the user to enter a start point and a distance to travel e.g. distance to travel = 6km and the site would then plot a 6km circular route which brings the user back to where they started without (if possible) backtracking on any of the route already travelled. I've been looking at Open Street Maps and Open Layers or Leaflet.js. I'm not sure if this explains it very well but I suppose the app might work by calculating a 3km "end" point with a 1.5km waypoint on the east and a 1.5km waypoint on the west. Prefer to use open source technology. I'm new to this so appreciate advice. Thanks!
[ -0.01585495099425316, 0.01499894168227911, 0.00041783449705690145, 0.00028253253549337387, -0.006002228241413832, -0.009239450097084045, 0.008236374706029892, 0.025657974183559418, -0.010162854567170143, -0.013075011782348156, -0.007189638912677765, 0.0025514913722872734, 0.01585950329899788...
[ 0.45056840777397156, -0.13388319313526154, 0.6055426001548767, 0.21445368230342865, 0.04515005275607109, 0.09601188451051712, 0.36907967925071716, -0.01609135791659355, -0.06021815538406372, -0.5339236259460449, 0.44217342138290405, 0.1004076674580574, -0.051578376442193985, -0.25472819805...
I konw that mass affects weight (force), so how does that relate to speed? F=ma. so how does all this affect speed?
[ -0.02256777137517929, 0.017068009823560715, -0.01623295247554779, 0.015061145648360252, 0.021893959492444992, -0.0039002709090709686, 0.011253707110881805, -0.028967760503292084, -0.03321678936481476, -0.00562897277995944, 0.011485150083899498, 0.040878862142562866, -0.03570741415023804, 0...
[ 0.23089385032653809, -0.044148631393909454, 0.2797321677207947, 0.09264250844717026, -0.14239515364170074, 0.40779611468315125, 0.01591886393725872, -0.18403029441833496, -0.5585138201713562, -0.3864747881889343, 0.17803701758384705, 0.3886268436908722, 0.2582806348800659, 0.11800853163003...
I cannot edit an ArcSDE database from multiple computers on a LAN network using a custom ArcGIS Engine application, I have tried both versioned and non- versioned feature classes, but they cannot be edited from more than one computer. Inside the custom ArcGIS Engine application, for every single operation (Create, Remove), I open and close an edit session. Any advice or experiences are welcome as why I may be unable to edit an ArcSDE database from multiple computers on a LAN network using a custom ArcGIS Engine application? I tried to edit the same feature classe from 2 computers in the same time , but when i closed the edit session in the second computer, it gimme this error `Your edits could not be saved because the target version has been updated since you started editing. These edits have been merged into your edit session. Review the changes and save again.` i'm editing feature classes like in this example : IWorkspaceEdit workspaceEdit = (IWorkspaceEdit)DB_Access.get_obj().ws; workspaceEdit.StartEditing(true); workspaceEdit.StartEditOperation(); IFeatureClass fc = DB_Access.get_obj().ws.OpenFeatureClass("points"); IFeature f = fc.CreateFeature(); f.set_Value(fc.Fields.FindField("name"), "name_of_point"); f.Shape = geometry; f.Store(); workspaceEdit.StopEditOperation(); workspaceEdit.StopEditing(true);
[ 0.011221347376704216, 0.009198448620736599, -0.04463117569684982, 0.028815539553761482, -0.02186116948723793, -0.033502668142318726, 0.011916056275367737, 0.024377938359975815, -0.02624361962080002, -0.01935630664229393, -0.004374191164970398, 0.03306567668914795, -0.012815344147384167, 0....
[ 0.5097589492797852, 0.17062565684318542, -0.01902160607278347, 0.11756160110235214, -0.4700039029121399, -0.17712973058223724, 0.38213658332824707, -0.0804450735449791, -0.10363907366991043, -0.6491648554801941, 0.07010293006896973, 0.5670496821403503, -0.2557573914527893, 0.15991985797882...
In this video, Leonard Susskind does a good job trying to explain succinctly the Higgs field and exactly how it gives elementary fermions mass, except for one point he seems to skip a few things. The weak hypercharge is endearingly called 'zilch' ( _0:46:00_ ). He goes on to explain how (e.g.) an accelerating electron (or other zilch-charged particle) can emit a $Z$ boson. He then postulates ( _0:48:16_ ) the existence of a field (none other than a 'ziggs boson condensate'), which (in a similar manner to a uniform electric field giving electric dipoles a potential energy - and therefore an extra mass - that depends on their orientation) gives fermions mass by the Dirac mechanism (explained earlier in the video). He adds how the ('zilch'-less) $Z$ boson also interacts with the 'ziggs' field to intermittently acquire 'zilch' (and becoming itself a 'ziggs'), and lose 'zilch', thereby acquiring mass as well. At that point, he jumps to the Higgs boson without apparently any connection to the mechanism just explained (he employs different names for the fields). He explains that the Higgs boson is a particular excitation mode of the Higgs field, but what is then the (general) 'ziggs' boson? Are Higgs and ziggs actually the same? The $Z$ boson has been known experimentally for decades, what about its weakly hypercharged product? Does the Higgs boson actually have nothing to do with the Higgs phenomenon, being merely a consequence of the theory that was begging to be discovered experimentally?
[ 0.003416936844587326, 0.002996018622070551, 0.0015267906710505486, 0.004425418563187122, -0.03271888941526413, -0.015463043004274368, 0.007988372817635536, -0.02130449190735817, -0.010805122554302216, -0.0021877288818359375, -0.003180817235261202, 0.012887032702565193, 0.0007201542612165213,...
[ 0.32210418581962585, -0.0041908021084964275, 0.4515973925590515, 0.31325918436050415, 0.057620055973529816, -0.35482823848724365, -0.00393639924004674, 0.04893578961491585, -0.16272202134132385, -0.08878359943628311, -0.23644979298114777, 0.04720568656921387, -0.5406789183616638, 0.6534714...
How does a mobile phone vibrate without any external force? By Newton's law, any body can't move without any external force
[ 0.003507083747535944, 0.021097492426633835, -0.011156030930578709, 0.019720720127224922, -0.042639877647161484, -0.026820523664355278, 0.025011945515871048, -0.06427299976348877, -0.03274785354733467, -0.06687108427286148, -0.05363587662577629, 0.07054103910923004, 0.02455434948205948, 0.0...
[ 0.3818742036819458, 0.13363371789455414, 0.45460617542266846, 0.12870994210243225, 0.16642363369464874, -0.164626345038414, 0.37623608112335205, -0.22673454880714417, -0.16807495057582855, 0.27962273359298706, 0.04346843063831329, 0.5325261950492859, -0.5301227569580078, -0.147193714976310...
First of all, Happy New Year! I am thinking about making my blog accessible and enjoyable on mobile devices (as many handsets as possible). I am using Wordpress 3.0.4. What I've found is couple of cool plugins like WPTouch (for touch enabled devices), The Wordpress Mobile Edition and the Wordpress Mobile Pack. I am wondering which one is the best to use if I can use only one. If I can use more or (even all these three) are they compatible? Does anyone have any experiences in this topic? If so I would much appreciate some advices or best practice links. Happy 2011, thanks in advance, Rob
[ 0.0037593687884509563, 0.004807949531823397, -0.01076221838593483, 0.012955093756318092, -0.021541614085435867, 0.014059750363230705, 0.00714527815580368, 0.018936268985271454, -0.014647334814071655, -0.03413917124271393, 0.0008483272977173328, 0.0004731027875095606, 0.014811975881457329, ...
[ 0.2869298756122589, 0.3146805763244629, 0.46138080954551697, 0.09534143656492233, -0.22824053466320038, -0.3776812255382538, -0.0621502622961998, 0.6694183945655823, -0.3909384310245514, -0.568204402923584, -0.06598962843418121, 0.611039936542511, 0.07224051654338837, 0.04744578152894974, ...
I have done some interpolation in Geostatistical Analyst in ArcGIS and I got the Standard Deviation (SD) a bit different from the SD that I calculated in Excel Spreadsheet. What algorithm does each package use? Why are the results different?
[ -0.032590046525001526, 0.01888151466846466, -0.021221298724412918, 0.03756621852517128, -0.006863203831017017, 0.019370602443814278, 0.015008565038442612, 0.006829338148236275, -0.02656334824860096, -0.0699172243475914, 0.018433397635817528, 0.015448296442627907, -0.0200002733618021, 0.004...
[ 0.43379610776901245, -0.33088189363479614, 0.1268358826637268, 0.2052927315235138, -0.1542655974626541, 0.018477099016308784, 0.20974737405776978, -0.39634019136428833, -0.05804534628987312, -0.4845932424068451, 0.339997798204422, 0.13162554800510406, 0.015745677053928375, -0.1306721717119...
I was drawn to the phrase, “Seeing right through them“appearing in the New York Times (October 5) article written by Daniel Goleman under the title, “Rich people just care less.” It begins with the following sentence: > “Turning a blind eye. Giving someone the cold shoulder. Looking down on > people. **Seeing right through them**. These metaphors for condescending or > dismissive behavior are more than just descriptive. They suggest -- the > social distance between those with greater power and those with less — a > distance that goes beyond the realm of interpersonal interactions and may > exacerbate the soaring inequality in the United States. http://opinionator.blogs.nytimes.com/2013/10/05/rich-people-just-care-less/?hp When we were listening to teacher’ words in class, or a boss's instructions in office, didn’t we see right through him or her in order to be attentive? Is “to see right through people” perceived as a condescending / dismissive or rude behavior? I happened to find the same “See right through them” phrase in the English version of the teaching of Morihei Ueshiba (1883-1969), the founder of Aikido – martial art for self defense derived from Jujutsu. He says: > “If your heart is large enough to envelope your adversaries, you can **see > right through them** and avoid their attacks. And once you envelope them, > you will be able to guide them along the path indicated to you by heaven and > earth.” It appears to me Goleman uses “Seeing right through people” in “condescending and dismissive” manner, and Ueshiba uses it in different, rather positive way. What implications does the expression “Seeing right through them” have?
[ 0.006946551147848368, 0.009930778294801712, -0.0023546202573925257, 0.004885888658463955, -0.01013108715415001, -0.033585429191589355, 0.00904313288629055, -0.012770643457770348, -0.010184317827224731, 0.0086602121591568, -0.02156083658337593, 0.010256689041852951, 0.0029817933682352304, 0...
[ 0.7146502733230591, 0.34432101249694824, 0.18821482360363007, 0.014036082662642002, -0.5302413702011108, 0.19252246618270874, 0.07345348596572876, 0.3202138841152191, -0.1633588671684265, -0.0713096410036087, -0.1191469356417656, 0.20008043944835663, -0.24072891473770142, 0.227090954780578...
I'm writing 'cause I would like to know if there's an automatic way (like something to put in the preamble) to make the subscript behave as double subscript: for example if I write something_a it should behave as something_{_a}
[ 0.005648583639413118, 0.0010152601171284914, -0.0038256095722317696, 0.024971595034003258, -0.013563713990151882, 0.015196488238871098, 0.01162020955234766, 0.03909943625330925, -0.024386776611208916, 0.005117898341268301, -0.015231338329613209, 0.0011981133138760924, -0.010946530848741531, ...
[ 0.4653284251689911, 0.5452885031700134, 0.04775289073586464, -0.41523051261901855, -0.312658429145813, 0.24910160899162292, 0.31574538350105286, 0.3994177579879761, -0.25151199102401733, -0.6759692430496216, 0.02230077236890793, 0.058657482266426086, -0.3767959475517273, -0.118164308369159...
Is there any significance to the Blue Hallway that appears sometimes before the 'two doors' room at the start? And has anyone else seen this hallway and found out that significance?
[ -0.024155639111995697, 0.0172593891620636, -0.006158705800771713, 0.010633193887770176, -0.053929757326841354, 0.029957765713334084, 0.012181432917714119, 0.019803861156105995, -0.03335433080792427, 0.002981670433655381, -0.031384848058223724, 0.025792492553591728, 0.004932112991809845, 0....
[ 0.4011065661907196, 0.13777334988117218, -0.2988875210285187, -0.02168717235326767, 0.07707900553941727, -0.1031874343752861, 0.2720739245414734, 0.1821649819612503, -0.635069727897644, -0.4104638695716858, 0.17639455199241638, -0.05643612891435623, 0.11039505153894424, 0.46632564067840576...
It is a theorem that any arbitrary unitary transformation in SU(2) can be factored into the following form: $ O = U_X(\theta) U_Y(\phi) U_X(\delta) $ Where $U_X$ is a Bloch sphere rotation. I believe it is possible to fix one of these angles, leaving only two angles as degrees of freedom. So, I am saying that, given that X and Y might be different, and the fixed angle is some number like $\frac{\pi}{2}$ any unitary can be written in a form something like: $ O = U_X(\theta) U_Y(\frac{\pi}{2}) U_X(\delta) $ Does this sound right? Perhaps I am working on the assumption that, if a pure input state is fixed, then you only need two degrees of freedom to map that input state to any other pure output state. This thinking comes from the fact that the pure states are all on the surface of the Bloch sphere. Here is a quote from Wikipedia: > Since polarization states are defined by two degrees of freedom, for example > azimuth angle and ellipticity angle of the polarization state, such a > polarization controller needs two degrees of freedom. The same holds for the > task of transforming an arbitrary polarization into a fixed, known one.
[ -0.015476470813155174, 0.020248452201485634, -0.02110922336578369, 0.006237152963876724, -0.00718872994184494, -0.00865976233035326, 0.008316967636346817, -0.0016393489204347134, -0.011874886229634285, -0.0008187558269128203, -0.014123484492301941, 0.0044310009106993675, -0.03242193162441254...
[ 0.349027544260025, -0.15717239677906036, 0.6346330046653748, -0.260529100894928, -0.04119476303458214, 0.053664032369852066, -0.10891496390104294, -0.5669636726379395, -0.026114722713828087, -0.29202577471733093, -0.3122371733188629, 0.34826529026031494, -0.3096792697906494, 0.339334636926...
I currently have a test class with a `setUp` method for creating a temp file and a `tearDown` method for deleting a temp file. However, only some of the the `test_` methods in that class require this functionality. Is possible to either ignore those 2 methods for the test cases which do not require it or to only call those 2 methods for test cases which do require it? It feels a bit 'wrong' to call these 2 methods for every test even if the temp file is not getting used in each one.
[ 0.017389696091413498, 0.02746439538896084, -0.00617696437984705, 0.0027353945188224316, 0.005141907371580601, 0.04077732935547829, 0.00865820050239563, 0.002847779542207718, -0.01567823812365532, -0.00918904971331358, -0.012475571595132351, 0.006354728247970343, -0.0010454021394252777, 0.0...
[ 0.4594177305698395, 0.08781563490629196, 0.08058831840753555, 0.11750798672437668, -0.18270428478717804, -0.3071500062942505, 0.30353760719299316, -0.3034040629863739, -0.038417987525463104, -0.800109326839447, -0.016846099868416786, 0.41684630513191223, -0.12860842049121857, 0.35548773407...
Could you please advise whether studentized residuals are meaningful when computed on a robust linear regression model using an M-estimator? I'd like to use it to detect outliers by doing something like this: rfit = rlm(y~x, data=d) pt(rstudent(rfit), df=nrow(d)-3) Is this reasonable? I'd be quite happy with a rather crude measure and I'd rather err on the conservative side. I'm also wondering whether I should run some kind of diagnostic of the general goodness of fit on this robust model before doing this.
[ 0.013744564726948738, 0.013763867318630219, -0.016325972974300385, 0.0051630279049277306, -0.02638547122478485, -0.003198597114533186, 0.006885930895805359, -0.004116096533834934, -0.008809189312160015, -0.013858120888471603, 0.004894989542663097, 0.005110609345138073, 0.010762946680188179, ...
[ 0.3551765978336334, -0.17585738003253937, 0.26577940583229065, 0.1511564701795578, 0.06987761706113815, -0.04281873255968094, 0.0006292274338193238, -0.42006292939186096, -0.058066632598638535, -0.22046443819999695, 0.2425192892551422, 0.4214157164096832, 0.12448930740356445, 0.30872955918...
I am working on my thesis analysis, and I have some error data that's right- skewed. I log-transformed it and ran glm on it (gaussian, identity in R) weighted by sample size, and my data is still over-dispersed. I keep seeing comments about using the quasi family, but I'm not sure I completely understand what I'm doing. I tried it out, and my question is, can I use quasi with the log-transformed data and still set my variance equal to my mean (or mean squared), or is that like a double variance stabilizer? If I run the data with quasi, link identity, variance = mean in R, I get approximately the same results, except the data appears to by neither under nor over-dispersed. Thanks!
[ 0.017699293792247772, 0.00981490034610033, -0.002268638461828232, 0.0039544496685266495, -0.0017543064896017313, 0.017171984538435936, 0.009013578295707703, -0.005795200355350971, -0.013380609452724457, -0.04128117114305496, -0.003643548581749201, 0.009443764574825764, -0.010589849203824997,...
[ 0.3056159019470215, -0.24588166177272797, 0.40551427006721497, 0.06105557084083557, -0.3348492681980133, -0.11396020650863647, 0.03397611156105995, -0.2734001576900482, -0.1185726523399353, -0.07324955612421036, 0.13719876110553741, 0.16305847465991974, -0.19315201044082642, 0.383197098970...
I just used the osgeo4w exe to install QGIS and all the other software included in the basic setup. When I open QGIS, I get the following error: Plugin error Unable to load GdalTools plugin. The required "osgeo [python-gdal]" module is missing. Install it and try again. I opened the Python.exe in the osgeo4w bin folder and am able to import gdal just fine. Any idea what's wrong here?
[ -0.029715092852711678, -0.009350716136395931, -0.004385457839816809, 0.02263640984892845, 0.005178360268473625, 0.030826084315776825, 0.012218328192830086, 0.024828752502799034, -0.02100254036486149, -0.02473740465939045, 0.0010857594897970557, 0.012648557312786579, -0.012047534808516502, ...
[ 0.27910953760147095, 0.05183872580528259, 0.16225875914096832, -0.2632969319820404, -0.19353510439395905, -0.2728619873523712, 0.6530055403709412, 0.18341952562332153, -0.19887468218803406, -0.6773771047592163, -0.12417753040790558, 0.9198009371757507, -0.47235044836997986, -0.047216836363...
According to the RHEL 6 documentation about the Relatime Drive Access Optimization: > The kernel used in Red Hat Enterprise Linux 6 supports another alternative — > relatime. Relatime maintains atime data, but not for each time that a file > is accessed. With this option enabled, atime data is written to the disk > only if the file has been modified since the atime data was last updated > (mtime), or if the file was last accessed more than a certain length of time > ago **(by default, one day)**. Emphasis mine -- this documentation suggests that the access time updating algorithm can be changed to use a different interval than 24 hours, but I have no idea how to actually accomplish this. There apparently used to be a sysctl option named `fs.relatime_interval` as mentioned here, but I cannot find it in my Centos 6.5 kernel (2.6.32-431.) Is it possible to make this value something arbitrary like 12 hours, 6 hours, etc?
[ 0.0021202419884502888, 0.02600645273923874, -0.01877644844353199, 0.0069314842112362385, 0.0382484570145607, -0.02998434193432331, 0.009565267711877823, -0.0015686386032029986, -0.011311243288218975, 0.013649309985339642, -0.015277085825800896, 0.005751527845859528, 0.009392057545483112, 0...
[ -0.053794801235198975, -0.12484315782785416, 0.6525957584381104, 0.09459161013364792, 0.022368423640727997, 0.011668093502521515, 0.2853470742702484, -0.3705471456050873, -0.12214379757642746, -0.6575419902801514, -0.5496595501899719, 0.8384335041046143, -0.22566884756088257, 0.19614119827...
I had a issue with a custom taxonomy and a custom post type using the same url and that is now fixed but it has created another bug. not showing the posts tagged with the custom taxonomy. my url i am using for my post type is /news and then i have a taxonomy which i need to have /news/cars I can get a all the posts for the custom post type to output on the index template. I also have the single taxonomy page working but my query doesnt show the posts that are tagged with the custom taxonomy for example all posts that are cars. here is my code i have so far : function register_custom_post_types() { $args = array( 'labels' => array( 'name' => __( 'News' ), 'singular_name' => __( 'News' ) ), 'menu_position' => 5, 'public' => true, 'query_var' => 'news', 'has_archive' => 'news', 'hierarchical' => true, 'taxonomies' => array('news_categories'), 'publicly_queryable' => true, 'capability_type' => 'post', 'rewrite' => array('with_front' => false, 'slug' => 'news/%news_categories%'), 'supports' => array('title', 'editor', 'thumbnail', 'revisions', 'page-attributes') ); register_post_type( 'news', $args); } add_action( 'init', 'register_custom_post_types' ); add_action( 'init', 'create_my_taxonomies', 0 ); function create_my_taxonomies() { register_taxonomy( 'news_categories', 'news', array( 'labels' => array( 'name' => 'News Categories', 'add_new_item' => 'Add New Category', 'new_item_name' => "New Category" ), 'query_var' => true, 'show_ui' => true, 'show_tagcloud' => false, 'hierarchical' => true, 'rewrite' => array( 'with_front' => false, 'slug' => 'news' ) ) ); } add_filter('post_type_link', 'news_term_permalink', 10, 4); function news_term_permalink($post_link, $post, $leavename, $sample) { if ( false !== strpos( $post_link, '%news_categories%' ) ) { $glossary_letter = get_the_terms( $post->ID, 'news_categories' ); $post_link = str_replace( '%news_categories%', array_pop( $glossary_letter )->slug, $post_link ); } return $post_link;
[ -0.0003842882579192519, 0.0092411944642663, 0.002976099494844675, 0.020590834319591522, 0.011248691938817501, 0.006224364507943392, 0.008312074467539787, 0.02685561776161194, -0.014963999390602112, -0.0216409619897604, -0.009980895556509495, 0.013616885989904404, -0.003336055437102914, 0.0...
[ 0.939257800579071, 0.5797505378723145, 0.22875282168388367, 0.3319544494152069, -0.11329063773155212, 0.08258939534425735, 0.0837482139468193, 0.10906460136175156, -0.15635505318641663, -0.5930010080337524, 0.5657707452774048, 0.6529099345207214, -0.08045805245637894, 0.26201242208480835, ...
I remember travelling in the Frosted Hills and encountering a group of injured Zeraphis and their ambassador Lord Taldimut. He offered two quests, which I completed in due time, the Zeraphi Envoy, and the Lost Key. After I finished the Lost Key, he said: > This is it! The Golden Gear! I thank you on behalf of all the Zeraphi > people. This might just have tipped the battle in our favor! I can only > offer you this as payment, but if you are ever in Zeryphesh, come find me > and I will grant you a proper reward! I've already reached Zeryphesh, and I still haven't found him. Does he actually appear as an NPC and give a reward? Does he only come after certain missions or during a certain time? Or does he not show up? (I'm currently on the quest where you have to complete favors for the Djinn)
[ 0.021938461810350418, 0.014232633635401726, 0.0007288753404282033, -0.0064022536389529705, -0.02224382758140564, 0.026984650641679764, 0.007064192555844784, -0.008158969692885876, -0.01933467760682106, -0.008007733151316643, -0.006156841292977333, 0.011904057115316391, -0.021986771374940872,...
[ -0.025192001834511757, 0.3288891017436981, -0.18437784910202026, 0.46961086988449097, -0.2611096501350403, 0.12878216803073883, 0.1544051319360733, 0.16012927889823914, 0.06206929683685303, -0.6072829365730286, -0.613073468208313, -0.054824553430080414, 0.18266218900680542, 0.2746725380420...
Some years ago I used MuPAD as a free computer algebra system. However since 2005 it's not free anymore. So, are there any good free or even better open- source computer algebra systems for linux coming close to commercial products like mathematica or maple?
[ -0.0061728134751319885, 0.010345770046114922, -0.014347312040627003, -0.007501174695789814, -0.0008920810068957508, -0.019075319170951843, 0.01153924036771059, -0.007780733983963728, -0.019083188846707344, -0.02191876620054245, -0.004757423419505358, 0.01567198894917965, 0.00299534248188138,...
[ 0.23253563046455383, 0.26459622383117676, 0.08366722613573074, 0.2797594368457794, -0.0843418538570404, 0.0811418890953064, -0.11483413726091385, 0.6380752921104431, -0.3913766145706177, -0.5248416066169739, 0.4934066832065582, 0.40270230174064636, 0.18643896281719208, 0.27562135457992554,...
I'm using the manage_users_columns to display a custom field I create in the usermeta database called company. My code is as follows: function mysite_column_company( $defaults ) { $defaults['mysite-usercolumn-company'] = __('Company', 'user-column'); return $defaults; } function mysite_custom_column_company($value, $column_name, $id) { if( $column_name == 'mysite-usercolumn-company' ) { return get_usermeta($id, 'company'); } } add_action('manage_users_custom_column', 'mysite_custom_column_company', 15, 3); add_filter('manage_users_columns', 'mysite_column_company', 15, 1); I'd like to get to display two other custom fields, but can't figure out the proper function. Duplicating this one doesn't work and I've not had any luck adding new functions within this one for each unique column, mostly because I can't seem to figure out how to properly define the $defaults variable. Any suggestions?
[ 0.027676530182361603, 0.011423119343817234, 0.00036627406370826066, 0.0034239296801388264, -0.02559996396303177, -0.0042847986333072186, 0.007155191153287888, -0.018179427832365036, -0.006939325015991926, 0.010455939918756485, -0.011662298813462257, 0.006610729265958071, 0.021955234929919243...
[ 0.2724283039569855, 0.4171542525291443, 0.6216080188751221, -0.5592100620269775, 0.3024097681045532, 0.339231938123703, -0.042259085923433304, -0.12635838985443115, 0.013527415692806244, -0.3626822233200073, 0.1673635095357895, 0.29674017429351807, -0.23283319175243378, 0.23423315584659576...
I've now completed the game, but I never found keys to open the 3 locked doors on Level 12: The Prison. They all have keyholes, but I've found no keys for them. ![enter image description here](http://i.stack.imgur.com/pr73O.jpg) How can I open these? What's behind them?
[ 0.0003356185043230653, 0.00044626553426496685, -0.003375774947926402, 0.006756362039595842, -0.004920409992337227, -0.006779195740818977, 0.00613490492105484, -0.0067778355441987514, -0.021580711007118225, 0.004613182973116636, -0.014947870746254921, 0.007083564531058073, -0.0217693746089935...
[ 0.23267871141433716, 0.10395006090402603, 0.34943723678588867, 0.1321081817150116, 0.49706488847732544, -0.20838546752929688, 0.7101746201515198, -0.4889551103115082, -0.5265052318572998, -0.47371476888656616, -0.17130064964294434, 0.2827169895172119, 0.15131230652332306, 0.113922230899333...
I have profiled my blog with the P3 Plugin and it pointed out that Yoast SEO slows down the site a bit. I then found out that the biggest performance penalty is paid when you "force rewrite" the titles[1], which I am doing. Now I do have quite an aggressive caching turned on with WP Super Cache. Does it still matter if I force-rewrite the titles then? Or in my case it doesn't change a thing since pages are pre-generated and served as static content? [1] http://wordpress.org/support/topic/plugin-wordpress-seo-by-yoast- performance-issues-does-it-do-any-good-now
[ -0.003514633048325777, 0.01259971410036087, -0.005455242935568094, 0.02076372690498829, -0.00010711420327425003, 0.003396580694243312, 0.008036606013774872, -0.006698525510728359, -0.009644851088523865, 0.006524552591145039, -0.008265702053904533, 0.010558192618191242, -0.007935561239719391,...
[ 0.36796656250953674, 0.20629839599132538, 0.5374047160148621, 0.04537684842944145, -0.1904740333557129, -0.25271373987197876, 0.2955716848373413, -0.15267235040664673, -0.3006318509578705, -0.3775523006916046, -0.22924679517745972, 0.6274581551551819, -0.060645267367362976, 0.1032385751605...
I want to write a server side application which allows several users to exchange files (not above 3MB) in the following way: user A connects to (server) S. User B connects to S. User C connects S. User A sends a file. Users B and C "see" that a file was/is being uploaded and start downloading it. My main concern is responsiveness. I want the file to arrive at B and C not long after A started uploading. I thought of doing it using HTTP: user A sends raw bytes to the server and the server saves the file locally. At this point users B and C see that a file was uploaded and start downloading it. My question: is HTTP the best way to go here? Or should I write an app that uses TCP sockets, and write my own protocol? If I go with sockets. How can I estimate the RAM requirements of the VPS so that I can allow N file-transfer connections simultaneously?
[ -0.018404042348265648, 0.007731840945780277, -0.0025321533903479576, 0.00004601897671818733, 0.027326125651597977, -0.009559123776853085, 0.005785780027508736, 0.0017724970821291208, -0.013102615252137184, -0.03998865932226181, -0.010240213945508003, 0.012251442298293114, 0.01832547783851623...
[ 0.20824109017848969, 0.15304304659366608, 0.6745454668998718, 0.008660322055220604, -0.19056552648544312, -0.007311901077628136, 0.03136102482676506, -0.15837855637073517, -0.224928617477417, -0.8363375663757324, -0.04728949815034866, 0.45312735438346863, -0.15152688324451447, 0.2494524568...
I have been using moderncv and I was thinking of adding my _stackoverflow_ and _github_ profiles to my resume, unfortunately there doesn't seem any macros for that (is there?) I wasn't using `\homepage{..}` so I added one of the profiles there. If I add another `\homepage{..}`, the first one disappears so I tried adding the other one with `\extrainfo{..}`. It looks ok this way but the problem is first link is clickable while the second one isn't. So my question is: * how can I add two `\homepage{..}` 's * and/or how can I make `\extrainfo{..}` clickable * and/or how hard it is to maybe create a macro specific to online profiles like `\profilepage{..}` (I don't have any experience with macros).
[ 0.02612319216132164, 0.007958678528666496, 0.006176288239657879, 0.013503668829798698, 0.02516922727227211, -0.00009184516966342926, 0.007157342042773962, -0.0005994368111714721, -0.01140507124364376, -0.0032962472178041935, -0.0014378102496266365, 0.007310011424124241, -0.004283898044377565...
[ 0.2738484740257263, -0.0147981783375144, 0.6375988721847534, -0.10405246168375015, 0.03555623069405556, -0.10510394722223282, 0.09622115641832352, 0.1204892098903656, -0.30082857608795166, -0.7986034154891968, 0.14863340556621552, 0.7083829045295715, -0.2554011642932892, 0.1447341740131378...
Anyone knows a way of using shortcodes on terms description? This is possible?
[ 0.030243277549743652, 0.0162664745002985, -0.025093307718634605, 0.026813989505171776, -0.020829761400818825, -0.0023812337312847376, 0.01729441061615944, 0.009450670331716537, -0.03080444596707821, 0.0038985460996627808, -0.04409882798790932, 0.03685665875673294, 0.016257712617516518, 0.0...
[ 0.9017342925071716, 0.37282195687294006, 0.04652683064341545, 0.4103001356124878, 0.3198905289173126, -0.25768083333969116, 0.19420939683914185, 0.33669793605804443, -0.36919113993644714, -0.3849445879459381, 0.1190812736749649, 0.2215915322303772, 0.011329992674291134, 0.00118301226757466...
I have a home directory in a disk that has a lot of corrupted files. I am trying to save most files as possible. I am trying to create a backup of my home directory on an external drive using tar -cvf backup /home/myHome and also have tried rsync -vfh /home/myHome /media/backup Both methods stop when they found a corrupt file. Is there a way to force one or the other to continue copying and ignore bad files, copying everything that is possible to copy and show me a log at the end of the files that were not copied? Thanks.
[ 0.007947609759867191, 0.015230640769004822, -0.015020444989204407, 0.023396220058202744, -0.013634431175887585, -0.0039023421704769135, 0.009772513061761856, -0.00012830772902816534, -0.019327666610479355, -0.026842869818210602, -0.005028621293604374, 0.0026139728724956512, -0.00161981023848...
[ 0.5010650753974915, 0.3315851390361786, 0.0694398283958435, 0.4512411952018738, 0.12014440447092056, -0.0526546873152256, 0.3119641840457916, 0.3436547815799713, -0.2717881202697754, -0.5907564163208008, 0.1901009976863861, 0.7392879128456116, -0.04489852860569954, 0.36084824800491333, 0...
As I noticed in the documentation `VertexConnectivity` is defined as the following > The vertex connectivity of a graph $g$ is the smallest number of vertices > whose deletion from $g$ disconnects $g$. Now for a `CycleGraph` I expect one needs to delete at least **two vertices** to make the resulting graph disconnected. Please correct me if I am driven by a wrong intuition here. opt = {VertexLabels -> "Name", ImagePadding -> 10}; graph1 = CycleGraph[8, opt]; graph2 = UndirectedGraph[Graph[{1 -> 3, 1 -> 7, 1 -> 10, 2 -> 5, 2 -> 6, 3 -> 6, 3 -> 10,4 -> 7, 4 -> 8, 4 -> 9, 5 -> 9, 6 -> 10, 7 -> 8, 8 -> 9}],opt]; Row[{graph1, graph2}] ![enter image description here](http://i.stack.imgur.com/2oFzG.png) ## **Strange!** Now Mathematica is returning this _not so intuitive_ values for the `graph1` and `graph2`. VertexConnectivity /@ {graph1, graph2} > {1,1} How can one make any of these above graphs disconnected by deleting just **one vertex**? Can someone shade some light on this issue? `EdgeConnectivity` though gives understandable output for both the graphs.
[ -0.019112221896648407, 0.004669914022088051, -0.0005194335244596004, 0.01147364266216755, 0.003344207303598523, -0.02858944609761238, 0.008497032336890697, -0.003334803506731987, -0.011729961261153221, 0.00367959076538682, -0.00401582196354866, 0.010110102593898773, -0.02005792036652565, 0...
[ -0.21423201262950897, -0.2556336224079132, 0.40045034885406494, -0.03812340274453163, -0.2104138433933258, -0.01666605845093727, 0.1413394957780838, -0.15611214935779572, -0.44417133927345276, -0.6828302145004272, -0.10513779520988464, 0.2340722382068634, -0.8611454367637634, 0.08968248218...
I'm using my university's LaTeX class package to write my thesis, but I'm running into an issue with line breaks in the title. I'm required to break my title manually in a certain spot, but the following: `\title{An AUV Simulator for Incorporating \\ Physical Feedback}` gives the following error after running `pdflatex`: ! Use of \@icentercr doesn't match its definition. <argument> \def l.59 \titlepage Instead of `\\`, I've also tried `\newline` and `\par`, but nothing seems to work (no errors, but no spacing either). I've also done what was suggested here, but it gives the same error. I'm not really sure how the class package does titles, and am generally have no experience with advanced formatting. Is there a way to accomplish a line break here? (Note: `\newline` works, but the first line will be off-center relative to the rest of the page.)
[ -0.010524828918278217, 0.0014389411080628633, -0.0038670338690280914, 0.01984010823071003, -0.0031593958847224712, 0.015257314778864384, 0.0077538397163152695, 0.01544689480215311, -0.011214610189199448, -0.00823131948709488, -0.019654590636491776, 0.011117193847894669, -0.004090993199497461...
[ -0.12231753021478653, 0.40561527013778687, 0.7064065933227539, -0.12881284952163696, 0.2315896600484848, -0.26070713996887207, 0.5376835465431213, -0.08528171479701996, 0.1679162234067917, -0.3559388518333435, -0.24202163517475128, 0.2619812488555908, 0.18490850925445557, 0.061713594943284...
We know that the Reynolds Number is $R_e=\frac{\rho vd}{\eta}$ ($\rho$ density, $v$ velocity, $d$ diameter and $\eta$ the viscosity of fluid). We also know that an ideal fluid has no viscosity which means that $\eta=0$ and what we get is $R_e=\frac{\rho vd}{0}$. Does this mean that the ideal fluid doesn't have a Reynolds Number? In my book it says that every fluid has it's own Reynolds Number. Can anybody help me? Thank you!
[ 0.013637734577059746, 0.0109828170388937, -0.010436462238430977, 0.0068797641433775425, -0.0017398321069777012, -0.0330292209982872, 0.00700928270816803, -0.009041030891239643, -0.0052758860401809216, -0.013282941654324532, -0.00336527731269598, 0.007954152300953865, -0.013660000637173653, ...
[ -0.1824563890695572, 0.031236302107572556, 1.0110580921173096, 0.14373664557933807, 0.06757141649723053, 0.23312829434871674, -0.14024756848812103, -0.20160509645938873, -0.14832180738449097, -0.5065569877624512, -0.17142130434513092, 0.7147458791732788, -0.21738991141319275, 0.54474359750...
How can I install a driver on any Linux? I have XG-762 Sagem adapter on USB with the ZyDAS 1211B chipset, and I don't have drivers for it. What should I do? PS: Despite it should be in my file system. It isn't...
[ 0.005407153628766537, 0.00275682401843369, -0.010666096583008766, 0.03598189353942871, -0.003040789859369397, -0.0010224219877272844, 0.011173604056239128, -0.021355055272579193, -0.029338251799345016, -0.043478723615407944, 0.0003046875062864274, -0.0009138888563029468, -0.03071190975606441...
[ 0.3003712296485901, 0.2842746675014496, 0.1652420610189438, 0.36499202251434326, 0.002447685459628701, -0.15470263361930847, 0.1588701456785202, 0.2394925206899643, 0.35674962401390076, -0.5224680304527283, 0.05609292536973953, 0.6647798418998718, -0.3170294165611267, 0.1464986801147461, ...
An XML sitemap generator plugin for WordPress puts the following strings in my _robots.txt_ file: User-agent: * Disallow: /wp-admin/ Disallow: /wp-includes/ But other WordPress blogs have lots of tags included in it. Also, my XML file's sitemap name looks like this: `sitemap.xml.gz`. How do I add other possible parameters in my _robots.txt_ file? Is the XML file `sitemap.xml.gz` name valid, and what is the difference between a `.xml` and `.gz` extension?
[ -0.011750966310501099, 0.011489031836390495, 0.005629810504615307, 0.03142665699124336, 0.006421325728297234, 0.015861839056015015, 0.009784223511815071, 0.012540534138679504, -0.021309738978743553, -0.007501392625272274, -0.013504687696695328, 0.005999715067446232, -0.002824189141392708, ...
[ 0.30432751774787903, 0.04358715936541557, 0.656009316444397, 0.09308690577745438, -0.2755919396877289, -0.09777092188596725, -0.11836288124322891, -0.024582013487815857, -0.008954284712672234, -0.5288936495780945, 0.042010072618722916, 0.5083090662956238, -0.3366289734840393, -0.0842137485...
Suppose you have an ext3 partition which was unfortunately formated as ext4 partition (and where now are some but not a lot new files on it). Is there any way to recover (some) files from the old ext3 partition?
[ 0.024164529517292976, 0.02725907228887081, -0.017866650596261024, 0.023618822917342186, -0.027563756331801414, 0.0007236459059640765, 0.007946613244712353, 0.0005342700169421732, -0.027327263727784157, -0.021904129534959793, -0.001228142762556672, 0.027800651267170906, 0.007203327491879463, ...
[ -0.09857127815485, -0.13425301015377045, -0.08239925652742386, 0.18780460953712463, 0.2229946404695511, 0.015025811269879341, 0.08144040405750275, 0.05699342489242554, -0.11792885512113571, -0.7634692788124084, -0.3134884834289551, 0.27843594551086426, -0.2676210403442383, 0.02963551133871...
I'm using the `algorithmic` environment for typesetting algorithm in my document, but I'd like to customize some aspect and fix problems. I'm using the `algpseudocode` package with `noend` option, but it produces an ugly vertical gap when the end lines are suppressed (see picture below between the sixth and the seventh lines). How can I fix this problem? Second, I'm using `function` command to produce two distinct procedures in the same `algorithm` environment, but I'd like to have only the function body to be line-numbered and the line count should restart from 1 for the Reduce function. Third and last, I've separated the two functions using the `\Statex` command, but I'd like to have a less vertical space width than a whole white line. This is my code. \begin{algorithm}[H] \caption{Calcola il grado di ogni vertice del grafo $G = (V,E)$} \label{degree} \begin{algorithmic}[1] \Function{\textsc{Map}$( \left \langle u;v \right \rangle )$}{} \State \textbf{emit} $\left \langle u;v \right \rangle$ \EndFunction \Statex \Function{\textsc{Reduce}$( \left \langle u;\Gamma(u) \right \rangle )$}{} \State $d(u) \gets 0$ \ForAll{$v \in \Gamma(u)$} \State $d(u) \gets d(u) + 1$ \EndFor \State \textbf{emit} $\left \langle u;d(u) \right \rangle$ \EndFunction \end{algorithmic} \end{algorithm} ![enter image description here](http://i.stack.imgur.com/69f3F.jpg) Thank you!
[ -0.001369974692352116, 0.009471716359257698, -0.006556923966854811, 0.017411157488822937, -0.026174617931246758, -0.0053107901476323605, 0.007616168819367886, 0.005680783651769161, -0.01269222330302, -0.008552569895982742, -0.016835296526551247, 0.009972316212952137, -0.0065288119949400425, ...
[ -0.15604864060878754, -0.006797534413635731, 0.38601142168045044, -0.06495289504528046, -0.12424331158399582, 0.14152617752552032, 0.3117402493953705, -0.003862963058054447, -0.1730196326971054, -0.7382919192314148, 0.03169931843876839, 0.47585809230804443, -0.37335869669914246, -0.1483851...
Given two samples of arrival rate of a process (they should end up something like a Poisson distribution): * One over a long period (something like several days) * Another from a shorter period time (maybe 30 minutes) How can I compare the shorter distribution to see if it varies significantly from the longer distribution? Ideally using an existing function in R? I believe a Q-Q Plot is showing this visually, but I am looking for a single number that would represent how far the shorter sample is from the longer sample. Also, perhaps better suited for another question. I am a bit curious about how I might say what amount of time would be a representative sample for the longer period of time. For reference the longer period plot looks something like the following, but it might very somewhat as I decide which attributes I might want to filter out: ![enter image description here](http://i.stack.imgur.com/ethnm.png)
[ 0.002659332938492298, 0.02417725697159767, -0.017311999574303627, 0.015523572452366352, -0.005515333730727434, -0.008079643361270428, 0.008208239451050758, 0.005123174749314785, -0.015069651417434216, -0.031643494963645935, -0.009265025146305561, 0.008986534550786018, 0.011665493249893188, ...
[ 0.29571714997291565, -0.665421187877655, 0.24901176989078522, 0.1759251058101654, -0.26594433188438416, 0.2584940195083618, -0.10896867513656616, -0.17152875661849976, -0.46960946917533875, -0.6704913377761841, 0.3724251687526703, -0.1350291669368744, 0.1332814246416092, 0.4774546623229980...
I have a problem to load a kml map using OpenLayers. I used this in this example: http://www.openlayers.org/dev/examples/sundials.html I put the following script to load the kml var wms = new OpenLayers.Layer.WMS('OpenLayers WMS','http://vmap0.tiles.osgeo.org/wms/vmap0', { layers: 'basic' }, {}); var gphy = new OpenLayers.Layer.Google("Google", { type: G_PHYSICAL_MAP }); var geomor = new OpenLayers.Layer.Vector("KML", { projection: map.displayProjection, strategies: [new OpenLayers.Strategy.Fixed()], protocol: new OpenLayers.Protocol.HTTP({ url: "/Shapes/Geomorfologia/est_temp2.kml", format: new OpenLayers.Format.KML({ extractStyles: true, extractAttributes: true }) }) }); map.addLayers([wms, gphy, geomor]); Only this kml not loading the folder where is the kml, this application within the Visual Studio folder. Why kml will not be changed. Alquem could tell me what's wrong.
[ 0.010651949793100357, 0.0012090951204299927, -0.007424470968544483, 0.02081003040075302, -0.024928051978349686, 0.006709768436849117, 0.008418131619691849, 0.00616737874224782, -0.01513120997697115, -0.01912478543817997, -0.006383041851222515, 0.010500960983335972, -0.01659560576081276, 0....
[ 0.134015291929245, 0.06658411026000977, 0.8551436066627502, 0.07232308387756348, -0.16692499816417694, 0.027437008917331696, 0.06671258807182312, -0.33362987637519836, -0.2841731309890747, -0.9900888800621033, -0.03033428080379963, 0.6681140065193176, -0.5289485454559326, 0.415650188922882...
As I understood, there is a coincidence of the weak coupling constant $g$ calculated in two different ways: 1) The muon lifetime $\tau_{\mu}$ is related to $g$ by the formula $(m_{\mu} c^2)^5 \tau_{\mu}= (8 \pi)^3 12 \hbar (\frac{m_W c^2}{g})^4$, where $m_{\mu}$ and $m_W$ are the masses of the muon and the W-boson. Putting in $m_W= 80.385 \frac{GeV}{c^2}$, $m_{\mu}= 105.65837 \frac{MeV}{c^2}$ and $\tau_{\mu} = 2.1969811 \ ms$ yields $g=0.65224$. 2) If for the Weinberg angle $\theta_W$ holds $\cos \theta_W = \frac{m_W}{m_Z}$, $\sin \theta_W = \frac{\sqrt{4 \pi \alpha}}{g}$, then with $1/\alpha = 137.035999$, and the Z boson mass $m_Z =91.1876 \frac{GeV}{c^2}$ it follows $g=0.64144$. The agreement is about $1.7$ percent. Is this the most precise (quantitative) test of electroweak unification, or are there other tests with a higher precision? (I do $not$ mean tests of QED such as the Lamb shift or the anomalous magnetic momentum of the electron)
[ -0.013478121720254421, 0.011858337558805943, -0.007348859682679176, 0.009373451583087444, 0.011124632321298122, -0.022379029542207718, 0.004995678551495075, -0.012510035187005997, -0.007266150787472725, 0.0053495075553655624, 0.003615159308537841, 0.010985622182488441, -0.01015415694564581, ...
[ -0.03327165171504021, -0.19703368842601776, 0.8096214532852173, -0.2948543429374695, 0.0566234327852726, 0.3376513123512268, -0.08675511181354523, -0.26005688309669495, -0.17337661981582642, -0.25172582268714905, -0.2882736623287201, 0.4834006130695343, -0.25944432616233826, 0.988018512725...
In particular:- * does a sunderer/S-AMS pilot get XP for providing a spawn point? * does an engineer get XP for resupplying ammo? * under what circumstances does a combat medic (or engineer) gain xp for healing (or repairing)?
[ 0.0051416135393083096, 0.025562012568116188, 0.001370945363305509, 0.019044654443860054, -0.023536087945103645, -0.021504273638129234, 0.012430774047970772, 0.03334765508770943, -0.0219785887748003, 0.007735135033726692, -0.004617841448634863, 0.031409867107868195, 0.009424478746950626, 0....
[ 0.2909584641456604, 0.3360026478767395, -0.02366713248193264, 0.43562525510787964, -0.1492585837841034, 0.15899404883384705, 0.15342038869857788, -0.6617623567581177, -0.2519204914569855, -0.34171178936958313, -0.43138429522514343, 0.4414038062095642, 0.30578556656837463, -0.22605632245540...
Sorry for the long question, but I want to be sure to define my problem clearly. For my thesis (in aerospace engineering, so not much programming experience) I want to develop a (prototype) mobile application which will facilitate the reporting of safety related events in African aviation. Although the content of my thesis focuses on the more technical and sociological aspects of safety (management) I decided to include some practical design for demonstrative purposes. Considering the criteria I will present below, my first guess is to develop a mobile app using PhoneGap or something similar. But life can't be that simple, so I'll need more than just this single framework. My question to you is, what a convenient and effective approach would be to satisfy the following description? Quite simply this is what I envision (in layman's terms): * A user (pilot, ground crew, etc.) experiences or notices a safety related event * The user fills out a form (Air Safety Report) about the event on his/her phone * Additional information (pictures/videos) are attached * Once a stable internet connection comes available (we're in Africa) the report/form is submitted to a server. Important considerations here are: * The limited availability of internet -> I need a native/self-contained solution/app so the reporter does not have wait with reporting, and which stores data and submits once possible. * I don't want - nor should - exclude potential users (for multiple reasons) -> The app should be available on as many platforms as possible (especially BlackBerry and Symbian are still very popular in Africa!). * I am doing this as a form of development aid and want it to be cheaply available -> open source preferably. Altogether, the following criteria can be defined: It should * be open-source * support multiple-platforms * allow file attachments (pictures mainly) * be self contained/native * support/allow web services * provide geolocation and date/time support (automate report whereabouts) * allow camera use for direct attachment (optional, but preferable) * identify an internet connection and submit the data once possible * store data until submission Besides PhoneGap (which doesn't support image library browsing? -> apparently is possible, see supporting documentation), I have considered the following frameworks: * Mobl (not tested for all major platforms, no date/time support?) * QuickConnect (does not support BlackBerry and Symbian) * Rhodes (meets my requirements as for as I can see) * Appspresso (doesn't support BB and Symbian, no date picker) I've read the following thread on stackoverflow: * how can I send the form data from a native jquery mobile app to a remote server? * Phonegap - Choose Image From Gallery
[ -0.0030210791155695915, 0.0147984828799963, 0.0058736419305205345, 0.000040869752410799265, -0.025002341717481613, 0.005219718907028437, 0.00676262192428112, 0.041290733963251114, -0.011122625321149826, -0.02999138832092285, -0.00133635220117867, 0.008930991403758526, 0.001998322317376733, ...
[ 0.4563678503036499, 0.4275517165660858, -0.20514045655727386, -0.0007512522861361504, 0.2955692708492279, -0.21343378722667694, 0.052681077271699905, -0.056432466953992844, -0.47303685545921326, -0.30427324771881104, 0.4486260414123535, 0.47744473814964294, -0.3992261290550232, -0.05755053...
How I finish this effect in my paper, 4 rows figures and 8 eights figures, especially a red line is in middle of the whole figure. Please. Can you help me finish? ![enter image description here](http://i.stack.imgur.com/F7RGz.png)
[ 0.011977702379226685, 0.009373844601213932, -0.020311899483203888, 0.04204510152339935, -0.024839041754603386, 0.00035070828744210303, 0.011398635804653168, 0.0023664284963160753, -0.02466818317770958, -0.0020883246324956417, -0.033268485218286514, -0.005556452553719282, 0.013163219206035137...
[ 0.2448393702507019, 0.2701840400695801, 0.7568477392196655, 0.014290855266153812, -0.013653763569891453, 0.3216867446899414, 0.3881474435329437, -0.11750173568725586, -0.6473121047019958, -0.626899778842926, 0.2212011069059372, 0.05610775947570801, 0.36334028840065, 0.05323527380824089, ...
I am trying to compile a .tex file using pdflatex and I get the following two kinds of errors: Package natbib Warning: Citation `give-citation' on page 14 undefined on input line 41. ! Undefined control sequence. <argument> ...e \NAT@cmt \fi \if \relax \NAT@date \relax \else \NAT@@close \... l.46 ... information is available \cite{} It works smoothly with pdflatex on my mac. Another error that I get on this linux box (again using pdflatex) is: ! LaTeX Error: Unknown graphics extension: .eps. See the LaTeX manual or LaTeX Companion for explanation. Type H <return> for immediate help. .... l.12 ...\includegraphics[width=4.2in]{tree.eps} These errors appear quite frequently, they are all over the document. Is it possible to fix them?
[ 0.00828848872333765, 0.0024940725415945053, -0.005778319668024778, 0.02671782672405243, 0.03380428999662399, 0.008608901873230934, 0.008134262636303902, -0.0039693196304142475, -0.017662879079580307, -0.01767652854323387, -0.016359327360987663, -0.0027205776423215866, -0.010635890066623688, ...
[ -0.10661444813013077, 0.4333137273788452, 0.5356647372245789, -0.05512356013059616, -0.06543924659490585, -0.37552377581596375, 0.695536196231842, -0.15057644248008728, 0.009520993568003178, -0.7758235931396484, 0.03046279400587082, 0.9300834536552429, -0.4351591467857361, 0.15844096243381...
I tried "$x" and "$y" for coordinates but it always give me zero "0" values too, and what I really want is to display the coordinates like in this image. ![enter image description here](http://i.stack.imgur.com/Z4848.jpg)
[ -0.028329452499747276, 0.008792930282652378, -0.018320053815841675, 0.010243953205645084, -0.03157512843608856, -0.009676303714513779, 0.0082547003403306, -0.004595739766955376, -0.03377645090222359, -0.02219671569764614, -0.022743238136172295, -0.003453848185017705, -0.020641736686229706, ...
[ 0.2081374228000641, 0.018524618819355965, 0.3467578589916229, 0.018502313643693924, 0.15558619797229767, 0.07889638096094131, 0.15474532544612885, -0.03633838891983032, -0.33801424503326416, -0.7234260439872742, 0.24978353083133698, 0.2784280776977539, -0.1424110233783722, 0.22918607294559...
Quite often I find myself referring in text to objects which have strict naming conventions e.g. Java objects, processes, services. I find myself in a dilemma over how to pluralize these names. It seems wrong to pollute the meaningful name by adding an 's' - or worse, a plural that modifies the rest of the name like SomethingUtility -> SomethingUtilities. What if someone does a text search looking for the name? Or cuts and pastes from my comment to search through a log file? I have sometimes resorted to an ungrammatical apostrophe, but this can confuse plural with ownership. Is there any better solution?
[ 0.004095922224223614, 0.013138066977262497, -0.007492242380976677, 0.01205132994800806, 0.005646828096359968, 0.00794014148414135, 0.006230844650417566, 0.01947135105729103, -0.012635360471904278, -0.007149805314838886, 0.00004175836511421949, 0.009068180806934834, -0.000631378497928381, 0...
[ 0.3078509569168091, 0.4161427617073059, 0.019436445087194443, -0.08934324979782104, -0.6497764587402344, -0.1291273832321167, 0.6019493937492371, 0.3300574719905853, -0.1636345386505127, -0.36495113372802734, -0.37001684308052063, 0.06781761348247528, -0.19807583093643188, 0.40681818127632...
_I have several questions. Perhaps it would be better to separate them into different posts. However, given their relative closeness to each other, I think putting it all in one place would be better. On suggestion, I will modify this post._ I am reading Penrose's paper on the conformal treatment of infinity (find here). The basic concept that underlies this is the fact that in order to study asymptotic behaviour of a space-time with metric ${\tilde g}_{\mu\nu}$, we may instead study a space-time that is conformally related to it by defining an unphysical metric on a compact manifold $g_{\mu\nu} = \Omega^2 {\tilde g}_{\mu\nu}$. He then goes on to say that the asymptotic properties of fields can then be investigated by studying local behaviour of fields at infinity on this unphysical manifold **provided that the relevant concepts can be put into a conformally invariant form** 1. What are the relevant conformally invariance concepts that one can study? What I can think of is the causal structure of space-time, gravitational waves (since they are described by a conformally invariant Weyl tensor). What else is there? 2. I have also often heard that massless fields satisfy conformally invariant equations in curved spacetimes. (see this question). Indeed Penrose claims that this can be done if "" **interpreted suitably** ". What does he mean by this? Further, since massless particles can only reach $\mathscr{I}^\pm$, does his formalism only apply to null infinity? 3. What about massive particles? Surely, the equations for such particles are not going to be conformally invariant. Further, these particles would reach $i^\pm$. One can't apply the above formalism to massive particles right? Is there an alternative way of constructing the asymptotic structures of $i^\pm$?
[ 0.005049431696534157, 0.0268564410507679, -0.0018173402640968561, 0.02527840994298458, -0.01086121704429388, 0.006973086856305599, 0.0061666411347687244, 0.008088326081633568, -0.011466939002275467, 0.0159846693277359, 0.0005541086429730058, 0.009033742360770702, -0.009944687597453594, 0.0...
[ 0.27502867579460144, -0.021649613976478577, 0.5053592920303345, 0.05952073261141777, 0.021876735612750053, -0.015667801722884178, 0.013487560674548149, 0.1320616453886032, -0.21731135249137878, -0.6783532500267029, 0.06114295870065689, 0.17749302089214325, -0.1597164273262024, 0.5306397676...
We have an existing Silverlight application that loads a ArcGIS MapService and display it to the user. We want to add the ability to also add a WMS layer on top of the existing displayed map. I am not too familiar about WMS layers and the samples i found on ArcGIS are only loading ArcGIS WMS layers. We have updated one of our ArcGIS map services and enabled the WMS capability. How do i add the WMS layer to the map? Would i access it the same as a ArcGISDynamicMapServiceLayer? using a url like this? `http://OurArcGISServer/arcgis/services/Map1/MapServer/WMSServer` Any input is greatly appreciated.
[ 0.0035850410349667072, -0.0038830216508358717, -0.006663668900728226, 0.02176481857895851, 0.018301645293831825, -0.0034923222847282887, 0.010195321403443813, 0.01707502081990242, -0.017509879544377327, -0.027361806482076645, -0.004445631522685289, 0.020462922751903534, -0.03426600247621536,...
[ 0.6942554116249084, 0.32405561208724976, 0.3082321584224701, 0.1803504377603531, -0.10750064253807068, -0.37665751576423645, 0.11778034269809723, -0.07220103591680527, -0.04984332248568535, -0.763292670249939, 0.11215759068727493, 0.3216288685798645, -0.1971145123243332, 0.349470853805542,...
I have setup my WordPress site with Yoast. However, it doesn't seem to appear in any of the search engines. In Google Webmaster Tools I don't see a sitemap. Is the plugin supposed to submit the sitemap and should it appear in Google Webmaster Tools? How do I know if it has been submitted to other search engines?
[ -0.010499952360987663, 0.0005597660783678293, -0.003013416426256299, 0.02206021174788475, 0.014392679557204247, 0.016343003138899803, 0.010395491495728493, 0.013425030745565891, -0.022210538387298584, -0.025711657479405403, -0.004579232539981604, 0.019276361912488937, -0.020569415763020515, ...
[ 0.5692622065544128, 0.35829034447669983, 0.4240853488445282, 0.038720954209566116, -0.029845602810382843, 0.12095386534929276, 0.07215223461389542, 0.23384621739387512, 0.056577976793050766, -0.6272000670433044, 0.11634889245033264, 0.3847633898258209, -0.24994300305843353, 0.2505490183830...
I liked this explanation of hard and soft/symbolic links: http://unix.stackexchange.com/a/22406/42226 Now the reason why I googled about soft and hard links is simply because I was organizing my folders and files and in one point I wanted one file to be referenced to from another directory. So I thought in shortcuts/symlinks, but would like to be able to rename the file or move it to another folder and still have the symlinks pointing to that file. Is that possible? Hard link isn't an option since I want the same functionality also for folders (and furthermore I want one reference -the first hard link- to be always the "most important").
[ 0.017596060410141945, 0.015584784559905529, -0.005787780974060297, 0.021201258525252342, 0.0011383425444364548, 0.0020005092956125736, 0.008667832240462303, 0.01598847098648548, -0.019890017807483673, -0.008639441803097725, -0.0032755485735833645, 0.005761489737778902, -0.0016985242255032063...
[ 0.14102689921855927, -0.025397511199116707, 0.10225289314985275, 0.07524197548627853, -0.12318766862154007, 0.110344298183918, -0.17907977104187012, 0.5182546377182007, -0.6240739226341248, -0.7264748215675354, -0.016856752336025238, 0.5746878981590271, -0.08011359721422195, 0.237675786018...
If I try to connect to a certain server with ssh example.com ssh aborts with `Too many authentication failures` because I have more than 5 ssh keys in my .ssh folder and it loops through all those keys with none of them fitting (detailed description here) How can I tell ssh not to look in my keys and just give me the password prompt?
[ 0.006939935032278299, 0.009538835845887661, -0.003903266042470932, 0.03879234567284584, -0.002551131183281541, 0.0063814022578299046, 0.007374780718237162, -0.02467179298400879, -0.022463487461209297, 0.007832529954612255, -0.00799283105880022, 0.002542833099141717, -0.014644325710833073, ...
[ 0.2665967047214508, 0.3153380751609802, 0.3250153064727783, 0.1458660513162613, 0.04027511551976204, -0.0540635846555233, 0.7532371282577515, 0.08940205723047256, -0.2851003110408783, -0.8034809827804565, -0.10305023938417435, 0.395754873752594, -0.2928203046321869, 0.4138237535953522, -...
What is a term to describe someone who acts abnormally, as if afflicted by mental illness? For example, they talk much too loudly, have conversations with themselves, and exhibit blatantly abnormal behaviors and emotions that appear nonsensical.
[ -0.04027678817510605, 0.018327772617340088, 0.01748616062104702, 0.030166327953338623, -0.027601541951298714, -0.04988349974155426, 0.012946803122758865, -0.004899519495666027, -0.024491095915436745, 0.014248551800847054, -0.01689130812883377, -0.004032103344798088, 0.009021377190947533, 0...
[ 0.6457703113555908, 0.1770944595336914, -0.3491290509700775, 0.18658572435379028, 0.2676590085029602, -0.32938921451568604, 0.7313182950019836, 0.30100858211517334, -0.36090341210365295, -0.24669411778450012, -0.0873897597193718, 0.15096727013587952, -0.4456443786621094, 0.26490318775177, ...
I apologize if this is a neophyte or duplicate question because I was unable to find it. A comment on the question Problem with \leq displaying as \leqslant, mentions that "you can include mathabx with the mathx option". How exactly does one pass the mathx option to the mathabx package, or more generally pass options to LaTeX packages? In addition to Google, I have also experimented with each of the following, all of which resulted in error messages. \usepackage{mathabx}{mathx} \usepackage{mathabx}[mathx] \usepackage[mathx]{mathabx}
[ 0.01964748464524746, 0.010276172310113907, -0.012631971389055252, 0.0006726908031851053, -0.0042711058631539345, -0.005208789836615324, 0.009407229721546173, 0.010317177511751652, -0.013294035568833351, -0.03078729659318924, -0.0075816623866558075, 0.006076838821172714, -0.01957090198993683,...
[ 0.0518023706972599, 0.3993107080459595, 0.04236011207103729, -0.051938217133283615, -0.24936924874782562, -0.17266139388084412, 0.6109777092933655, 0.2746298015117645, -0.09845665842294693, -0.4446980655193329, -0.18570084869861603, 0.34759703278541565, -0.4258539080619812, 0.2408217638731...
Does gravity slow the speed that light travels? Can we actual measure the time it takes light from the sun to reach us? Is that light delayed as it climbs out of the sun's gravity well?
[ -0.010079126805067062, -0.0025463863275945187, 0.007345512043684721, -0.003639038186520338, -0.04244661703705788, -0.04015803709626198, 0.01814546249806881, -0.015595989301800728, -0.03563765808939934, -0.025488106533885002, -0.005329668056219816, 0.026386786252260208, -0.004191978834569454,...
[ 0.3691559433937073, -0.047014255076646805, 0.4761604070663452, 0.38557931780815125, 0.11400691419839859, 0.12328867614269257, 0.20504756271839142, 0.06920851767063141, -0.5131062269210815, -0.35641998052597046, 0.4652499854564667, 0.17580126225948334, 0.13876111805438995, 0.259634196758270...
In my exams a type of question is coming like - "Change into verb forms" and questions are given below. Are there any methods of forming verbs, or do we have to learn all the verb, adjective, adverb, noun formations?
[ 0.007086873985826969, 0.004791300278156996, -0.01060000341385603, 0.023630857467651367, -0.027870941907167435, -0.011291914619505405, 0.012280047871172428, 0.03313571959733963, -0.028561847284436226, 0.009585550054907799, -0.010987089946866035, -0.0000836399631225504, 0.017208801582455635, ...
[ 0.2508639991283417, 0.2303493320941925, 0.1474270075559616, 0.06990925222635269, 0.048377737402915955, 0.011186192743480206, 0.6715633273124695, 0.22104889154434204, -0.18290473520755768, -0.3659715950489044, 0.05609317496418953, 0.2616950273513794, 0.5017150044441223, 0.1984950751066208, ...
I've been wondering exactly why the elements are distributed the way they are on Earth. The heavier elements have their origins in the centers of stars, or in supernovae. After the death of the stars, you end up with a dust cloud containing the heavier elements. Later, planets form out of these along with new stars. If the heavier elements were randomly distributed in these explosions, how do we end up with special deposits of minerals on Earth? We have mines and special areas where the concentrations of different materials are higher. For example, we have iron, gold, nickel and uranium mines. Why aren't these materials uniformly distributed? Why isn't the Earth composed of a substance which is a uniform mixture of all the heavy elements? Why does it seem there is sometimes a preference for like materials to clump together? The nickel and iron mostly make up the Earth's core, which I guess makes sense in terms of them being fairly heavy while comparatively abundant relative to heavier elements, but what about everything near the surface that we observe?
[ 0.012442883104085922, 0.010091607458889484, -0.00875429529696703, 0.0036045683082193136, -0.012575612403452396, -0.009133676066994667, 0.007972157560288906, 0.012850074097514153, -0.017042305320501328, 0.0016990676522254944, 0.009922970086336136, 0.018661275506019592, -0.003640098264440894, ...
[ 0.9468021988868713, 0.6251958012580872, -0.4314977526664734, 0.5383960604667664, 0.33820945024490356, -0.288353830575943, 0.029007792472839355, 0.2877104580402374, -0.4395868480205536, -0.21247638761997223, -0.11574950069189072, 0.02407371811568737, -0.27297767996788025, 0.6986607909202576...
I need to add an option in my theme options panel to set a specific custom menu's links all to nofollow. Does wp_nav_menu() allow this or does it require a custom walker to manually ad rel="nofollow"?
[ 0.031035492196679115, 0.017315838485956192, -0.010992304421961308, 0.015730861574411392, 0.02368018962442875, -0.005215938203036785, 0.012254931963980198, 0.007594010327011347, -0.023232948035001755, -0.017208103090524673, -0.009974847547709942, 0.009896902367472649, 0.006814436987042427, ...
[ 0.3828945457935333, -0.0994303897023201, 0.5681710839271545, -0.060518402606248856, 0.49946895241737366, 0.020681682974100113, 0.11057368665933609, 0.15831570327281952, -0.4470377564430237, -0.3514209985733032, -0.10656065493822098, 0.2761915922164917, -0.49097320437431335, 0.2434442937374...
I have a model $f(x|\theta)$ ($\theta$ is a vector) for which I want to specify a prior $\pi(\theta)$. I only know that $\theta$ is in some interval. There are ways to specify an ignorance prior (Jeffery's, transformation group, reference prior etc.). Not knowing how to choose, I have simulated $f(x|\theta=\theta_0)$ many times and tried to estimate with each prior candidate, so that I have a distribution $D(\hat{\theta_0}|prior)$. Based on these distributions, I could choose the "optimal prior" using some criterion (e.g. MSE). Is it reasonable at all to proceed this way?
[ 0.008631739765405655, 0.009465975686907768, -0.007329467684030533, 0.016465019434690475, 0.01009365450590849, -0.0010716569377109408, 0.008832834661006927, -0.004430665634572506, -0.011596264317631721, -0.003211298957467079, -0.011385714635252953, 0.006947968155145645, -0.025425724685192108,...
[ -0.20352615416049957, 0.11200028657913208, 0.3811109960079193, -0.003758152248337865, 0.11249196529388428, 0.5615808963775635, 0.03819938376545906, -0.3574458658695221, -0.06893804669380188, -0.6922466158866882, 0.22117377817630768, 0.5205416679382324, -0.45747461915016174, 0.5275600552558...
I've previously converted simple shape files to WKT so that I can store them in RDMS that has spatial capabilities. I'm a new user to QGIS, which I'm using to browse a shape file. But now I have a shape file like this: ![](http://i.imgur.com/3CD6swi.jpg) Not only are there polygons, but there are labels on those polygons. Really what I'd like is to strip off the polygons and save them with their associated label. So something like: row1 ("POLYGON((1 1,5 1,5 5,1 5,1 1))) 'high' row2 ("POLYGON((2 2,5 1,5 5,1 5,2 2))) 'medium' row3 ("POLYGON((5 5,5 1,7 7,1 5,5 5))) 'high' row3 ("POLYGON((8 8,5 1,5 5,1 7,8 8))) 'low' I'm a little stumped at the best way to creatively slice up my shape file, ideally by some of the attributes defined in the accompanying dbf file. Ideally I'd like to: 1)export, as WKT, all of the polygons with 'high' 2)export, as WKT, all of the polygons with 'low' etc, etc.
[ 0.003938748501241207, 0.0032886064145714045, 0.004913129843771458, 0.03178289532661438, -0.011444613337516785, 0.010539008304476738, 0.007050841115415096, 0.014315856620669365, -0.01859593763947487, 0.007604774087667465, -0.003059233771637082, 0.007783140987157822, -0.006791434250771999, 0...
[ 0.3008060157299042, 0.12023379653692245, 0.24640002846717834, -0.0365099161863327, -0.3083808720111847, 0.18897101283073425, 0.05845721811056137, 0.2426215261220932, -0.16361239552497864, -1.0409324169158936, 0.06025530397891998, 0.18070536851882935, -0.0320778451859951, 0.3915809690952301...
I one of my headerfiles, I used the command `\setmathfont[]{XITS Math}`, once I tried to compile the document as always with `xelatex`, I get the following error: ! Undefined control sequence. l.87 \setmathsfont []{XITS Math} Also, for some reason, XeLaTeX seems to hang when I comment out that line. This may or may not be relevant: $ xelatex --version XeTeX 3.1415926-2.4-0.9998 (TeX Live 2012/Debian) kpathsea version 6.1.0 Copyright 2012 SIL International and Jonathan Kew. There is NO warranty. Redistribution of this software is covered by the terms of both the XeTeX copyright and the Lesser GNU General Public License. For more information about these matters, see the file named COPYING and the XeTeX source. Primary author of XeTeX: Jonathan Kew. Compiled with ICU version 49.1 [with modifications for XeTeX] Compiled with zlib version 1.2.7; using 1.2.7 Compiled with FreeType2 version 2.4.9; using 2.4.9 Compiled with fontconfig version 2.9.0; using 2.9.0 Compiled with libpng version 1.2.49; using 1.2.49 Compiled with poppler version 0.18.4
[ 0.014327684417366982, -0.009950490668416023, -0.01897568255662918, 0.011746646836400032, 0.014809874817728996, -0.007061450742185116, 0.007407680153846741, -0.0023599322885274887, -0.015779707580804825, -0.013355711475014687, -0.01159067451953888, -0.0023616463877260685, -0.00073454546509310...
[ 0.267964243888855, 0.39648324251174927, 0.5321667194366455, -0.16228654980659485, -0.10325277596712112, -0.14086247980594635, 0.15204662084579468, 0.017996734008193016, -0.02506842464208603, -0.18874941766262054, 0.0006750249885953963, 0.5101689100265503, -0.3874650001525879, 0.30149748921...
Could anybody help me here, please? I’d like to know the grammatical form of the verb “can" in the following examples: > 18y old Tim is asked by a relative what he’s going to do with his life. > After some hesitation, he finally answers: “Well, I COULD help my father on > the farm - but I’m not sure if I want to be a farmer. I COULD also go to > university but I don’t like studying… or I COULD train really hard and > become a football pro. I don’t now yet, too many options…” Ten years later, > Tim is lying on the street, an empty wine bottle in his hand, and he’s > saying: “What ever happened to me,… I COULD HAVE become a football pro, if > only I had tried harder,... or I COULD HAVE helped my father - but now he’s > sold the farm.” It's like conditionnel 1 and conditionnel 2 in French ("je pourrais / j’aurais pû”), but how is it called in English? Thank you!
[ 0.004159586038440466, 0.00011371442815288901, -0.010304020717740059, 0.005828863009810448, -0.014329898171126842, -0.008713632822036743, 0.005197310354560614, 0.0005541780265048146, -0.009820405393838882, 0.02969425916671753, 0.0002481315459590405, 0.0009010985377244651, 0.011343134567141533...
[ 0.3448878228664398, 0.5345633625984192, 0.33678996562957764, 0.02775506116449833, 0.3758769631385803, 0.5819211602210999, 0.35412487387657166, -0.021271217614412308, -0.4433518052101135, -0.39494240283966064, 0.15749986469745636, 0.380769282579422, 0.28859052062034607, -0.13672949373722076...
I have pandas DataFrame with these columns: point datetime value 1 Point(-24.6064453125 26.3683280945) 2014-09-11 19:00:00 24.885258 2 Point(-24.6472167969 26.4629898071) 2014-09-11 19:00:00 24.854557 3 Point(-24.6881408691 26.5576820374) 2014-09-11 19:00:00 24.822819 ... with types: string, datetime and float. What would be the easiest way to migrate this data to a new table in PostGIS enabled database? Pandas has `to_sql` function (working through sqlalchemy connector, which is based on psycopg2), but I don't think there is an option for declaring geometry/geography data type.
[ -0.011493071913719177, 0.018294982612133026, -0.007845386862754822, 0.007314515300095081, 0.012383239343762398, -0.0011752694845199585, 0.006140438839793205, 0.024753941223025322, -0.007264104206115007, 0.01658158004283905, -0.007968166843056679, 0.00610067555680871, -0.0016875818837434053, ...
[ 0.013090870343148708, 0.19190290570259094, 0.5595389008522034, 0.015840407460927963, 0.021875126287341118, 0.3542708158493042, -0.031860850751399994, -0.39241930842399597, -0.14329347014427185, -0.5154960751533508, 0.15117494761943817, 0.34928321838378906, -0.11674179881811142, 0.026387818...
This may see like a stupid question, but I'm new to QGIS, so please bear with me. I'm having a problem when saving as .dxf. I export a building layer (polygons) as a .dxf file However in AutoCAD it turns up as a hatch instead of as a closed polyline which is what I'm looking for. I can of course recreate the boundary in AutoCAD, but I would rather just save the data correctly in QGIS. Can anyone help? thank you in advance & best regards Martin
[ -0.0275388415902853, 0.008860702626407146, -0.0026608104817569256, 0.01705533266067505, -0.00426194304600358, -0.018556367605924606, 0.008653489872813225, 0.03448798879981041, -0.017756910994648933, -0.01924040913581848, -0.012281459756195545, 0.00904218852519989, -0.013207433745265007, 0....
[ 0.40701907873153687, 0.35406360030174255, 0.31362640857696533, 0.1927148997783661, -0.10595354437828064, -0.005597020499408245, 0.20210352540016174, -0.024748794734477997, -0.3117215633392334, -0.9691075682640076, -0.1228826493024826, 0.32812565565109253, -0.0094170942902565, 0.06411740183...
There are a lot of algorithms which will tell you whether or not a given point is found inside a polygon. I'm looking to write an algorithm which, given a non-convex polygon, will return a point which is inside the polygon. I don't need the point to be in any specific location inside the polygon, but I prefer to receive a point which isn't very close to an edge, but that is not a deal-breaker. It's there merely to mark that polygon's planar straight-line graph (PSLG) as an internal shape for use with Shewchuck's Triangle library for some complicated constrained Delaunay triangulations. My initial thinking is: 1. Compute the bounding box 2. Cast a ray from one corner in the direction of the opposite corner, or from the center of a bounding box edge to the opposite edge center. 3. Then, a point exactly between the first and second intersection will be inside the polygon. Is there a better approach?
[ -0.006846011150628328, 0.013260727748274803, -0.0004908177070319653, 0.025037162005901337, 0.007827412337064743, -0.0024131175596266985, 0.008078893646597862, -0.0031677945517003536, -0.01591132953763008, 0.010310550220310688, -0.0019013965502381325, 0.010063747875392437, 0.00466030836105346...
[ -0.08546119928359985, 0.15313474833965302, 0.258974552154541, 0.2591651678085327, -0.1145058125257492, 0.32842788100242615, -0.15461485087871552, 0.28077155351638794, -0.39588841795921326, -0.8137692213058472, -0.31810811161994934, -0.08498499542474747, -0.050036121159791946, 0.24677766859...
I'm currently searching for the right term for someone who * Organizes or * Promotes Parties, "Clubbings", Events or similar. I understand that there's a difference between someone who _promotes_ a party and someone who makes the organization, but maybe there's some common word as the vast majority of those people are one person companies.
[ -0.028825173154473305, 0.004106256645172834, 0.00916630681604147, 0.01733831875026226, 0.008941438980400562, 0.002232026308774948, 0.011114283464848995, 0.015138085000216961, -0.019415486603975296, 0.02300165966153145, -0.0018818322569131851, 0.007683595642447472, 0.038590166717767715, 0.0...
[ 1.0410447120666504, 0.21926623582839966, -0.25528284907341003, -0.025650762021541595, -0.027485398575663567, -0.49706220626831055, 0.11522773653268814, 0.5479135513305664, -0.6903476715087891, -0.06260485202074051, -0.18625091016292572, 0.20679768919944763, 0.04419058561325073, 0.361017048...
I'm using samba4 as a dc with sssd to talk to a separate LDAP server. When I try to login to a win7 machine that's added as a domain member samba is stripping off the domain name and substituting the samba domain instead. Errors that I see in log.samba: [2014/07/18 06:46:28.177400, 3] ../source4/auth/ntlm/auth.c:270(auth_check_password_send) auth_check_password_send: Checking password for unmapped user [ldapdom]\[user]@[win7host] auth_check_password_send: mapped user is: [sambadom]\[user]@[win7host] [2014/07/18 06:46:28.178098, 3] ../source4/auth/ntlm/auth_sam.c:61(authsam_search_account) sam_search_user: Couldn't find user [user] in samdb, under DC=dom,DC=server,DC=edu [2014/07/18 06:46:28.178184, 2] ../source4/auth/ntlm/auth.c:420(auth_check_password_recv) auth_check_password_recv: sam_ignoredomain authentication for user [sambadom\user] FAILED with error NT_STATUS_NO_SUCH_USER I can't tell if I need some sort of username mapping or a change to either the smb.conf or sssd.conf. Username mapping would seem odd since I essentially want it to map the original value to itself. CentOS 6.4 x64 Samba4 4.1.9 sssd
[ -0.008397519588470459, -0.005210101138800383, -0.005575357470661402, 0.017494574189186096, -0.012718437239527702, 0.029363755136728287, 0.007763161323964596, 0.026016993448138237, -0.014667980372905731, -0.0015183426439762115, -0.011641865596175194, 0.000040744198486208916, -0.00930795539170...
[ 0.028459986671805382, 0.3150845766067505, 0.32255029678344727, -0.04568947106599808, -0.21866227686405182, 0.08869452774524689, 0.33386892080307007, -0.2623580992221832, -0.41970667243003845, -0.7598296403884888, 0.09273563325405121, 0.16852886974811554, -0.49603942036628723, 0.36084502935...
I wanted to know: does Sahara give you a prize for buying a specific amount of her items, or does she give you absolutely NO award for buying her items?
[ -0.013185633346438408, 0.015156629495322704, 0.0019673092756420374, 0.016405554488301277, -0.016861893236637115, -0.036299195140600204, 0.016332954168319702, -0.018597597256302834, -0.03497091680765152, -0.03369085490703583, -0.033173568546772, 0.00007685867603868246, -0.023018747568130493, ...
[ 0.20470868051052094, -0.22705130279064178, -0.3337770700454712, 0.5214171409606934, -0.3815867602825165, -0.007025879807770252, 0.2749311327934265, 0.014854839071631432, -0.32950910925865173, 0.419124573469162, 0.1956295371055603, 0.23078973591327667, -0.17812441289424896, 0.24349799752235...