text
stringlengths
23
30.4k
embeddings_A
list
embeddings_B
list
So... I just got Hitman: Absolution for PS3. I have started playing a few missions, but I feel like I am missing something. Do I feel like I am missing something because I haven't played the other Hitman games OR of my failure to understand the story due to some other reason? **Should I play the other Hitman games to understand the story?**
[ 0.0030563767068088055, 0.014533141627907753, 0.008804318495094776, 0.01794201135635376, 0.022440919652581215, -0.015505819581449032, 0.009669746272265911, -0.021810298785567284, -0.024957170709967613, 0.0024386802688241005, -0.008441902697086334, 0.020759008824825287, -0.008110061287879944, ...
[ 0.26228249073028564, 0.007385818287730217, 0.29590389132499695, 0.26244550943374634, -0.42742958664894104, -0.33044004440307617, 0.3879508078098297, -0.2773295044898987, -0.5106136798858643, -0.14657065272331238, -0.0061357938684523106, 0.6425203680992126, 0.0033560586161911488, -0.1330583...
I'm looking for help on installing Zorin OS 6.3 32bit. The program gets stuck on the "preparing to install" phase. ![ss of error](http://i.stack.imgur.com/4722W.png) _original URL:http://imgur.com/PulGPMD_ I've tried loading it in both "Live System", and "Graphics Safe Mode". Both get stuck at the same screen. Alternatively, I've tried the "Load the Installer Directly", right after startup, but I only receive an error. It's being run side by side a no longer working version of Windows 7 x32bit. When I try to boot from my hard drive "Drive Read Error Press Ctrl+alt+delete to re- reload" appears. Not sure if that's relevant, but figured it might be.
[ -0.01337639894336462, -0.001007034326903522, -0.007324088364839554, 0.019742123782634735, -0.017781665548682213, -0.003918756265193224, 0.006170556880533695, -0.0017041480168700218, -0.016104383394122124, 0.013037554919719696, -0.017968662083148956, 0.00619890633970499, 0.015172078274190426,...
[ 0.5018659830093384, 0.05100662261247635, 0.34361231327056885, -0.2656717300415039, -0.09092635661363602, 0.05616484954953194, 0.24626167118549347, 0.25279560685157776, -0.08187106996774673, -0.881070613861084, -0.002629608381539583, 0.7778401374816895, -0.2399028092622757, 0.21029424667358...
> **Possible Duplicate:** > LaTeX conditional expression I've already taken a look at ‎books‎ such as "‎Te‎X for the Impatient", but its examples aren't clear enough to me. What‎ book/tutorial ‎do‎ you recommend for learning conditional tests ‎in ‎‎‎LaTeX‎?‎
[ 0.02475151978433132, 0.007112438790500164, -0.011248372495174408, 0.027416417375206947, 0.028347037732601166, 0.04128124564886093, 0.00978959258645773, -0.0037558532785624266, -0.03961637243628502, 0.008865994401276112, -0.008009924553334713, 0.0067580281756818295, -0.009906645864248276, -...
[ 0.23100538551807404, -0.20982562005519867, -0.03468243032693863, 0.024247851222753525, -0.04470305144786835, 0.4245602488517761, 0.4327150583267212, 0.17516115307807922, 0.009705400094389915, -0.5857614874839783, 0.08497525751590729, 0.6891865730285645, -0.17556609213352203, -0.60249269008...
Everything goes ok, but when Eola begins fighting, a bit after the game will crash. I was able to prevent the crash by selecting her in console and typing: setunconscious 1 When the fight ends I type setunconscious 0 If I do not do that, it is 100% guaranteed that the game will crash. This tip can probably be used on similar situations, despite this being the first time I've see it happen. Is there any fix for this?
[ 0.022472908720374107, 0.00577189028263092, -0.01386390533298254, -0.0123185645788908, 0.009425712749361992, -0.011828621849417686, 0.0069353291764855385, 0.014508897438645363, -0.019083231687545776, 0.02385805733501911, -0.004927236121147871, 0.021064097061753273, 0.0018785124411806464, 0....
[ 0.10646484047174454, -0.3558892607688904, 0.5791158080101013, 0.18419194221496582, -0.17844094336032867, -0.20179039239883423, 0.3581770062446594, -0.09027807414531708, -0.011324690654873848, -0.1736813634634018, -0.056941110640764236, 0.3395529091358185, -0.12633968889713287, 0.0982144176...
I am applying for a job in finance where I think I am just a beginner. So there are a lot of things I will learn on the job. I want to express that I am ready for it and to make clear that I am an career entrant. I thought about something like: "I am willing to enhance my knowledge in finance." I am not happy with this and I do not know how to express what I want in a nice sentence. So I want to state that I am ready to expand my knowledge where it is necessary and that I am eager to learn new technologies (which are necessray for this) job on the job.
[ 0.008592034690082073, 0.013070622459053993, -0.014382355846464634, 0.013968468643724918, 0.0004525585100054741, 0.010088235139846802, 0.005483157001435757, 0.0034505054354667664, -0.014953719452023506, 0.014992578886449337, -0.004518470726907253, 0.011630279012024403, -0.0036972351372241974,...
[ 0.8933348655700684, 0.49622318148612976, 0.010077106766402721, -0.08407513797283173, 0.5751702785491943, -0.17968560755252838, -0.006753908004611731, 0.30178460478782654, -0.16463755071163177, -0.6554951071739197, 0.22316573560237885, 0.6641813516616821, 0.5178269147872925, 0.1012695953249...
I just rsync-ed 2,000,000 files (3TB) from one RAID to another. I want to make sure my data is intact. `rsync -c` takes a really long time. `diff` doesn't show me what it's doing. Is there an alternative that's (a) faster, and (b) will show me progress while it's comparing? (I'm on Mac, and `brew search diff` gives me `apgdiff colordiff diffstat diffutils fmdiff libxdiff open-vcdiff podiff rfcdiff vbindiff bsdiff diffpdf diffuse dwdiff kdiff3 ndiff perceptualdiff rdiff-backup tkdiff wdiff ` ... would one of these do the job?)
[ 0.0016940883360803127, 0.0042147887870669365, -0.027208523824810982, 0.011603916063904762, -0.004466029815375805, -0.008172126486897469, 0.007010220550000668, 0.0065999021753668785, -0.021661438047885895, -0.008928370662033558, 0.0071740904822945595, 0.0019836826249957085, 0.0097837680950760...
[ 0.35283440351486206, 0.3440774977207184, 0.5523344874382019, -0.14492003619670868, 0.0474882647395134, -0.0685044527053833, 0.5321842432022095, -0.46992260217666626, -0.28103703260421753, -0.6190248727798462, 0.010855553671717644, 0.5311261415481567, -0.053173333406448364, 0.05711874738335...
As far as I know, there aren't a lot of differences between JDK 6 and JDK 7. At least, I haven't yet come across a difference (I coded with JDK 7 and with JDK 6). A lot of computers run JRE 6 or JRE 7. Compiling my program using JDK 7 will narrow the number of potential computers that can run my application. Is it generally a good idea to always program using JDK 6 and compile my program for JRE 6? Is this something common? Or is it generally a bad idea?
[ -0.00448865070939064, 0.003492937423288822, -0.02089529111981392, 0.00393647002056241, 0.011869193986058235, 0.01917988993227482, 0.007135249674320221, -0.006123412400484085, -0.015549946576356888, -0.03966226428747177, -0.0032965028658509254, 0.01588170789182186, 0.010145608335733414, 0.0...
[ 0.48980215191841125, 0.14035797119140625, 0.09517588466405869, -0.14425161480903625, -0.1206926703453064, -0.17533141374588013, 0.3002801537513733, 0.1630566567182541, -0.11149344593286514, -0.5165603756904602, 0.17447708547115326, 0.9741464853286743, -0.08133267611265182, 0.04073810204863...
My client's site is in French and we want to translate the canned emails (such as user registration and password lost) to French. What are the hooks/filters for these? The default language is french, i.e. define('WPLANG', 'fr_FR');
[ -0.01350595336407423, 0.018902909010648727, 0.0010924600064754486, 0.010504994541406631, -0.0117320092394948, -0.01331143919378519, 0.012332501821219921, 0.023121951147913933, -0.01907825469970703, 0.01252834778279066, -0.018438367173075676, 0.00542635889723897, 0.006681871134787798, 0.006...
[ 0.1907339245080948, 0.11486925929784775, 0.3842000365257263, -0.20113347470760345, -0.04318632557988167, 0.0785590410232544, 0.41932541131973267, 0.5398556590080261, 0.09740552306175232, -0.6611568331718445, -0.37930071353912354, 0.608273446559906, -0.47079479694366455, -0.3901506364345550...
I'm using a GLM with logistic link function to try to predict Y (0 or 1) as a function of a ton of predictor variables (A, B, C, etc.). Some of the predictor variables (A*, B*, C*, etc.) have been shown in other studies to be significant predictors. I want to show essentially that Y is unrelated to all of the other predictor variables, and I thought the simplest way to do this would be to run the full model (Y ~ .) and the null model (Y ~ A* + B* + C* + ...), and then use anova() to compare the two and show that they aren't different (i.e. have equal predictive power). However, anova() only outputs p-values (type I error), but I need a type II error rate here (since I want to show that the models are the same, I need a false negative rate for that). Any ideas on how to approach this?
[ 0.01499241590499878, 0.017628664150834084, -0.01585501804947853, 0.0025878739543259144, -0.002638575155287981, -0.001359674846753478, 0.009464945644140244, 0.0014165232423692942, -0.01317383162677288, -0.039566509425640106, -0.0017318171449005604, 0.0031521935015916824, -0.005659605376422405...
[ 0.17613522708415985, 0.06882829964160919, 0.2560991644859314, 0.1709563285112381, -0.15390747785568237, 0.11725663393735886, 0.030459631234407425, -0.2807866334915161, -0.11958497762680054, -0.29205769300460815, 0.048781462013721466, 0.45041000843048096, -0.14291945099830627, 0.48570087552...
For a trivial example, suppose that `\f(x)` is a quadratic curve defined as follows \def\f(#1){-#1*(#1-8)/4} For each element `\ix` in `{0,1,2,3,4,5,6,7,8}` there is a point `(\ix,\f(\ix))` that will be the center of a square of 1 cm square. How to express the bottom-left & top-right points of `\psframe` in terms of `\ix` and `\f(\ix)`? NOTE: Please don't suggest other simpler solutions such as using `\rput` to put the squares because I am interested in PSTricks point expression. RPN notation should be avoided if it is possible. I give you the skeleton (not working) as below to save your time. \documentclass[pstricks,border=12pt]{standalone} \usepackage{pst-plot}%loads multido automatically \begin{document} \def\f(#1){-#1*(#1-8)/4} \begin{pspicture} \multido{\ix=0+1}{9}{ % the following expression is wrong! \psframe[dimen=middle](*\ix-0.5 {\f(x)-0.5})(*\ix+0.5 {\f(x)+0.5}) } \psplot[algebraic]{0}{8}{\f(x)} \end{pspicture} \end{document} The output I want to get is as follows: ![enter image description here](http://i.stack.imgur.com/sYlX7.png)
[ -0.005058995448052883, 0.004927007481455803, -0.007157132960855961, 0.005760421510785818, -0.004022108856588602, -0.018661029636859894, 0.005682101473212242, 0.004572609439492226, -0.010768473148345947, -0.013166755437850952, -0.003888398641720414, 0.0011377658229321241, -0.01202691625803709...
[ -0.3857063055038452, -0.02438628487288952, 0.6372644305229187, 0.015963401645421982, -0.08310069143772125, 0.5354475975036621, -0.3214873969554901, -0.3045409917831421, -0.19338782131671906, -0.67722487449646, -0.2866453528404236, -0.14691048860549927, -0.021645771339535713, 0.356703788042...
Functional Requirements state "WHAT" needs to be done from a user's perspective Functional Specification state "HOW" it needs to be done from technical standpoint. Do QA teams focus and Functional Requirements whereas development team focus primarily on functional specification though they need to have an understanding of functional requirements? I assume QA would sign off the product as long as the functional requirements are met irrespective of whether the functional specification is followed or not. Is my assumption correct?
[ -0.0008594269747845829, 0.018318992108106613, 0.0022913513239473104, -0.0024084276519715786, -0.007754228077828884, 0.007788184564560652, 0.010689415968954563, 0.03335868567228317, -0.016059543937444687, -0.014010492712259293, -0.017101606354117393, 0.023003943264484406, 0.001707107410766184...
[ 0.4623083472251892, 0.37373173236846924, 0.27754122018814087, -0.2999926805496216, -0.1334887146949768, 0.127452090382576, 0.18847939372062683, -0.33244362473487854, 0.19177083671092987, -0.36145496368408203, -0.4076698422431946, 0.7604199647903442, 0.06817363947629929, -0.0326546728610992...
I was wondering if anyone knew of any noteworthy SEO tracking web apps (either free or paid). I.e. i want to be able to track page rankings, rankings against keywords, trends over time etc. I've seen a couple, but most seem pretty amateur. I was hoping that there might be a gem out there somewhere? PS. I am aware of the google tools (webmaster and analytics). What I am interested in, is a web solution that will help me pro-actively keep on top of my sites SEO strategy, as well as pro actively alert me to opportunities to increase / enhance exposure.
[ -0.004456584807485342, -0.0015482284361496568, 0.0047206394374370575, 0.006803406402468681, 0.012169213965535164, 0.012512004002928734, 0.005124310031533241, 0.005027735140174627, -0.016354797407984734, -0.004667936824262142, 0.000193367013707757, 0.005078610964119434, 0.006174641661345959, ...
[ 0.7277712821960449, 0.3299579918384552, 0.022052252665162086, 0.23535361886024475, -0.20753316581249237, -0.11684421449899673, 0.3056718111038208, 0.31618091464042664, -0.17001405358314514, -0.48000746965408325, 0.0913897380232811, 0.5800071358680725, 0.35207217931747437, -0.38065454363822...
The following code shows that a `\color{red}` command increases the height of a tabular cell. I want to use a switch, not a command as `\textcolor{red}{text}`. \documentclass{article} \usepackage{xcolor} \usepackage{float} \begin{document} \begin{table}[H] \begin{tabular}{*{2}{p{0.45\textwidth}}} \hline \color{red} header & header \tabularnewline \hline % content & content \tabularnewline content & content \tabularnewline content & content \tabularnewline \hline \end{tabular} \end{table} \end{document} ![enter image description here](http://i.stack.imgur.com/MVFpZ.png) How can I work around this problem?
[ -0.004776865243911743, 0.005429947283118963, -0.0023697749711573124, 0.01525593176484108, 0.0032636350952088833, -0.0022370624355971813, 0.006769371218979359, -0.0029953252524137497, -0.009328911080956459, -0.007054866757243872, -0.012928910553455353, 0.0012360253604128957, 0.000964184000622...
[ -0.27738282084465027, 0.2504529356956482, 0.7953981757164001, -0.08851903676986694, 0.017202772200107574, 0.12872503697872162, -0.04411058872938156, -0.5255299806594849, -0.5681959390640259, -0.5323604345321655, 0.05707497522234917, 0.5587759613990784, -0.1013057753443718, -0.0069518615491...
We are supposed to give a matrix representation of $L\cdot S$ for an electron with $l=1$ and $s=\frac{1}{2}$. I read $L\cdot S$ as $L \otimes S$. Is this correct? Then we would have e.g. for $L\otimes S (|1,1\rangle \otimes |1/2,1/2\rangle) = L |1,1\rangle \otimes S|1/2,1/2\rangle $ $= \sqrt{2} \hbar |1,1\rangle \otimes \sqrt{\frac{3}{4}} \hbar |1/2,1/2 \rangle = \sqrt{\frac{3}{2}}\hbar^2 |1,1\rangle \otimes |1/2,1/2\rangle $. Is this correction correct? In that case should I proceed in this way with all the other basis vectors and write the eigenvalues down the diagonal in a matrix?
[ -0.0054123373702168465, 0.0009180436609312892, -0.022472204640507698, 0.004463772755116224, 0.014057543128728867, -0.006589766591787338, 0.00756719009950757, -0.011411767452955246, -0.008044164627790451, 0.013266090303659439, -0.004677926655858755, -0.0003411115612834692, -0.0337388962507247...
[ -0.13958247005939484, -0.08538047224283218, 0.5504052042961121, -0.3925369679927826, 0.13224220275878906, 0.4541670083999634, -0.19329437613487244, -0.40828534960746765, -0.12738271057605743, -0.7201120853424072, -0.15713831782341003, 0.33123430609703064, -0.5545991659164429, 0.04688170179...
I'm trying to query all posts by an author and paginate them on the author.php page. I've tried messing with the WP default blogs per page settings and that doesn't help. I've looked at other posts on here and could find a solution either. This works on my category pages and custom template pages (with the exception of the author attribute). Here's what I have for my query: if (get_the_author_meta( 'ID' )) { $authorID = get_the_author_meta( 'ID' ); } $paged = (get_query_var('paged')) ? get_query_var('paged') : 1; $args = array( 'post_type' => 'post', 'author'=>$authorID, 'orderby'=> 'date', 'showposts'=>'5', 'paged'=>$paged ); $query = new WP_Query( $args ); I even tried query_posts() with no luck. The query itself works, but the pagination breaks. Any ideas?
[ 0.013256072998046875, 0.01012977585196495, -0.005524907261133194, 0.01306937262415886, -0.013446198776364326, 0.023535355925559998, 0.005547357257455587, -0.007263103500008583, -0.013368528336286545, -0.000027163885533809662, -0.004183927085250616, 0.012334048748016357, -0.002980098128318786...
[ 0.13363173604011536, 0.1725868582725525, 0.6988968253135681, 0.1241159737110138, -0.25272923707962036, -0.010538728907704353, 0.21197344362735748, -0.20732684433460236, -0.10801967233419418, -0.7394241094589233, 0.24190421402454376, 0.5301074981689453, -0.30309051275253296, 0.2695610225200...
I've bought an HTC Desire HD yesterday with Android 2.2 Froyo and HTC Sense. So I migrated my calendar items manually and it seemed to pick up some connections from my Facebook and my Twitter, which was nice.. But: From time to time the calendar app just becomes really really slow for some reason, the sync icon is running, and the items just won't open. When this happens I keep staring at a blank screen for about 20 seconds which is really annoying... When I check out what he is syncing its always either Facebook, or the Google calendar. I already tried to switch these things off, but it does not seem to help :( ... It even goes to the point where the whole calendar app just crashes, or even takes HTC Sense with it.. How can I make the calendar fast? Thanks in advance!
[ -0.026501361280679703, -0.007923487573862076, -0.006622768938541412, 0.01525807473808527, -0.012585979886353016, -0.008413868956267834, 0.006518255919218063, 0.023966211825609207, -0.00911588966846466, -0.01796165481209755, -0.02408444695174694, 0.003616932313889265, -0.010282550007104874, ...
[ 0.3896414041519165, -0.14263726770877838, 0.8995397686958313, 0.3099450170993805, -0.17556433379650116, 0.31182122230529785, 0.014893782325088978, -0.04269171878695488, -0.06192798912525177, -0.8125340342521667, 0.4106806218624115, 0.39813539385795593, -0.23852553963661194, 0.0642103180289...
My wife recently went to install a game on her phone (Sims Freeplay) from Google Play. She tells me the phone had enough memory for the install (~15MB free for a 13MB app), which is why she tried. (As an aside, I'm assuming it downloads the package, then installs from that, so you need at least twice the final space to unpack the thing correctly.) It downloaded successfully, but when the phone tried to install the app the phone unexpectedly self-rebooted. After booting, the game wasn't installed and the phone is reporting low memory, with <1MB of phone memory free. She's using an app to check this usage (called DiskUsage), which is reporting "System Data" as using 28MB - she's sure it was much smaller when she checked the space before attempting the installation. Assuming she's right, I'm thinking maybe the downloaded package is still sitting around somewhere consuming space. Is this possible (given it crashed mid-install)? If so how can we get rid of it?
[ -0.04059460386633873, -0.009803073480725288, -0.00265333685092628, 0.00776943564414978, 0.019358040764927864, -0.004102032631635666, 0.00821662973612547, 0.0034210928715765476, -0.01321356650441885, -0.015875814482569695, -0.013613581657409668, 0.006048042792826891, 0.035440608859062195, 0...
[ 0.21229758858680725, 0.34770387411117554, 0.2292536050081253, 0.17278249561786652, 0.6679291725158691, 0.25684818625450134, 0.13776420056819916, -0.11176783591508865, -0.24079985916614532, -0.6154140830039978, 0.4844067692756653, 0.7629548907279968, -0.11793883144855499, -0.000306697678752...
I've taken a few probability classes and now understand how to calculate some statistical measures like mean and confidence intervals. What I don't know is the what, when, and why of using these measures for specific situations. I'm hoping to put together a good collection of each of these measures, what they're used for, and what situations these are good to use. Specifically I'm looking for these (but not limited to): * Mean (average) * Standard Deviation * Variance * Confidence Intervals * Median
[ 0.010923472233116627, 0.0017516663065180182, -0.017482586205005646, 0.007764232344925404, 0.014556897804141045, 0.00699556665495038, 0.007395394612103701, -0.030895350500941277, -0.018900733441114426, -0.03763654828071594, 0.006420515477657318, 0.001335872570052743, -0.008894596248865128, ...
[ 0.4912579655647278, -0.13185332715511322, 0.07835157215595245, 0.31113511323928833, 0.01899431087076664, 0.23286309838294983, 0.09930232167243958, 0.06021202728152275, -0.09976467490196228, -0.5133618116378784, 0.20154069364070892, 0.4125693440437317, 0.07584234327077866, 0.092946723103523...
I have HM Strength and HM Cut. I would love to fly around this new beautiful world but I'm curious when it will be possible. Any help?
[ 0.03154044970870018, 0.023369701579213142, -0.016165923327207565, 0.02212444320321083, -0.019875138998031616, 0.004490641411393881, 0.010718660429120064, 0.013442027382552624, -0.033654458820819855, 0.013829171657562256, -0.007727786432951689, 0.02732658199965954, -0.001583980629220605, 0....
[ 0.7555274367332458, 0.20791001617908478, 0.10081078857183456, 0.16160500049591064, 0.31640541553497314, 0.20695164799690247, 0.06639278680086136, 0.16583265364170074, -0.35162752866744995, -0.767845630645752, 0.25148439407348633, 0.3809475600719452, 0.35551202297210693, 0.07621067017316818...
So, this is a strange occurrence, which seems similar to other questions but which I've not been able to find a reference for. Maybe my Google-Fu is not strong right now. It seems to be a common complaint that user's root apps (sudo'd apps like Synaptic) are falling back on default themes. Usually the solution is given as "copy themes from `~/.themes` to `/usr/share/themes` or `/root/.themes`," for example. This hasn't worked for me, because there's nothing listed in `~/.themes` for me. Then I figured out that changing the _controls_ setting in the Appearance dialog breaks root app theming instantly. See the following screenshots. **Before** ![Default Mint-X theme/controls selected](http://i.stack.imgur.com/5H4tN.png) **After** ![One click later, root app theme broken](http://i.stack.imgur.com/RyzAV.png) _Installing_ another themeset (such as `grey-theme` from Synaptic) appears to break this somehow too, although it goes away when uninstalled. I have not tested this extensively and it may be a fluke related to the controls changing. On a hunch, I tried running `sudo mate-appearance-properties` and/or `gksudo mate-appearance-properties`. Notably, nothing changed when I selected different controls, not even the root app. Also, `sudo` had only the default theme selected, but `gksudo` had " _Custom Theme_ " selected, which was identical to the default and also could not be changed. Unsure what this means, if anything. So, long story short, I doubt this is a bug per se, but I would like to be able to change my default theme/controls and have root apps follow it somehow.
[ 0.007274439092725515, -0.004552900325506926, 0.006440884433686733, 0.015935273841023445, 0.0036941897124052048, -0.012511595152318478, 0.004922551568597555, 0.008360570296645164, -0.0097853047773242, -0.01206873171031475, -0.018523067235946655, 0.006842377595603466, -0.011692875996232033, ...
[ 0.3195119798183441, 0.09655231237411499, 0.5253867506980896, 0.40531012415885925, 0.005303981248289347, -0.2565430998802185, 0.2618728578090668, 0.40481409430503845, -0.3932740092277527, -0.5565250515937805, -0.1026582196354866, 0.4295632541179657, -0.18125683069229126, 0.3854435086250305,...
I am preparing a potential interview, which can involve detailed questions on linear regression and logistic regression. Besides going over the textbook, could you share some possible interview questions on these topics based on your industrial and education experiences? Thanks.
[ 0.040658801794052124, 0.02785649336874485, -0.017373640090227127, 0.014358364045619965, -0.012189456261694431, -0.012826171703636646, 0.01363627053797245, 0.02537066489458084, -0.03615251183509827, -0.04704691842198372, 0.0016090743010863662, 0.015734493732452393, 0.007648523431271315, 0.0...
[ 0.3442985415458679, 0.2080184519290924, -0.06128066033124924, 0.4248066544532776, 0.1269393265247345, -0.23148682713508606, 0.15397775173187256, 0.5310878753662109, 0.21583318710327148, -0.5409534573554993, 0.3071746826171875, 0.11665644496679306, 0.5510411262512207, 0.2395971566438675, ...
I want to change the login URL per a client request. so, instead of mysite.com/wp-login.php or mysite.com/wp-admin it is this: mysite.com/someotherpagename.php or mysite.com/someotherpagename I do not want to use /login. It's too obvious, per the client. Thanks in advance.
[ -0.00024770511663518846, 0.00535436999052763, 0.00009729492012411356, 0.016823910176753998, -0.019667668268084526, -0.0016445720102638006, 0.011151877231895924, -0.01029704138636589, -0.028068073093891144, -0.014500442892313004, -0.008408967405557632, 0.006989424582570791, 0.0101006263867020...
[ 0.5170851945877075, 0.21948087215423584, 0.5180245041847229, -0.056851960718631744, -0.013924894854426384, -0.11218201369047165, 0.1962565779685974, 0.7224225401878357, -0.4104270935058594, -0.6844894289970398, 0.21267427504062653, -0.011224509216845036, 0.10719257593154907, 0.221088975667...
I have the following scenario. Domain Name : **mydomain.com** I have registered a country level tdl with our ISP. country tdl : **www.mydomain.com.qa** ISP has created a CNAME to my **www.mydomain.com.qa** to point to **mydomain.com**. The problem is, when I enter **www.mydomain.com.qa** on browser's address bar, it shows my default page of the webserver instead of showing **mydomain.com**. Can anybody suggest how to resolve this issue?
[ -0.008930807933211327, 0.006552140694111586, 0.004183211829513311, 0.010705267079174519, -0.031062038615345955, 0.02122860588133335, 0.0097140371799469, 0.02501959353685379, -0.015278534963726997, -0.015733305364847183, -0.015078796073794365, 0.010996758937835693, -0.014025909826159477, 0....
[ 0.4553780257701874, 0.11032209545373917, 0.8842501640319824, -0.1194726973772049, 0.0922483280301094, -0.4944758117198944, 0.363427996635437, 0.4011119604110718, -0.2315167635679245, -0.6978033185005188, -0.055713824927806854, 0.005825601518154144, 0.04400310292840004, 0.5974380373954773, ...
I am on arch-linux and just upgraded to the latest version of `gnome- terminal`. I used to have both the terminal colour and theme set by `gnome- tweak-tool` having `Global Dark Theme` turned on. Does anyone know how to change this back without downgrading my terminal. You can see here what the terminal now looks like, and a window that looks correct. ![example showing the problem](http://i.stack.imgur.com/SJzz7.png) I downgraded my terminal to the last one that did not do this, so the problem is somewhere between: gnome-terminal 3.12.0-1 and 3.10.2-1 (3.10.2-1 is the working one, these are from the arch package manager).
[ 0.01187102496623993, -0.005055747460573912, -0.009199058637022972, 0.00716294115409255, -0.023106155917048454, -0.022736575454473495, 0.005957446061074734, -0.007193983998149633, -0.012021122500300407, -0.025033893063664436, -0.014084065333008766, 0.003412432037293911, -0.005071971565485001,...
[ 0.006328575778752565, 0.4483519196510315, 0.8594995141029358, -0.259820818901062, 0.025823304429650307, -0.10657358914613724, 0.49543389678001404, 0.7239810824394226, -0.44705304503440857, -0.9291380643844604, 0.021817674860358238, 0.5145713090896606, -0.06612148135900497, 0.54263842105865...
I just created a widget where I can see all the searches like in the screenshot. I use following filter regexp to get all urls I want see: `q=([^&]+)` But it's still too much information, because I do not want see the whole url. I just want see the things between `()` like `iphone` or `samsung`. Like in other languages you can slice out a part of a string with regexp and address those parts in example with $1, $2, etc. Is there any possibility to just get them into the widget table or will I've to do it with events or something similar? ![That is what I see.](http://i.stack.imgur.com/ym3F7.png) **Edit:** As you see on this screenshot, I get the whole URL. But I just want a result like that: What I get: What I want: ------------------------------ --------------------- | /product/search?q=eos+650D | | eos+650D | ------------------------------ --------------------- | /product/search?q=iphone | | iphone | ------------------------------ --------------------- But it looks like this regexp filter `q=([^&]+)` is really just for filtering the URLs and not for better result display. So I've to do it with events or dimensions.
[ 0.012157551944255829, 0.0053793685510754585, -0.0022148971911519766, -0.006103556603193283, 0.009394245222210884, 0.016982778906822205, 0.0074915518052875996, 0.020851287990808487, -0.021935369819402695, 0.0030722017399966717, -0.004981852602213621, -0.0005003992700949311, -0.000970022752881...
[ 0.28209540247917175, -0.027845265343785286, 0.3040446639060974, 0.06515035033226013, 0.08217012137174606, 0.04326307028532028, 0.1007983461022377, 0.02615152671933174, -0.17629152536392212, -0.6714704036712646, -0.2067156434059143, 0.29811152815818787, -0.3160502314567566, 0.35750210285186...
I've been testing some dictionaries the last few days and right now I've my SDCard with plenty of directories with many MB of data that belongs to those dictionaries. How can I find out for a particular application (may be other than a dictionary) the list of directories it's using?
[ -0.016154494136571884, -0.011518309824168682, -0.024489330127835274, 0.03502468392252922, -0.015657152980566025, 0.012802853249013424, 0.008996637538075447, 0.03667576611042023, -0.022100357338786125, 0.00029525658464990556, 0.004771354608237743, 0.006280962843447924, 0.03870481997728348, ...
[ 0.4130934178829193, 0.22223089635372162, 0.1556236892938614, 0.31687504053115845, 0.18680210411548615, 0.07541239261627197, 0.2791360020637512, 0.11968231946229935, -0.0828588679432869, -0.5709001421928406, 0.024395626038312912, 0.28126803040504456, 0.23661720752716064, 0.1813187152147293,...
I've tasked myself with writing more letters (of the dead tree variety); because of this, I was intrigued by a certain take on complimentary closes, whilst enjoying a re-read of Dracula: > Your loving > > Mina And: > ...no more just at present from your loving > > Lucy Is there a name for this (admittedly subtle) variation on the usual complimentary close, wherein the signature is an extension of the close itself? I don't recall ever having seen this used in modern writing, but I'm curious as to when it fell out of favour. Naturally, it's rather hard to search for this sort of thing.
[ 0.011433985084295273, 0.0043393406085669994, -0.011588240042328835, 0.0129134152084589, 0.01912684179842472, -0.001334026688709855, 0.007761002518236637, -0.002202923409640789, -0.015290861949324608, 0.00239582359790802, -0.02028285712003708, -0.0019381820457056165, 0.011078513227403164, 0...
[ 0.18699978291988373, 0.09419690072536469, 0.42176952958106995, -0.3340015113353729, -0.41262689232826233, 0.2374839335680008, 0.29021766781806946, -0.03478957712650299, -0.28346917033195496, -0.385871559381485, -0.014331698417663574, -0.13421712815761566, -0.09099868685007095, 0.3849658966...
I have a linux router (d-link 320-nru) with tc installed. I have some users' computers to fairly share bandwidth between them. Also there is such complication as inconstant uplink bandwidth. How can I configure tc to meet all these requirements? Thanks in advance.
[ 0.010646586306393147, 0.01035220269113779, -0.02919774129986763, 0.020709551870822906, 0.01454838551580906, -0.013606718741357327, 0.012074708007276058, -0.03386969491839409, -0.03375120833516121, -0.052341051399707794, 0.005440993700176477, 0.009695030748844147, -0.010469005443155766, 0.0...
[ 0.5669529438018799, 0.17472435534000397, 0.4482104778289795, 0.45446857810020447, 0.009494555182754993, -0.1719719022512436, 0.09534543007612228, 0.25978273153305054, -0.10168085247278214, -0.6598235368728638, 0.1543349325656891, 0.4984118938446045, -0.0973825454711914, 0.10503838956356049...
Recently my server crashed. So I reinstalled Plesk and uploaded all databases to `var/lib/mysql` and site files to `backup/vhosts`. How do I get it all back to normal?
[ 0.03365060314536095, 0.02920595183968544, 0.020217886194586754, 0.007410339079797268, -0.00666078832000494, 0.0012895470717921853, 0.012889274396002293, 0.013491340912878513, -0.022395189851522446, 0.001776231685653329, -0.0043053715489804745, 0.02544436976313591, 0.018282609060406685, 0.0...
[ 0.2646021246910095, 0.033918771892786026, 0.6401614546775818, 0.024300523102283478, 0.12929503619670868, -0.11779499799013138, 0.4180174767971039, 0.45710626244544983, -0.1934323012828827, -0.5091749429702759, 0.07219795882701874, 0.32726946473121643, -0.19209083914756775, 0.37157183885574...
Yesterday I bought a brand new samsung galaxy s3. WHen I hold home button and lock button, it doesnt shot the screen. how to fix it?
[ -0.02305709384381771, -0.002168981358408928, -0.010689185000956059, 0.00564276659861207, -0.07288545370101929, -0.01946284994482994, 0.013449129648506641, 0.014202577993273735, -0.012852581217885017, -0.01931650936603546, -0.028437769040465355, 0.030933018773794174, 0.006344134453684092, 0...
[ -0.11001890152692795, 0.30202656984329224, 0.841781735420227, 0.09033828228712082, -0.03982528671622276, -0.07864658534526825, 0.48065948486328125, -0.12904223799705505, -0.23457717895507812, -0.3982493281364441, 0.17802107334136963, 0.46936866641044617, -0.22961121797561646, -0.0306190438...
Here are the differential equations that set's up the 11 coupled oscillators. new = Join[ Table[x[i]''[t] == - x[i][t] + 0.1*(x[i + 1][t] - 2*x[i][t] + x[i - 1][t]), {i, 1, 9}], {x[0]''[t] == -x[0][t], x[10]''[t] == x[9][t], x[0][0] == 1, x[0]'[0] == 1, x[1]'[0] == 0, x[1][0] == 0}, Table[x[i][0] == 0, {i, 2, 10}], Table[x[i]'[0] == 0, {i, 2, 10}]] Here are the solutions. Solt = NDSolve[new, Table[x[i], {i, 0, 10}], {t, 25}] Here are the individual plots. Table[Plot[Evaluate[x[i][t] /. Solt], {t, 0, 25}, PlotRange -> All], {i, 0, 10}] I am trying to figure out how to make a graph so along the x-axis are my i's from 0 to 10, and I can watch the wave move along each oscillator as time moves on. I keep getting errors in which it floods my notebook and doesn't stop unless I close the kernel. This is what I have so far, and I'm not sure how to incorporate time into this. Plot[Evaluate[x[i][t] /. Solt], {i, 0, 10}] EDIT Coupled in a circle Stew = Join[ Table[x[i]''[t] == - x[i][t] + 0.1*(x[i + 1][t] - 2*x[i][t] + x[i - 1][t]), {i, 1, 9}], {x[10]''[t] == - x[10][t] + 0.1*(x[0][t] - 2*x[10][t] + x[9][t]), x[0]''[t] == - x[0][t] + 0.1*(x[1][t] - 2*x[0][t] + x[10][t])}, {x[0][0] == 0, x[0]'[0] == 0, x[1][0] == 1, x[1]'[0] == 0.5}, Table[x[i][0] == 0, {i, 2, 10}], Table[x[i]'[0] == 0, {i, 2, 10}]]; The Dsolve Loin = NDSolve[Stew, Table[x[i], {i, 0, 10}], {t, 6.28}] The individual graphs Table[Plot[Evaluate[x[i][t] /. Loin], {t, 0, 6.28}, PlotRange -> All], {i, 0, 10}] How would I go about putting the i=0 to 10 around in a circle?
[ 0.006355778314173222, 0.008040039800107479, -0.0116085484623909, 0.005878480151295662, -0.017688795924186707, -0.0026840469799935818, 0.0040349713526666164, -0.002384212566539645, -0.008714154362678528, -0.010497553274035454, -0.000862765358760953, 0.007874686270952225, -0.008299829438328743...
[ -0.13372525572776794, -0.08818446099758148, 0.9143107533454895, 0.20355933904647827, -0.3830898106098175, 0.5028586983680725, 0.17768383026123047, -0.9036852717399597, -0.23658514022827148, -0.09972460567951202, 0.1688673198223114, 0.8083170056343079, -0.5837493538856506, 0.603991687297821...
I have free demo use of my application for 10 days then use automatically disable. Not delete. After he pay us for using that application for 1 year we can extend that time duration. If there is no any plugin, any snippet suggestion ?
[ 0.007713119499385357, 0.02407543547451496, -0.006397368386387825, 0.005233291070908308, -0.018314402550458908, -0.011305801570415497, 0.011238266713917255, 0.01708071120083332, -0.01718399114906788, -0.003631996689364314, -0.03143120929598808, 0.01778622902929783, 0.010032104328274727, 0.0...
[ 0.616832435131073, -0.05808098241686821, 0.3415886163711548, 0.35095009207725525, 0.11114117503166199, -0.3173227310180664, 0.5638169646263123, 0.049003515392541885, 0.026992538943886757, -0.6600862145423889, -0.23650847375392914, 0.5343565940856934, -0.10199137777090073, -0.04393015056848...
I want to check if my server has RAID configured both soft and hard. I tried some commands on my server (Suse Enterprise Edition 11): **ms:~ # cat /proc/mdstat Personalities : unused devices: <none> ms:~ #** I tried to execute `raiddev /dev/md0` but there is no raiddev command available. Is there any other command that will let me check the raid level of the server?
[ 0.02416219748556614, 0.00399048300459981, -0.011175602674484253, -0.0016691444907337427, 0.003993863705545664, 0.005183339584618807, 0.01039710734039545, -0.008481311611831188, -0.0193124208599329, -0.01678324304521084, -0.002611892530694604, 0.003966652322560549, 0.009412710554897785, 0.0...
[ 0.2866266965866089, -0.19784589111804962, 0.2269369214773178, -0.22686649858951569, -0.30319392681121826, 0.03371609374880791, 0.4825253188610077, -0.06754975020885468, -0.31268638372421265, -0.8962063193321228, 0.08570583909749985, 0.5818372964859009, -0.1518147736787796, -0.0013187087606...
I am owner of a blog/website and I am using this code for tracking post views in site : function getPostViews($postID){ $count_key = 'post_views_count'; $count = get_post_meta($postID, $count_key, true); if($count==''){ delete_post_meta($postID, $count_key); add_post_meta($postID, $count_key, '0'); return "0 View"; } return $count.' Views'; } function setPostViews($postID) { $count_key = 'post_views_count'; $count = get_post_meta($postID, $count_key, true); if($count==''){ $count = 0; delete_post_meta($postID, $count_key); add_post_meta($postID, $count_key, '0'); }else{ $count++; update_post_meta($postID, $count_key, $count); } } I want to fetch most viewed posts today, last week, last month to show in sidebar. Is there any way to sort posts based on post views. The post views is stored in a custom field post_views please guide me. Rias
[ 0.004274341277778149, 0.009911278262734413, -0.00008771533612161875, 0.013360686600208282, 0.015784945338964462, 0.0058262040838599205, 0.006498164962977171, -0.018973827362060547, -0.010668626986443996, -0.009338108822703362, -0.0075126634910702705, 0.008195577189326286, -0.0108291432261466...
[ 0.5450932383537292, 0.12391402572393417, 1.0164117813110352, -0.1190614327788353, -0.21872305870056152, 0.1325484812259674, 0.35742563009262085, -0.20013098418712616, 0.22728098928928375, -0.6844726204872131, -0.056476183235645294, 0.5146973729133606, -0.15802504122257233, 0.19977578520774...
I am employed in a very brilliant company with a true intention of doing XP. Communication is good and management is open to constructive discussion but due to pressing time constraints, some certain things are considered too RUP to be discussed. At the moment I'm a little troubled with the volume of change that becomes necessary while implementing the stories. I believe many of these discoveries (which take time and effort of course) are the responsibilities of the story writers (customers, end users and product owners) and not the developers. To put is short, user-stories are too conceptual and just convey the underlying intention but lack enough details (specially pre-conditions and post- conditions, relevance to other stories, dependencies and alike). The developer is expected to fill in the blanks at his own discretion by the virtue of XP developers being designers and analysts at the same time. The problem is many of these blanks are discovered after some wrong assumptions have made made their way into evaluation time and code since noticing added complexities emerge than initially anticipated. Even then finding the right thing to fill in takes time which is - to various degrees - considered as deviation from the initial estimations. I'm looking for a constructive way of conveying these implications to the management in a way that would not pose me as someone who is trying to unnecessarily complicate things. I'm new and as yet I have not established much credibility. You insights are most welcome. Closely related and somehow gives an answer : How much detail about a user story can a developer expect?
[ -0.001392776146531105, 0.01509552076458931, -0.007315474096685648, 0.009439467452466488, 0.019463593140244484, -0.012506447732448578, 0.006649007089436054, 0.017207155004143715, -0.010042006149888039, -0.01714928075671196, -0.011337267234921455, 0.013289570808410645, 0.007646534591913223, ...
[ 0.6712996959686279, 0.6915908455848694, 0.07634589076042175, 0.11296629905700684, 0.19484500586986542, -0.12829266488552094, 0.3722151219844818, 0.24047628045082092, -0.09379363059997559, -0.3074749708175659, -0.09961408376693726, 0.7319856286048889, 0.4806731641292572, -0.0848011896014213...
I understand that matter will gravitate toward matter. (ex: Earth gravitates a satellite toward it, and the satellite toward Earth.) Does this always apply, regardless of distance? Take two atoms, the farthest apart in the universe. Do they gravitate toward each other, but at an infinitesimal rate? (And actual position displacement is likely in a different direction, due to stronger/closer forces)
[ -0.010189047083258629, 0.007684257812798023, -0.0010132258757948875, 0.011098247952759266, -0.01704101450741291, -0.007295018527656794, 0.009845522232353687, -0.004144292324781418, -0.01511008758097887, -0.006234761793166399, -0.0064711482264101505, 0.009924476034939289, 0.014286291785538197...
[ 0.22612732648849487, -0.037453532218933105, 0.22048941254615784, 0.11090810596942902, -0.04042781516909599, 0.46853750944137573, -0.4574713408946991, -0.12795911729335785, -0.19448471069335938, -0.5292324423789978, 0.12521617114543915, 0.146534264087677, -0.08708076179027557, 0.20641635358...
What is a correct way to say it: _I hope you live for many years._ or _I hope you will live for many years_? I am really confused by it since I've seen quotes of famous writers that said _live_ but an English teacher I know says it is not the correct way. Edit: a quote by F. Scott Fitzgerald: I hope you live a life you're proud of. If you find you're not, I hope you have the strength to start all over again.
[ 0.025720415636897087, 0.005325007252395153, -0.021946676075458527, 0.008905123919248581, -0.0029672896489501, -0.028834400698542595, 0.006627680733799934, 0.008323743008077145, -0.017264917492866516, -0.013621983118355274, 0.011914009228348732, 0.009322195313870907, 0.011995282024145126, 0...
[ 0.5129265189170837, 0.33318477869033813, -0.29404589533805847, -0.16278181970119476, 0.240085631608963, -0.013120840303599834, 0.8092103004455566, 0.6213740110397339, -0.0909515768289566, -0.6994410753250122, -0.1747536063194275, 0.16362138092517853, 0.019531145691871643, 0.266067832708358...
In quantum mechanics, any density matrix (or density operator) is Hermitian. Observables are also represented by Hermitian operators. So it follows that a density matrix can also be interpreted as an observable. What is the physical meaning of this observable?
[ -0.01176980696618557, 0.020338576287031174, -0.009423194453120232, 0.03363353759050369, -0.015741387382149696, -0.04615689814090729, 0.01531771570444107, -0.003550378605723381, -0.01758265122771263, 0.0015894586686044931, -0.029602792114019394, 0.02130437269806862, 0.008391289971768856, -0...
[ 0.07342350482940674, -0.030658943578600883, 0.1202305257320404, -0.17419126629829407, -0.3448924720287323, -0.1644035130739212, -0.1463860273361206, -0.16761291027069092, 0.008957674726843834, -0.2876114845275879, -0.40434178709983826, 0.21069510281085968, -0.742439866065979, 0.52003401517...
Has any body used ITinNodeCollection to generate Voronoi/Thiessen polygons. I have a point featureclass for which I want to create Voronoi/Thiessen polygons. Does this require 3d analyst license? I appreciate your help. Thanks.
[ 0.020692378282546997, -0.0008954523946158588, -0.00470552546903491, 0.0337647907435894, -0.010829214937984943, 0.0037437535356730223, 0.014906252734363079, 0.029882214963436127, -0.03007414937019348, 0.023724475875496864, -0.0005277750315144658, 0.03167980909347534, -0.03484700247645378, -...
[ 0.4389135241508484, 0.39407584071159363, 0.13123483955860138, 0.2369588017463684, -0.023569250479340553, -0.08134070038795471, 0.20512305200099945, 0.17154720425605774, -0.11372862756252289, -0.5066831707954407, 0.10414347052574158, 0.37272152304649353, 0.169972762465477, 0.094018116593360...
I just bought a Nexus 4, and this is like my first Android phone. Can someone please help me to increase the call volume? It is really low on my phone. I tried to increase it but the maximum is too low for me.
[ 0.002258478896692395, -0.009451303631067276, -0.0122280428186059, 0.012967388145625591, -0.05404358357191086, -0.02741776779294014, 0.007958181202411652, 0.010437043383717537, -0.030153946951031685, -0.02873038686811924, -0.013800610788166523, 0.024119587615132332, -0.0058111692778766155, ...
[ 0.24630150198936462, -0.012166417203843594, 0.5938903093338013, 0.20675714313983917, 0.17829708755016327, 0.09826315194368362, 0.44533777236938477, 0.21631649136543274, -0.4187014400959015, -0.6902839541435242, 0.3610796332359314, 0.7487365007400513, -0.6934775114059448, -0.011083050630986...
I was removed from all groups before, when I changed my number. Will the same happen if I change my phone? This time, I have the same number but a different phone.
[ 0.017778335139155388, 0.017923444509506226, -0.012564040720462799, 0.03780188411474228, 0.010126814246177673, 0.004312791861593723, 0.012295051477849483, 0.03973683714866638, -0.037816550582647324, -0.059753160923719406, -0.008723529987037182, 0.03908262401819229, 0.014116207137703896, 0.0...
[ 0.3182362914085388, 0.07173927873373032, 0.37835076451301575, -0.14601871371269226, -0.2016904205083847, -0.04517585411667824, 0.47835245728492737, -0.1540992110967636, -0.3405638337135315, -0.4575043320655823, 0.1715470552444458, 0.30166512727737427, -0.48048296570777893, 0.18656376004219...
For example, "This doesn't cause any crash but later it can cause unknown memory-crashes." I see here that the subject 'it' is not immediately after 'but'. Do I still need to use a comma before 'but' as it links another subject?
[ -0.02462306246161461, 0.010390295647084713, -0.018026158213615417, 0.031236808747053146, 0.018014416098594666, 0.020247546955943108, 0.011656026355922222, -0.010744267143309116, -0.019273262470960617, 0.0017541260458528996, -0.0020810302812606096, 0.006202383898198605, 0.022529244422912598, ...
[ -0.04880411922931671, -0.1181860864162445, 0.2313145250082016, 0.11137013882398605, -0.35153552889823914, -0.3699721097946167, 0.13599787652492523, 0.021574851125478745, -0.30715930461883545, -0.41023311018943787, -0.4996902048587799, 0.41126206517219543, -0.001242366386577487, 0.159467056...
I just installed MiKTeX 2.9 x64 on my windows 8.1 laptop. I use Texmaker as a TeX editor. I'm trying to compile a document with these packages: \usepackage[utf8]{inputenc} \usepackage[spanish]{babel} \usepackage[spanish]{layout} when the compiler is running, it appears the MiKTeX Package Manager's message saying that the file `spanish.ldf` is missing and it will install the `babel- spanish` package. So, I click yes, then the package is installed, but I get the following error: (and hundred more) ! Undefined control sequence. l.117 \StartBabelCommands on line 117 but my file only has 67 lines!. On the other hand, if I compile the file clicking on "no" to the package manager, the file compile normally. How can I install the `babel-spanish` in order to compile normally my files? I tried to install this package from differents repositories and I still get the error.
[ 0.01486628595739603, 0.0008351001888513565, 0.0031662965193390846, 0.01614774391055107, 0.015690090134739876, 0.022436536848545074, 0.00837293453514576, -0.0012670084834098816, -0.015132579952478409, -0.006563705392181873, -0.018893690779805183, 0.008519411087036133, 0.011446928605437279, ...
[ -0.017156245186924934, 0.47570469975471497, 0.45693498849868774, -0.20807711780071259, 0.23592157661914825, -0.23443837463855743, 0.615058183670044, 0.21372222900390625, 0.22368384897708893, -0.5927204489707947, -0.009260809049010277, 0.85704106092453, -0.47960418462753296, 0.1165319308638...
I get output like this .. === Stratified cross-validation === === Summary === Correctly Classified Instances 85 53.125 % Incorrectly Classified Instances 75 46.875 % Kappa statistic 0.0625 Mean absolute error 0.4688 Root mean squared error 0.6847 Relative absolute error 93.75 % Root relative squared error 136.9306 % Coverage of cases (0.95 level) 53.125 % Mean rel. region size (0.95 level) 50 % Total Number of Instances 160 What is the generalization error here?
[ 0.0037051166873425245, 0.011958341114223003, -0.01373366266489029, 0.024327091872692108, 0.0073177870362997055, 0.01346316933631897, 0.0068199667148292065, 0.003408463206142187, -0.012938104569911957, 0.03313858062028885, -0.006082464009523392, 0.006341590080410242, -0.016946639865636826, ...
[ 0.027536196634173393, 0.13081492483615875, 0.4182928502559662, 0.1047583818435669, -0.1424482762813568, 0.25313499569892883, 0.6753069758415222, -0.22779759764671326, 0.057773396372795105, -0.6314970254898071, 0.11442076414823532, 0.38123375177383423, -0.008704550564289093, 0.1591827869415...
I have a site based on WordPress. This site has some functions for member only. One of the functions is a external script and I believe anyone can access this without authentication. I can add .htaccess to this script to prevent unauthorized access. But then my users have to login again and it's kind of inconvenient. Is there any solution to add some check to my script that use the same WordPress authentication mechanism, that is, if user already login, he have full access to the script. If not, he will be redirected to WordPress login page. **EDIT** : My script is written in PHP
[ 0.015615839511156082, 0.004468092229217291, 0.007049635984003544, 0.02661159262061119, 0.008545750752091408, -0.008592173457145691, 0.007693866267800331, -0.005381853319704533, -0.017986012622714043, -0.008093884214758873, -0.021025020629167557, 0.011420140042901039, -0.0021846650633960962, ...
[ 0.4817085266113281, 0.1218273863196373, 0.13684380054473877, 0.10072240978479385, -0.08571820706129074, -0.25017526745796204, 0.3676106035709381, 0.18174977600574493, -0.24951006472110748, -0.6875877976417542, 0.2479865550994873, 0.35329917073249817, -0.20890642702579498, 0.577533483505249...
I understand the basics of confidence intervals, the central limit theorem, etc, to be able to know things like given N samples of random variable, we're 68/95/99.7 percent sure the variable is within these two values. I'm wondering about how this generalizes to cases where we're trying to derive a relationship between two variables. Ie, if we're trying to measure how some variable Y depends on a variable X, one thing I guess we could do is set X to 1, then get a 95% confidence interval for X = 1. Then set X += 0.1 and do that again, and again again etc, to derive a 95% "confidence relationship", or something. This seems kind of sloppy though. Is there some other statistical mechanism designed for doing things like this (ie, finding a relationship between X and Y that we are 95% sure is the right relationship). edit: Some more info on what I'm looking for: Say I have a bunch of (x,y) pairs. I'm trying to express the y's as a function of x. x values might repeat. I also don't necessarily care about expressing the relationship algebraically. So, the naive thing to do would be to average over the ys for every possible x value we have data for, and plot that. (so if the dataset was {(1,2), (1,3), (2, 2)}, I'd plot the points {(1,2.5), (2,2)}). This totally disregards data about how much y data we got for each x value, though. It also seems to be ignoring the fact that we suspect the data to be correlated in some way. Like if our data points were {(1 + 1e-10, 4), (1, 17)}, my naive approach would plot {(1,17), (1 + 1e-10, 4)}. But considering how close 1 and 1e-10 are, we'd think that the values 4 and 17 should be close to each other. So maybe something like {(1, 10.5), (1 + 1e-10, 10.5)} would be more accurate. Or, in another sense, if our dataset was {(1,16), (1+1e-10,16.2}), the fact that we got two values so close to 16 both times should seem to "inspire our confidence". But since the x values are technically different we won't notice this.
[ 0.011861019767820835, 0.008399281650781631, -0.018710855394601822, 0.0052388980984687805, 0.0037220269441604614, -0.009126203134655952, 0.00679495744407177, -0.014283371157944202, -0.01505756564438343, -0.026356356218457222, 0.00017054204363375902, 0.007046018727123737, -0.006629279814660549...
[ 0.3680228888988495, 0.07255422323942184, -0.014002283103764057, 0.31360819935798645, 0.056969910860061646, 0.09392864257097244, 0.471064031124115, -0.27649664878845215, 0.02206723764538765, -0.30938875675201416, 0.361142098903656, 0.5884592533111572, 0.050638098269701004, 0.184592813253402...
I already found the console and tried things like `fps 1`, `show_fps`, `/fps` and `/fps on`, but nothing worked yet. I guess there is some command to display how many fps I have but **how** is it named? I use no mods or something.
[ 0.005661393981426954, 0.0026715511921793222, 0.0002885200665332377, -0.0031183939427137375, -0.017462925985455513, -0.031751349568367004, 0.008802182972431183, -0.022654831409454346, -0.03511888161301613, 0.026916585862636566, 0.0009876692201942205, 0.006988635752350092, -0.00438915984705090...
[ 0.354775071144104, 0.03728574886918068, 0.16587918996810913, 0.12116993963718414, -0.11877446621656418, 0.0940740779042244, 0.43794873356819153, 0.49071836471557617, -0.04295419901609421, -0.4329674541950226, 0.21842621266841888, 0.7590360045433044, -0.020286977291107178, 0.300054371356964...
I would like to know if it's a good practice to use transient api to cache social counters for every post. Loading post pages get quite slow while processing and getting all social APIs.
[ 0.0303265992552042, 0.02654222585260868, 0.00389179028570652, 0.023213708773255348, 0.05608941242098808, -0.01347415056079626, 0.013756904751062393, -0.03143275901675224, -0.031088486313819885, -0.05795131251215935, -0.001786976121366024, 0.020288199186325073, -0.011682686395943165, 0.0159...
[ 0.7604662775993347, 0.13008363544940948, 0.15084770321846008, 0.3693501949310303, -0.15778447687625885, 0.1319592446088791, 0.1393180787563324, 0.4555571973323822, -0.17535042762756348, -0.6733696460723877, 0.5057196021080017, 0.263869971036911, -0.12205059826374054, 0.3085474669933319, ...
I have searched for a solution for this problem for a few days now and cannot seem to find anything close to my situation (server is always NAT'd and clients are sometimes NAT'd). I am trying to create a VPN server (CentOS 6.5) to host a VPN so I can stay connected. My server sits behind a Cisco 3825 who gets it's WAN IP via Verizon FiOS (DHCP). The server's IP is 10.10.250.211/24 and I would like VPN'd users to use IP addresses in the 10.10.251.0/24 subnet. The traffic makes it into my network via Cisco static NAT (so I know the router isn't the issue): ip nat inside source static udp 10.10.250.211 500 interface GigabitEthernet0/0 500 ip nat inside source static udp 10.10.250.211 4500 interface GigabitEthernet0/0 4500 ip nat inside source static udp 10.10.250.211 1701 interface GigabitEthernet0/0 1701 /etc/ipsec.conf: version 2.0 # conforms to second version of ipsec.conf specification config setup interfaces=%defaultroute klipsdebug=none nat_traversal=yes nhelpers=0 oe=off plutodebug=none plutostderrlog=/var/log/pluto.log protostack=netkey virtual_private=%v4:10.10.250.0/24 conn L2TP-PSK authby=secret pfs=no auto=add keyingtries=3 rekey=no type=transport forceencaps=yes right=%any rightsubnet=10.10.251.0/24 rightprotoport=17/0 left=%defaultroute leftprotoport=17/%any dpddelay=10 dpdtimeout=90 dpdaction=clear /etc/ppp/options.xl2tpd: ipcp-accept-local ipcp-accept-remote ms-dns 10.10.250.213 noccp auth crtscts idle 1800 mtu 1410 mru 1410 nodefaultroute debug lock proxyarp connect-delay 5000 logfile /var/log/ppp.log /etc/xl2tpd/xl2tpd.conf: [global] listen-addr = 10.10.250.211 ; ; requires openswan-2.5.18 or higher - Also does not yet work in combination ; with kernel mode l2tp as present in linux 2.6.23+ ipsec saref = yes ; Use refinfo of 22 if using an SAref kernel patch based on openswan 2.6.35 or ; when using any of the SAref kernel patches for kernels up to 2.6.35. ; saref refinfo = 30 ; force userspace = yes ; ; debug tunnel = yes [lns default] ip range = 10.10.210.100-10.10.251.199 local ip = 10.10.251.1 ; require chap = yes refuse pap = yes require authentication = yes name = h3rrmillerVPN ppp debug = yes pppoptfile = /etc/ppp/options.xl2tpd length bit = yes Unfortunately the pluto logs are somewhat cryptic (to me, anyway): packet from 198.228.227.72:35280: ignoring Vendor ID payload [MS NT5 ISAKMPOAKLEY 00000008] packet from 198.228.227.72:35280: received Vendor ID payload [RFC 3947] method set to=109 packet from 198.228.227.72:35280: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-02_n] meth=106, but already using method 109 packet from 198.228.227.72:35280: ignoring Vendor ID payload [FRAGMENTATION] packet from 198.228.227.72:35280: ignoring Vendor ID payload [MS-Negotiation Discovery Capable] packet from 198.228.227.72:35280: ignoring Vendor ID payload [Vid-Initial-Contact] packet from 198.228.227.72:35280: ignoring Vendor ID payload [IKE CGA version 1] "L2TP-PSK"[1] 198.228.227.72 #1: responding to Main Mode from unknown peer 198.228.227.72 "L2TP-PSK"[1] 198.228.227.72 #1: OAKLEY_GROUP 20 not supported. Attribute OAKLEY_GROUP_DESCRIPTION "L2TP-PSK"[1] 198.228.227.72 #1: OAKLEY_GROUP 19 not supported. Attribute OAKLEY_GROUP_DESCRIPTION "L2TP-PSK"[1] 198.228.227.72 #1: transition from state STATE_MAIN_R0 to state STATE_MAIN_R1 "L2TP-PSK"[1] 198.228.227.72 #1: STATE_MAIN_R1: sent MR1, expecting MI2 "L2TP-PSK"[1] 198.228.227.72 #1: NAT-Traversal: Result using RFC 3947 (NAT-Traversal): both are NATed "L2TP-PSK"[1] 198.228.227.72 #1: transition from state STATE_MAIN_R1 to state STATE_MAIN_R2 "L2TP-PSK"[1] 198.228.227.72 #1: STATE_MAIN_R2: sent MR2, expecting MI3 "L2TP-PSK"[1] 198.228.227.72 #1: Main mode peer ID is ID_IPV4_ADDR: '172.20.10.2' "L2TP-PSK"[1] 198.228.227.72 #1: switched from "L2TP-PSK" to "L2TP-PSK" "L2TP-PSK"[2] 198.228.227.72 #1: deleting connection "L2TP-PSK" instance with peer 198.228.227.72 {isakmp=#0/ipsec=#0} "L2TP-PSK"[2] 198.228.227.72 #1: transition from state STATE_MAIN_R2 to state STATE_MAIN_R3 "L2TP-PSK"[2] 198.228.227.72 #1: new NAT mapping for #1, was 198.228.227.72:35280, now 198.228.227.72:58327 "L2TP-PSK"[2] 198.228.227.72 #1: STATE_MAIN_R3: sent MR3, ISAKMP SA established {auth=OAKLEY_PRESHARED_KEY cipher=aes_256 prf=oakl ey_sha group=modp2048} "L2TP-PSK"[2] 198.228.227.72 #1: Dead Peer Detection (RFC 3706): not enabled because peer did not advertise it "L2TP-PSK"[2] 198.228.227.72 #1: the peer proposed: 74.107.96.224/32:17/0 -> 172.20.10.2/32:17/0 "L2TP-PSK"[2] 198.228.227.72 #1: NAT-Traversal: received 2 NAT-OA. using first, ignoring others "L2TP-PSK"[2] 198.228.227.72 #2: responding to Quick Mode proposal {msgid:01000000} "L2TP-PSK"[2] 198.228.227.72 #2: us: 10.10.250.211[+S=C]:17/%any "L2TP-PSK"[2] 198.228.227.72 #2: them: 198.228.227.72[172.20.10.2,+S=C]:17/0===10.10.251.0/24 "L2TP-PSK"[2] 198.228.227.72 #2: transition from state STATE_QUICK_R0 to state STATE_QUICK_R1 "L2TP-PSK"[2] 198.228.227.72 #2: STATE_QUICK_R1: sent QR1, inbound IPsec SA installed, expecting QI2 "L2TP-PSK"[2] 198.228.227.72 #2: netlink_raw_eroute: WARNING: that_client port 0 and that_host port 58327 don't match. Using that_ client port. "L2TP-PSK"[2] 198.228.227.72 #2: Dead Peer Detection (RFC 3706): not enabled because peer did not advertise it "L2TP-PSK"[2] 198.228.227.72 #2: transition from state STATE_QUICK_R1 to state STATE_QUICK_R2 "L2TP-PSK"[2] 198.228.227.72 #2: STATE_QUICK_R2: IPsec SA established transport mode {ESP/NAT=>0x3d587759 <0x58e5ad8a xfrm=AES_128 -HMAC_SHA1 NATOA=172.20.10.2 NATD=198.228.227.72:58327 DPD=none} "L2TP-PSK"[2] 198.228.227.72 #1: the peer proposed: 74.107.96.224/32:17/0 -> 172.20.10.2/32:17/0 "L2TP-PSK"[2] 198.228.227.72 #1: NAT-Traversal: received 2 NAT-OA. using first, ignoring others "L2TP-PSK"[2] 198.228.227.72 #3: responding to Quick Mode proposal {msgid:02000000}
[ 0.003335898742079735, -0.007708318531513214, -0.015750078484416008, -0.0002907856833189726, -0.018597466871142387, -0.0018118247389793396, 0.006573233753442764, 0.007858457043766975, -0.013037617318332195, 0.015147751197218895, -0.009284993633627892, 0.005211124662309885, -0.0094120185822248...
[ 0.13868418335914612, 0.04848708212375641, 0.7906680107116699, 0.04231933131814003, 0.05159294605255127, -0.025646276772022247, 0.13626348972320557, 0.30385518074035645, -0.32942667603492737, -0.6678798794746399, -0.06955739110708237, 0.3409799039363861, -0.1893056035041809, 0.3408532440662...
I've bought a ps4 with some games. But the "laggy" cutscene behavior occured only with two of them so far: Tomb Raider and Watch Dogs. To be more precise, the initial cutscenes are "laggy": the one which lara gets hit on the head at the beach and the one which shows the bikers chasing Aiden (Tomb Raider and WD respectively) But the games runs smoothly as they should. The only problem is with cutscenes. And i haven't reached a point on those games to see if the problem persists with anoter cutscenes. What could be causing this? I know that the ps4 does a mandatory installation...could this be the hdd? Should i format it or something?
[ -0.009647786617279053, 0.005095330998301506, -0.0009417622350156307, 0.011347205378115177, 0.004851804114878178, -0.008280202746391296, 0.008199205622076988, -0.005596861243247986, -0.014975694939494133, 0.03567270189523697, -0.012175993993878365, 0.010787919163703918, 0.01700582355260849, ...
[ 0.2670271396636963, -0.09767063707113266, 0.1598968654870987, 0.36033785343170166, -0.31994518637657166, -0.14549694955348969, 0.03835644945502281, 0.10228154808282852, -0.19437944889068604, -0.019902421161532402, 0.03335915505886078, 0.9893554449081421, 0.069260373711586, -0.2409102171659...
I hear people complain: > my biggest gripe with google spatial apis is that they hide the fact that > latitude/longitude is not a universal coord system My question is **Is latitude / longitude a universal coordinate system?** What are the implications if it is not?
[ -0.007057292852550745, 0.003553188405930996, -0.001381859416142106, 0.02894808165729046, 0.0054367841221392155, 0.017156194895505905, 0.013661474920809269, 0.02344680018723011, -0.02729584276676178, -0.05960109829902649, -0.00650713499635458, 0.02612984925508499, 0.006179820280522108, 0.01...
[ 0.6130709648132324, 0.3831165134906769, 0.6097345352172852, 0.2162056267261505, 0.011106782592833042, -0.1522718220949173, 0.3211868107318878, 0.2680128812789917, -0.388454407453537, -0.5895423293113708, -0.05280774459242821, 0.17652907967567444, -0.028371987864375114, 0.4256736934185028, ...
I'm attempting to install Spacewalk 1.8 with mrepo and create a mirror of RHEL 5&6 packages to use for frequently spun up development VMs. There are several guides that indicate I should use mrepo with gensystemid to register individual repositories as systems on RHN and then use the entitlements granted to my rhel systems. However gensystemid fails because /usr/share/rhn/RHN-ORG-TRUSTED-SSL-CERT is not present. It's not clear to me where I get this cert, whether it is included with an official rhel install, or if it is pulled in after using rhn_register, or if this the "key" per se that you purchase with the Satellite product that allows you to mirror? Several unhelpful how-to's say it's included with CentOS and it is not, nor is it present in rhn-tools rpm.
[ 0.007799600251019001, 0.006580276880413294, -0.004200908355414867, 0.015570630319416523, -0.030857913196086884, 0.01848173886537552, 0.011344676837325096, -0.010806845501065254, -0.010490618646144867, -0.02519455924630165, -0.0027383947744965553, 0.010515223257243633, 0.0005284333601593971, ...
[ 0.44304561614990234, 0.2560584545135498, 0.5686368942260742, 0.13176323473453522, 0.040840551257133484, -0.05370132625102997, 0.13529343903064728, -0.21056675910949707, -0.4639403223991394, -0.586494505405426, -0.0012010905193164945, 0.3555455207824707, -0.22443372011184692, 0.058127500116...
I have a set of general conditions A, B, C, D, E and F to solve a problem. Each of these conditions lead me to different number of experiments according to a parameter that i change in each general condition: * condition A has 6 experiments * condition B has 24 experiments * condition C has 28 experiments * condition D has 5 experiments * condition E has 25 experiments * condition F has 8 experiments Each experiment returns a 10-dimensional vector of accuracy used in a cross- validation machine learning approach. Now I want to compare the conditions in pairs: I want to know if its better to use condition A than B, condition C than D and Condition E than F. Which statistical test can I use in this case? My suspicion is a repeated- measure based statistical one.
[ 0.005459798965603113, 0.029542790725827217, -0.02663087472319603, 0.010319911874830723, 0.011293241754174232, 0.007978534325957298, 0.010295660234987736, -0.015233960002660751, -0.012539529241621494, -0.0331958532333374, -0.004012422636151314, 0.010218147188425064, -0.006829702295362949, 0...
[ 0.3579632639884949, 0.3904643654823303, 0.18048615753650665, 0.07647641748189926, 0.2596108317375183, 0.6280717253684998, 0.38470277190208435, -0.7734919786453247, -0.179904043674469, -0.5759740471839905, -0.10342598706483841, 0.4047028124332428, -0.04190053418278694, 0.4943208694458008, ...
I am trying to get multiple shortest path using A* Algorithm available in PgRouting. I referred Solution given by UnderDark ( Link of the Stakoverflow Post ) . But it is not giving me proper Output. MY Query (Simple Query - Without Multiple Path Detection) SELECT * FROM shortest_path_astar(' SELECT gid as id, source::integer, target::integer, length(the_geom)::double precision as cost, x1, y1, x2, y2 FROM connectivity', 110, 111, false, false); Output of this Query: vetex_id | edge_id | cost 110 | 1 | 0.01235 111 | -1 | 0 Now Query as suggested by Underdark on Given link: SELECT source, target, (SELECT SUM(cost) FROM ( --- Complete Above Query Here ) ) AS foo ) AS cost FROM connectivity; Output of this Query: source | target | cost 9 | 8 | 0.01235 ... |... | ... (More 211 records ) `Note: In my test table (connectivity) I have 212 Records` So this query is giving me all the records as output without any path relativity. I do not know where I am doing wrong. any help / suggestion will be helpful for me.
[ 0.008306995965540409, 0.012644123286008835, -0.0069774240255355835, 0.0044548530131578445, -0.020440319553017616, 0.02015640400350094, 0.007089930586516857, 0.004739807453006506, -0.01420817244797945, -0.014885714277625084, -0.0008683984633535147, 0.013957113958895206, 0.00014574173837900162...
[ -0.044957175850868225, 0.10504806786775589, 0.6311677098274231, 0.0881098210811615, -0.14729024469852448, 0.343384325504303, 0.21208705008029938, -0.1162487044930458, -0.25607582926750183, -0.7220504879951477, 0.06587640196084976, 0.37016329169273376, -0.29201337695121765, 0.39578905701637...
My thesis compiles okay but the TOC/List of Figure/List of Table section texts are going out of the margin. The margin is working for the texts in rest all the chapters. I have used `\usepackage[textwidth=13cm,textheight=20cm]{geometry}`. \documentclass[12pt]{report} \usepackage{dcolumn}% Align table columns on decimal point \usepackage{bm}% bold math \usepackage{MnSymbol} \usepackage{html} \usepackage{url} \usepackage[none,dcucite,abbr]{harvard} \usepackage{graphicx} \usepackage{amsmath} \usepackage{uathesis} \usepackage{lscape} \usepackage{longtable} \usepackage{cite} \usepackage{epsfig} \usepackage{upgreek} \usepackage[textwidth=13cm,textheight=20cm]{geometry} \newtheorem{theorem}{Theorem} \newtheorem{acknowledgement}[theorem]{Acknowledgement} \newtheorem{algorithm}[theorem]{Algorithm} \newtheorem{axiom}[theorem]{Axiom} \newtheorem{case}[theorem]{Case} \newtheorem{claim}[theorem]{Claim} \newtheorem{conclusion}[theorem]{Conclusion} \newtheorem{condition}[theorem]{Condition} \newtheorem{conjecture}[theorem]{Conjecture} \newtheorem{corollary}[theorem]{Corollary} \newtheorem{criterion}[theorem]{Criterion} \newtheorem{definition}[theorem]{Definition} \newtheorem{example}[theorem]{Example} \newtheorem{exercise}[theorem]{Exercise} \newtheorem{lemma}[theorem]{Lemma} \newtheorem{notation}[theorem]{Notation} \newtheorem{problem}[theorem]{Problem} \newtheorem{proposition}[theorem]{Proposition} \newtheorem{remark}[theorem]{Remark} \newtheorem{solution}[theorem]{Solution} \newtheorem{summary}[theorem]{Summary} \newenvironment{proof}[1][Proof]{\textbf{#1.} }{\ \rule{0.5em}{0.5em}} %\input{tcilatex} \begin{document} \tableofcontents \listoffigures \listoftables \bodyoftext \doublespacing \chapter{Introduction} \chapter{other chapters} \chapter{Summary and Conclusions} \clearpage\addcontentsline{toc}{chapter}{Bibliography} \bibliographystyle{unsrt} \bibliography{aaa} \bigskip \end{document}
[ 0.0024529260117560625, -0.0014382220106199384, -0.01663030870258808, 0.017349116504192352, 0.0016513755545020103, 0.008652938529849052, 0.008082624524831772, 0.005478433333337307, -0.012155982665717602, -0.0170583538711071, -0.018499333411455154, -0.0006326012080535293, -0.009319535456597805...
[ -0.12271885573863983, 0.5188185572624207, 0.8457998633384705, -0.178824782371521, -0.011725480668246746, 0.08524706214666367, 0.3769693374633789, -0.25462087988853455, -0.12289517372846603, -0.3893064856529236, 0.11873143911361694, 0.4571864902973175, 0.4173606038093567, 0.1703565269708633...
Does anyone know of any good example code illustrations for the rulefit Rule Based Learning Ensembles package? The documentation is incredibly lacking. I was guided to the package by this paper. If anyone who is familiar with the model could give a brief interpretation of how the Rule based method works (and why it works well or not), it would also be appreciated. The results in the paper look very promising. Also, are there any other alternative ensemble packages that perform as well or have similar functionality? I'm familiar with Caret, although, I don't know if any of the Caret methods are necessarily comparable to the above. For instance, Caret uses linear correlation for the variable importance selection, whereas the Rule Based Method seems to use interaction statistics and partial dependence functions and plots to determine importance.
[ -0.008002493530511856, 0.01104634627699852, -0.015396482311189175, 0.022997260093688965, 0.030257046222686768, 0.0035103727132081985, 0.006149514578282833, -0.0003325416473671794, -0.01400888804346323, 0.007548018358647823, -0.011787078343331814, 0.010252087377011776, 0.00776283722370863, ...
[ 0.4535468816757202, -0.05706845596432686, -0.14342443645000458, 0.17674435675144196, 0.22288832068443298, -0.08833885937929153, 0.14891625940799713, -0.015104562975466251, -0.17052778601646423, -0.5041618943214417, 0.2705228924751282, 0.6372252702713013, 0.09683816879987717, -0.05849497392...
OS: FreeBSD9 64 Bit MTA: EXIM4 with TLS with Self Signed Certificate. I am using telnet myserver.com 25 EHLO dummy@dummy.com STARTTLS It says TLS go ahead And then I issue MSG FROM: me@me.com It says ** 554 Security Error ** _It first says tls is ok go ahead then produces error!!!!!_ Can someone figure out What is the problem. And/Or provide necessary solution. If more information is required please let me know I tested my tls here https://www.wormly.com/test_smtp_server, It produced: Resolving hostname... Connecting... SMTP -> FROM SERVER: SMTP -> FROM SERVER: SMTP -> ERROR: EHLO not accepted from server: SMTP -> FROM SERVER: SMTP -> ERROR: HELO not accepted from server: Message sending failed.
[ -0.00517700333148241, -0.01351175643503666, -0.006903939414769411, -0.000329849892295897, -0.005975991487503052, 0.031067606061697006, 0.00922437198460102, 0.0149669349193573, -0.01748666912317276, -0.03818192332983017, -0.016534797847270966, 0.0016169670270755887, 0.010638143867254257, 0....
[ 0.3972015380859375, 0.1947583705186844, 0.3412976562976837, 0.05547745153307915, -0.16750891506671906, 0.029536057263612747, 0.5335063338279724, 0.1651052087545395, -0.3263203799724579, -0.7044315934181213, -0.4368079900741577, 0.46992024779319763, -0.33015817403793335, -0.0624404400587081...
I often hear about how general relativity is very complicated because of all forms of energy are considered, including gravitation's own gravitational binding energy. I have two questions: 1. In general relativity, objects following the motion of gravitation should simply be travelling by geodesics. In such 'free fall', why would there be any 'binding energy'? 2. From Einstein's field equations, $$R_{\mu \nu} - {1 \over 2}g_{\mu \nu}\,R = {8 \pi G \over c^4} T_{\mu \nu},$$ isn't the curvature only coupled to the energy-momentum tensor? As far as I understand, potential energy is not included inside the energy-momentum tensor.
[ 0.006904738023877144, 0.019795913249254227, -0.0020396385807543993, 0.008128251880407333, 0.0030899825505912304, -0.0036918718833476305, 0.009304545819759369, -0.0185876302421093, -0.011373801156878471, -0.022203173488378525, -0.0025000493042171, 0.017527850344777107, -0.01161782257258892, ...
[ 0.12023940682411194, 0.021826570853590965, 0.3513800799846649, 0.21803537011146545, -0.0416036956012249, -0.0426509752869606, -0.28119519352912903, -0.14996908605098724, -0.16951577365398407, -0.34886032342910767, -0.1830223947763443, 0.2498200535774231, -0.37373194098472595, 0.71483594179...
When you're playing "random 2v2" (random partner, not race), can you be paired against a team of two players that are in a pre-arranged team and not randomly assigned?
[ -0.0030092026572674513, 0.019433144479990005, 0.00759231997653842, -0.0031767881009727716, -0.010393111035227776, 0.011646864004433155, 0.01602017879486084, 0.008069191128015518, -0.019082169979810715, 0.03339989483356476, -0.028972690925002098, 0.03818820044398308, -0.00067073421087116, 0...
[ 0.12659437954425812, -0.5897098779678345, -0.26330071687698364, 0.31610968708992004, -0.025075826793909073, -0.16637644171714783, -0.20110484957695007, -0.10194680094718933, -0.25199806690216064, -0.44480058550834656, 0.08436533063650131, 0.1723146289587021, -0.36539116501808167, -0.541950...
I've been looking around at this for solutions but strangely I can't seem to find one that appears decent - maybe I haven't looked hard enough. What I want is a notification on my dashboard when posts are pending review. I can't believe this doesn't seem to be core behaviour for WP. I so far can't seem to find any recent plugins for this, maybe it's a setting I haven't noticed on WP. Any help gratefully appreciated. Jonathon
[ -0.003131855046376586, 0.0011119443224743009, -0.003595061367377639, 0.028879165649414062, 0.02219386398792267, -0.004618580918759108, 0.0035524615086615086, -0.001606179866939783, -0.011735105887055397, 0.022741343826055527, -0.005206785164773464, 0.006120037753134966, -0.006435632705688476...
[ 0.6687771677970886, 0.3697185814380646, 0.06363941729068756, 0.10966281592845917, -0.08259593695402145, -0.040893204510211945, 0.5498309135437012, 0.8145018815994263, -0.3619144558906555, -0.5059871077537537, 0.03246954828500748, 0.06018717586994171, -0.13061001896858215, -0.15731598436832...
In the wordpress theme I am developing I inserted some code to display links to the previous and next posts which appears at the bottom of each post. The code looks like this.. <?php previous_post_link(); ?> <?php next_post_link(); ?> The problem is that these arrows << and >> appear next to them. Does anyone know how to get rid of them? Any help will be appreciated.
[ 0.0026174590457230806, 0.003511646296828985, -0.004754027351737022, 0.02274801768362522, -0.007825750857591629, -0.005785747896879911, 0.006730805616825819, 0.017152082175016403, -0.017604010179638863, 0.023318203166127205, -0.030201461166143417, 0.007755665108561516, 0.005746536422520876, ...
[ 0.05941102281212807, -0.003962945193052292, 0.7152774333953857, 0.10303841531276703, -0.16933783888816833, 0.07599025964736938, 0.22445173561573029, 0.21448782086372375, -0.16358442604541779, -0.798839271068573, 0.10569349676370621, -0.08040887117385864, -0.359868586063385, 0.3125089406967...
I have the sentence "The role has gotten easier and easier to play, and been forced upon me more frequently than not." I was wondering if "and been" is correct to use in this case, or if "has" must be used again i.e. "and been" or "and has been?" Thanks!
[ -0.004805308301001787, 0.022655362263321877, -0.03295576572418213, 0.016013866290450096, 0.0028980483766645193, -0.008847911842167377, 0.010804742574691772, -0.02835848554968834, -0.022343697026371956, -0.04004944488406181, -0.004820639733225107, 0.010672552511096, 0.016344211995601654, 0....
[ -0.2607068419456482, 0.05228239297866821, 0.4178048074245453, -0.2085389792919159, 0.0014402266824617982, 0.040863145142793655, -0.12848784029483795, 0.2520720660686493, -0.3478868305683136, -0.4788261353969574, -0.011337644420564175, 0.5204669237136841, 0.4743717908859253, 0.2957909107208...
I just ran the OSGeo4W installer to update my previous installation and I noticed that it was updating my QGIS installation from 1.8 to 2.0. I haven't been able to find any information on the web about QGIS 2.0 being released yet. The installer ran without any issues, but now that it is complete I cannot find executables for QGIS. The folder for QGIS exists within the `apps` folder under `C:/OSGeo4W/`, but there are no executables in there. I've already completely uninstalled all OSGeo4W applications and reran the installer using the Express Install option, but I had the same result. How can I use QGIS and has version 2.0 been released yet?
[ -0.02864721231162548, -0.007996347732841969, -0.0028215842321515083, 0.029054168611764908, 0.02068796381354332, 0.019762692973017693, 0.0099895428866148, 0.01729045808315277, -0.02081320248544216, -0.011615081690251827, -0.00498526357114315, 0.015504450537264347, -0.020488642156124115, -0....
[ 0.5755854845046997, 0.2215256541967392, 0.43429890275001526, -0.1489751785993576, -0.0880507230758667, -0.1988501399755478, 0.47689124941825867, 0.12486983090639114, -0.14972899854183197, -0.7306005954742432, -0.10360229015350342, 0.8200842142105103, -0.3396403193473816, 0.0405775234103202...
There is clear meaning of _Pearson product-moment correlation coefficient_ : > it is cosine of angle between two vectors based on variables. Also there are 12 other ways to clearify the meaning of _Pearson correlation_. Is there any similar for _Polychoric correlation coefficient_? Besides just "correlation between two ordinal variables" and without complicated formulas.
[ -0.0021041198633611202, 0.029404008761048317, -0.0054717594757676125, 0.02797522582113743, -0.015425552614033222, -0.013079863972961903, 0.010080034844577312, -0.028051411733031273, -0.017968105152249336, 0.016784369945526123, -0.014188612811267376, 0.018463337793946266, 0.02014649473130703,...
[ 0.4444843530654907, -0.03284508362412453, 0.45718684792518616, 0.30641669034957886, -0.4319460093975067, -0.13641300797462463, -0.24433617293834686, -0.06152508407831192, -0.1899995505809784, -0.3192533552646637, 0.3282455801963806, 0.40124383568763733, -0.15623098611831665, 0.333937019109...
So, I was doing something like this: `\cfoot{\color{color2}\itshape John~~Doe}` to get my name centered on bottom of each page. Now I'd like to have it in one line with page number. Now it looks like that: PAGE/PAGE John Doe And I'd like to look it like this: John Doe PAGE/PAGE My name, and numbered page in one line. With `\lfoot` or `\cfoot` I can't achieve that. Also, how to start page numbering from page 2? `\setcounter{page}{2}` does not works. Thanks in advance. Minimal working example (`\newpage`'s were added ONLY to force page numbers - I don't use them in original document!): \documentclass[11pt,a4paper]{moderncv} \moderncvstyle{classic} \moderncvcolor{orange} \usepackage[scale=0.75]{geometry} \usepackage{fontspec} \firstname{John} \familyname{Doe} %\title{} \address{street}{postcode city} \mobile{+4~(11)~311~111} %\phone{+2~(345)~678~901} %\fax{+3~(456)~789~012} \email{mail@yahoo.com} %\homepage{www.johndoe.com} \extrainfo{additional info} \photo[64pt][0.4pt]{picture.png} %\quote{} \cfoot{\color{color2}\itshape John~~Doe} \begin{document} \section{Expirience} \subsection{My work} \cvcomputer{One}{Bla, Bla, Bla}{Two}{Bla,Bla,Bla} \newpage \section{Foreign Languages} \cvitemwithcomment{English}{Fluent}{Some description here} \newpage \section{Other Skills} \cvitemwithcomment{Computers}{Very good}{Some description here} \end{document}
[ 0.00581252109259367, 0.01583792269229889, -0.015508539043366909, 0.014166331849992275, -0.00905837956815958, 0.00895149540156126, 0.00945274718105793, -0.0014084549620747566, -0.015050213783979416, 0.012875750660896301, -0.005189971532672644, 0.003910152241587639, 0.00712514016777277, 0.00...
[ 0.13118109107017517, 0.3255002498626709, 0.5767683982849121, 0.025416376069188118, -0.25640642642974854, 0.34989801049232483, -0.2381744682788849, 0.13282817602157593, -0.5703598856925964, -0.9096173644065857, 0.13015194237232208, -0.010426357388496399, -0.36090970039367676, 0.112680509686...
There is already one answered the question at Best way to use a shell with Cygwin in Windows 7. That is to start Cygwin with `C:\cygwin\bin\mintty.exe -` With that `$SHELL = /bin/sh` How can I launch mintty with `/usr/bin/bash` as default shell?
[ -0.011540893465280533, 0.011590656824409962, -0.020542891696095467, 0.008249837905168533, 0.0016400297172367573, -0.013929941691458225, 0.010889180935919285, -0.004696019925177097, -0.01828804612159729, -0.030887501314282417, -0.020704219117760658, 0.004708923865109682, -0.026425499469041824...
[ 0.23453351855278015, 0.09671010822057724, 0.4697873890399933, 0.12636850774288177, -0.0032253277022391558, -0.4329352080821991, 0.014275704510509968, 0.22163799405097961, 0.04853031784296036, -0.23807744681835175, -0.07949831336736679, 0.7605064511299133, -0.059728045016527176, 0.188316240...
Consider this below clause(or whatever it is called) > This is correct grammer. My friend and I are having an argument about the correctness of the below two sentences which refer to the above clause. > 1. I see only a spelling mistake. > 2. I see only one spelling mistake. > She says both are correct, but I feel only the second one is correct. Who is right? Edit: Is a spelling mistake also a grammatical mistake?
[ 0.012283585034310818, 0.03279499337077141, -0.015765821561217308, 0.021163491532206535, 0.024748893454670906, -0.013713473454117775, 0.010771295987069607, -0.02606792189180851, -0.016677049919962883, 0.001437688828445971, -0.0007114349282346666, -0.00971092376857996, -0.012236101552844048, ...
[ -0.11117683351039886, 0.43738481402397156, 0.040373172610998154, -0.03572217375040054, -0.22928008437156677, -0.009528747759759426, 0.525862455368042, -0.4195324778556824, -0.3159612715244293, -0.376361608505249, 0.25770002603530884, 0.31522491574287415, -0.29550257325172424, 0.06687655299...
I have had facebook synced perfectly almost a year now. Yesterday I was at a concert and my S3 got all buggy from the heat. Today I had to resync all my accounts to my contact list, but it won't add facebook. Here is what happens: 1. I hit "Add Account" 2. I hit "Facebook" (which is there and not highlighted just like you would expect) Then the screen goes black (as if loading facebook) for two seconds, before going back to the "Accounts" menu with no facebook account. I have tried opening facebook first and once it sent me to facebook before again doing nothing. I have tried restarting my phone. I have added another account (Podio) which worked fine. Help?
[ 0.0010299843270331621, -0.0064468239434063435, 0.0005230706883594394, 0.014574088156223297, 0.0037711842451244593, -0.01611941307783127, 0.005674668587744236, 0.013339965604245663, -0.01279861107468605, -0.017398903146386147, -0.009027154184877872, 0.015103258192539215, -0.002907436573877930...
[ 0.4180576801300049, -0.03401640057563782, 0.47898101806640625, -0.14465969800949097, -0.1876150369644165, 0.12242703884840012, 0.5940582752227783, 0.38407203555107117, -0.47158536314964294, -0.6881096959114075, 0.25723734498023987, 0.45054954290390015, -0.26657477021217346, 0.2638021111488...
My phone turns on, I hear notifications when I receive SMS (as sound works) but the screen is damaged. Nothing displays at any time and I cannot interact with touch. I have replaced my damaged phone and wish to factory reset it. How can I do this without the screen displaying? I don't think I can load an app without being able to interact with the display.
[ -0.00648803636431694, -0.0009663620148785412, -0.000070769659942016, -0.0017862431704998016, -0.024474311619997025, -0.023757461458444595, 0.009345224127173424, -0.0012116534635424614, -0.020009079948067665, -0.012659371830523014, -0.025334656238555908, 0.014509845525026321, 0.00306857586838...
[ 0.33512991666793823, 0.4938404858112335, 0.5236343741416931, -0.3451869785785675, 0.048450615257024765, -0.01853717491030693, 1.09065842628479, 0.11437520384788513, -0.2963158190250397, -0.5239779949188232, 0.18121856451034546, 0.590216338634491, -0.1564858853816986, 0.41971707344055176, ...
> **Possible Duplicate:** > Could I buy a domain name to increase traffic to my site like this? Let's assume we have a generic website " **MyWebsiteAboutGamesAndThings.com** ". We discover that a lot of people are searching for " **Make games** ". We find that the domain name **MakeGames.com** is free so we register it with the assumption that a keyword specific domain such as this will rank favorably for the search term we have discovered. We now have two options with the domain name: * We can 301 redirect **MakeGames.com** to **MyWebsiteAboutGamesAndThings.com** * Or we can make a simple webpage on it that intends to feed visitors to **MyWebsiteAboutGamesAndThings.com**. The reason for registering **MakeGames.com** was to target that specific search, however I have a feeling that you lose the benefit of the keywords in the domain by 301ing it. Is it better to make a sort of landing website for **MakeGames.com**? If you 301 the domain to **MyWebsiteAboutGamesAndThings.com** will it now rank better for the search **Make games**? Which is a better method?
[ -0.013464625924825668, 0.0012053963728249073, 0.007743632420897484, 0.0023185566533356905, 0.009849917143583298, -0.008534318767488003, 0.007201792672276497, 0.010648950934410095, -0.01326393336057663, 0.0063487933948636055, -0.01063617691397667, 0.016212671995162964, 0.01705784909427166, ...
[ 0.7597876787185669, 0.11474048346281052, 0.1291830688714981, 0.43390700221061707, 0.20083175599575043, 0.07206852734088898, 0.07114318013191223, 0.09327994287014008, -0.47214049100875854, -0.6268523335456848, 0.3485056459903717, 0.41019389033317566, -0.0948072299361229, 0.5001047253608704,...
> A value that is _expected_ might be nounized to an _expectee_ awaited by the > _expector_. Is there a formal term corresponding to that for the adjective _actual_? > A value that is _actual_ might be nounized to a _what_ received by the > _whom_. My suggestions for _what_ are either _actual_ or _actuee_ and for _whom_ they're _actuator_ or _actuator_. Am I anywhere close to being correct? Is it understandable to NSEs (if yet a bit artificial and formal)? What'd be a better way to express these formally?
[ 0.008521412499248981, 0.0132764196023345, 0.014903875067830086, 0.01786835491657257, 0.010870790109038353, -0.021753540262579918, 0.008277732878923416, -0.0047873989678919315, -0.01136310026049614, 0.011269433423876762, -0.007171096745878458, 0.009895969182252884, 0.01520885992795229, 0.00...
[ -0.034959279000759125, -0.15782645344734192, 0.4459270238876343, 0.07229289412498474, -0.46559372544288635, 0.34275364875793457, 0.4435668885707855, -0.14613503217697144, -0.05519520118832588, -0.4819453954696655, -0.013400939293205738, 0.46108946204185486, 0.12155868113040924, -0.20466583...
I am new to LaTeX. I searched ctan.org, but didn't find what I want. I want to typeset a chemical reaction equation. The `=` mark seems too short. I need a long one. I tried to add an option `2cm` to get a long `=` but failed. Here is the raw code: \documentclass{beamer} \begin{document} \begin{frame} $H_{2}O =[2cm] H_{2}\uparrow+O_{2}\uparrow$ \end{frame} \end{document} The output screenshot is here: ![Screenshot](http://i.stack.imgur.com/nEvzv.png) In my opinion, the font doesn't look good. Would you give me some advice? For example, is there some font which would be appropriate for typesetting chemical reaction equations?
[ 0.002031224314123392, 0.01218643132597208, -0.0019250297918915749, 0.009307058528065681, 0.0034333132207393646, 0.003833124181255698, 0.0084514319896698, 0.013750549405813217, -0.014058523811399937, -0.011274481192231178, -0.0015722726238891482, -0.00008299585897475481, 0.002868247451260686,...
[ 0.2547098696231842, 0.1404951810836792, 0.6395970582962036, -0.2163282334804535, 0.28089264035224915, -0.04367118328809738, 0.020200179889798164, -0.4580821096897125, 0.07139192521572113, -0.6160075664520264, -0.18982848525047302, 0.5091120004653931, -0.3395610749721527, 0.1684908568859100...
I tried adding my corporate Exchange user to my KitKat tablet, so I can sync email and calendar. The security policy page that came up states the admins in my company will be allowed to block my device, remotely erase it, change my password, etc. - if I accept the policy. Proceeding without accepting the policy is impossible (obviously). Since this is a private device that I intend to use personally as well, my question is: can I create a separate user on the device, create the Exchange account under that user, accept whatever policy is required - but not have it applied to the entire device, only to that specific user profile? Or is a policy applied to one user applicable to the entire device?
[ 0.011058498173952103, 0.0001334487460553646, -0.0067688655108213425, 0.015061160549521446, -0.012064512819051743, 0.0030087244231253862, 0.0084915142506361, -0.0003977857995778322, -0.01318824477493763, -0.01936684362590313, -0.004902328364551067, 0.014822667464613914, -0.012095438316464424,...
[ 0.2444850653409958, 0.3123709261417389, 0.2772200405597687, -0.06365706026554108, 0.20437827706336975, 0.28603604435920715, -0.02811829186975956, 0.049410805106163025, -0.06557665765285492, -0.2879202961921692, 0.09083189815282822, 0.20731984078884125, -0.47870302200317383, 0.0977593660354...
First off, I am a WordPress newbie so please be gentle! :) Long story short, I have been given a project in mid-flight. Development computer has Wordpress 3.8.1 setup with Microsoft Webmatrix. A purchased theme with customizations and content exists on this box. The client has an already purchased domain and hosting package from network solutions. Network solutions states they using Wordpress 3.5.1. When it comes to deploying from local to the web to the Network Solutions site, will I run into major pain points with the site being on a newer version? Maybe I'm being overly paranoid but I don't want to tell my developer to keep working on this site if we have to blow everything away and revert to an older version or work entirely online within Network Solutions. **UPDATE: Network Solutions does allow you to upgrade Wordpress to the latest stable version which meant I didn't have to worry about this anymore.**
[ -0.012029631063342094, -0.010714315809309483, -0.005476972088217735, 0.019618358463048935, -0.010278796777129173, 0.007953174412250519, 0.008274763822555542, 0.01180504634976387, -0.0124838687479496, -0.031841620802879333, -0.010067906230688095, 0.013585196807980537, 0.007479229476302862, ...
[ 0.29833534359931946, 0.05191706493496895, 0.5487862229347229, -0.01089148223400116, -0.10301783680915833, 0.20782655477523804, 0.13730010390281677, 0.13989458978176117, -0.16853837668895721, -0.6662841439247131, 0.07575742900371552, 0.481143981218338, -0.010572553612291813, 0.3263193666934...
I have a few products with no images but they template leaves a big empty space where the image should be, I was wondering if there was a way to hide this product gallery section only IF there are NO images uploaded to that product. Thanks Badger
[ 0.018524719402194023, 0.00007258196274051443, 0.00975926872342825, 0.04814901575446129, -0.017561307176947594, 0.004449900239706039, 0.010157131589949131, -0.02185751311480999, -0.022807477042078972, 0.01168935839086771, -0.026226937770843506, 0.006091313902288675, 0.00339046330191195, 0.0...
[ 1.017185091972351, 0.3465306758880615, -0.07931052893400192, 0.13334712386131287, -0.46480464935302734, 0.2687603235244751, 0.502193033695221, 0.6955966949462891, -0.2600114047527313, -0.32784321904182434, 0.555452823638916, 0.7403962016105652, 0.4931544065475464, 0.49529731273651123, -0...
I want to print list of numbers from 1 to 100 and I use a for loop like the following: number=100 for num in {1..$number} do echo $num done When I execute the command it only prints {1..100} and not the list of number from 1 to 100.
[ -0.0027054608799517155, 0.006505539640784264, -0.018486451357603073, 0.004312233068048954, -0.01758469082415104, 0.006897814106196165, 0.01154884323477745, -0.021001815795898438, -0.028514115139842033, -0.030763093382120132, -0.013947664760053158, 0.00801458302885294, -0.03975103050470352, ...
[ 0.2833820879459381, 0.16951608657836914, 0.20750674605369568, -0.37400293350219727, -0.017430486157536507, 0.12224697321653366, 0.39667651057243347, -0.6209645867347717, -0.2013847380876541, 0.04486658051609993, 0.33295437693595886, 0.2210618108510971, -0.24506798386573792, 0.2160023599863...
In Starcraft/BroodWar is there a way to change the default hotkeys? I'm almost completely sure that there is no way in-game to change them. But is there a mod or any 3rd party software that can modify them?
[ 0.019620897248387337, -0.0034246500581502914, -0.009939541108906269, 0.018744254484772682, 0.0302862711250782, -0.017497843131422997, 0.01109632570296526, -0.02287713624536991, -0.029172981157898903, 0.06600354611873627, 0.003924139775335789, 0.043049268424510956, 0.004758720751851797, 0.0...
[ 0.32434067130088806, -0.01350152026861906, 0.1716669648885727, 0.20709359645843506, 0.09536398202180862, -0.13833045959472656, -0.27491962909698486, 0.2740916609764099, -0.3294209837913513, -0.433397114276886, 0.07655380666255951, 0.4819521903991699, -0.13396653532981873, -0.05590609833598...
I'm trying to stick a post in the search results for the value and meta_key meta_value. If the search results in the post is meta_key - sticky and meta_value - fatured, the display on top of these posts. This is my search.php : <?php get_header(); ?> <?php if (is_paged()) $is_paged = true; ?> <div class="wrapper" > <div class="clearfix container_border"> </div> <?php $is_search = 0; global $wpdb; $totalpost_count = 0; $all_pids = $wpdb->get_var("SELECT group_concat(ID) FROM $wpdb->posts where post_status='publish'"); $all_pids_arr = explode(',',$all_pids); if($_REQUEST['srch_location']) { $is_search = 1; $srch_location = $_REQUEST['srch_location']; //$location_pids = $wpdb->get_var("SELECT group_concat(tr.object_id) FROM $wpdb->term_taxonomy tt join $wpdb->term_relationships tr on tr.term_taxonomy_id=tt.term_taxonomy_id where tt.term_id=\"$srch_location\""); $location_pids = $wpdb->get_var("select group_concat(post_id) from $wpdb->postmeta where meta_key like 'add_location' and meta_value like \"$srch_location\""); $location_pids_arr = explode(',',$location_pids); $all_pids_arr = array_intersect($all_pids_arr,$location_pids_arr); } if($is_search && !$all_pids_arr) { $all_pids_arr[0] = 'nopost'; } if($_REQUEST['srch_property_id']) { $post_ids_str = $_REQUEST['srch_property_id']; $sub_cat_sql .= " and p.ID in ($post_ids_str) "; }else { if($all_pids_arr) { $post_ids_str = implode(',',$all_pids_arr); if($post_ids_str) { $sub_cat_sql .= " and p.ID in ($post_ids_str) "; } } } $featurecat = get_cat_id_from_name(get_option('ptthemes_featuredcategory')); if($featurecat) { $srch_feature_pids = $wpdb->get_var("SELECT group_concat(tr.object_id) FROM $wpdb->term_taxonomy tt join $wpdb->term_relationships tr on tr.term_taxonomy_id=tt.term_taxonomy_id where tt.term_id in ($featurecat)"); $srch_feature_pids = ''; } $blogcat = get_cat_id_from_name(get_option('ptthemes_blogcategory')); $blogcatcatids = get_sub_categories($blogcat,'string'); if($blogcatcatids) { $srch_blog_pids = $wpdb->get_var("SELECT group_concat(tr.object_id) FROM $wpdb->term_taxonomy tt join $wpdb->term_relationships tr on tr.term_taxonomy_id=tt.term_taxonomy_id where tt.term_id in ($blogcatcatids)"); } if($srch_blog_pids && $srch_feature_pids) { $sub_cat_sql .= " and p.ID not in ($srch_blog_pids,$srch_feature_pids) "; }elseif($srch_blog_pids && $srch_feature_pids=='') { $sub_cat_sql .= " and p.ID not in ($srch_blog_pids) "; }elseif($srch_blog_pids=='' && $srch_feature_pids) { $sub_cat_sql .= " and p.ID not in ($srch_feature_pids) "; } $srch_sql = "select p.* from $wpdb->posts p $post_meta_join where p.post_status='publish' and p.post_type='post' $sub_cat_sql"; if($srch_feature_pids) { $feature_srch_sql = "select p.* from $wpdb->posts p where p.post_status='publish' and p.post_type='post' and p.ID in ($srch_feature_pids)"; $srch_sql = " select * from (($feature_srch_sql) union ($srch_sql))"; } $totalpost_count = $wpdb->get_var("select count(p.ID) from $wpdb->posts p $post_meta_join where p.post_status='publish' and p.post_type='post' $sub_cat_sql"); global $posts_per_page,$paged; if($paged==''){$paged=1;} $startlimit = $posts_per_page*($paged-1); $srch_sql .= " order by p.ID desc limit $startlimit , $posts_per_page"; $post_info = $wpdb->get_results($srch_sql); ?> <div class="contentarea"> <h5> <?php if($_REQUEST['s'] == 'viewmore') { _e(LATEST_PROPERTIES_TEXT); } elseif(is_category() && $_REQUEST['search']=='') { echo single_cat_title(); }else { //echo __(SEARCH_TEXT). get_search_param(); echo get_search_param(); } ?></h5> <?php if($post_info) { ?> <ul class="display "> <?php $count=0; foreach($post_info as $post_info_obj) { $count++; $post = $post_info_obj; get_property_info_li($post); if($count%3==0) { ?> <li class="blank"></li> <?php } } ?> </ul> <?php }else { _e(NO_PROPERTY_AVAILABLE_MSG); if($_POST['search']=='search') { echo get_search_param(); } } ?> <?php if($post_info) { ?> <div class="pagination"> <?php if (function_exists('wp_pagenavi')) { ?><?php wp_pagenavi(); ?><?php } ?> </div> <?php }?>
[ -0.01765740104019642, 0.015134459361433983, 0.010326934047043324, 0.020597299560904503, 0.0034954766742885113, 0.01287478394806385, 0.009022541344165802, -0.0161326564848423, -0.014495206996798515, 0.008258465677499771, -0.008006146177649498, 0.007497224025428295, -0.018978215754032135, 0....
[ 0.37316977977752686, -0.05403916910290718, 0.6897810101509094, -0.14191794395446777, -0.004107260145246983, -0.019018519669771194, 0.0872497409582138, 0.12185201048851013, -0.22216235101222992, -0.8735490441322327, -0.04901560768485069, 0.09353649616241455, -0.5291005969047546, 0.187844425...
Based on code the awesome G.M. provided in an earlier question (Use Custom Fields to Create Guest Author Link), I now have a custom function that uses custom fields to support guest authors. The code changes the post author to display the guest author name and changes the author link to the guest author URL of my choosing. This is the code I've added to my theme's function.php file to support guest authors: //Guest Author Handling - Uses custom fields for guest author name and URL. Changes the guest author name display and changes the author link to guest author URL of your choosing add_filter( 'get_the_author_user_url', 'guest_author_url' ); add_filter( 'the_author', 'guest_author_link' ); add_filter( 'get_the_author_display_name', 'guest_author_name' ); function guest_author_url($url) { global $post; $guest_url = get_post_meta( $post->ID, 'guest-url', true ); if ( filter_var($guest_url, FILTER_VALIDATE_URL) ) { return $guest_url; } elseif ( get_post_meta( $post->ID, 'guest-author', true ) ) { return '#'; } return $url; } function guest_author_link($name) { global $post; $guest_url = get_post_meta( $post->ID, 'guest-url', true ); $guest_name = get_post_meta( $post->ID, 'guest-author', true ); if ( $guest_name && filter_var($guest_url, FILTER_VALIDATE_URL) ) { return '<a href="' . esc_url( $guest_url ) . '" title="' . esc_attr( sprintf(__("Visit %s&#8217;s website"), $guest_name) ) . '" rel="author external">' . $guest_name . '</a>'; } elseif( $guest_name ) { return $guest_name; } return $name; } function guest_author_name( $name ) { global $post; $guest_name = get_post_meta( $post->ID, 'guest-author', true ); if ( $guest_name ) return $guest_name; return $name; } After playing with this, it occurred to me that it's possible to simplfy the guest author URL entry. The URL for my guest authors will always be formatted as follows: http://yoursite.com/author/[AuthorName] Since the URL will always follow this same pattern, what I need to know is how would I modify G.M.'s code so that I only need to enter the [AuthorName] in the custom field for the guest authors URL? I hope this makes sense...
[ 0.018143149092793465, 0.007816744968295097, -0.007931709289550781, 0.021329740062355995, 0.009245350956916809, 0.0014372927835211158, 0.009692557156085968, -0.016522394493222237, -0.017316769808530807, 0.00014246528735384345, -0.0015786486910656095, 0.013497426174581051, 0.02054944634437561,...
[ 0.4541963040828705, 0.09072654694318771, 0.6028003096580505, 0.2692756950855255, -0.14156652987003326, -0.25097063183784485, 0.07900150120258331, -0.05703370273113251, -0.2690562307834625, -0.32710614800453186, -0.01576133817434311, 0.5448018908500671, -0.07408278435468674, 0.2113196700811...
I need to make a custom slider with images (same images on all pages) in a template I am crafting for a client. Each of the slides could have a link to one of the pages. How and where should I store images, so the client will have no problem adding them or changing links?
[ 0.024828338995575905, 0.008690474554896355, 0.0064538526348769665, 0.017684150487184525, 0.0024901125580072403, 0.0025335669051855803, 0.008247345685958862, 0.007451391313225031, -0.024745658040046692, -0.019451303407549858, -0.002274279249832034, 0.006285047624260187, 0.009331882931292057, ...
[ 0.698686420917511, 0.1985320746898651, 0.5824803709983826, 0.21196280419826508, 0.0669013261795044, 0.19339020550251007, -0.2388085275888443, -0.17707931995391846, -0.3372962176799774, -0.5746414661407471, 0.44422832131385803, 0.3346984386444092, -0.23865963518619537, 0.456398069858551, ...
I heard someone use the British slang "dogs body". What does it mean?
[ -0.10323739051818848, 0.034090422093868256, -0.0015844814479351044, 0.0422302708029747, -0.049572598189115524, 0.008999226614832878, 0.01696067675948143, 0.037751078605651855, -0.0234836395829916, -0.05737749859690666, 0.0030118641443550587, 0.03400946408510208, 0.0901988074183464, 0.01226...
[ 0.9801943898200989, 0.36131855845451355, 0.011078515090048313, -0.21725130081176758, -0.09784001857042313, -0.2099335491657257, 0.5772139430046082, -0.061816226691007614, -0.3699886202812195, 0.08127187937498093, -0.130952849984169, 0.05203852429986, -0.21989330649375916, 0.020915554836392...
I'm working through a migration from Drupal to a WP multisite network, I've got a post meta field that contains the URL of the old post (mostly to maintain social counts and such) that I'm using for the permalink structure. I'm doing this with the following: add_filter('post_link', 'migration_permalinks', 10, 3 ); function migration_permalinks( $permalink, $post, $leavename ) { if( ($post->post_type == 'post' ) && ( $the_link = get_post_meta( $post->ID, '_the_old_site_permalink', true) ) ) { $url = parse_url( $the_link ); $permalink = site_url() . $url['path']; } return $permalink; } The issue I run into is that if the post title contains an apostrophe, `’` (not a straight single quote `'`), the post URL does not work. Encoded, that converts to `%E2%80%99` . In the browser address bar, if I change the apostrophe to a straight single quote, the URL works and the post displays. The existing Drupal site contains the apostrophe (`%E2%80%99`) without issue. Maintaining the URL as is is pretty important because I don't want any existing links to the page to 404 when the migration occurs. Why is the link not working when it contains an apostrophe?
[ -0.0006205744575709105, 0.01464977115392685, 0.007261850871145725, 0.013187624514102936, 0.022805102169513702, -0.009310894645750523, 0.008523521944880486, 0.008464718237519264, -0.012007059529423714, -0.014083507470786572, -0.0004800293827429414, 0.0008881049579940736, -0.004332059063017368...
[ 0.06228766217827797, 0.06677257269620895, 0.6606242060661316, -0.41769710183143616, 0.12880003452301025, 0.2063562422990799, 0.1329280436038971, -0.12982776761054993, 0.042316269129514694, -0.7622395157814026, 0.05741341784596443, -0.07624609023332596, -0.5154072642326355, 0.55443400144577...
I'm just wondering if you have an application where you define a class that defines some user configurable settings (from an xml file, or a GUI), should you design it so that it follows SOLID as much as possible like the rest of the application? More specifically, how do you change the available settings without violating the **Open Closed Principle?** Because I'm having trouble picturing this. I can see the addition of settings you can extend the class, but with the removal of settings, or change of a setting there would be no way to modify the configuration class so that it follows SOLID. Or is it possible somehow? I can also imagine trickle down effect from those settings changes, that require classes that depend on them to be modified as well.
[ 0.0016509476117789745, 0.02165961265563965, -0.004155517090111971, 0.002296886406838894, -0.004026077687740326, -0.003684672759845853, 0.006848690565675497, 0.010372343473136425, -0.01560373418033123, -0.005768997594714165, -0.005168850068002939, 0.013843998312950134, 0.007829461246728897, ...
[ 0.6073930263519287, 0.04365777224302292, 0.09976334124803543, 0.12221501022577286, 0.1593765914440155, -0.03705206513404846, 0.17169015109539032, -0.08424744755029678, -0.31022775173187256, -0.5227921009063721, 0.027719492092728615, 0.539737343788147, -0.36260125041007996, 0.18785892426967...
I want to make my labels' size changed automatically with the scale of the map,so in the menu of "Layer labeling settings",I set the Text size" in map Unit".However,the labels in the selected layer disappeared.I turned off the "Scale-based visibility" and tried to set the size from 0.05 to 2000,but the labels still refused to meet me :P What should I do? ![enter image description here](http://i.stack.imgur.com/dwPP8.png) ![enter image description here](http://i.stack.imgur.com/mud9A.png)
[ 0.0025362297892570496, 0.006661202292889357, -0.005746685899794102, 0.00530019449070096, 0.010322600603103638, -0.01268064696341753, 0.005778446793556213, -0.0023703896440565586, -0.014426636509597301, 0.005527691915631294, -0.005365223623812199, 0.006740153767168522, -0.0039872536435723305,...
[ 0.37386423349380493, -0.1314372718334198, 0.673195481300354, -0.1907888501882553, -0.0009263045503757894, 0.14429649710655212, 0.35758137702941895, -0.5208861827850342, -0.5303515791893005, -0.7854305505752563, 0.28879666328430176, 0.38958707451820374, 0.1023087128996849, 0.151168704032897...
Hello fellow typesetters. I've messed with bibliographies before, creating a bibliography file for every folder where there were .tex files that needed them. Recently though, I ventured to create a central bibliography that would be easier to manage. Since I work on the move, I'd like to put this in my cloud. For my desktop computer though, this cloud is best accessed using the local network, since it doesn't require sending the file over and back via FTP. My question is for a Windows computer: How can I reference my .bib file on a network drive? Normally, the file would be called: \\MEROVINGIAN\Public\LaTeX\bibliography.bib This should reference the file through the local network. However, backslashes don't work with BibTeX, so I tried using this command in TeXstudio for BibTeX: bibtex.exe % --include-directory="//MEROVINGIAN/Public/LaTeX" and then referencing the bibliography with the following command in my document: \bibliography{bibliography} However, it doesn't seem to work. None of the references to the bibliography could be found. When I load up NetDrive and map the cloud to a drive letter it does work: bibtex.exe % --include-directory="M:/LaTeX" However, this then goes through NetDrive and subsequently through FTP. My question is as follows: How would I configure BibTeX to find this bibliography file without having to go through NetDrive? Is there a format that needs to be used to access these folders, or is it simply impossible? Edit: I believe these are called UNC paths.
[ -0.0020261830650269985, 0.003158337902277708, -0.009810976684093475, 0.01589760184288025, 0.0029567216988652945, -0.01220980379730463, 0.006784961558878422, 0.020648522302508354, -0.018431685864925385, -0.03067111410200596, 0.0036321748048067093, 0.006089035887271166, -0.006503248121589422, ...
[ 0.12907278537750244, 0.3026556670665741, 0.5484266877174377, 0.08474063128232956, -0.20144295692443848, 0.12920454144477844, 0.2890876829624176, 0.42993760108947754, -0.4696578085422516, -0.8711465001106262, 0.0375712551176548, 0.21099485456943512, -0.3254545331001282, 0.23133298754692078,...
I have a character in Diablo 3 and am able to create games with the options of starting from a lot of quests. However, when I try to join a public game, I notice that my options are severely limited. I am wondering why that is? For example, I am able to create a public game with a quest from Act 4. However, when I try to join a public game in act 4, that option is unavailable. I am wondering how does Blizzard decide when you can join/create a game with certain quest.
[ -0.01974593475461006, 0.0022204038687050343, -0.011855686083436012, -0.007470874115824699, 0.02698718197643757, 0.010154280811548233, 0.010426001623272896, 0.01297689788043499, -0.01945711299777031, 0.015088689513504505, -0.0126276770606637, 0.021996470168232918, 0.005251395981758833, 0.02...
[ 0.15970779955387115, 0.005775332450866699, -0.028099652379751205, -0.03300357982516289, -0.5189602971076965, -0.011059140786528587, 0.5986451506614685, -0.045454274863004684, -0.38944798707962036, -0.5264093279838562, 0.25081866979599, 0.49775421619415283, 0.09295458346605301, -0.035159502...
A word for something that can't be undone or undo ?
[ 0.013682975433766842, 0.016000088304281235, -0.020774124190211296, 0.0522032268345356, 0.0210808664560318, -0.019136784598231316, 0.024836312979459763, 0.012008067220449448, -0.020084014162421227, 0.05264430120587349, -0.06544940173625946, 0.02385997772216797, 0.0117384297773242, 0.0200781...
[ 0.596478283405304, -0.13879719376564026, -0.29492586851119995, 0.012944686226546764, -0.32049185037612915, 0.24214588105678558, 0.2792254686355591, -0.01945936307311058, -0.42499643564224243, 0.11734914034605026, -0.038003575056791306, 0.20421555638313293, -0.4641953408718109, 0.3735574185...
Output of `uname` -a on my RHEL 5.4 machine is: Linux <machine name> 2.6.18-164.el5 #1 SMP Tue Aug 18 15:51:48 EDT 2009 x86_64 x86_64 x86_64 GNU/Linux Does it mean that hardware is 64 bit (going by perhaps first x86_64) and OS is also 64-bit going by last x86_64? Also, what are these so many instances of x86_64? Can I install 64-bit vm over 32-bit OS and vice versa?
[ -0.003996592015028, 0.01705051213502884, -0.007255519274622202, 0.005387016572058201, -0.036086566746234894, -0.0028850631788372993, 0.006216665729880333, 0.010857782326638699, -0.008449959568679333, -0.01692761853337288, -0.006475822534412146, 0.0009092864347621799, 0.007674692198634148, ...
[ -0.01012332458049059, -0.06858178228139877, 0.47586590051651, -0.04721122980117798, -0.2840389311313629, 0.19164074957370758, -0.1229763925075531, -0.038437370210886, -0.2518705725669861, -0.62284916639328, -0.08966538310050964, 0.7613607048988342, -0.21095970273017883, 0.29455357789993286...
I recently purchased a domain through a registrar which I won't name here. Within the first five minutes of logging in, I found a severe vulnerability that allows me access to all registration details of all users. Simply put, I do not trust this registrar with any kind of business. But I'm unable to transfer the domain because, for some reason, it has to exist in its current state for 60 days. We're planning to launch the site this weekend - we can't wait 60 days. But I can not trust this registrar: if I found such a severe vulnerability in the first few minutes, how many more similar un- trustables will I find in those 60 days? Is there a higher authority to whom I can submit a case to get my domain transferred to a different registrar?
[ -0.009907053783535957, 0.013238371349871159, 0.0041317446157336235, 0.010587909258902073, 0.004739208146929741, 0.008320996537804604, 0.005514064338058233, -0.004657551180571318, -0.008713873103260994, -0.012924352660775185, -0.020324714481830597, 0.012519355863332748, 0.011660687625408173, ...
[ 0.7137346863746643, 0.11580085754394531, 0.6248357892036438, 0.2213972806930542, 0.2854015827178955, -0.2828556001186371, 0.40178078413009644, 0.24277205765247345, -0.28517967462539673, -0.35403335094451904, 0.28473183512687683, 0.1374862641096115, -0.06328070908784866, 0.8853794932365417,...
I'm currently writing an arcobjects app that performs an arbitrary set of geoprocessing tasks on a group of shapefiles and/or dbf tables. My current implementation uses Directory.GetFiles to build an array of file paths which I feed to various GP execution functions. I'm wondering if ListDatasets can perform the same task. I need to search an entire directory tree starting at a given root folder for all files which match a given search string. I am not certain if ListDatasets will do this (I have a devil of a time deciphering the SDK documentation). In short, which would prove more efficient and effective and recursively searching a directory tree for .shp and .dbf files, `Directory.GetFiles`, `GP.ListDatasets`, or some heretofore unmentioned method?
[ -0.01282482035458088, 0.013772380538284779, -0.013197628781199455, 0.004221063107252121, 0.02133222669363022, 0.004745324142277241, 0.007436022162437439, 0.029680466279387474, -0.010933876037597656, -0.0019160658121109009, 0.0022131940349936485, 0.014207087457180023, 0.0010110973380506039, ...
[ 0.4176770746707916, 0.31294122338294983, 0.32650279998779297, -0.07632674276828766, -0.180512934923172, 0.0018636042950674891, 0.08391260355710983, -0.2019989937543869, 0.007688841316848993, -0.9550667405128479, 0.3031981587409973, 0.4656639099121094, -0.3137965202331543, 0.116698920726776...
Please see this article: here I don't know enough EMFT to comment on this but I am working on other tasks for a wireless speaker system prototype called "Busker's Friend". Completing my Schaum's EMFT book is on my list. So please comment on the truth or invalidity of this article and give an explanation. Thank you! This is physics related and I have a good reason for not posting on electronics.stackexchange. So please be kind. Thanks. :)
[ 0.003875375259667635, 0.00798998773097992, 0.00011516576341819018, 0.015568084083497524, 0.00434567267075181, -0.012913917191326618, 0.005930784624069929, -0.004206651821732521, -0.01438384409993887, -0.020622462034225464, -0.0028192929457873106, 0.020406967028975487, -0.00003353046486154198...
[ 0.2409580945968628, 0.6285381317138672, 0.1708398163318634, 0.07313301414251328, 0.0017438328359276056, -0.32427337765693665, 0.2824040353298187, 0.25400134921073914, -0.09397611021995544, -0.37787291407585144, 0.11722307652235031, 0.6823872327804565, 0.1509423851966858, -0.067539006471633...
The possible problems for a space ship near an active galactic nucleus are the following: tidal forces, hot plasma accreting into the black hole and "jetting" out, radiation, and magnetic fields. If I am missing something, please feel welcome to tell me. * * * **Tidal forces** Suppose we have a ship which has dimensions $< 1\, \rm km$, then tidal forces would surely not pose a problem as this length is much smaller than the Schwarzschild radius of Sgr A* which is presumed to be $r_s \approx 10^7 \rm km$. (You can also verify the negligibility explicitly through the equation of geodesic deviation in a Kerr metric.) * * * **Plasma** With plasma, I am not entirely sure. The accretion disc has many various models including "slim" and "thin" versions, which have a negligible extent beyond the equatorial plane, but also "thick" ones. Nevertheless, even with the "thick" Polish donut models, we see that the main density is concentrated around the equatorial plane (see image below) and there might be pathways passing through reasonably low densities of plasma. On the other hand, there surely is a means of transport of the plasma towards the axis to create jets. So for a perfectly axially symmetric situation, the horizon should be "screened" by a considerable flow of plasma from the disc towards the jets. On the "other other" hand, the typical length-scale of the accretion disc-hole system are proportional to the Schwarzschild radius $ \sim 10^7\rm km$ and this is also the typical length of the accuracy of the models we should expect. For the spaceship to safely pass, we would need an underdensity say of the lenght-scale $\sim100 \, \rm km$ i.e. of order $10^{-5}$ with comparison to the length-scale of the system. Even though a mere "bubble" in a plasma would surely not guide the space ship all the way towards the black hole, it provides a possibility to cross thin regions where the model predicts the plasma to be denser. ![thick disk simulation](http://i.stack.imgur.com/ybDqo.jpg) * * * **Radiation** We know Sgr A* to be a source of radiation all across the spectrum including X-rays. However, we see the radiation already vastly shielded by dust and factors such as the inclination of the disc with respect to us also play a role. Hence, estimating the actual intensity is not straightforward. If the radiation is too strong, the ship gets heated up beyond sustainability, not to mention the damage done on the pilots by the not-shielded X-rays. Once again, the jets can change the game - for sufficiently rapid jets, the radiation will gain much higher energies via inverse Compton scattering thus posing a much larger threat for the spaceship. Once again, I am not able to quantitatively assess the importance of such an effect especially due to the fact that from e.g. this article it would seem that Sgr A* is very much gamma- ray quiet. (Which does not mean the ship and it's crew wouldn't get fried by the "weak" gamma ray production from the vicinity of the jet.) Nevertheless, we should not forget that the ship would still be very close to a black hole and the radiation is not homogenous/isotropic. The strongest radiation comes from the parts closest to the inner edge of the disc at $\sim 3 r_s$. If the pathway was taken away from the equatorial plane, much of the radiation could be devoured or deflected by the black hole's shadow. * * * **Magnetic fields** Accretion is believed to be accompanied by _very_ strong magnetic fields. The two effects which could cause a problem are quickly varying fields causing electromagnetic induction on the electronic devices on the ship, and the effects of a very strong magnetic field on e.g. ferromagnetic parts of the spaceship. The variance of the field could be induced either by the ship moving through a space-varying field or by the inherent time-variance of the field itself. Once again, the variations of the field will be modeled on idealized models with length- and time-scales appropriate for the system itself, not the time- and length-scales relevant for the moving ship. There could be a "noise" in the field either spatial or temporal (on scales $\ll r_s$ or $\ll r_s/c$) strong enough to cause critical damage to the ship through induction. * * * It would be a too broad question to theorize whether a ship could be able to pass through such conditions (I might ask this on Space.SE using the answers to this question). For now, I will ask only about the _conditions_ relevant for a space ship approaching the Black-hole horizon. **QUESTION 1:** What are the quantitative observational constraints on the density and temperature of plasma around Sgr A* and its spatial configuration? Is there a very-low-plasma-density pathway into the black hole allowed by the current knowledge? **QUESTION 2:** What is the intensity of different types of radiation in the very vicinity of Sgr A*? What would be the optimal angle to approach the presumed black hole in it's center to maximally evade the radiation? **QUESTION 3:** What is the presumed structure of the magnetic field in Sgr A*, it's strength and derivatives in both time and space? Is there a possibility of a "strong noise" in the magnetic field say at the time-scales $10^{-4}-10 \, \rm s$ and length-scales larger than $1\, \rm cm$? For the last question we do not know what the viable velocity of the space- ship would be, but assume at least $100 \, \rm m \cdot s^{-1}$. For the moment, let us neglect the fact that the proper time of the background (defined by the quasi-exact Killing vector of the close-to-Kerr space-time) and the space-ship time might differ notably and always describe the situation from the point of view of an observer at spatial infinity. * * * This question is a followup trying to clear up the discussion of this question concerning itself with a spaceship plunging into an (im)possible worm-hole. EDIT: I have added a note on the gamma rays produced by the black hole as suggested by Kyle Kanos.
[ -0.008230487816035748, 0.005923496559262276, 0.010781008750200272, 0.009588086977601051, 0.014099027030169964, 0.0016481293132528663, 0.009098142385482788, -0.013032050803303719, -0.012872070074081421, -0.014683798886835575, -0.004265150520950556, 0.02593550831079483, -0.013203660026192665, ...
[ 0.01587250456213951, 0.06605900079011917, 0.8206990361213684, 0.04022279009222984, 0.30798161029815674, -0.21792404353618622, -0.05124218016862869, 0.09989633411169052, -0.6784611344337463, -0.5891584157943726, -0.32189813256263733, 0.5210695266723633, 0.09740033000707626, 0.33245146274566...
Creating a `wp-content/themes/your-theme/templates/loop/add-to-cart.php` file with: <?php /** * Custom Loop Add to Cart. * * Template with quantity and ajax. */ if ( ! defined( 'ABSPATH' ) ) exit; // Exit if accessed directly. global $product; ?> <?php if ( ! $product->is_in_stock() ) : ?> <a href="<?php echo apply_filters( 'out_of_stock_add_to_cart_url', get_permalink( $product->id ) ); ?>" class="button"><?php echo apply_filters( 'out_of_stock_add_to_cart_text', __( 'Read More', 'woocommerce' ) ); ?></a> <?php else : ?> <?php $link = array( 'url' => '', 'label' => '', 'class' => '' ); switch ( $product->product_type ) { case "variable" : $link['url'] = apply_filters( 'variable_add_to_cart_url', get_permalink( $product->id ) ); $link['label'] = apply_filters( 'variable_add_to_cart_text', __( 'Select options', 'woocommerce' ) ); break; case "grouped" : $link['url'] = apply_filters( 'grouped_add_to_cart_url', get_permalink( $product->id ) ); $link['label'] = apply_filters( 'grouped_add_to_cart_text', __( 'View options', 'woocommerce' ) ); break; case "external" : $link['url'] = apply_filters( 'external_add_to_cart_url', get_permalink( $product->id ) ); $link['label'] = apply_filters( 'external_add_to_cart_text', __( 'Read More', 'woocommerce' ) ); break; default : if ( $product->is_purchasable() ) { $link['url'] = apply_filters( 'add_to_cart_url', esc_url( $product->add_to_cart_url() ) ); $link['label'] = apply_filters( 'add_to_cart_text', __( 'Add to cart', 'woocommerce' ) ); $link['class'] = apply_filters( 'add_to_cart_class', 'add_to_cart_button' ); } else { $link['url'] = apply_filters( 'not_purchasable_url', get_permalink( $product->id ) ); $link['label'] = apply_filters( 'not_purchasable_text', __( 'Read More', 'woocommerce' ) ); } break; } // If there is a simple product. if ( $product->product_type == 'simple' ) { ?> <form action="<?php echo esc_url( $product->add_to_cart_url() ); ?>" class="cart" method="post" enctype="multipart/form-data"> <?php // Displays the quantity box. woocommerce_quantity_input(); // Display the submit button. echo sprintf( '<button type="submit" data-product_id="%s" data-product_sku="%s" data-quantity="1" class="%s button product_type_simple">%s</button>', esc_attr( $product->id ), esc_attr( $product->get_sku() ), esc_attr( $link['class'] ), esc_html( $link['label'] ) ); ?> </form> <?php } else { echo apply_filters( 'woocommerce_loop_add_to_cart_link', sprintf('<a href="%s" rel="nofollow" data-product_id="%s" data-product_sku="%s" class="%s button product_type_%s">%s</a>', esc_url( $link['url'] ), esc_attr( $product->id ), esc_attr( $product->get_sku() ), esc_attr( $link['class'] ), esc_attr( $product->product_type ), esc_html( $link['label'] ) ), $product, $link ); } ?> <?php endif; ?> And in your theme `functions.php` : <?php function cs_wc_loop_add_to_cart_scripts() { if ( is_shop() || is_product_category() || is_product_tag() || is_product() ) : ?> <script> jQuery(document).ready(function($) { $(document).on( 'change', '.quantity .qty', function() { $(this).parent('.quantity').next('.add_to_cart_button').attr('data-quantity', $(this).val()); }); }); </script> <?php endif; } add_action( 'wp_footer', 'cs_wc_loop_add_to_cart_scripts' ); ?> This code works fine but whenever I add more more then one quantity, it adds only one product to cart means only quantity is not adding. Any help would be greatly appreciated.
[ 0.0022621843963861465, 0.019960928708314896, 0.0030975702684372663, 0.01791081577539444, 0.010411897674202919, 0.0005163277965039015, 0.006684678606688976, -0.0038153354544192553, -0.012867795303463936, -0.002793441526591778, -0.015475974418222904, 0.0010336687555536628, -0.00995809398591518...
[ 0.14403069019317627, -0.04006138816475868, 0.6829643249511719, 0.04843252897262573, 0.10017754882574081, 0.18605589866638184, -0.030379602685570717, -0.35028254985809326, -0.4434049427509308, -0.5007484555244446, -0.20316430926322937, 0.2898314297199249, -0.37925323843955994, 0.16475926339...
Basically, I want to make api calls using an SDK I am writing. I have the following classes: * Car * CarData (stores input values needed to create a car like model, make, etc) Basically to create a car I do the following: [Car carWithData: cardata onSuccess: successHandler onError: errorHandler] that basically is a factory method that creates instance of Car after making an API call request and populating the new Car class with the response and passes that instance to the successHandler. So "Car" has the above static method to create that car, but also has non- static methods to edit, delete cars (which would make edit, delete API calls to the server) So when the Car create static method passes a new car to the successHandler by doing the following: successHandler([[Car alloc] initWithDictionary: dictionary) The success handler can go ahead and use that new car to do the following: [car update: cardata] [car delete] considering the new car object now has an ID for each car that it can pass to the update and delete API calls. My questions: * Do I need a cardata object to store user inputs or can I store them in the car object that would also later store the response from all of the api calls? * How can I improve this model? With regards to CarData, note that there might be different inputs for the different API calls. So create function might need to know model, make, etc, but find function might need to know the number of items to find, the limit, the start id, etc.
[ -0.019048795104026794, 0.014588074758648872, -0.010368426330387592, -0.006149054504930973, -0.00954815186560154, 0.0025626099668443203, 0.007875495590269566, 0.011390425264835358, -0.012640339322388172, -0.001060502603650093, -0.0073401788249611855, 0.011262859217822552, 0.005774984136223793...
[ 0.39752665162086487, 0.24329957365989685, 0.1786971241235733, 0.31922447681427, -0.09980499744415283, 0.4785611033439636, -0.08384137600660324, -0.15378552675247192, 0.08508824557065964, -0.29149699211120605, 0.20281492173671722, 0.647517740726471, 0.30165547132492065, 0.2071155309677124, ...
I'm not sure about the pluralization here: > menu of services includes or > menu of services include Which is correct?
[ 0.024157043546438217, -0.0017274849815294147, -0.017219221219420433, 0.046183180063962936, 0.007670024875551462, -0.04007652774453163, 0.02118692174553871, 0.007928167469799519, -0.04482497647404671, 0.01146125141531229, -0.059885140508413315, 0.02279420755803585, -0.035762473940849304, 0....
[ 0.23874326050281525, 0.19368013739585876, 0.027633538469672203, -0.3429538905620575, -0.3316894471645355, 0.41877293586730957, 0.5096177458763123, 0.21668507158756256, -0.28199565410614014, -0.535820484161377, -0.29406628012657166, 0.48034194111824036, -0.3641611337661743, 0.06613367050886...
I'm trying to create a form with gravity forms. I need to merge two fields to one Excerpt. I can't give the Excerpt field a merge tag. So I'm trying to use an custom field. My question is now: what is the WP Parameter to map the Excerpt to the post. I've tried: the_ excerpt, _excerpt, post_excerpt, etc. But it never appears in the Excerptfield...
[ 0.009521927684545517, -0.0009496095590293407, -0.009877758100628853, 0.022423183545470238, -0.0011594269890338182, 0.021128173917531967, 0.0094180041924119, 0.012173589318990707, -0.021796513348817825, -0.014721409417688847, -0.0032594739459455013, 0.013036752119660378, 0.0001547458086861297...
[ 0.46599990129470825, -0.396791934967041, 0.7276512384414673, 0.08755825459957123, -0.21729910373687744, 0.0035866580437868834, 0.04652903601527214, -0.18904343247413635, -0.09628847986459732, -0.8232366442680359, 0.17443116009235382, -0.0036236417945474386, -0.3804498016834259, 0.265214204...
I've been looking at online master's degree programs in software engineering. Have you completed a master's program online? If so, where? Did you like it? Was it relevant or out of date?
[ 0.008756008930504322, 0.01084068138152361, -0.025589538738131523, -0.010045373812317848, -0.005121960770338774, 0.04926624521613121, 0.010522134602069855, -0.003289356827735901, -0.04422646388411522, -0.03114570677280426, 0.021501099690794945, 0.02422165870666504, 0.0025226424913853407, -0...
[ 0.6105457544326782, -0.0033198401797562838, -0.038962215185165405, 0.3074931800365448, 0.2591436207294464, -0.06666263937950134, -0.06906287372112274, 0.5868805646896362, -0.5853481888771057, -0.1920872926712036, 0.38491925597190857, 0.2142631858587265, 0.37105029821395874, -0.133163526654...
I just started using LaTeX some weeks ago and I'm getting better and better. But now as I try to make up my own `.bbx` style for `biblatex` I'm at a point where I'm stuck: I'm having trouble editing the `online` Driver for the bibliography to fit my needs. I want to mention, if the article I found online also was published in a book or journal. This is how I got so far: \DeclareBibliographyDriver{online}{% \usebibmacro{bibindex}% \usebibmacro{begentry}% \usebibmacro{author/editor+others/translator+others}% \setunit{\labelnamepunct}\newblock \usebibmacro{title}% \newunit \printlist{language}% \newunit\newblock \usebibmacro{url+urldate}% \usebibmacro{addendum+pubstate}% \setunit{\addspace}% \iffieldundef{crossref}{}{\usebibmacro{in:}}% \usebibmacro{bybookauthor}% \setunit*{\addslash} \newblock \usebibmacro{byeditor+others}% \setunit{\addcolon\space}\newblock \usebibmacro{maintitle+booktitle}% \newunit\newblock \usebibmacro{journal+issuetitle}% \printfield{edition}% \newunit\newblock \printfield{note}% \newunit\newblock \usebibmacro{publisher+location+date}% \newunit\newblock \usebibmacro{chapter+pages}% \newunit\newblock \iftoggle{bbx:isbn} {\printfield{isbn}} {}% \newunit\newblock \printfield{doi}% \newunit\newblock \printfield{eprint}% \usebibmacro{finentry}} As you see I tried working with `\iffieldundef{crossref}`, so that whenever the `field{crossref}` is defined the `bibmacro{in:}` is shown and left out if the field is not defined. But with the definition above it doesn't print “In:” at all! Also I would like to modify the text that is printed by `bibmacro{in:}`. For example if `field{crossref}` is defined using english: Huber, Gerald (2013): Having trouble with biblatex. www.goforlatex.com. [Visited on 02/19/2013] **Also in**: Jacobs, Daniel (ed.) (2013): Solvingall the biblatex problems. Vienna: Latex University Press, pp. 1-999. on the other hand using german: Huber, Gerald (2013): Having trouble with biblatex. www.goforlatex.com. [Besucht am 19.2.2013] **Auch in**: Jacobs, Daniel (ed.) (2013): Solving all the biblatex problems. Vienna: Latex University Press, S. 1-999. No `field{crossref}` should simply result in: Huber, Gerald (2013): Having trouble with biblatex. www.goforlatex.com. [Visited on 02/19/2013] So on the one hand I need to make the `bibmacro{in:}` dependent on the `field{crossref}` and modify it to show “Also in:” when using english or “Auch in:” when using german. I don't know wheter or not it was a good idea trying to solve the first problem with `\iffieldundef` and concerning the language problem, I have no idea where to start. So I'm reaching out for your help, big cyberspace-(bib)latex-community. Here I have an example of what I have so far: \documentclass[a4paper,12pt]{scrreprt} \usepackage[ngerman]{babel} \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage[babel]{csquotes} %%%bibliography entries \begin{filecontents}{literatur.bib} @online{huber.12o,% author = {1 Crossref}, title = {Having trouble with biblatex}, year = {2013}, url = {www.goforlatex.com}, urldate = {2013-02-19}, crossref = {huber.12}, } @article{huber.12,% author = {1 Crossref}, title = {Having trouble with biblatex}, year = {2013}, journal = {The Great Biblatex Journal}, volume = {3}, number = {11}, pages = {1-999}, } @online{huber.13o,% author = {2 Crossref}, title = {Having trouble with biblatex}, year = {2013}, url = {www.goforlatex.com}, urldate = {2013-02-19}, crossref = {huber.13}, } @inbook{huber.13,% crossref = {jacobs.13}, title = {Having trouble with biblatex}, author = {2 crossref}, pages = {1--999}, year = {2013}, } @book{jacobs.13,% editor = {Daniel Jacobs}, year = {2013}, address = {Vienna}, publisher = {Latex University Press}, title = {Solving all the Latex problems}, edition = {5} } @online{nocrossref,% author = {Crossref No}, title = {Some articles are just available in the www}, url = {www.whereisthecross.com}, urldate = {2013-02-19}, year = {2013}, } \end{filecontents} \usepackage[citestyle=authoryear,bibstyle=authoryear,sorting=nyt,maxbibnames=3,bibencoding=utf8,backend=biber]{biblatex} %%% here is where it gets interesting... \DeclareBibliographyDriver{online}{% \usebibmacro{bibindex}% \usebibmacro{begentry}% \usebibmacro{author/editor+others/translator+others}% \setunit{\labelnamepunct}\newblock \usebibmacro{title}% \newunit \printlist{language}% \newunit\newblock \usebibmacro{url+urldate}% \usebibmacro{addendum+pubstate}% \setunit{\addspace}% \iffieldundef{crossref}{}{\usebibmacro{in:}}% \usebibmacro{bybookauthor}% \setunit*{\addslash} \newblock \usebibmacro{byeditor+others}% \setunit{\addcolon\space}\newblock \usebibmacro{maintitle+booktitle}% \newunit\newblock \usebibmacro{journal+issuetitle}% \printfield{edition}% \newunit\newblock \printfield{note}% \newunit\newblock \usebibmacro{publisher+location+date}% \newunit\newblock \usebibmacro{chapter+pages}% \newunit\newblock \iftoggle{bbx:isbn} {\printfield{isbn}} {}% \newunit\newblock \printfield{doi}% \newunit\newblock \printfield{eprint}% \usebibmacro{finentry}} \bibliography{literatur.bib} \begin{document} \nocite{huber.12o} \nocite{huber.13o} \nocite{nocrossref} \printbibliography \end{document} Hope someone can help me out, any hint is appreciated...
[ 0.0028692129999399185, 0.007135226856917143, -0.012532737106084824, 0.015332639217376709, 0.003404337912797928, 0.009788185358047485, 0.006994451396167278, 0.0016995270270854235, -0.015073218382894993, -0.02332352288067341, -0.0036527325864881277, 0.004147167783230543, -0.01165024284273386, ...
[ 0.4835643768310547, 0.12467973679304123, 0.6700389981269836, -0.15719091892242432, 0.1455358862876892, -0.07871775329113007, -0.08502040803432465, 0.13279616832733154, -0.40104004740715027, -0.4448744058609009, 0.36580348014831543, 0.4291025400161743, -0.6894027590751648, 0.251774966716766...
i'm trying to understand a constant 0x9e3779b9 what kind of data is this? it's not binary, not decimal, what is this? it's a constant used on the TEA algorithm it says it's derived from the golden number but the golden number is 1.618?
[ -0.01679232157766819, 0.01540064625442028, -0.009937701746821404, 0.009666440077126026, -0.034956611692905426, -0.0165813360363245, 0.010950921103358269, -0.034398552030324936, -0.01919168047606945, -0.03100673481822014, -0.012743284925818443, 0.0008759716292843223, 0.008170193992555141, 0...
[ 0.24415990710258484, 0.4194152057170868, 0.368767112493515, 0.3455941379070282, -0.06931985169649124, -0.1641763597726822, -0.027914321050047874, 0.19404719769954681, -0.3842911720275879, -0.07077489048242569, 0.25225725769996643, -0.14106827974319458, -0.09344353526830673, 0.8126258850097...
Some games (like Mass Effect 2) build on the saved games from earlier versions by allowing the player to keep going with the choices they'd made earlier. Assassin's Creed isn't big on non-linearity, but my ACII character does have nice armor, weapons, and some cash. Will AC: Brotherhood notice that saved game and pick up information from it or does Brotherhood always start from scratch?
[ 0.008297442458570004, 0.029434796422719955, -0.00012252920714672655, 0.01605374366044998, 0.020149823278188705, -0.028182201087474823, 0.009975558146834373, -0.02908247336745262, -0.02197992242872715, 0.026958486065268517, -0.024673692882061005, 0.03778836131095886, 0.00890262145549059, 0....
[ 0.5221417546272278, -0.3598984479904175, -0.16585688292980194, 0.40700504183769226, -0.18143396079540253, 0.18290866911411285, -0.13196821510791779, -0.1635819524526596, -0.2091568559408188, -0.27404293417930603, 0.23268136382102966, 0.2074682116508484, -0.1462477147579193, -0.197551101446...