text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
I'm trying to execute a script in each shutdown and reboot but the script never runs. 1. I created the script baseRhel64 and save it to `/etc/rc.d/init.d` 2. I did `chkconfig --add baseRhel64` 3. I included in the script #chkconfig --list # chkconfig: 06 10 10 4. I verified that the S10 script where created under `/etc/rc0.d/S10baseRhel64` and `/etc/rc6.d/S10baseRhel64` The following is my script: #!/bin/sh #chkconfig --list # chkconfig: 06 10 10 start(){ echo "`basename $0` start" touch /root/installscripts/test1 } stop(){ echo "`basename $0` stop" touch /root/installscripts/test2 touch /root/installscripts/"`basename $0`" } case "$1" in start) start;; stop) stop;; *) echo $"Usage: $0 {start|stop}" RETCAL=1 esac exit 0 | [
0.015498476102948189,
0.005493754521012306,
-0.010530337691307068,
0.006935587152838707,
-0.006807154975831509,
0.010587078519165516,
0.00737713323906064,
0.018681831657886505,
-0.015848450362682343,
-0.004446925595402718,
-0.019732220098376274,
0.012280933558940887,
-0.0337495431303978,
0... | [
0.6796963810920715,
0.37103334069252014,
0.36763593554496765,
-0.0048192148096859455,
0.018468452617526054,
-0.08336246758699417,
0.39598140120506287,
-0.2500290274620056,
-0.04139194265007973,
-0.5646564364433289,
0.039830297231674194,
0.6276522874832153,
-0.296766072511673,
0.24620164930... |
NOTE: Edited changes to bibliographystyle --> now \bibliographystyle{acm} \documentclass{report} \usepackage[latin1]{inputenc} \usepackage[T1]{fontenc} \usepackage{textcomp} \usepackage{fixltx2e} \usepackage[left=2cm,right=2cm,top=1cm,bottom=1.5cm]{geometry} \usepackage{color} \usepackage{amsmath} \usepackage{graphicx} \usepackage{caption} \usepackage{natbib} \title{Literature Review: \\ \textbf{Farms: \\ Supply}} \author{ Jane } \date{\today} \begin{document} \maketitle \chapter*{\textit{Introduction}} Some Text\cite{Nielsen2011}. More text \cite{Byrne2003}. \bibliographystyle{acm} \bibliography{literature_review_references} \end{document} The citation is causing me trouble. I have another citation in a different section that worked the first time but after adding the second citation and running again, I now get the error stated at the top. I have my compile settings as: Latex Bibtex Latex Latex View Pdf (evince) Ok so I have also removed the PdfLaTeX command in my compile line. The .bbl file looks like: \begin{thebibliography}{1} \bibitem{Byrne2003} {\sc Byrne, B.~W., and Houlsby, G.~T.} \newblock Foundations for offshore wind turbines. \newblock {\em Philosophical Transactions: Mathematical, Physical and Engineering Sciences 361}, 1813 (2003), pp. 2909--2930. \bibitem{Nielsen2011} {\sc Nielsen, J.~J., and Sorensen, J.~D.} \newblock On risk-based operation and maintenance of offshore wind turbine components. \newblock {\em Reliability Engineering and System Safety 96\/} (2011), 218--229. \end{thebibliography} Ok I changed the bibliography style so now it gets past Process started. New errors! * citations undefined (again, this was working originally but now are undefined?) Package natbib Error: Bibliography not compatible with author-year citations. | [
0.0031873215921223164,
0.01572497934103012,
-0.004389009438455105,
0.02397540770471096,
-0.01522059366106987,
-0.014547457918524742,
0.007716010790318251,
0.0026258439756929874,
-0.008874623104929924,
0.014878209680318832,
-0.006412875838577747,
0.0005453517660498619,
-0.0038571588229388,
... | [
-0.48039549589157104,
-0.0853700116276741,
0.4012289345264435,
0.05038920417428017,
0.030596068128943443,
0.19291484355926514,
-0.05124545842409134,
0.018615087494254112,
-0.33484968543052673,
-0.8580800294876099,
-0.2328537404537201,
0.2341172844171524,
-0.4283217489719391,
-0.21805883944... |
What strategy can I use so that I can pass wave 33 in Samurai vs. Zombies when I have maxed farmers and level 4 hero? I have been on this level for very long and I can't pass the stage. | [
-0.002578615676611662,
0.005506944842636585,
-0.01570400781929493,
-0.0043285908177495,
0.00942536722868681,
0.016446871683001518,
0.012674571946263313,
-0.012064125388860703,
-0.027970686554908752,
-0.029212625697255135,
-0.02341208979487419,
0.02396886609494686,
-0.024431580677628517,
0.... | [
-0.15442728996276855,
-0.10131712257862091,
0.39424529671669006,
-0.1701515167951584,
-0.3468508720397949,
0.12471312284469604,
0.3172342777252197,
-0.12517236173152924,
0.06782940030097961,
-0.9022557139396667,
0.37971031665802,
0.5759021639823914,
0.38795730471611023,
0.0406896248459816,... |
I am conducting a research paper on the physics of skiing, specifically how ski parameters affect the ski's ideal carve. I have come across this paper (abstract link to arXiv paper), which is incredibly relevant but I am having trouble deciphering its text. From the very start, chapter 2.1, they claim > According to figure 1, an approximate formula for the smooth curve described > by the edge of the ski is $y\approx x^2/(2R_{sc})-d$. This relation implies > $d\approx C^2/(8R_{sc})$. Therefore, we approximately have > $$\frac{d^2y}{dx^2}\approx\frac1R_{SC}$$ How did they arrive at these conclusions? What math is going on here? | [
0.004135242197662592,
0.013209905475378036,
-0.009740668348968029,
0.027338502928614616,
-0.013126815669238567,
-0.024603914469480515,
0.0065367696806788445,
-0.015461365692317486,
-0.015028256922960281,
-0.020009096711874008,
-0.006243602838367224,
0.002257245359942317,
-0.01542542036622762... | [
0.07947888970375061,
0.16734135150909424,
0.523743748664856,
-0.04465186968445778,
-0.04151950404047966,
0.22035297751426697,
0.2764197289943695,
-0.2336864024400711,
-0.19810478389263153,
-0.4692019522190094,
0.1035798117518425,
0.10548905283212662,
0.2907550036907196,
0.01159024611115455... |
I have the following problem. A working verbose x11 forward request looks like this: remote server release: Red Hat Linux release 6.2 (Zoot) sh-4.1$ ssh -1 -vvv -l username -o "ForwardX11 yes" server.name netscape OpenSSH_5.8p1, OpenSSL 0.9.8r 8 Feb 2011 debug2: ssh_connect: needpriv 0 debug1: Connecting to server.name [192.131.79.70] port 22. debug1: Connection established. debug1: identity file /home/username/.ssh/identity type 0 debug1: identity file /home/username/.ssh/identity-cert type -1 debug1: Remote protocol version 1.5, remote software version 1.2.26 debug1: no match: 1.2.26 debug1: Local version string SSH-1.5-OpenSSH_5.8 debug2: fd 3 setting O_NONBLOCK debug1: Waiting for server public key. debug1: Received server public key (768 bits) and host key (1024 bits). debug1: Server host key: RSA1 0d:02:ad:64:5f:af:fa:e1:c6:ae:50:1a:91:7e:ab:f9 debug3: load_hostkeys: loading entries for host "server.name" from file "/home/shayre/.ssh/known_hosts" debug2: key_type_from_name: unknown key type '1024' debug3: key_read: missing keytype debug3: load_hostkeys: found key type RSA1 in file /home/username/.ssh/known_hosts:1 debug3: load_hostkeys: loaded 1 keys debug3: load_hostkeys: loading entries for host "192.131.79.70" from file "/home/shayre/.ssh/known_hosts" debug2: key_type_from_name: unknown key type '1024' debug3: key_read: missing keytype debug3: load_hostkeys: found key type RSA1 in file /home/username/.ssh/known_hosts:1 debug3: load_hostkeys: loaded 1 keys debug1: Host 'proxy-bt-2' is known and matches the RSA1 host key. debug1: Found key in /home/username/.ssh/known_hosts:1 debug1: Encryption type: 3des debug1: Sent encrypted session key. debug2: cipher_init: set keylen (16 -> 32) debug2: cipher_init: set keylen (16 -> 32) debug1: Installing crc compensation attack detector. debug1: Received encrypted confirmation. debug1: Trying RSA authentication with key '/home/username/.ssh/identity' debug1: Received RSA challenge from server. Enter passphrase for RSA key '/home/username/.ssh/identity': debug1: Sending response to host key RSA challenge. debug1: Remote: RSA authentication accepted. debug1: RSA authentication accepted by server. Authenticated to server.name ([192.131.79.70]:22). debug2: x11_get_proto: /usr/bin/xauth -f /tmp/ssh-aPHVuHic3976/xauthfile generat Warning: untrusted X11 forwarding setup failed: xauth key data not generated Warning: No xauth data; using fake authentication data for X11 forwarding. debug1: Requesting X11 forwarding with authentication spoofing. debug2: fd 3 setting TCP_NODELAY debug3: packet_set_tos: set IP_TOS 0x10 debug1: Sending command: netscape debug1: Entering interactive session. debug2: fd 0 setting O_NONBLOCK debug1: fd 0 clearing O_NONBLOCK debug1: Received X11 open request. debug2: fd 4 setting O_NONBLOCK debug3: fd 4 is O_NONBLOCK debug1: channel 0: new [X11 connection from server.name port 1874] debug2: client_check_window_change: changed debug2: client_check_window_change: changed the non working output is as follows: Red Hat Linux release 9 (Shrike) $ ssh -1 -vvv -l username -o "ForwardX11 yes" server.address netscape OpenSSH_5.8p1, OpenSSL 0.9.8r 8 Feb 2011 debug2: ssh_connect: needpriv 0 debug1: Connecting to server.address [192.65.220.98] port 22. debug1: Connection established. debug1: identity file /home/username/.ssh/identity type 0 debug1: identity file /home/username/.ssh/identity-cert type -1 debug1: Remote protocol version 1.99, remote software version OpenSSH_3.7.1p2 debug1: match: OpenSSH_3.7.1p2 pat OpenSSH_3.* debug1: Local version string SSH-1.5-OpenSSH_5.8 debug2: fd 3 setting O_NONBLOCK debug1: Waiting for server public key. debug1: Received server public key (768 bits) and host key (1024 bits). debug1: Server host key: RSA1 53:40:10:d1:86:92:1e:d5:a1:70:0f:b0:4f:c8:67:eb debug3: load_hostkeys: loading entries for host "ext-proxy-2.ftel.co.uk" from file "/home/shayre/.ssh/known_hosts" debug2: key_type_from_name: unknown key type '1024' debug3: key_read: missing keytype debug3: load_hostkeys: found key type RSA1 in file /home/username/.ssh/known_hosts:1 debug3: load_hostkeys: loaded 1 keys debug3: load_hostkeys: loading entries for host "192.65.220.98" from file "/home/shayre/.ssh/known_hosts" debug2: key_type_from_name: unknown key type '1024' debug3: key_read: missing keytype debug3: load_hostkeys: found key type RSA1 in file /home/username/.ssh/known_hosts:1 debug3: load_hostkeys: loaded 1 keys debug1: Host 'server.address' is known and matches the RSA1 host key. debug1: Found key in /home/username/.ssh/known_hosts:1 debug1: Encryption type: 3des debug1: Sent encrypted session key. debug2: cipher_init: set keylen (16 -> 32) debug2: cipher_init: set keylen (16 -> 32) debug1: Installing crc compensation attack detector. debug1: Received encrypted confirmation. debug1: Trying RSA authentication with key '/home/username/.ssh/identity' debug1: Received RSA challenge from server. Enter passphrase for RSA key '/home/username/.ssh/identity': debug1: Sending response to host key RSA challenge. debug1: Remote: RSA authentication accepted. debug1: RSA authentication accepted by server. Authenticated to server.address ([192.65.220.98]:22). debug2: x11_get_proto: /usr/bin/xauth -f /tmp/ssh-kOywxfZB5752/xauthfile generate 127.0.0.1:0.0 MIT-MAGIC-COOKIE-1 untrusted timeout 1200 2>/dev/null Warning: untrusted X11 forwarding setup failed: xauth key data not generated Warning: No xauth data; using fake authentication data for X11 forwarding. debug1: Requesting X11 forwarding with authentication spoofing. debug2: fd 3 setting TCP_NODELAY debug3: packet_set_tos: set IP_TOS 0x10 debug1: Sending command: netscape debug1: Entering interactive session. debug2: fd 0 setting O_NONBLOCK debug1: fd 0 clearing O_NONBLOCK bash: line 1: netscape: command not found Transferred: sent 1096, received 716 bytes, in 0.0 seconds Bytes per second: sent 26095.4, received 17047.7 debug1: Exit status 127 I can see my client machine is forwarding x11 traffic and I am able to establish a normal ssh session ok, just when I request the netscape from this particular release it fails? the working release Red Hat Linux release 6.2 (Zoot) gives me: debug1: Received X11 open request. the non-working release Red Hat Linux release 9 (Shrike) exits with: bash: line 1: netscape: command not found apart from the obvious is there another browser supported on this release, and how can I get this to work? | [
0.022072669118642807,
-0.006859961431473494,
-0.006033249199390411,
0.006621254608035088,
-0.0175174567848444,
-0.00305688614025712,
0.007647356018424034,
0.00673260074108839,
-0.014680649153888226,
-0.019905967637896538,
-0.005088081583380699,
0.00824461504817009,
-0.0051164207980036736,
... | [
-0.3134719431400299,
0.551312267780304,
0.8523362874984741,
-0.39340049028396606,
-0.36571627855300903,
0.12861888110637665,
0.20350728929042816,
-0.550622820854187,
-0.23604954779148102,
-0.2847841680049896,
0.1616590917110443,
0.3083817660808563,
0.18400830030441284,
0.388394296169281,
... |
I have heard it many times in movies and shows. I think it means "I do not feel very good" or "I do not feel as good as you think", but why do we use "that" here, and is it correct? | [
-0.02517675794661045,
0.017366774380207062,
-0.006834167055785656,
0.014317619614303112,
-0.01129260566085577,
-0.021684866398572922,
0.010442836210131645,
-0.00706646079197526,
-0.023424046114087105,
0.03276485949754715,
-0.029544811695814133,
0.0349707193672657,
0.026405777782201767,
0.0... | [
0.42903396487236023,
0.1441672146320343,
0.10832180827856064,
-0.042634639889001846,
-0.32467904686927795,
-0.4897134006023407,
0.7945062518119812,
0.33788642287254333,
0.0719652771949768,
-0.20104211568832397,
-0.057152822613716125,
0.41673219203948975,
-0.043259017169475555,
0.5441168546... |
I'm using this code to draw curly brackets \documentclass[]{scrartcl} \usepackage{tikz} \usetikzlibrary{fit,calc,positioning,decorations.pathreplacing,matrix} \begin{document} \begin{tikzpicture}[decoration={brace}][scale=2] \node [draw] (A) {A}; \node [fit=(A)] (fit) {}; \draw [decorate,line width=1pt] (fit.south west) -- (fit.north west); \end{tikzpicture} \end{document} How can I make this brace span the entire height of the "A" character ? | [
0.006817574612796307,
-0.0015984394121915102,
-0.001967137213796377,
0.01768280193209648,
-0.0010751974768936634,
-0.003166795475408435,
0.007027441635727882,
-0.004169161431491375,
-0.012286709621548653,
-0.006153947673738003,
-0.004610684234648943,
-0.006589422933757305,
-0.018372215330600... | [
0.5379586219787598,
-0.052603013813495636,
0.7960096597671509,
-0.28248053789138794,
-0.06032970920205116,
0.41103819012641907,
0.3343975245952606,
-0.5857514142990112,
-0.2848556935787201,
-0.7623745203018188,
0.48513051867485046,
0.10185594111680984,
-0.01070637721568346,
0.1744144111871... |
I don't like how there is often a slight misalignment between assignment statements in an `algorithmic` environment. For example, the first version of the algorithm given below looks strange to me. I would like to be able to align the two statements, but the commented version in my example fails to even compile. I could always manually add some space to make things line up (see the last version below), but this is obviously not a good way to go about doing things for a number of reasons. Does anyone have a good way of making the assignment statements to line up? \documentclass{article} \usepackage{algorithm} \usepackage{algorithmicx} \usepackage{amsmath} \begin{document} \begin{algorithm} \caption{Compute some stuff} \begin{algorithmic} \State \( x \gets 1 \) \State \( m \gets 2 \) \end{algorithmic} \end{algorithm} %\begin{algorithm} % \caption{Compute some stuff} % \begin{algorithmic} % \begin{align*} % x & \gets 1 \\ % m & \gets 2 % \end{align*} % \end{algorithmic} %\end{algorithm} \begin{algorithm} \caption{Compute some stuff} \begin{algorithmic} \State \( \hspace{3pt} x \gets 1 \) \State \( m \gets 2 \) \end{algorithmic} \end{algorithm} \end{document} | [
0.006564659532159567,
0.017930766567587852,
-0.013967983424663544,
0.01866401731967926,
-0.004835905507206917,
0.011743124574422836,
0.005721283610910177,
-0.003013975452631712,
-0.01593834161758423,
0.020826375111937523,
-0.009153769351541996,
0.016680803149938583,
-0.0008609220385551453,
... | [
0.47749826312065125,
0.18615318834781647,
0.18744800984859467,
0.0873059555888176,
0.04685482755303383,
-0.42338210344314575,
0.1911187767982483,
-0.002003022935241461,
-0.24286942183971405,
-0.4908171594142914,
0.2560902535915375,
0.4003467559814453,
-0.331938773393631,
0.2310292273759842... |
I open a page and make an edit and the update button becomes grayed out. I have to copy all page content, refresh the page and paste it back in the page, only then does the update button become clickable. It does update the page so there is no issue with that. Only when making the initial edit is in not clickable. Update: I was able to recreate the issue, **but not faithfully**. It seems to occur when the page is trying to auto save. It would begin to autosave and then hang. I don't even know where to begin with this one. Has anyone encountered something similar? | [
-0.007461266592144966,
0.00587516650557518,
-0.0011224709451198578,
0.01635975018143654,
-0.020909644663333893,
-0.024197962135076523,
0.006357438862323761,
0.01444132998585701,
-0.019053859636187553,
0.02202332392334938,
-0.020840156823396683,
0.008370105177164078,
-0.011105630546808243,
... | [
0.3286360800266266,
0.2059202939271927,
0.6544241905212402,
-0.05746859312057495,
0.0103791868314147,
-0.23441828787326813,
0.5883838534355164,
0.31335580348968506,
-0.24323660135269165,
-0.7459742426872253,
-0.23799832165241241,
0.258523166179657,
-0.2245265692472458,
0.4303639531135559,
... |
At WP Front-end I create a upload form for users. That form has three input fields (title, tags and upload). So I want to add " **Title - Tag - file_name** " as `alt` attribute for images during the file upload. How can I do with my code below? if ($gui == 'upload') { if (!function_exists('wp_generate_attachment_metadata')){ require_once(ABSPATH . "wp-admin" . '/includes/image.php'); require_once(ABSPATH . "wp-admin" . '/includes/file.php'); require_once(ABSPATH . "wp-admin" . '/includes/media.php'); } if ($_FILES) { foreach ($_FILES as $file => $array) { if ($_FILES[$file]['error'] !== UPLOAD_ERR_OK) { return "Upload Err: " . $_FILES[$file]['error']; } $attach_id = media_handle_upload( $file, $newPost ); } } if ($attach_id > 0){ //and if you want to set that image as Post then use: update_post_meta($newPost,'Uploaded_Image_ID',$attach_id); } } | [
-0.01341463252902031,
0.011316838674247265,
-0.00011866592103615403,
0.017791252583265305,
0.015052800066769123,
0.008212345652282238,
0.007713219150900841,
0.002752225613221526,
-0.01228366233408451,
0.013084281235933304,
-0.01729123666882515,
0.010716267861425877,
0.007618066854774952,
0... | [
0.1472005695104599,
-0.11369818449020386,
0.8761653900146484,
-0.1838008612394333,
-0.11110813170671463,
-0.08657844364643097,
-0.04399052634835243,
-0.6683308482170105,
-0.07373935729265213,
-0.8607074618339539,
-0.20466472208499908,
0.8607114553451538,
-0.15992195904254913,
-0.0258436594... |
I recently attempted to run a boot up script which failed I didn't know why it failed so I posted a question about that here. chkconfig bootup script failing, script runs good manually Where the user schaiba suggested I add `>> /tmp/test.log 2>&1` at the end of my start up script to log and figure why it doesn't start. I realized that it has the following problem. > sudo: sorry, you must have a tty to run sudo failed I looked in my script and found this line, pretty much the important line that starts up my java program. (which is the `$cmd`) javaCommand="java" # name of the Java launcher without the path javaArgs="-classpath .:recaptcha4j-0.0.8.jar:mysql-connector-java-3.0.17-ga-bin.jar ChatServer" # arguments for Java launcher javaCommandLine="$javaCommand $javaArgs" # command line to start the Java service application cmd="nohup $javaCommandLine >>$serviceLogFile 2>&1 & echo \$! >$pidFile" # Don't forget to add -H so the HOME environment variable will be set correctly. sudo -u $serviceUser -H $SHELL -c "$cmd" || return 1 When looking what to do with that error about sudo failing.. Lots of solutions state I should disable the Defaults `require-tty` in `etc/sudoers` which another article wrote would give me a security vulnerability where SSH users could see my root password, even though SSH already logs in root haha, anyways I don't want to disable any security. My guess is I need to add that -H to some Home environment variable which I have no idea what that means anyone know what I have to do there? Also `$SHELL` variable isn't found anywhere in the script, `$serviceUser` is another word for `root`. | [
0.00015358952805399895,
-0.0027656634338200092,
0.009842246770858765,
0.019668886438012123,
-0.03320981189608574,
0.001460982020944357,
0.01075541228055954,
0.008878608234226704,
-0.02292444370687008,
0.018306560814380646,
-0.029673218727111816,
-0.0027748055290430784,
-0.007996744476258755,... | [
0.19678212702274323,
0.17106306552886963,
0.08412541449069977,
-0.05328873172402382,
-0.39300402998924255,
-0.05413127690553665,
0.9177961945533752,
0.15089008212089539,
-0.05207328870892525,
-0.7237670421600342,
0.1812722235918045,
0.3863532543182373,
-0.0941900834441185,
0.24363178014755... |
Suppose that on the first floor of a building, there are three separate apartments numbered 1, 2 and 3 respectively. How can I refer to one of them when writing a postal address? I am wondering if there is any need to mention words like "apartment" or "unit" before the apartment number. Which one is more appropriate to use? > "Apartment 2, first floor" > "Unit 2, first floor" > "2, first floor" > "2 first floor" > "2 1st floor" I also found this entry on Wikipedia which suggests that the words "lot" and "unit" can be used before the apartment number in Malaysia. (My question is more about American English though): > Salutation, Name of recipient > Name of Position & Department (if applicable) > Company name > **Unit/Lot number** , Building name/Commercial area > Lot number (for building), Street name > Postcode Post office/Mail centre > State (optional) > Country > > _example_ : > > **Lot** 323, 1st Floor, Bintang Commercial Centre Any advice would be appreciated. | [
0.00025408901274204254,
0.01617243140935898,
-0.0027286007534712553,
0.018028438091278076,
-0.015457360073924065,
0.023750247433781624,
0.010980425402522087,
-0.010460108518600464,
-0.015778474509716034,
-0.005619221366941929,
-0.0021018898114562035,
-0.0009553576819598675,
0.018571605905890... | [
0.5059438347816467,
0.07983848452568054,
0.18039311468601227,
0.09038018435239792,
0.4319024682044983,
0.5655004382133484,
0.04991302639245987,
-0.56638503074646,
-0.13189361989498138,
-1.0172102451324463,
-0.12502668797969818,
0.14564362168312073,
-0.10548292100429535,
0.19584575295448303... |
I am trying to understand a GWAS paper which reports the effect of a SNP on outcome as a % variation, e.g., 15% variation in xxx levels. The only information reported are the beta, standard error and p-values. I tried to read the paper but there is no mention of how this is calculated. The paper in question is Genome-wide association meta-analysis for total serum bilirubin levels (Hum. Mol. Genetics, 2009). Paragraph in question = These SNPs accounts for approximately 17.5, 18.1 and 16.7% of the variation... If anyone has any clue, please help. I'll be most appreciative. | [
0.00855838693678379,
0.010958679020404816,
0.0006866584299132228,
0.03183222562074661,
0.01769992522895336,
0.0044133467599749565,
0.010642105713486671,
-0.011739389970898628,
-0.013041841797530651,
-0.026145610958337784,
-0.012533536180853844,
0.010633420199155807,
-0.021013403311371803,
... | [
0.5481494069099426,
-0.18727126717567444,
0.08698131144046783,
-0.0017536511877551675,
-0.447086364030838,
0.39291366934776306,
0.2529255151748657,
-0.24740418791770935,
-0.19611714780330658,
-0.32659971714019775,
0.3732631504535675,
-0.07263531535863876,
-0.11606666445732117,
0.5228595137... |
I took quantum mechanics from our school's electrical engineering department. It was a grad level class designed for students working in device physics, thus it covered a lot of materials: from the basics (Schrodinger's equation, tunneling, the harmonic oscillator), to statistical physics (variational methods, Fermi-Dirac, Bose-Einstein, and Boltzmann distribution functions), as well as some solid state physics basics (simple models for metals, semiconductors). I then went on to take solid state physics, which used Ashcroft&Mermin, and Lundstrom. Now I no longer plan to work in device physics for my phD, but I still want to have a good understanding of QM and Solid state physics. I was working through the Griffith text, hoping to graduate toward the Shankar text when I came across Dirac's book. It seemed really elegant and focuses on intuition first. I was wondering if anyone would recommend going through Dirac's text before going to Griffith's? It makes more sense to me but most curriculums never even touches Dirac's book. Thanks, Al | [
0.011274438351392746,
0.008205944672226906,
-0.0002949379850178957,
-0.0009357728995382786,
0.006701510399580002,
-0.016003098338842392,
0.008661588653922081,
0.004559015389531851,
-0.01793770119547844,
-0.006290828809142113,
-0.012917302548885345,
0.013101287186145782,
-0.007368377409875393... | [
0.24228046834468842,
0.190282940864563,
0.0024439420085400343,
-0.03964559733867645,
-0.11935451626777649,
0.10221624374389648,
-0.16669629514217377,
0.04228854924440384,
-0.030302569270133972,
-0.3908286988735199,
0.16651374101638794,
0.43356654047966003,
0.041975218802690506,
0.226775631... |
I read somewhere that part of Minkowski's inspiration for his formulation of Minkowski space was Poincare's observation that time could be understood as a fourth spatial dimension with an imaginary coefficient. Clearly, taking the Euclidean norm of the vector $$(i \Delta t, \Delta x, \Delta y, \Delta z)$$ gives the correct spacetime interval (assuming appropriate units), but I don't really know where it goes from there (possibly something to do with Moebius transforms?) I think this is mentioned in Taylor and Wheeler's book, but I may have read it elsewhere. After the historical note, the author (whoever it was) said it was "preferable" to use Minkowski geometry straight off, rather than mucking about with time as an imaginary space coordinate. Could anyone elaborate on Poincare's formulation? Why is Minkowski's methodology better? | [
-0.006809143349528313,
0.017129210755228996,
-0.00971097219735384,
0.004580358508974314,
0.029972005635499954,
-0.0006741881370544434,
0.007072380743920803,
-0.004667687229812145,
-0.011069577187299728,
0.02573910728096962,
-0.016437260434031487,
0.011266939342021942,
-0.010520078241825104,
... | [
0.14749081432819366,
-0.29318201541900635,
0.17903396487236023,
-0.015219186432659626,
-0.06129971146583557,
0.6647884249687195,
-0.07655281573534012,
-0.18429657816886902,
-0.04683459550142288,
-0.7233939170837402,
-0.08323381096124649,
0.2646593153476715,
-0.16961386799812317,
0.67181950... |
In comparison with an open source license and requesting donations, is a free- for-open-source-projects (or free for non-commercial developers) closed source and otherwise commercial project likely to get more license fees? Or just to alienate potential users? Assume the project has value to programmers - I'm looking for generalizations here, though specific examples comparing existing projects will be very interesting. What I have in mind involves code generating programming utilities. And one issue I can think of, either way, is a near total inability to enforce any license restrictions. After all, I can't go around the internet demanding that everyone show me their source code just in case! | [
-0.006103023421019316,
0.009745285846292973,
0.0046706488355994225,
0.016024399548768997,
-0.02684021182358265,
-0.0012655691243708134,
0.007409991696476936,
0.01979510299861431,
-0.01885470189154148,
-0.014143900945782661,
-0.009070087224245071,
0.01632111892104149,
0.014086758717894554,
... | [
0.7329644560813904,
0.3086550533771515,
-0.036475472152233124,
0.3406817615032196,
-0.21313592791557312,
-0.48366397619247437,
0.03734421730041504,
-0.010118428617715836,
-0.19381634891033173,
-0.22010910511016846,
0.01595347374677658,
0.7136737704277039,
-0.1622677445411682,
0.17505249381... |
Does google first search as per the user given order of keywords in search bar? For example : I typed word1 word2 word3 word4 word5 in google search bar. Now my question is does google first search titles matching the order. Two case: 1. If it DOES so and unable to search for any webpage with that keyword order, does it skip to broad search? 2. If it DOES NOT, and there is page with that same order keyword typed, would it fetch that and display first, asumming that only my webpage have the order in which the user asked for in Google. Any thoughts? | [
-0.01031461264938116,
0.007829586043953896,
-0.0037666356656700373,
0.019017253071069717,
-0.01979644037783146,
0.00933479517698288,
0.00899810902774334,
0.013198776170611382,
-0.019174696877598763,
0.0014071527402848005,
-0.005870096385478973,
0.01871594786643982,
0.010157369077205658,
0.... | [
-0.18751750886440277,
0.0714949518442154,
0.2090771347284317,
-0.18387818336486816,
-0.27781233191490173,
-0.05260239914059639,
-0.15527643263339996,
-0.14434054493904114,
-0.264687716960907,
-0.4768258333206177,
0.06726387143135071,
0.40724995732307434,
-0.16497544944286346,
0.23940189182... |
These questions have been asked before, but since they are not answered I will try to pick up the ends here. 1) First, as stated here: QGIS 2.2 saving layer as CSV with GEOMETRY=AS_WKT not working in Qgis 2.2 it is not possible to save a shapefile as .csv including the coordinates using the function _GEOMETRY=AS_WKT_. The issue http://hub.qgis.org/issues/9597 is set to closed. Does it mean that it has been fixed? And if so, how to make it work again? 2) Second, if you want to save a shape file as .csv with coordinates through PyQgis it can be done with: http://gis.stackexchange.com/questions/61857/how-to-write-vector-layer-to-csv-file-with-geometry-by-python as stated here: How to write vector layer to csv file with geometry by PYTHON? But again, this does not save the coordinates, and I can't find any solution to this either. So right now, I still have to use an old version on QGIS. IS there any solutions to these, available at the moment? | [
-0.0011498830281198025,
0.005573003552854061,
-0.0069854045286774635,
0.03805955499410629,
-0.009391644969582558,
0.01630070060491562,
0.009631436318159103,
0.014393435791134834,
-0.01687408797442913,
0.0030172017868608236,
-0.01026572659611702,
0.02062016911804676,
-0.01569187082350254,
0... | [
0.18274030089378357,
0.013978804461658001,
0.8361088037490845,
-0.0391460582613945,
-0.06180034205317497,
-0.1356479376554489,
-0.047877173870801926,
-0.2050628513097763,
-0.15324915945529938,
-0.7510196566581726,
-0.15137843787670135,
0.19963879883289337,
-0.39332225918769836,
-0.04372280... |
I am looking to find the ACF of a time series, but after it is differenced. $y_t = a_1y_{t-1} + \epsilon_t , \mid a_1 \mid < 1$ I understand that to find the ACF this process needs to be stationary, but I am struggling to difference the equation due to the constant $a_1$ Any help would be greatly appreciated. | [
-0.011569982394576073,
0.021238328889012337,
-0.021821770817041397,
0.01593218743801117,
-0.01216957625001669,
-0.024594640359282494,
0.008799808099865913,
0.004367679823189974,
-0.02181716449558735,
0.0019483063369989395,
-0.0007375659770332277,
0.006156132556498051,
-0.03646192327141762,
... | [
0.1744707226753235,
-0.016488295048475266,
0.4208766222000122,
-0.05114635452628136,
-0.026655957102775574,
0.23031765222549438,
-0.23275232315063477,
0.02388714626431465,
0.036112524569034576,
-0.5381279587745667,
0.15728768706321716,
0.5083999037742615,
-0.24106276035308838,
0.4949591457... |
I am using standard WP gallery to display images. I have added caption to every image, image title was not changed so it is same like file name. When I open one image in gallery it shows caption text in span class "TitleText". This is ok but with some images it shows the same text and it display span TitleText and span CaptionText. You can see it here: http://www.tandem.sk/galeria/automobilizmus/cerpacie-stanice/ If you click on first image caption is showing twice, however second image is working fine. The text you see is set as caption in gallery. Any ideas how to fix this? | [
-0.012853826396167278,
0.008384635671973228,
-0.012642014771699905,
0.03012748807668686,
0.00498778885230422,
0.013190876692533493,
0.010249532759189606,
0.03161843493580818,
-0.013839073479175568,
0.012428848072886467,
-0.025166992098093033,
0.004325253423303366,
0.01362865325063467,
0.02... | [
0.8004959225654602,
0.19158460199832916,
0.31800225377082825,
-0.05773179605603218,
-0.31348279118537903,
0.03977562487125397,
-0.01236752700060606,
-0.3993781805038452,
-0.4238283336162567,
-0.6106633543968201,
0.3708982467651367,
0.7847110629081726,
-0.041732266545295715,
0.1426517218351... |
Can somebody please explain to me how the following statement is true? > The Riemann curvature tensor $R^c_{dab}$ is given by the Ricci identity > $$(\nabla_a\nabla_b-\nabla_b\nabla_a)V^c\equiv R^c_{dab}V^d$$ where > $\nabla_a$ denotes the covariant derivative. It is linear in $V^c$, hence > may be shown by the Quotient theorem to be a tensor. Now, I can see that the $R^c_{dab}$ is a tensor by construction -- based on the LHS of the Ricci identity. However, I don't understand how the linearity in $V^d$ comes to play. * * * Also, it is given that for covectors, the Ricci identity takes the form > $$(\nabla_a\nabla_b-\nabla_b\nabla_a)V_c\equiv -R^d_{cab}V_d$$ How does this follow from the Ricci identity for (contravariant) vectors? If I write $$(\nabla_a\nabla_b-\nabla_b\nabla_a)V_c=(\nabla_a\nabla_b-\nabla_b\nabla_a)(g_{cd}V^d)$$ and in GR, the Levi-Civita connection has that the metric is covariantly constant, we have $$(\nabla_a\nabla_b-\nabla_b\nabla_a)(g_{cd}V^d)=g_{cd}(\nabla_a\nabla_b-\nabla_b\nabla_a)V^d\\\=g_{cd}R^d_{eab}V^e=R_{ceab}V^e=R^d_{cab}V_d$$ Where has my minus sign gone? I have read that you can the Ricci identity for covectors by arguing using the fact that the Levi-Civita connection is symmetric, but I don't know how they mean. Thanks in advance for any help! | [
-0.0029738163575530052,
0.0032752705737948418,
0.002489716513082385,
0.012230582535266876,
-0.0129717867821455,
-0.004776008427143097,
0.005261862650513649,
0.000027033034712076187,
-0.008809574879705906,
-0.03396115079522133,
-0.00246357056312263,
0.010591764003038406,
-0.010706193745136261... | [
0.1138189509510994,
-0.21665190160274506,
0.9864516258239746,
-0.1422518640756607,
-0.11569145321846008,
0.07375433295965195,
0.02707844413816929,
-0.16470244526863098,
-0.15310630202293396,
-0.2699311673641205,
-0.03148740530014038,
0.8200477957725525,
-0.3923637568950653,
0.4755295217037... |
I want to put output files (`.pdf`) into a `_out` dir and auxiliary files such as `.synctex.gz` or `.toc` into a `_aux` subdir. From the help it seems that the `-d` option is what allows one to do it, but if I try to use it I get a Segmentation fault at the end. Am I using it wrong? echo '\documentclass{article} \begin{document} a \clearpage a \end{document}' > a.tex mkdir _aux _out pdflatex -interaction=nonstopmode -synctex=1 -output-directory _aux "a.tex" synctex view -i "5:1:a.tex" -o "_aux/a.pdf" echo $? gives the good output (page 2) and `0` exit status but: mv _aux/a.pdf _out/ synctex view -i "5:1:a.tex" -o "_out/a.pdf" -d "../_aux/" echo $? gives the good output + a line: Segmentation fault (core dumped) and `139` exit status and: synctex view -i "5:1:a.tex" -o "_out/a.pdf" -d "_aux/" echo $? gives just: This is SyncTeX command line utility, version 1.2 Segmentation fault (core dumped) and `139` exit status. Version: 1.2 on Ubuntu 12.10 from `texlive-full` package. | [
-0.017466610297560692,
0.0042869094759225845,
-0.010363949462771416,
0.019262416288256645,
0.0192263163626194,
-0.007430368103086948,
0.008360099978744984,
0.010188021697103977,
-0.017595896497368813,
-0.01178320124745369,
-0.014655284583568573,
-0.001442150678485632,
-0.01404627040028572,
... | [
0.19334295392036438,
0.057032328099012375,
0.330236554145813,
-0.2992834448814392,
0.10139882564544678,
-0.005905332043766975,
0.411537766456604,
-0.5475824475288391,
-0.07934632152318954,
-0.3942241072654724,
-0.03265853598713875,
0.7091081738471985,
-0.4648336172103882,
0.007203025277704... |
I have a Android Samsung Galaxy S Captivate SGH-I896 and I have some problems with the sound. My provider sent me a new one and expect the old one back in it's original state. So I have one that is rooted and unlocked and a clean one. Can someone help me unroot it knowing I already did a factory reset on it. I'm trying to root it again (thinking the unroot will work) and I'm trying the same way I did it the first time but it doesn't work. Is it possible to format the old one with some image made from the new one? Any suggestions? | [
0.012084399349987507,
-0.0005872070323675871,
-0.00536537729203701,
0.021873164921998978,
0.0006018972489982843,
0.002345214132219553,
0.005364555865526199,
0.009524254128336906,
-0.012912502512335777,
-0.005667387507855892,
-0.013437886722385883,
0.00884358212351799,
-0.005510750226676464,
... | [
0.20784257352352142,
0.10033393651247025,
0.3612823784351349,
0.15225297212600708,
0.16423000395298004,
0.15645013749599457,
0.5161510705947876,
-0.453955739736557,
-0.40982452034950256,
-0.6760055422782898,
-0.08175007998943329,
0.36044296622276306,
-0.1920727640390396,
0.3016934692859649... |
While reading a popular author I came across: > "I left my clubs at home, so I can't play." > > "Oh? I didn't know (that) you **golf**." This sounds wrong, but presumably it was passed by at least one editor. To me, the past tense "golfed" would be more appropriate. But if this way works too, what sort of construct is it? | [
-0.008150089532136917,
0.013941499404609203,
0.00040506175719201565,
0.01099769864231348,
0.03485436737537384,
-0.0038288559298962355,
0.009117942303419113,
-0.012024876661598682,
-0.01409166306257248,
-0.00003597552495193668,
0.009571418166160583,
0.005264611449092627,
0.018209518864750862,... | [
0.6333892941474915,
0.33825504779815674,
-0.05121057108044624,
-0.05576076731085777,
-0.3133355975151062,
-0.22727642953395844,
0.10080831497907639,
0.20641984045505524,
-0.47326022386550903,
-0.18617184460163116,
0.26758745312690735,
0.394197553396225,
-0.06911023706197739,
-0.04874420538... |
I'm wondering how I can change the network admin add "new user" or "new site" form so that I can set the password instead off having it randomly generated and e-mailed to the new user. There seems to be no real option or plugin for this oddly enough. I've noticed that `user-new.php` is different for the main site vs inside /multisite but don't really know what to change. I want to make it more like the add user form on a normal installation. | [
-0.004922517109662294,
0.005625115241855383,
-0.0048527163453400135,
0.0026692261453717947,
-0.02845713496208191,
-0.011543002910912037,
0.007889900356531143,
-0.006900382228195667,
-0.018960896879434586,
-0.0009187170071527362,
-0.010575337335467339,
0.01442129909992218,
0.00952753983438015... | [
0.8302421569824219,
0.1450493484735489,
0.30419686436653137,
0.15819409489631653,
-0.15324611961841583,
-0.15364646911621094,
0.024349650368094444,
0.34151315689086914,
-0.3362608253955841,
-0.6002167463302612,
0.4334361255168915,
0.3644781708717346,
-0.015812816098332405,
0.43778020143508... |
Is there a site that tracks lightning strikes at a local level in real-time? Google results for "lightning tracker" yield several sites, but none do this. It seems easy enough to set up. Have volunteers report: * their location * the time they saw the strike * the direction of the strike, to the nearest 45 degree (eg, "southwest" or "east") * the delay between the lightning and the thunder (ie, the distance in fifths of a mile) * optional: description of the lightning strike Of course, portions of the report could be automated (eg, geolocated tweets). It would then be easy to map lightning strikes at a given time, watch a storm move, etc, down to the several-hundred-feet level. Since this seems easy to do, I'm sure someone has done it, no? | [
0.016670970246195793,
0.013177918270230293,
-0.001666223630309105,
0.004182838834822178,
0.03513219952583313,
0.0031867967918515205,
0.007520340383052826,
0.03196951374411583,
-0.030095094814896584,
0.009785966947674751,
0.006549344398081303,
0.023487882688641548,
-0.004705492407083511,
0.... | [
0.43634623289108276,
-0.039453182369470596,
0.40422308444976807,
0.3026801645755768,
-0.020907575264573097,
-0.10133656859397888,
0.6470423340797424,
-0.014770056121051311,
-0.25049397349357605,
-0.4278678596019745,
0.018036875873804092,
0.14280574023723602,
0.12171903252601624,
-0.2591235... |
After some profiling I've found that my program spends most of its time in the simplification step. The program uses my simplification function many times (over different data). My calculation is not possible without simplifications (expressions quickly become too long). Here is a representative sample of of my data. (pastebin, 6413 lines of plain text) The expressions that I want to simplify contain one parameter `r`. They are linear combinations of the following functions: { ExpIntegralEi[2 r], ExpIntegralEi[-2 r], ExpIntegralEi[4 r], ExpIntegralEi[-4 r], f000000[r], f000001[r], Log[r], Log[4 r] }. The coefficients of linear combination contain exact numbers, powers of `r` and Exp[-2 r], Exp[2 r], Exp[4 r], Exp[-4 r] **How can I make the simplification step in my program faster?** * * * Below I summarize the ideas that I have tried. No. Simplification function Timing ByteCount 1 Collect[sample, r, Simplify] 0.512032 19 784 2 Collect[sample, _[r], Simplify] 0.348022 18 632 3 Collect[sample, _[_. r], Simplify] 0.228014 13 136 4 Collect[sample, _[_. r], Collect[#, r, Simplify]& ] 0.17601 20 536 5 Collect[sample /. logrule, _[_. r], Simplify] 0.284018 11 624 6 Collect[sample , _[_. r], Simplify[ExpandAll[ Simplify[#]]]& ] 0.284018 11 624 The first column of the table gives the number of the approach (for convenient reference), the second gives the simplification function, the third column gives the value given by Timing[] (for every `Timing[]` a fresh kernel session was used), the fourth column gives the `ByteCount[]` of the result. All calculations were done for the sample available via the link above. The definition of `logrule` is: logrule = {Log[4] -> 2 Log[2], Log[4 r] -> 2 Log[2] + Log[r], Log[16] -> 4 Log[2]}; Simplificatiom methods `5` and `6` indeed give the same `Timing` and `ByteCount`, this is not a mistake. From the table above it can be seen that method no. `4` is the fastest one, methods `5` and `6` give the best compression (I believe that `5` should be generally better). Nevertheless, considering both speed and compression efficiency I conclude that my best function is method no. `3`. **How can this method be improved? Or maybe should I replace it with some other method?** | [
0.028068847954273224,
0.025965027511119843,
-0.015641257166862488,
0.006048609036952257,
-0.0061719901859760284,
-0.00234987773001194,
0.006657566875219345,
-0.006493783090263605,
-0.014452962204813957,
-0.04455295577645302,
-0.007271544076502323,
0.006212794687598944,
-0.010298414155840874,... | [
-0.006927224807441235,
0.13868501782417297,
0.4918679893016815,
-0.30383238196372986,
0.09655404835939407,
0.5290213227272034,
0.37876757979393005,
-0.6992631554603577,
-0.1788175106048584,
0.10127806663513184,
-0.015962548553943634,
0.5434409379959106,
-0.5222708582878113,
-0.064820446074... |
Like say your a teacher and for the summer you need a job... to "supplement" your income? I'm helping a friend with her résumé (under reason for applying) and any help would be appreciated. **Definition of supplement: Add an extra element or amount to.** Doesn't seem right, so I'm just making sure. | [
-0.013262036256492138,
-0.0017001390224322677,
0.010496960952877998,
0.025138236582279205,
0.024676037952303886,
-0.01687604747712612,
0.00941081065684557,
0.013578590005636215,
-0.02559857815504074,
0.020080285146832466,
0.0029684363398700953,
0.0019374712137505412,
0.008366851136088371,
... | [
0.84217768907547,
-0.04097426310181618,
0.015420973300933838,
0.10626287758350372,
0.029454227536916733,
-0.07039425522089005,
0.15781311690807343,
0.25932931900024414,
-0.2891269028186798,
-0.18440474569797516,
0.10371367633342743,
0.4124026596546173,
0.09554950892925262,
-0.2939678728580... |
I am drawing E-R diagrams using the tikz-er2 package. In the `begintikzpicture` environment, there is an optional argument `node distance`. This is global for all the nodes. Now, If I give a small distance, The nodes intersect and if I give an average distance, the nodes appear too far. How do I set the distance for individual nodes? The `tikz-er2` package **sty** file and **user manual** can found here. Here is a MWE: \documentclass[12pt,draft]{article} \usepackage{tikz-er2} \begin{document} \tikzstyle{every entity} = [draw=blue,fill=blue!20,text=red] \tikzstyle{every attribute} = [fill=yellow!20] \tikzstyle{every relationship} = [fill=red!20] \tikzstyle{every edge} = [link] \begin{tikzpicture}[node distance=8em] \node[entity] (student) {Student}; \node[attribute] (stud-id) [left of=student] {\key{student-id}} edge (student); \node[attribute] (name) [above left of=student] {Name} edge (student); \node[attribute] (fname) [above left of=name] {First-Name} edge (name); \node[attribute] (lname) [above right of=name] {Last-Name} edge (name); \node[attribute] (cgpa) [above right of=student] {CGPA} edge (student); \node[relationship] (joins) [right of=student] {joins} edge [total] (student); \node[entity] (pgroup) [right of=joins] {Project-Group} edge [<-] (joins); \end{tikzpicture} \end{document} I have enabled the draft mode. It seems there is a bad box. How do I remove that too?  | [
-0.0047398959286510944,
0.007253235671669245,
-0.01747729815542698,
0.020223041996359825,
-0.01745736226439476,
0.0103249317035079,
0.009877797216176987,
0.006421139929443598,
-0.02009701542556286,
-0.016934432089328766,
-0.0019454644061625004,
0.011402145028114319,
-0.004084868356585503,
... | [
0.0030727239791303873,
0.23474198579788208,
0.6488383412361145,
-0.26488596200942993,
0.023974590003490448,
0.01794891618192196,
0.10903380066156387,
-0.30422356724739075,
-0.01560747716575861,
-0.95661461353302,
0.20514383912086487,
0.3851287364959717,
-0.3151591122150421,
-0.109816841781... |
How to scrape headlines from New York Times, Wall Street Journal main pages to create datasets similar to this service? Importing HTML from nyt.com (HTML4) results in a `String` and the markup is not preserved. Is there a workaround? For wsj.com (XHTML) is either not valid (is there a W3C XHTML validator?) or a problem for XML`Parser. Any clues? In[362]:= StringQ@Import["http://nyt.com","HTML"] Out[362]= True Import["http://wsj.com","XML"] During evaluation of In[361]:= XML`Parser`XMLGet::nfprserr: Attribute 'property' is not declared for element 'meta' at Line: 11 Character: 71 in /tmp/m00009067531/wsj. During evaluation of In[361]:= XML`Parser`XMLGet::prserr: Expected an attribute name at Line: 50 Character: 45 in /tmp/m00009067531/wsj. During evaluation of In[361]:= Import::fmterr: Cannot import data as XML format. >> Out[361]= $Failed | [
-0.017043765634298325,
-0.0009256280027329922,
-0.0005580928409472108,
0.015683764591813087,
-0.0021877014078199863,
0.03249520808458328,
0.010156006552278996,
0.022684376686811447,
-0.017960108816623688,
-0.01898014545440674,
-0.00813609641045332,
0.013334563001990318,
-0.004826291929930448... | [
0.288530558347702,
-0.00645893020555377,
0.058290302753448486,
0.20085686445236206,
0.05602838099002838,
-0.16088193655014038,
0.08268403261899948,
-0.166739821434021,
-0.14043648540973663,
-0.5958837866783142,
0.19035476446151733,
0.3214310109615326,
-0.0861087366938591,
-0.14633104205131... |
I have insert 14.485 lines on MySQL like this: INSERT INTO `bairros` (`id`,`cidade_id`,`descricao`) VALUES (1,8891,'VILA PELICIARI'); INSERT INTO `bairros` (`id`,`cidade_id`,`descricao`) VALUES (2,8891,'VILA MARIANA'); ... It took around 5 minutes. I had to insert in another table 16.021 lines, same database, so for test I did this: INSERT INTO `bairros` (`id`,`cidade_id`,`descricao`) VALUES (1,8891,'VILA PELICIARI'),(2,8891,'VILA MARIANA'); ... It took just a few seconds. What is the difference, for the database, between the scripts? And why one is faster than the other? | [
0.00795100536197424,
0.016477486118674278,
-0.012660847045481205,
0.011753092519938946,
-0.01100197434425354,
0.004545064643025398,
0.006779317744076252,
0.011564225889742374,
-0.01025419868528843,
-0.007592661771923304,
-0.010496465489268303,
0.007908379659056664,
-0.016713999211788177,
0... | [
-0.20644207298755646,
0.19139814376831055,
0.6962993741035461,
0.2954231798648834,
0.29259228706359863,
0.5589765906333923,
0.19837673008441925,
-0.043329961597919464,
0.16680672764778137,
-0.6726268529891968,
0.06457656621932983,
0.12161397933959961,
-0.14514124393463135,
-0.1224490404129... |
I am at the part of the tutorial where I am supposed to leave Velika, but I don't have money to pay for the pegasus - slightly over 10 silvers. I took the tutorial armor crafting quest and spent all my money on materials which I have used up crafting the training armor. That training armor is not sellable and I ran out of cash because I could make enough of them to turn in for the quest. Am I stuck now? Is there another way (hopefully repeatable) to make money in Velika? Or do I have to sell the literal clothes off my character's back to get out? | [
0.040656834840774536,
0.03328920155763626,
-0.0009629654232412577,
-0.0037115635350346565,
-0.03464829549193382,
0.0019628696609288454,
0.010152374394237995,
-0.004996220115572214,
-0.02462005987763405,
0.007296767085790634,
-0.03236203268170357,
0.02263324335217476,
-0.018978971987962723,
... | [
0.411404550075531,
-0.05004839599132538,
0.19058926403522491,
0.18433284759521484,
-0.423408180475235,
0.537566065788269,
0.3450441360473633,
-0.08330265432596207,
-0.3093336820602417,
-0.4875727891921997,
0.1093486100435257,
0.2184220552444458,
0.32203495502471924,
0.24619750678539276,
... |
From my understanding about Linear Interpolation especially of that of Newton's Divided Difference Polynomial, we could get the value of `F(x)` from the given set of coordinates. But would it be possible of the if `x` values itself is an expression, like `0.25X` and having a constant value of `0.5`. I don't know if I get the correct understanding of the NDDP: Linear Interpolation correctly given the scenario. Given the set of coordinates. How would I find the value for `X = 0.4x` where `x = is integer constant to the plane` X Y 0 0.25 0.25x 0.5 0.5x 1 0.75x 2 x 4 | [
-0.01750427857041359,
0.01767006888985634,
-0.02199413999915123,
-0.0008960324339568615,
-0.014878585934638977,
-0.024918388575315475,
0.007169428281486034,
0.014468982815742493,
-0.014332592487335205,
-0.009496601298451424,
0.001800232334062457,
0.020219359546899796,
-0.01840352825820446,
... | [
-0.05792083963751793,
-0.1419522613286972,
0.1936825066804886,
0.00015855388483032584,
0.017462167888879776,
0.10419110208749771,
-0.06047076731920242,
-0.16959130764007568,
-0.07814633101224899,
-0.41712436079978943,
0.1365971714258194,
0.6816591024398804,
-0.13437286019325256,
0.38909986... |
I've come across two pointer declarations that I'm having trouble understanding. My understanding of precedence rules goes something like this: Operator Precedence Associativity (), [ ] 1 Left to Right *, identifier 2 Right to Left Data type 3 But even given this, I can't seem to figure out how to evaluate the following examples correctly: ### First example float * (* (*ptr)(int))(double **,char c) My evaluation: 1. `*(ptr)` 2. `(int)` 3. `*(*ptr)(int)` 4. `*(*(*ptr)(int))` Then, 1. `double **` 2. `char c` ### Second example unsigned **( * (*ptr) [5] ) (char const *,int *) 1. `*(ptr)` 2. `[5]` 3. `*(*ptr)[5]` 4. `*(*(*ptr)[5])` 5. `**(*(*ptr)[5])` How _should_ I read them? | [
-0.0051223281770944595,
0.009896798059344292,
-0.01620013266801834,
0.007473584730178118,
0.0011865897104144096,
0.0260433629155159,
0.007169962860643864,
0.005958281923085451,
-0.015840481966733932,
0.011340026743710041,
0.005828845780342817,
-0.0001695534447208047,
-0.0008042661938816309,
... | [
-0.09694964438676834,
0.04884300008416176,
0.30529680848121643,
-0.26749783754348755,
-0.11709930747747421,
0.1974160075187683,
-0.04280779883265495,
-0.5479096174240112,
-0.14962607622146606,
-0.35725727677345276,
0.20399457216262817,
0.3857342302799225,
-0.14062310755252838,
-0.137669056... |
I open a file using vim in ubuntu, and this is displayed at the bottom of the screen. "file.py" [noeol] 553L, 16620C What does noeol indicate ? | [
-0.04069752246141434,
0.01565631479024887,
-0.020805692300200462,
0.031335391104221344,
-0.07754021137952805,
0.007497437298297882,
0.01696314476430416,
0.02509082481265068,
-0.022270208224654198,
-0.005080035421997309,
-0.024162715300917625,
-0.009738771244883537,
0.006945890840142965,
0.... | [
0.3464069962501526,
0.4737716317176819,
0.6973850131034851,
-0.19566182792186737,
0.07694073021411896,
-0.07550067454576492,
0.19922317564487457,
0.36060845851898193,
-0.1245126947760582,
-0.5625718832015991,
-0.45914342999458313,
0.05706744268536568,
-0.15332727134227753,
0.39751863479614... |
I have circular data such that multiple human participants were, each shown a color from a color wheel, asked to remember it for a "retention interval", then report it back by clicking a color wheel. Each participant was tested many times in each of two conditions testing conditions (short retention interval vs long retention interval). I have written and evaluated code that, for a given participant and condition, will model their data (represented as degrees of error between stimulus and report) as a mixture of a Von Mises distribution and a uniform distribution, estimating the proportion of Von Mises as well as the concentration of the Von Mises (the code also optionally permits estimating the location of the Von Mises, if one is unwilling to assume unbiased error, but for now I am willing to make this assumption for this data). However, the code was really hacked together from help that others provided me (see author's note in the previously linked paper) and I really don't understand the EM algorithm enough to extend it to achieve either of the following: 1) Permit estimation of data from multiple conditions from the same participant whilst enforcing a shared value of one or more parameters. This would allow me to evaluate, within each participant, the degree to which various hypotheses about the differences between conditions (no difference, difference on all parameters, difference on just the mixing proportion, difference on just the Von Mises concentration) are supported by the data. I'd likely want to use AIC-corrected (or cross-validated, if AIC is unavailable) likelihoods for comparing these hypotheses. 2) Implement (1) plus a hierarchical strategy that permits simultaneous estimation of parameters across the entire set of observed participants. I don't fully understand the concept of shrinkage yet, but I gather that it might usefully apply here, making (2) a more powerful approach than simply applying (1) to each participant individually and aggregating the results. Can anyone provide assistance with (1) or (1)&(2)? -- Following a request to clarify the model, here is my attempt. I'm more familiar with representing models algorithmically than symbolically, so I'll use R code. The data from the ith participant in the jth condition is assumed to be generated by a process like: library(CircStats) data_ij = c( rvm( n_ij*rho_ij , pi , k_ij ) , runif( n_ij*(1-rho_ij) , 0 , 2*pi ) ) data_ij = data_ij[ sample(n_ij) ] #shuffle to make mixture identity latent where `n_ij` is the number of observations made (determined by the experiment, so not a to-be-estimated parameter), `rho_ij` is the proportion of observations sampled from the Von Mises, and `k_ij` is the concentration of the Von Mises. With 2 or more conditions measured within each participant, rho_ij = rho_i + conditionEffectOnRho_i*condition_j k_ij = k_i + conditionEffectOnK_i*condition_j where `conditionEffectOnRho` and `conditionEffectOnK` are vectors supplying the effect on `rho` and `k`, respectively, in each condition. I don't have strong a priori expectations for the distribution of `rho_i` and `k_i`, nor am I sure that it is appropriate to restrict the variance of condition effects across participants to zero (indeed, the covariance of the effects might be of theoretical interest), hence the `_i` associated with `conditionEffectOnRho` and `conditionEffectOnK`. | [
0.010769512504339218,
0.009547138586640358,
-0.01822816953063011,
0.0027007493190467358,
0.012804190628230572,
0.014713961631059647,
0.008958812803030014,
0.0014446517452597618,
-0.011120370589196682,
0.000736036105081439,
-0.015334149822592735,
0.007847018539905548,
-0.005580412223935127,
... | [
0.21185703575611115,
0.11256131529808044,
-0.0561128668487072,
0.08011680096387863,
-0.0026628177147358656,
0.8635979294776917,
-0.00991613045334816,
-0.24737539887428284,
-0.1083342507481575,
-0.3918546140193939,
0.21947695314884186,
0.31245043873786926,
-0.050145018845796585,
0.569618999... |
Does anyone know any sources for free European city boundary data? Thanks! | [
-0.0005486389272846282,
0.05725610628724098,
0.003511944552883506,
0.061676986515522,
0.07758038491010666,
-0.09945300966501236,
0.017856383696198463,
-0.012411427684128284,
-0.05351530760526657,
0.12988874316215515,
0.05816454440355301,
0.06405788660049438,
-0.04719839617609978,
-0.044748... | [
0.6366754770278931,
0.13898883759975433,
0.38735026121139526,
0.3267515003681183,
-0.03710810840129852,
-0.6922856569290161,
0.5683894157409668,
0.6479358673095703,
-0.3937756419181824,
-0.015555634163320065,
-0.1298278421163559,
-0.028412094339728355,
0.1791774034500122,
0.017175780609250... |
My recent assignment is to model a multi-user geodatabase for the whole world using the ESRI platform. This geodatabase contains administrative layers and topographic layers. The schema is to be designed in such a way that users can update different countries simultaneously. My first problem is that every country has different administrative structures. For example, Country A has 2 administrative levels, Country B has 3 administrative levels and Country C has again 2 administrative levels. Now when I merge each set of admin levels into layers (i.e. all level 1 areas in one layer, level 2 in a second layer and level 3 in a third layer) and establish a topology there is a gap between Country C and Countries A and B in admin level 3 because that level is not available in Countries A and B. The goal is for the topology to not have a gap. Am I organizing this data in the wrong way? Any ideas on this will be highly appreciated. | [
-0.016953079029917717,
0.022178519517183304,
-0.006593784317374229,
0.02893739938735962,
0.011776622384786606,
0.006505126599222422,
0.008724759332835674,
0.018443964421749115,
-0.0155565170571208,
0.01347285695374012,
-0.008224938064813614,
0.011913565918803215,
-0.0028871120885014534,
0.... | [
0.107387013733387,
0.22209814190864563,
0.5098912119865417,
0.03063107654452324,
0.03003549948334694,
0.0555557906627655,
0.20038382709026337,
-0.08952885866165161,
-0.1771126538515091,
-0.8661544919013977,
-0.25592362880706787,
-0.019846076145768166,
0.07857430726289749,
0.318121552467346... |
How to make a random function $\eta(t)$ to insert in a differential equation for NDSolve? Edit: example: to solve equations like $\frac{dx}{dt}=\eta(t)$ | [
0.018662944436073303,
0.017745045945048332,
-0.03598504140973091,
0.003123718313872814,
-0.009718907065689564,
-0.015333220362663269,
0.013088760897517204,
0.0341818705201149,
-0.027098914608359337,
-0.01746804267168045,
-0.016300968825817108,
0.00843516644090414,
-0.030991049483418465,
0.... | [
0.0031603274401277304,
-0.42844679951667786,
-0.1716848611831665,
0.20658652484416962,
0.24673834443092346,
0.16341263055801392,
-0.07377887517213821,
-0.3407578766345978,
0.1406223624944687,
-0.5152045488357544,
-0.03783682361245155,
0.4885070323944092,
-0.5737888813018799,
-0.12164337188... |
Recently some of my friends and I grew nostalgic about the original Diablo and realized we've never truly tried the multiplayer aspect of that game. So we set up a modern VPN LAN solution via tuungle (using the appropriate IPX Wrappers) but the problem is, while any of us can manage to create session in Diablo FULL, others cannot find a way to join my IPX/LAN game. **There are no "join" buttons in the game itself.** I've read somewhere that Spawn version is a trial version but can be used to join Multiplayer games. If they can, is it possible for them to play the entire game on LAN or is trial restricted from choosing some classes/accessing some areas? Or are there some other ways to join? | [
-0.019650906324386597,
0.002670841757208109,
-0.006700374186038971,
-0.0046477909199893475,
-0.012301085516810417,
0.0019712778739631176,
0.007395608350634575,
0.017537379637360573,
-0.017862919718027115,
0.010306261479854584,
-0.0036779565270990133,
0.016248954460024834,
0.01450695283710956... | [
0.5851801037788391,
-0.08484791964292526,
0.286797434091568,
0.0500040203332901,
-0.21301965415477753,
-0.39634838700294495,
0.3277357816696167,
0.054673802107572556,
-0.3806057274341583,
-0.5191155076026917,
0.2682543694972992,
0.582396388053894,
0.12013933807611465,
0.15832598507404327,
... |
When I turn the camera on with my new Nexus 7 gen 2 it locks up and won't start. This happens half the time. What can I do to fix this or should we send it back? | [
-0.006195141933858395,
0.007777382154017687,
-0.0028090067207813263,
0.017704548314213753,
-0.03705413639545441,
0.016638819128274918,
0.009590597823262215,
0.04425088316202164,
-0.02297540195286274,
0.03435730189085007,
-0.023649679496884346,
0.002826153300702572,
-0.010911018587648869,
0... | [
0.09868787974119186,
0.02716195583343506,
0.46183982491493225,
0.03852846845984459,
0.013024216517806053,
-0.30286967754364014,
0.6762696504592896,
0.1586771458387375,
-0.3846328854560852,
-0.4490937888622284,
0.11757008731365204,
0.47358107566833496,
-0.44185134768486023,
0.20289140939712... |
I am looking for some (preferably free) GIS software that will enable me to export multi-layered world maps (made up of shape-files and kml files and so on) as hi resolution (21,000 pixels wide), single layered tiff, jpeg or other raster format files, so that i can then edit them in photoshop. I am an artist, and new to GIS. I have recently started running the free Mapwindow program on my computer. This has enabled me to start constructing some of the world maps that I'm looking for, though I cant seem to find a way of exporting them as images from that program. | [
-0.009114572778344154,
0.005597793962806463,
0.000759833084885031,
0.010891390964388847,
0.01659451425075531,
0.007946199737489223,
0.007199903484433889,
0.023297052830457687,
-0.02078578621149063,
-0.007542711682617664,
0.013383022509515285,
0.007699577137827873,
0.011762003414332867,
-0.... | [
0.6947324872016907,
0.3889813721179962,
0.33797091245651245,
-0.07818561047315598,
-0.09663474559783936,
-0.07804039120674133,
-0.03069901466369629,
0.13997821509838104,
-0.11533163487911224,
-1.102333664894104,
0.15920335054397583,
0.32870322465896606,
0.12099076807498932,
-0.319503843784... |
A normal electrometer filled with air will repel like it should do for electrostatic demonstration, but what if it is filled with water instead or even oil, what will happen? My guess is that the water is charged too, making the net repelling force equal to zero. But what will happen if it is filled with oil or another liquid? | [
0.019484540447592735,
0.033608824014663696,
0.006195174530148506,
0.016505062580108643,
-0.028482675552368164,
-0.03489120304584503,
0.012138977646827698,
-0.03132419288158417,
-0.028433620929718018,
-0.007769581396132708,
0.0008986029424704611,
0.03848366066813469,
0.007853302173316479,
-... | [
0.3839316964149475,
-0.1357765942811966,
0.46983247995376587,
0.5212129950523376,
-0.1909036934375763,
-0.25018373131752014,
0.05896772816777229,
-0.3966424763202667,
-0.16640464961528778,
-0.2417837232351303,
0.2755697965621948,
0.4455675482749939,
-0.3378382921218872,
0.251525342464447,
... |
I will build a 16 hour (2 day) course on QGIS for ArcGIS users. The target users that breathe and sleep ArcGIS. As we're trying to reduce our costs, we're thinking of switching some people to QGIS and having a floater ArcGIS license rather than have everyone with it. I've used QGIS extensively and very familiar with it (versions 0.7 to 1.8) Are there any course materials that are specific to "QGIS to ArcGIS" or that target advanced operations in QGIS? | [
-0.002557553816586733,
-0.002853659214451909,
-0.024685246869921684,
0.013854997232556343,
-0.0034141752403229475,
0.006118997931480408,
0.008474734611809254,
0.019916027784347534,
-0.0216872189193964,
-0.03867589682340622,
-0.004794684704393148,
0.015443197451531887,
-0.010217685252428055,
... | [
0.6964120268821716,
-0.007009381894022226,
0.35791823267936707,
-0.060312896966934204,
-0.04700564593076706,
-0.2693299651145935,
0.11686469614505768,
-0.25472474098205566,
0.41790395975112915,
-0.5005733370780945,
-0.13738664984703064,
0.5001947283744812,
0.02496766485273838,
-0.362162411... |
I'm about to take a degree in Linguistics and Japanese. I want to explain my thought that linguistics supplements Japanese, and Japanese supplements linguistics. Is there a single verb that implies two entities support each other in this way? "Symbiotic" is a useful adjective, but it's not quite appropriate, because it suggests that the two entities depend on one another for survival. I'm merely trying to express that the two entities enhance or add to one another in some way. I was thinking that a verb like "cosupplement" and "cosupport" would work, but they don't appear to exist as words. > Linguistics and Japanese [mutually add to/supplement] one another. A noun that refers to two entities adding to one another would also work nicely, as in: > Linguistics and Japanese are [mutually supplementary]. | [
0.008896817453205585,
0.025370921939611435,
0.00550642516463995,
0.023705752566456795,
-0.0032939566299319267,
-0.016058016568422318,
0.00877656415104866,
0.01568741723895073,
-0.016742419451475143,
0.007807660847902298,
-0.025713086128234863,
0.012568904086947441,
0.01705821603536606,
0.0... | [
0.5225563049316406,
0.29413142800331116,
0.432852178812027,
-0.03045053966343403,
0.11922606080770493,
-0.12157741189002991,
0.18434755504131317,
0.2047465294599533,
-0.07668924331665039,
-0.4395933747291565,
0.15621314942836761,
0.4376797676086426,
0.3370944559574127,
-0.22734466195106506... |
The HTC photo-album takes up an entire screen; is there a way to make a shortcut, or is there a smaller widget or app, that would open up the all- photos view? The easiest method I've found is to go in via a camera, view the albums, and select "All photos" Surely there must be a better way? I'm on the HTC Incredible running Froyo (2.2) | [
-0.0006915718549862504,
-0.012935380451381207,
-0.013395308516919613,
0.012389912270009518,
-0.023086797446012497,
0.0029649778734892607,
0.008359908126294613,
0.025724424049258232,
-0.018173085525631905,
0.0011163591407239437,
-0.019037438556551933,
-0.0008593201055191457,
0.014151726849377... | [
0.3700685203075409,
0.1298648864030838,
0.8561766743659973,
0.22099299728870392,
-0.06210041046142578,
-0.5568327903747559,
0.14699450135231018,
-0.3114393353462219,
-0.3871303200721741,
-0.08610367774963379,
-0.11364046484231949,
0.7508673667907715,
-0.4617695212364197,
-0.296726346015930... |
I would like to add footnotes, with the text of the .bib file, not only the name of the cite. Found some solutions using jurabib, but my other packages aren't compatible. How could I fix this? My tex file: \documentclass[]{report} \usepackage{times} \usepackage[dutch]{babel} \usepackage{cite} \usepackage{comment} \usepackage[hidelinks]{hyperref} %Makes Table of Contents clickable \begin{document} \includecomment{comment} %change exclude to include for showing in pdf too % Article top matter \title{X} \author{XX} \date{\today} \maketitle \tableofcontents %Begin text Text\footnote{name1} more text %End text \bibliography{pws}{} \bibliographystyle{plain} \end{document} My bib file: @misc{name1, author = "Name, Name", title = "", date = "17-05-2014", howpublished = "conference" } | [
0.013877367600798607,
0.005964930634945631,
-0.006050987169146538,
0.025284070521593094,
0.015146974474191666,
0.006089611444622278,
0.007395740132778883,
0.023637352511286736,
-0.016770578920841217,
-0.010189004242420197,
-0.0025861081667244434,
0.00006362283602356911,
-0.015162108466029167... | [
-0.025446003302931786,
0.3334140181541443,
0.5339510440826416,
-0.13732334971427917,
0.056834157556295395,
0.007162361405789852,
0.4601656198501587,
-0.05212731659412384,
-0.09352639317512512,
-0.7221927642822266,
0.2094673067331314,
0.3461826741695404,
-0.42553868889808655,
0.056127261370... |
While running a flash point my group decided to suicide instead of making the long lonley run back to the start point. We were wondering if dying naked was less expensive than wearing your gear. Does the damage to gear from death apply only to equipped gear or to gear in your bags as well. | [
-0.028131160885095596,
0.028945904225111008,
0.005834946408867836,
-0.0000624645035713911,
-0.03389957174658775,
-0.023629026487469673,
0.011341847479343414,
0.053449276834726334,
-0.0239734910428524,
0.01103085558861494,
0.0016884312499314547,
0.031289055943489075,
0.005402848124504089,
0... | [
0.8728646636009216,
-0.039467621594667435,
-0.38294216990470886,
0.5455517768859863,
-0.18736885488033295,
0.05819309875369072,
0.5634952187538147,
-0.39971932768821716,
-0.18012408912181854,
-0.13590730726718903,
0.3154148757457733,
0.38851407170295715,
-0.2338467836380005,
-0.58879435062... |
Is there a way to redirect all traffic, UDP and TCP, coming to and from eth1 and eth2 through a SOCKS proxy (Tor) which then passes it through eth0? eth0: Internet in - leads to the main router, then the cable modem eth1: A USB Ethernet port setup as a modem (I think that's the word I'm looking for, right?) eth2: A USB WiFi antenna setup as a WiFi hotspot Could I use something like iptables to directly route it through Tor or would I need an adapter like Privoxy? | [
-0.00001143268309533596,
0.01306675560772419,
-0.0081022959202528,
0.012881380505859852,
-0.029917119070887566,
-0.02328188717365265,
0.009650524705648422,
-0.0008017215877771378,
-0.015812132507562637,
0.002225855365395546,
-0.014285750687122345,
0.004675482399761677,
-0.012732721865177155,... | [
0.026808809489011765,
-0.0021680889185518026,
0.44607317447662354,
0.44858255982398987,
-0.12089105695486069,
-0.04769688844680786,
-0.025696728378534317,
0.05014539510011673,
-0.4448254108428955,
-0.6349217891693115,
0.0199576485902071,
0.16988496482372284,
-0.21168872714042664,
-0.002426... |
What is the difference between the following two: > I didn't know if this community existed > > I didn't know if this community was here | [
-0.031731028109788895,
0.027187708765268326,
0.0458061546087265,
0.008467460982501507,
0.005648949183523655,
-0.015308374539017677,
0.024775266647338867,
-0.029468480497598648,
-0.008324915543198586,
0.04348374158143997,
-0.013070546090602875,
0.0034608503337949514,
0.031746186316013336,
0... | [
0.647834300994873,
0.31773456931114197,
0.14351752400398254,
-0.18988564610481262,
0.020660625770688057,
0.04968947917222977,
0.5204344987869263,
0.09785883128643036,
-0.2066458761692047,
-0.3878863751888275,
0.3585757613182068,
0.24656790494918823,
0.013514497317373753,
0.1373893320560455... |
I often have displayed equations in the middle of a paragraph, so that the following text of the equation does not start on a new line but follows with a space. In such situations, I would like to know how I can adjust the vertical space between the displayed equation and the following text. I've tried with the usual `\vspace{}` trick, by putting it just in front of the first word that follows the equation. However, it produces the wrong effect by increasing the vertical space between the first line of text after equation and the second. Any suggestion? Thanks! | [
-0.002620204584673047,
0.0018874985398724675,
-0.025093896314501762,
0.01207978930324316,
-0.0027988050132989883,
-0.023121614009141922,
0.009464634582400322,
0.005348734557628632,
-0.01640162244439125,
-0.007262360770255327,
-0.011409062892198563,
0.01042138785123825,
-0.005035384092479944,... | [
0.28289833664894104,
0.19912564754486084,
0.43356215953826904,
0.017357196658849716,
0.04462070018053055,
0.06208598241209984,
0.19512046873569489,
0.1878845989704132,
-0.26811227202415466,
-0.684210479259491,
0.0166852418333292,
0.2961076498031616,
-0.14404626190662384,
-0.068159684538841... |
In this chart I can determine the number of slices, but not the size of slices. How to determine the size of each slice ? Thanks for your help \documentclass{article} \usepackage[paperheight=30cm,paperwidth=35cm,margin=1in,heightrounded]{geometry} \usepackage{tikz} \usetikzlibrary{calc} \begin{document} \thispagestyle{empty} \fbox{\begin{tikzpicture}[scale=0.6] % pour rester dans la page % center \path (0,0) coordinate(A); % circles \foreach \rr in {12,...,25}{ \draw (A) circle (\rr); } \def\pp{4} % nb of slices % rays \foreach \aa in {1,...,\pp}{ \draw ($(A)+(\aa/\pp*180:12)$) -- ($(A)+(\aa/\pp*180:24)$); } \end{tikzpicture}} | [
0.0003688163124024868,
0.017283566296100616,
0.004448512569069862,
0.03493405878543854,
0.008692077361047268,
0.009883381426334381,
0.009005192667245865,
-0.012422564439475536,
-0.016867730766534805,
-0.026321977376937866,
-0.008887726813554764,
-0.007207193411886692,
-0.005998598411679268,
... | [
0.06246684864163399,
-0.2900301218032837,
0.8026676177978516,
0.01876995339989662,
-0.10193834453821182,
0.26230308413505554,
-0.0711977556347847,
-0.25156137347221375,
-0.3501950800418854,
-1.078111171722412,
-0.10108255594968796,
0.1613461971282959,
0.07241945713758469,
-0.17373670637607... |
In Peskin and Schroeder p. 669 it is argued that the axial current can be parametrized between the vacuum and an on-shell pion state as: $$<0|j^{\mu 5}(x)|\pi^b(p)>=-ip^\mu f_\pi \delta^{ab}e^{-ipx}$$ This is then described as a parametrization of the amplitude for the axial current to create a pion state from the vacuum. This interpretation puzzles me: isn't it the role of the creation operators of a given theory to do this? Does the above expression correspond to an actual physical process or is it just part of an amplitude in a Feynman diagram for an actual process? | [
-0.002359069650992751,
0.010925707407295704,
-0.003274894319474697,
0.01598786562681198,
0.005048554390668869,
-0.024773038923740387,
0.006785601377487183,
-0.009635087102651596,
-0.01318131573498249,
0.018402734771370888,
-0.0220610611140728,
0.014204290695488453,
-0.03569166362285614,
0.... | [
0.11869834363460541,
0.03266134858131409,
0.48942747712135315,
-0.22849000990390778,
0.02741193026304245,
0.6654489040374756,
-0.2509637475013733,
-0.5259628891944885,
0.040369149297475815,
-0.14235897362232208,
-0.02127131260931492,
0.375238299369812,
-0.2881534695625305,
0.74926465749740... |
So I have what might be a fairly basic question, but my understanding that in the quantization of the the string, or the 1-brane, there are conditions on the number of spacetime dimensions to ensure that the Lorentz algebra still holds. For the bosonic string we have D=26 and for the superstring we have D=10. Now I know that an equivalent condition on the critical dimension comes up by requiring that the conformal anomaly vanish. This is the prelude to my question: Are there any similar or equivalent restrictions on the number of spacetime dimensions are come up in the quantizing more general branes? Is it only the string because only the worldsheet theory has conformal invariance that we want to ensure holds in the quantized theory? If this statement is equivalent to the Lorentz algebra closing, why for a general p-brane do we not need to impose any conditions on the dimensions for this to hold? | [
0.0003192832227796316,
0.014365240931510925,
-0.010865867137908936,
0.0003540260950103402,
0.04209846258163452,
-0.013944289647042751,
0.01030554436147213,
0.00603822385892272,
-0.009230597876012325,
-0.01652236469089985,
-0.007725238800048828,
0.018465260043740273,
-0.0042737629264593124,
... | [
0.19680634140968323,
0.013749749399721622,
0.36676812171936035,
-0.37276628613471985,
-0.26315104961395264,
-0.23703216016292572,
-0.05835387110710144,
-0.45322665572166443,
-0.11464016884565353,
-0.27490293979644775,
-0.254377156496048,
0.5738108158111572,
-0.22092792391777039,
0.39142155... |
I noticed that if you make changes trough the new "customize" feature, when you navigate on a different page within the iframe preview document, your changes still apply, even if they are unsaved. It looks like WP is storing the temporary changes somewhere, and it applies them on the site if the site is viewed in "customize" mode. But how does the site know it's in customize mode? Because I don't see any query arguments appended to the links or anything like that. | [
-0.0019030371913686395,
0.0070880744606256485,
0.00694253109395504,
0.020478881895542145,
0.003743163077160716,
-0.006168920546770096,
0.007921276614069939,
-0.005751732271164656,
-0.01185898669064045,
0.02800818718969822,
-0.011648591607809067,
0.015020165592432022,
0.0020629202481359243,
... | [
0.6353847980499268,
0.034189622849226,
0.6210769414901733,
-0.024020306766033173,
0.16880835592746735,
0.04323993995785713,
0.19693681597709656,
0.02372818812727928,
-0.5505005121231079,
-0.3772537112236023,
0.19811612367630005,
0.3409045934677124,
-0.16187185049057007,
0.461060494184494,
... |
Will the PC version of _Lost Planet 2_ be able to play multiplayer against the Xbox Live community, like you could with the first version? | [
-0.0016633777413517237,
-0.004446410574018955,
0.00039466447196900845,
0.002562812762334943,
0.016542598605155945,
-0.018770433962345123,
0.013587390072643757,
-0.046059489250183105,
-0.024966642260551453,
-0.026695817708969116,
0.0029733898118138313,
0.06361854076385498,
0.01076264865696430... | [
0.3049440085887909,
-0.15568357706069946,
0.22844724357128143,
0.32967230677604675,
0.2666284143924713,
-0.6994903087615967,
0.1164252907037735,
0.23401235044002533,
-0.007916934788227081,
-0.6740794777870178,
-0.0015210954006761312,
0.5971825122833252,
-0.3599936366081238,
-0.109999395906... |
Can I merge contacts in CyanogenMod 7 like I was able to do in HTC Sense? Let's that I have my contact 'John Smith', then I import contacts from Facebook and a new John Smith appeared. I want these two contacts to be merged under one single contact. Can I do that somehow in CyanogenMod 7? | [
0.012968950904905796,
0.015109328553080559,
-0.012201732024550438,
0.02574944868683815,
-0.02408594824373722,
-0.012354522943496704,
0.01452729944139719,
-0.0025745066814124584,
-0.033888161182403564,
-0.01910800114274025,
-0.002288087969645858,
0.04083007946610451,
-0.010476833209395409,
... | [
0.38985416293144226,
0.11662992835044861,
0.2523673474788666,
-0.006528632715344429,
-0.04699213430285454,
0.07510768622159958,
0.008224740624427795,
0.3750990033149719,
-0.3505970239639282,
-0.6034159660339355,
-0.027823317795991898,
0.620657742023468,
-0.3322601616382599,
0.1404517441987... |
I am looking for the absolute fastest approach to adding +10,000 shapely geometry objects to a PostGIS database using geoalchemy2. I'm having no speed issues when uploading to a localhost postgres server, but when I upload directly to my remote server it takes 35x as long! I realize this might be a server question and not appropriate for this site, but curious if anyone has a successful workflow for: **Quickly uploading shapely geometries to a PostGIS database using geoalchemy2** | [
-0.014447294175624847,
0.010676544159650803,
-0.013182824477553368,
0.022643940523266792,
-0.017744001001119614,
-0.009709534235298634,
0.011285954155027866,
0.025745529681444168,
-0.015362315811216831,
-0.021898848935961723,
-0.010560230351984501,
0.01674727350473404,
-0.015153863467276096,... | [
0.35044044256210327,
0.1635291576385498,
0.5059635639190674,
0.1600254625082016,
-0.11646157503128052,
0.09487280249595642,
-0.013042235746979713,
0.19461821019649506,
-0.08061006665229797,
-1.005403757095337,
0.062238313257694244,
0.17462880909442902,
0.20183147490024567,
0.37090724706649... |
I've tried out the SMS feature of Hangouts app, which allows you to handle SMS within itself instead of the default one. I didn't like it for certain reasons, so I disabled it. However, for some reason, almost each time anyone sends me something via Hangouts (from a PC for example), I get it in both Hangouts and via SMS (with phone number that looks like "1000274" ). That's right - I get 2 messages - one for hangouts and one is an SMS message shown on the normal messaging app. Why does it happen? How can I stop it from happening? I hope it's a bug and that it will get fixed soon... :( NOTE: I have Android 4.3 on a Galaxy S3 I9300 device. | [
-0.01481098122894764,
-0.008763769641518593,
-0.016272641718387604,
0.005569516681134701,
-0.009130630642175674,
-0.009247257374227047,
0.0091054392978549,
0.009942885488271713,
-0.01432018168270588,
-0.03406735509634018,
-0.014853214845061302,
-0.006910954602062702,
0.005834478419274092,
... | [
0.5798175930976868,
0.2267761379480362,
0.3193798065185547,
-0.07002660632133484,
0.06285269558429718,
0.016595453023910522,
0.3854740858078003,
0.3128325641155243,
-0.6392664313316345,
-0.34758833050727844,
0.2576036751270294,
0.2201564908027649,
-0.3144344389438629,
0.16701845824718475,
... |
I created a Dashboard widget to send push notifications to an app, but the way I have it coded, whenever I make changes in the Plugin Editor and press Update File, the code gets ran and I get a blank push notification. I am sure my problem lies in this part of the plugin code: if('POST' == $_SERVER['REQUEST_METHOD']) { // process the form here } Any suggestions on how to fix this or a better way to do it? Here is the gist of my plugin code if('POST' == $_SERVER['REQUEST_METHOD']) { //do the stuff with the 'message' } // Function that outputs the contents of the dashboard widget function dashboard_widget_function() { echo '<form method="post"> <p>Message: <input type="text" name="message" /></p> <p><input type="submit" /></p> </form>'; } // Function used in the action hook function add_dashboard_widgets() { wp_add_dashboard_widget('push_notification', 'Push Notification', 'dashboard_widget_function'); } // Register the new dashboard widget with the 'wp_dashboard_setup' action add_action('wp_dashboard_setup', 'add_dashboard_widgets' ); | [
-0.014547862112522125,
-0.0032581002451479435,
0.00804186798632145,
0.01656440459191799,
-0.0005077244713902473,
-0.00661407271400094,
0.006752749904990196,
-0.006614240817725658,
-0.014413893222808838,
0.015791643410921097,
-0.009596044197678566,
0.011776261031627655,
-0.002357248216867447,... | [
0.1270085573196411,
0.2118772715330124,
0.49150291085243225,
-0.34707435965538025,
0.05524376034736633,
0.24231283366680145,
0.22794072329998016,
-0.005972050130367279,
-0.0035570808686316013,
-0.6315526366233826,
0.19389194250106812,
0.49457287788391113,
-0.5446877479553223,
0.01803699694... |
A quine is a computer program which takes no input and produces a copy of its own source code as its only output. There are many trivial quines in Mathematica: In[1]:= "Hello world" Out[1]= "Hello world" In[2]:= 3.14 Out[2]= 3.14 In[3]:= f[x] Out[3]= f[x] where `f` and `x` are undefined symbols; and some more In[4]:= Hold[N[\[Pi]]] Out[4]= Hold[N[\[Pi]]] These are all trivial. I was thinking that perhaps a more interesting challenge for Mathematica was a multiquine. This is a program A that outputs another program B, _distinct_ from A, such that when B is exectued, the output is A. Multiple levels of depth are also allowed: Thus one might have a program that when executed outputs a distinct program that when executed outputs another program distinct from the first two ... that when executed outputs the original program. There are also multiquines that output a distinct program _in a different language_ , such that when this program is executed, the output is the original program. **QUESTION** (Though more of a challenge): Can you come up with a multiquine for Mathematica? | [
-0.01504988968372345,
0.005187198519706726,
0.005874306429177523,
-0.0036458580289036036,
-0.016828959807753563,
-0.012678690254688263,
0.006063160952180624,
0.007304743863642216,
-0.010185268707573414,
-0.017055092379450798,
-0.010333596728742123,
-0.002128427382558584,
0.017003819346427917... | [
0.011028468608856201,
0.0934869572520256,
0.08995167911052704,
0.15641480684280396,
-0.1818973869085312,
0.5183311104774475,
0.270827054977417,
-0.16517023742198944,
0.026430612429976463,
0.03140072152018547,
-0.3464924097061157,
0.23182202875614166,
-0.42640256881713867,
0.160653859376907... |
I needed to type something like this:  But wanted to change the 'numbering' to 'Question No. x.' from just 'x.' How should I go about doing that? An alternative solution is also welcome if what I'm asking for just isn't possible in the exam class. Thanks. | [
0.008238894864916801,
0.00406472897157073,
-0.011316092684864998,
0.013030397705733776,
0.008450822904706001,
-0.007364127319306135,
0.006239673588424921,
0.016966968774795532,
-0.017215823754668236,
-0.009864097461104393,
-0.012026162818074226,
-0.0008492415654473007,
-0.003384894458577037,... | [
0.5838831067085266,
0.1554400771856308,
0.13973954319953918,
0.3262711465358734,
-0.38050830364227295,
-0.137556254863739,
0.2880411744117737,
0.14669884741306305,
-0.29611682891845703,
-0.4466980993747711,
0.03335315361618996,
0.23044352233409882,
0.10968025028705597,
-0.01494672801345586... |
I got a partition that is probably created by dd, and is about 200MB on a 4GB USB stick. I am failing to delete it using GParted (the option is greyed on the right-click menu).   | [
-0.02044595219194889,
0.00015931000234559178,
-0.015129615552723408,
0.0070385257713496685,
0.002299562329426408,
-0.015803342685103416,
0.004157665651291609,
0.018828893080353737,
-0.016390366479754448,
0.00975426658987999,
-0.0027445948217064142,
-0.0017204572213813663,
0.01609941385686397... | [
0.0838828906416893,
0.09623315185308456,
0.22711050510406494,
0.27296221256256104,
0.1998462677001953,
-0.035280629992485046,
0.35397350788116455,
0.10231788456439972,
-0.47960346937179565,
-0.7820645570755005,
-0.0030302032828330994,
0.423258900642395,
-0.07944658398628235,
0.445595830678... |
In index the `!` char is used for item separation. I have index strings which contain folders and files. These index entrys should look like \index{files!fonts/!fonts.tex} However I pass these to the index using this command: \newcommand*{\labelfile}[1]{\index{files!#1}} which results in \index{files!fonts/fonts.tex} thus the replacement of every `/` in #1 by `/!` is missing. How can I replace chars in a string in LaTeX ? EDIT: Using the code example from Heiko Oberdiek (which works!) I created the following approach to wrap all items in a `\texttt` command. Unfortunately my approach fails... \newcommand*{\labelfile}[1]{% \StrSubstitute{#1}{/}{,}[\TempIndexFileEntry]% \def\doc@IndexFileEntry\empty \def\do##1{% \edef\@tempa{\zap@space##1 \@empty}% \ifdefempty{\doc@IndexFileEntry}{ \renewcommand{\doc@IndexFileEntry}{\texttt{\@tempa}} }{% \renewcommand{\doc@IndexFileEntry}{\doc@IndexFileEntry/\texttt{\@tempa}} }% }% %% Process csv list with command \do (etoolbox) \docsvlist{\TempIndexFileEntry}% %% rename \let\TempIndexFileEntry\doc@IndexFileEntry \@bsphack \StrSubstitute{\TempIndexFileEntry}{/}{/!}[\IndexFileEntry]% \index{files!\IndexFileEntry}% \@esphack } with ! Use of \TempIndexFileEntry doesn't match its definition. \@xs@StrSubstitute@@ ...}\@xs@def \@xs@arg@ii {#2} \@xs@def \@xs@arg@iii {#3}... | [
0.0073028067126870155,
0.015461052767932415,
-0.007129126228392124,
0.022906288504600525,
-0.013754956424236298,
-0.007404963951557875,
0.007163058966398239,
-0.013415658846497536,
-0.01855803281068802,
-0.024129847064614296,
-0.015412918291985989,
0.001573373912833631,
0.011987024918198586,... | [
-0.2185002863407135,
0.18953627347946167,
0.3910442292690277,
-0.030546747148036957,
0.08377934992313385,
0.054928164929151535,
0.12902413308620453,
-0.22284618020057678,
-0.0968756452202797,
-0.4911879003047943,
-0.10015293955802917,
0.5254390239715576,
-0.4143463969230652,
0.178809672594... |
The question was inspired by this discussion: How to expand a function into a power series with negative powers? I am interested in asymptotic behavior of a function at infinity: f[r_]:=(0.04962 Exp[-2 r] (-1.000+r))/((0.06119+(Exp[-2 r])^(2/3))^2 r) or (TeX) $$ f(r)=\frac{0.04962 e^{-2 r} (r-1.000)}{\left(\left(e^{-2 r}\right)^{2/3}+0.06119\right)^2 r} $$ Tried `Series[f[r],{r,0,10}]` for expansion in **_negative powers_** at infinity (as suggested) and got: $$ -\frac{0.04406}{r}+0.02146+0.02106 r+0.004405 r^2-0.001355 r^3-0.001205 r^4-0.0003607 r^5-\left(8.402\times 10^{-6}\right) r^6+0.00004149 r^7+0.00001982 r^8+\left(3.921\times 10^{-6}\right) r^9-\left(6.018\times 10^{-7}\right) r^{10}+O\left(r^{11}\right) $$ Seems like the function decays faster than $1/r^n$ and the expansion is meaningless. But what does the term $-\frac{0.04406}{r}$ mean then? The function is strictly positive at infinity and I am kinda confused by that. Does this mean that the asymptotic form of the function is _something_ plus the term $-\frac{0.04406}{r}$ which effectively gives the observed behavior? Can someone clarify it? How can one explain the term $-\frac{0.04406}{r}$? | [
0.01675153151154518,
0.007533338386565447,
-0.001980723813176155,
0.01422329805791378,
-0.0012503275647759438,
-0.009347997605800629,
0.0050418796017766,
0.003793980460613966,
-0.012376098893582821,
0.01151476614177227,
-0.0049573518335819244,
0.004538119304925203,
-0.009370545856654644,
0... | [
-0.16397257149219513,
-0.19494488835334778,
0.4343068301677704,
-0.36746424436569214,
0.17775098979473114,
0.21563348174095154,
0.12543264031410217,
-0.6798616051673889,
0.2313246726989746,
-0.24953269958496094,
-0.22477251291275024,
0.9559043049812317,
-0.3326926827430725,
-0.101972110569... |
I recently finished act 3 on my newest character and hit level 51, and I was thinking it would help to deck my companion out with mods from the Makeb planetary comm vendor as well as myself. I had enough comms to take care of most of my slots, bit not enough to even start on my companion. Now that I'm 51 all the flashpoint rewards seem to be Basic comms rather than planetary comms. While those are helpful it would be nice to be able to pick up a bunch of planetary comms fast like I used to be able to by doing a flashpoint for its 5 + the conflicts daily 10 + the group finder 10. How can I quickly acquire planetary comms now? Techniques that are only doable once per day are fine as long as they're efficient. | [
0.015265065245330334,
0.018229514360427856,
-0.018700411543250084,
-0.0013579754158854485,
0.0016502626240253448,
0.01120857335627079,
0.006969034671783447,
0.0064474111422896385,
-0.018729625269770622,
0.011242344975471497,
-0.0038501922972500324,
0.02005503512918949,
-0.0030359639786183834... | [
-0.11731390655040741,
0.0865604355931282,
0.7276521921157837,
-0.036303311586380005,
-0.20794393122196198,
0.13478542864322662,
0.3312482535839081,
-0.11547105759382248,
-0.45623907446861267,
-0.7001291513442993,
0.09789455682039261,
0.5328679084777832,
0.3601343631744385,
0.05594273284077... |
I don't think atleast is an actual word, but I've found many instances of its usage. A simple google search for _atleast_ reveal 13,100,000 hits. What is the meaning of _atleast_ and is it different from _at least_? | [
-0.000968228850979358,
0.018644271418452263,
-0.0071442266926169395,
0.031126396730542183,
-0.021382872015237808,
-0.030308380722999573,
0.00938276294618845,
-0.024533726274967194,
-0.04017740488052368,
0.004500333219766617,
0.004143061116337776,
0.013833812437951565,
0.03289806842803955,
... | [
0.5243536233901978,
0.11381228268146515,
0.3407546281814575,
0.0665317177772522,
-0.5775131583213806,
0.16567549109458923,
0.6848166584968567,
0.2397095412015915,
-0.3570876121520996,
-0.06264281272888184,
-0.1457708477973938,
0.315970778465271,
-0.40637701749801636,
0.2138710767030716,
... |
I rooted my phone and then an update unrooted it. I described my story here . I am glad that the update unrooted it, I was also planning on doing it myself. What can I do to make sure that all traces of me having root access on the phone are erased? I'd like to do this, so that if I have to go in for warranty purposes, I don't have any problems. I'm guessing uninstalling z4root would be the first step. What else? | [
0.022867979481816292,
0.014303917065262794,
0.005430189426988363,
0.009731690399348736,
-0.02134336158633232,
0.008620894514024258,
0.005023070145398378,
0.006703725550323725,
-0.02196612022817135,
-0.014866615645587444,
-0.006513209547847509,
0.01045496016740799,
-0.00922610517591238,
0.0... | [
0.41504889726638794,
-0.07541436702013016,
0.2715199887752533,
0.2055637091398239,
0.30827596783638,
-0.39735883474349976,
0.5071300268173218,
-0.06836003810167313,
-0.32935309410095215,
-0.30305716395378113,
0.24691973626613617,
0.39439767599105835,
0.12014496326446533,
0.0765724182128906... |
What's the difference? > I'm having hard time figuring that out > > I'm having a hard time figuring that out According to Google both are used equally often. Does the article change meaning here? | [
-0.027073247358202934,
0.010920812375843525,
-0.04077247902750969,
0.03722800314426422,
-0.08169376105070114,
-0.030065501108765602,
0.015720423310995102,
-0.026507817208766937,
-0.024671336635947227,
-0.02046993188560009,
0.002196444198489189,
-0.006761979777365923,
0.037250880151987076,
... | [
0.6093252301216125,
0.1653355211019516,
0.2625216543674469,
-0.11888235062360764,
-0.4418002963066101,
0.13349804282188416,
0.18311087787151337,
0.16455240547657013,
-0.23743903636932373,
-0.37407928705215454,
0.44824668765068054,
0.3745276927947998,
-0.28422993421554565,
0.111151307821273... |
Is there a simple one-liner to convert a given base 10 integer to a string representation of this integer in some desired base 'b'? Is there a way to do this avoiding the subscript indicating the base, for example the small 2 if we ask for a binary string? The best thing for my purpose, despite the danger, would be have a function that takes a decimal integer and returns what appears to be the binary, ternary, etc. representation of this integer, but which Mathematica still treats as a decimal integer value. | [
0.014684250578284264,
0.020658034831285477,
-0.016138140112161636,
0.006662866100668907,
-0.01464330218732357,
-0.003944018390029669,
0.00631486251950264,
0.01219294872134924,
-0.01729607954621315,
-0.043542444705963135,
-0.006612504832446575,
0.0011571146314963698,
0.00040250702295452356,
... | [
0.010420803911983967,
0.1764483004808426,
-0.11608224362134933,
0.08857344090938568,
-0.13046376407146454,
0.16681741178035736,
0.05999506637454033,
-0.03536279872059822,
-0.21754032373428345,
-0.26901277899742126,
0.03919433429837227,
0.5029243230819702,
-0.18213781714439392,
-0.024252137... |
I've written a plugin that creates a feed reader like google reader in wordpress.  Currently I let folks reblog feed items by using the built in WP PressThis plugin url. I pass in parameters and WP formats a post for my users. I'd like to get more granular control - let my users specify a post format etc. I want to control the title, the body, and maybe even things like tags or format, but I want my users to have a chance to edit it all right before they hit publish to add in their wise thoughts. What's the WP way to do this in my plugin? | [
-0.0032856110483407974,
0.01122207473963499,
-0.001168536487966776,
0.020185615867376328,
0.014059079810976982,
0.00953960046172142,
0.00699243601411581,
0.01775244064629078,
-0.020434044301509857,
-0.015408424660563469,
-0.006150583736598492,
0.010262181982398033,
0.022947777062654495,
0.... | [
0.8167950510978699,
-0.031589556485414505,
0.5653381943702698,
0.11762132495641708,
-0.30880048871040344,
-0.1435243785381317,
-0.06785073131322861,
-0.06961428374052048,
-0.18780642747879028,
-0.57890784740448,
0.15628355741500854,
0.35564643144607544,
-0.26438450813293457,
0.119829826056... |
In previous generations, EVs only took effect upon level up, meaning that a level 100 Pokemon was stuck with the EVs you had previously trained it with. With the edition of Super Training, and specifically, 'Reset Bags', you can re-spec your EVs at any time. But do these new EVs still only take effect upon level up? Say I have a Gardevoir with mixed EVs on level 100, and I want to respec them into Special Attack & Speed for example, is this now possible? or will Resetting & Reapplying do nothing? | [
0.015597201883792877,
0.02068076655268669,
-0.003521345555782318,
0.009447375312447548,
0.01116175763309002,
0.01782785728573799,
0.008306705392897129,
-0.015482012182474136,
-0.014297285117208958,
0.009616991505026817,
-0.003633236512541771,
0.03516243398189545,
0.007523150183260441,
-0.0... | [
0.3485953211784363,
-0.43113645911216736,
0.3211325705051422,
0.16124014556407928,
-0.31330233812332153,
-0.0238961149007082,
0.33737897872924805,
-0.373677134513855,
-0.08319683372974396,
-0.14281441271305084,
-0.06282287836074829,
0.19453728199005127,
-0.2702440619468689,
-0.155125975608... |
I am having a debate at work regarding the correct choice: > No investor has ever lost its capital. or > No investor has ever lost their capital. It seems to sound awkward to say "its capital" as opposed to "their capital". We are referring to a single investor (in a context where _all_ investors are institutional entities). | [
0.04041038826107979,
0.0010920939967036247,
-0.030912375077605247,
0.009000374935567379,
0.024376198649406433,
-0.03408048301935196,
0.01402274239808321,
0.00133527594152838,
-0.020734107121825218,
-0.0695696622133255,
-0.010806857608258724,
0.02688872255384922,
0.01690024696290493,
0.0059... | [
0.3872528076171875,
0.11355337500572205,
0.2715035676956177,
0.03818366676568985,
0.033025164157152176,
0.012749083340168,
-0.3621283173561096,
0.23191243410110474,
-0.4954444468021393,
-0.36025696992874146,
-0.13357584178447723,
0.41465985774993896,
0.003019640687853098,
0.518548905849456... |
As Eclipse uses OSGi to handle modules and bundles, and therefore should be able to hot-deploy them, why does it suggest to restart after installing a plugin "for changes to take effect"? Has Eclipse hit some OSGi limitation, or is its architecture such that even with hot-deploy, it cannot guarantee that deployments went smoothly and hence takes the safest route by offering a restart? | [
-0.006410686764866114,
0.032683562487363815,
-0.007838955149054527,
0.024746207520365715,
0.013467871583998203,
0.006989095360040665,
0.010603959672152996,
0.0016207359731197357,
-0.015119995921850204,
-0.01317556481808424,
-0.021771270781755447,
0.036253415048122406,
0.010735672898590565,
... | [
0.3063235282897949,
-0.1862950325012207,
0.20269593596458435,
0.17361563444137573,
0.033042438328266144,
-0.47513920068740845,
0.13127762079238892,
-0.2412792593240738,
-0.25821173191070557,
-0.1673448532819748,
-0.31159552931785583,
0.5026984214782715,
-0.2143796682357788,
-0.447281062602... |
It seems that at Three Mile Island and in Fukushima, there was a build-up of radioactive gasses, and other gasses like hydrogen in the containment buildings. Wikipedia says: The containment [building] is the fourth and final barrier to radioactive release (part of a nuclear reactor's defence in depth strategy), the first being the fuel ceramic itself, the second being the metal fuel cladding tubes, the third being the reactor vessel and coolant system. How did radiation escape into the containment building in TMI if the reactor vessel wasn't breached? Wasn't it the same case in Fukushima that reactor vessels were intact? And where do these gasses come from? | [
-0.0067798662930727005,
0.01855308935046196,
0.014871596358716488,
0.004256065934896469,
0.002248927718028426,
-0.015031024813652039,
0.009317886084318161,
-0.034462884068489075,
-0.012974708341062069,
-0.018277499824762344,
-0.010034297592937946,
0.0038271788507699966,
0.005293217953294516,... | [
0.6396800875663757,
0.2604100704193115,
0.1280212104320526,
0.20908670127391815,
0.22809231281280518,
-0.04045331850647926,
0.5074191093444824,
-0.5334053039550781,
-0.27226394414901733,
-0.205022394657135,
0.21837563812732697,
0.12101946026086807,
-0.09390880167484283,
0.4378367066383362,... |
I realize this is subjective, but I thought it would be nice to talk about our favorite datasets and what we think makes them interesting. There is a wealth of data out there, and what with all of the APIs (e.g., Datamob) along with classic datasets (e.g., R data), I think this could have some very interesting responses. For example, I have always liked datasets like the "Boston Housing" dataset (unfortunate implications notwithstanding) and "mtcars" for their versatility. From a pedagogical standpoint, one can show the merits of a wide variety of statistical techniques using them; and Anderson/Fisher's iris dataset will always have a place in my heart. Thoughts? | [
0.007376964204013348,
0.0063685039058327675,
-0.01219197642058134,
-0.0007820390164852142,
0.0228374395519495,
0.0028612485621124506,
0.006071711890399456,
0.008147088810801506,
-0.012672096490859985,
-0.010206839069724083,
0.001378409331664443,
0.005574644543230534,
0.01210029423236847,
-... | [
0.7065175771713257,
0.1870795041322708,
-0.022162996232509613,
0.4616151452064514,
0.12215542048215866,
0.04149825870990753,
-0.2568778395652771,
0.10880772024393082,
-0.2880891263484955,
-0.5645257234573364,
-0.07937778532505035,
0.4658948481082916,
0.21952186524868011,
0.2916436195373535... |
I am trying to understand why I would use a near infrared as opposed to an infrared camera for data capture of vegetation health monitoring. | [
-0.013258220627903938,
0.021027477458119392,
-0.022034093737602234,
0.021271830424666405,
-0.0428604818880558,
0.03142017871141434,
0.020425396040081978,
0.02075864002108574,
-0.02235986478626728,
-0.023078985512256622,
-0.003462309017777443,
0.034955818206071854,
-0.04003503918647766,
0.0... | [
0.6139155030250549,
0.2504696547985077,
-0.047386132180690765,
0.03415513411164284,
0.2407396137714386,
-0.5417323112487793,
0.33538785576820374,
-0.004428075160831213,
-0.1003136932849884,
-0.5704317092895508,
0.28162863850593567,
0.48846691846847534,
-0.10884753614664078,
0.1848797351121... |
How can I draw a line from the origin to the edge of circle? \documentclass{article} \usepackage{tikz} \usepackage{fp} \usetikzlibrary{calc, arrows, fixedpointarithmetic} \begin{document} \begin{tikzpicture}[line cap = round, line join = round, >=triangle 45, scale = 1.25, fixed point arithmetic] \coordinate (O) at (0, 0); \coordinate (P1) at ($(O) + (-30:1.5cm and .75cm)$); \draw[-latex] (P1) arc (-30:310:1.5cm and .75cm) coordinate (P2); \draw[-latex] (O) -- (0, 1.75) node[above, scale = .75] {\(\mathbf{h}\)}; \draw ($(P1)!.5!(P2)$) circle (.18cm) coordinate (P3); %\draw[blue] (O) -- How to finish this? \end{tikzpicture} \end{document} I don't know what the last coordinate should be to complete the line. I have tried: ($(P3) + (-30:1.5cm and .75cm)$) and something else but I can't quite remember what else it was. Any ways, it didn't work. After that, I want to draw a dashed line from the first edge of the circle to the other edge in line with the solid blue line. | [
0.004782794509083033,
0.0074708471074700356,
0.0015016932738944888,
0.011591173708438873,
0.00033715227618813515,
-0.005787276197224855,
0.009292647242546082,
-0.0038302333559840918,
-0.01787160336971283,
-0.01510064210742712,
-0.008376069366931915,
0.0016638943925499916,
-0.0164612270891666... | [
0.021121302619576454,
-0.04519901052117348,
0.6710484623908997,
0.09306122362613678,
-0.18584571778774261,
0.36445143818855286,
0.23003248870372772,
-0.5526736378669739,
-0.051598094403743744,
-0.5870664119720459,
0.08076443523168564,
-0.045529283583164215,
0.1767590045928955,
0.0884880200... |
Would you agree that the present perfect is used more than the past tense by native speakers to emphasize the situation at hand? Some languages, like Arabic and Japanese, use the simple past much more. And the Japanese seldom use their version of the present perfect. So why are English speakers so enchanted with the present perfect? | [
0.007283939979970455,
0.02801499515771866,
-0.03649191930890083,
0.019173266366124153,
-0.027865028008818626,
0.006801031995564699,
0.013551788404583931,
-0.007438811473548412,
-0.02454189769923687,
-0.017421340569853783,
-0.01158667728304863,
0.004512782208621502,
0.013744588941335678,
0.... | [
-0.1448264718055725,
-0.014647665433585644,
-0.2792769968509674,
-0.13098718225955963,
-0.3940364122390747,
0.07587768137454987,
0.6430884599685669,
0.5105624198913574,
-0.005478560924530029,
-0.39477938413619995,
-0.3945416510105133,
0.6459043025970459,
0.04354831576347351,
-0.09927699714... |
I'm using the package `dtsyntax` to highlight code text based on listings. For in-line annotation they make a new command `\modelica{ }` based on lstinline as follows : \newcommand{\modelica}[1]{\lstinline[basicstyle=\ttfamily,language=modelica]|#1|} But when i use this as follows this is a test concerning highlighting \modelica{"modelica"} in LaTeX ... I get the following error which i can not trace back : Package Listings Error: lstinline ended by EOL. can anyone help me with this ? | [
0.017361246049404144,
0.012413768097758293,
-0.013815266080200672,
0.014208881184458733,
0.022159624844789505,
0.01320701651275158,
0.010838249698281288,
-0.0001164628192782402,
-0.011768233962357044,
0.008344728499650955,
-0.02264082059264183,
0.00012796418741345406,
-0.0075537171214818954,... | [
0.12689527869224548,
0.26481181383132935,
0.25735390186309814,
-0.3187236189842224,
-0.17857500910758972,
-0.020093059167265892,
0.408408522605896,
0.07480227202177048,
-0.19331595301628113,
-0.7785128355026245,
0.05623247101902962,
0.6036166548728943,
-0.24408653378486633,
0.2726000547409... |
I have a HTC One X (not rooted) running Android 4.1.1 but unfortunately it keeps disconnecting from my router when I am at home. The router in question is a Buffalo AirStation WZR-HP-G300NH2 running DD-WRT and the issue occurs even when I am less than one meter away from it. The phone shows the WiFi signal as full strength, but most of the time it will constantly disconnect and reconnect. The router is running in 802.11g/n mixed mode at 2.4GHz. Here's the thing: the phone was previously working fine with my network for several months yet nothing has changed that I am aware of and I have not updated the firmware of either device in some time. I still have no problems connecting to my neighbour's router (which I have been granted access to) from the same phone even though the signal is understandably poor. Also, no other devices have trouble connecting to my router. My router was configured to choose a channel automatically. Using inSSIDer on the phone I could see that the router was picking channel 6 - this seems to be logical as there does not seem to be anything else using it. I manually set the WiFi to channel 6 to be sure, but the issue remains. I'm having trouble working out whether the phone or the router is to blame. I'm guessing that there could be some kind of interference on the channel I'm using, but I don't want to change to another one because no other devices have a problem using it and I'm guessing that performance could be reduced due to interference from other networks. Any help in diagnosing the problem would be appreciated. Both devices have been restarted yet there was no improvement. | [
0.0026782029308378696,
-0.0027000869158655405,
-0.015000714920461178,
0.010800715535879135,
-0.057081181555986404,
-0.027925994247198105,
0.007972370833158493,
0.013382337987422943,
-0.008894490078091621,
-0.007377345114946365,
-0.0055753253400325775,
0.01179601065814495,
-0.0017492076149210... | [
0.3858693540096283,
0.26157689094543457,
0.7594583630561829,
0.32670214772224426,
-0.03672879561781883,
-0.4355056881904602,
0.4784790575504303,
0.10033132135868073,
0.02920088730752468,
-0.5841505527496338,
0.31085604429244995,
0.20805752277374268,
-0.18761783838272095,
0.0352593585848808... |
I saw this in an informal email but I don't understand it. Can you help me with the meaning and the context in which it is used (if used at all or if it simply a mistake from the sender) Thank you! | [
-0.060345448553562164,
0.01068886462599039,
0.008068382740020752,
0.048264048993587494,
0.027859773486852646,
-0.017156239598989487,
0.00522228516638279,
-0.014893288724124432,
-0.021911997348070145,
-0.010264551267027855,
-0.01070718839764595,
0.00989622063934803,
0.011504177935421467,
-0... | [
0.29496803879737854,
0.2400481253862381,
0.17229673266410828,
0.20055299997329712,
-0.21300309896469116,
-0.23216460645198822,
0.7260342836380005,
0.4307745099067688,
-0.0070471749641001225,
-0.30380779504776,
-0.12081252038478851,
0.12953044474124908,
0.193922221660614,
0.2827723324298858... |
Each new character created in Neocron is 'born' with a LAW Enforcer #1 CPU in one of their implant slots. Is this something that I should keep in or should I remove it? what does it do? | [
-0.004944872111082077,
0.03861454129219055,
-0.013821789994835854,
-0.005008448846638203,
-0.061247650533914566,
-0.02348511479794979,
0.01428390946239233,
-0.007018334232270718,
-0.02695210464298725,
0.031202692538499832,
-0.0087102260440588,
0.04619770124554634,
0.015557950362563133,
0.0... | [
0.34014540910720825,
0.10491868853569031,
0.2655203342437744,
0.4437921643257141,
0.30147600173950195,
0.12273725122213364,
-0.1638139933347702,
0.07362683862447739,
-0.03817197307944298,
-0.43062132596969604,
-0.013180808164179325,
0.2254260629415512,
-0.38571104407310486,
0.3344601392745... |
In our development process we have several branches of the same theme for each developer. Theme branches live in different directory names, so when doing `get_theme_mod()`, `get_option("stylesheet")` can return different values, should we for some reason (customer demo) be forced to activate another theme branch besides "master". This makes us have to reconfigure menus, widgets, customizer mods etc etc. In our case, it's very rare that theme mods need to be different across branches, so I would like `get_theme_mod()` to always access the same mods array. The obvious solution seems to be making "`pre_option_stylesheet`" filter as a standard part of the theme codebase, but I am not sure if it may have any unintended side effects. Grepping around, looks like this option is almost exclusively used in a few places in `wp-includes/theme.php`. Q: Is "`pre_option_stylesheet`" filter the way to go here? Or can you think of something else? | [
0.015951836481690407,
0.018070880323648453,
-0.002805737778544426,
0.00807026494294405,
0.0062963333912193775,
0.0039788200519979,
0.008234387263655663,
0.0018184559885412455,
-0.014568092301487923,
0.00994987040758133,
-0.027766596525907516,
0.0218830369412899,
0.006083360873162746,
0.013... | [
0.340598464012146,
-0.350881427526474,
0.5771650075912476,
0.20421989262104034,
0.15487131476402283,
0.027972858399152756,
-0.02963850647211075,
-0.4106009006500244,
-0.3225661814212799,
-0.461708664894104,
-0.22509558498859406,
0.47280359268188477,
-0.03811503201723099,
0.4679756462574005... |
I'm building a theme using the Theme Customizer and have a few options with color pickers. It seems as though once you set a color using the color picker, there's no way to clear/remove/ unset it, is that right? If you haven't saved your settings yet, you can hit the cancel button (though you'd lose all your other unsaved settings) -- but once it's saved, it's set in stone? Or am I missing something? | [
0.029598843306303024,
0.006464293226599693,
0.004653507377952337,
0.018068045377731323,
-0.01235272828489542,
0.006740008946508169,
0.006018553860485554,
-0.00613719318062067,
-0.016492584720253944,
0.01605364866554737,
0.0017057422082871199,
0.014447701163589954,
0.009698747657239437,
0.0... | [
0.573001503944397,
0.11386407166719437,
0.3659282922744751,
0.02091902308166027,
0.35021424293518066,
0.03180355951189995,
0.07064688950777054,
-0.07554057240486145,
-0.40741872787475586,
-0.3482194244861603,
0.19197116792201996,
0.2043062448501587,
-0.22917413711547852,
0.5333526134490967... |
1. In djvused bookmark format, to disable characters that have special meaning in djvused bookmark format: The format of `djvused` bookmark of a djvu file is for example: (bookmarks ("1 first chapter" "#10" ("1.1 first section" "#11" ("1.1.1 first subsection" "#12" )) ("1.2 second section" "#13" )) ("2 second chapter" "#14" ("2.1 first section" "#16" ) ("2.2 second section" "#13" )) ) ...where the main points are the paring of left and right parenthesis for tree-like organization of sections and chapters, double quote for each bookmark item and each page number is preceded by a `#`. How can I escape characters like `"`, `(` and `)` to not be interpreted as control characters in the titles of chapters and sections? e.g. The following examples will not be accepted by `djvused`: ("2.2 "Hello!" " "#13" ) ("2.2 f(g)" "#13" ) The command I use to embed bookmarks into a djvu file is `djvused in.djvu -e 'set-outline bmks' -s`, where `bmks` is the text file for bookmarks. 2. In djvused bookmark format, to enable characters that have special meaning in general text files: The character `\n` means new line. But if using it directly in djvu bookmark format, it will be shown as it is, not be interpreted as new line. For example: (bookmarks ("long title part 1 \n long title part 2" "#10" ) The long title will not be broken into two lines where `\n` is specified. | [
0.005539162550121546,
0.01913110911846161,
-0.006126976106315851,
0.022888265550136566,
0.002918195677921176,
0.019277364015579224,
0.009326591156423092,
0.010742202401161194,
-0.015622555278241634,
0.010287914425134659,
-0.019239842891693115,
0.0021837446838617325,
-0.004893263801932335,
... | [
0.37417933344841003,
0.04056687280535698,
0.43218979239463806,
-0.017001884058117867,
-0.06142525002360344,
-0.5441257357597351,
0.04469646140933037,
-0.6386745572090149,
-0.11413160711526871,
-0.33434227108955383,
-0.3953881561756134,
0.45409664511680603,
-0.5620965361595154,
0.0085066631... |
I need to generate hundreds of MBTiles from different shapefiles and same styles, so I created a xml from tilemill with a single shapefile and styling, using python landez library to generate MBTiles from this xml (keep changing datasource for each shapefile). Problem is landez is taking too long to create while tilemill is faster in exporting MBTiles, also through landez, few tiles are missing at different zoom levels (blank tile, so default color). I was just wondering if this can be done using tilemill, (tilemill seems to export only projects). PS: I want to do this process through terminal. | [
0.021436292678117752,
0.013341819867491722,
-0.0187330674380064,
0.009578069671988487,
-0.04288169741630554,
0.03213910758495331,
0.011846189387142658,
0.020535003393888474,
-0.019168823957443237,
-0.03384227305650711,
-0.002250018762424588,
0.011117236688733101,
-0.014966527000069618,
0.0... | [
0.29755815863609314,
0.18193718791007996,
0.26420876383781433,
0.055604785680770874,
-0.18561869859695435,
0.14996442198753357,
-0.017223702743649483,
-0.45996278524398804,
-0.0647258311510086,
-0.7600878477096558,
0.2606767416000366,
0.48774945735931396,
-0.2629041373729706,
0.02376812882... |
How can I account for a nonlinear independent variable in a logistic regression? For example, consider this data set: b1 b2 b3 b4 b5 1 1 0 0 20 0 1 0 0 20 1 0 0 0 13 1 1 1 0 8 1 1 0 1 5 0 1 0 1 4 1 1 0 1 5 0 0 0 0 8 0 0 1 0 13 1 0 0 1 20 1 0 0 0 29 1 1 0 1 40 1 0 0 0 53 0 1 1 0 68 Suppose I gathered some data like the above (suppose I had 10,000 rows) and I want to use them to predict future sets of similar data with an equation generated by a logistic regression. I get the above data. My independent variables are b1 through b5. By looking at the data, I can clearly see that independent variables b1 through b4 are binary (maybe they are male/female, rent/own, etc). However I see that b5 is clearly not binary. Upon graphing it and looking at the numbers I determine that the b5 variable has a bathtub shaped U curve with an equation of b5=(x-4)^2+4. I understand that using the b5 variable as it is would screw up my logistic regression - since any interval independent variables must have a linear relationship with the dependent variable. Yet, I must account for the variable b5 in my data set while doing a logistic regression for it to be accurate and predict well (right?). Given the above data, how would I go about making a predictive model with a logistic regression? Any references online or in books would greatly be appreciated as I am having trouble finding a good explanation of the above. Cheers =) EDIT: Sorry if my question is unclear. Consider this: Suppose we had the following data on many people: Y: Whether they had a stroke before the age of 50 or not 1\. Whether they own a pet or not. 2\. Whether they graduated college or not. 3\. Whether they are male or female. 4\. The number of hours they exercise every week. We want to train a logistic regression model based on historical data to predict if someone in the future with the same data will be more apt or not to have a stroke before the age of 50. The problem I am having is with the last variable. A continuous ratio variable like "number of hours exercised each week." Exercising more each week may have diminishing returns on health and thus having a stroke. For example decreasing an increasing rate. Or maybe (totally made up) exercising a little bit is really good for you, exercising a moderate amount is really bad for you, and exercising a lot is really good for you again. We might have a U shaped curve like (x-4)^2+4 for example when plotting health to exercise. I would image that having a curved, non-linear independent variable in a logistic regression like the one I described above would cause problems. Am I right? And if so, what kinds of things can you do to still include the variable in the logistic regression. Perhaps a transformation? Is that all? | [
0.011828258633613586,
0.011273548007011414,
-0.008680963888764381,
0.009373868815600872,
0.0007445462979376316,
-0.012077342718839645,
0.0050643617287278175,
0.002335580997169018,
-0.010637806728482246,
0.0049906508065760136,
-0.007935584522783756,
0.0025986703112721443,
-0.01191295310854911... | [
-0.3294870853424072,
0.2070903480052948,
0.1483868658542633,
-0.08258331567049026,
-0.3141879141330719,
0.5945569276809692,
0.25701940059661865,
-0.46791884303092957,
-0.29002222418785095,
-0.07167887687683105,
0.3540709614753723,
0.28880998492240906,
-0.29791003465652466,
0.38079917430877... |
I have a Custom Post Type and i want to rewrite the URL. nothing is working though. This is the adress i have now /?portfolio=adidas. When I rewrite the numerous ways i tried i always get this instead: feeds/active/ which seems to be a template for blogs in my theme. I think this can be due to me using the wp rss aggregator to pull rss feeds, who i call with a shortcode in my custom post, but seems to be stored somehow with the url Feeds/post_slug, and this is where i get directed. How can I change this? Tried pretty much every plugin on the market and all the code snippets i can found.. The custom post type is registered in my theme, and not in the functions. This is the code for that: $args = array( "singular_label" => __("Portfolio Category", "swiftframework"), 'public' => true, 'hierarchical' => true, 'show_ui' => true, 'show_in_nav_menus' => false, 'args' => array( 'orderby' => 'term_order' ), 'rewrite' => array('slug'=>'','with_front'=>false), 'query_var' => true ); register_taxonomy( 'portfolio-category', 'portfolio', $args ); add_action('init', 'portfolio_register'); function portfolio_register() { $labels = array( 'name' => _x('Portfolio', 'post type general name', "swiftframework"), 'singular_name' => _x('Portfolio Item', 'post type singular name', "swiftframework"), 'add_new' => _x('Add New', 'portfolio item', "swiftframework"), 'add_new_item' => __('Add New Portfolio Item', "swiftframework"), 'edit_item' => __('Edit Portfolio Item', "swiftframework"), 'new_item' => __('New Portfolio Item', "swiftframework"), 'view_item' => __('View Portfolio Item', "swiftframework"), 'search_items' => __('Search Portfolio', "swiftframework"), 'not_found' => __('No portfolio items have been added yet', "swiftframework"), 'not_found_in_trash' => __('Nothing found in Trash', "swiftframework"), 'parent_item_colon' => '' ); $args = array( 'labels' => $labels, 'public' => true, 'show_ui' => true, 'show_in_menu' => true, 'show_in_nav_menus' => false, 'menu_icon' => 'dashicons-format-image', 'rewrite' => array('slug'=>'','with_front'=>false), 'supports' => array('title', 'editor', 'thumbnail'), 'has_archive' => true, 'taxonomies' => array('portfolio-category') ); register_post_type( 'portfolio' , $args ); } add_filter("manage_edit-portfolio_columns", "portfolio_edit_columns"); function portfolio_edit_columns($columns){ $columns = array( "cb" => "<input type=\"checkbox\" />", "title" => "Portfolio Item", "description" => "Description", "portfolio-category" => "Categories" ); return $columns; } add_action("manage_posts_custom_column", "portfolio_custom_columns"); function portfolio_custom_columns($column){ global $post; switch ($column) { case "description": the_excerpt(); break; case "portfolio-category": echo get_the_term_list($post->ID, 'portfolio-category', '', ',,''); break; } } | [
0.004534454084932804,
-0.003419480286538601,
0.008971180766820908,
0.019007395952939987,
0.018665548413991928,
-0.003464288543909788,
0.007319590076804161,
0.016907233744859695,
-0.009934818372130394,
-0.012044933624565601,
-0.012575776316225529,
0.007205176632851362,
-0.00219224882312119,
... | [
0.7598583102226257,
-0.108117014169693,
0.6244499087333679,
-0.12323840707540512,
-0.30549299716949463,
-0.11454061418771744,
0.052018050104379654,
0.36032596230506897,
-0.4464167356491089,
-0.41348081827163696,
0.22481971979141235,
0.29263895750045776,
-0.4655880033969879,
0.4950907230377... |
I have defined the following functions: γ[r_, v_, rDet_] := Which[ Abs[r - v tDet] >= rDet, 0, r + v tDet <= rDet, π, True, ArcCos[((v tDet)^2 + r^2 - rDet^2)/(2 v tDet r)]]; ρ[r_, v_, v0_] := Exp[-(1/2) (r/rCloud)^2] r Exp[-(1/2) (v/v0)^2] v; ΔΦ[v_] := (2 π )/λ c (4 (v*t)^2)/(rBeam)^2; I want to weight `ΔΦ` by `ρ` and `γ`. I do so by defining a function `Awfc` that numerically integrates the product of the three functions over `v` and `r` and divides by the numerical integral over the two weighting functions: Awfc[rDet_?NumericQ, v0_?NumericQ] := NIntegrate[ γ[r,v,rDet]*ρ[r,v,v0]* Exp[I ΔΦ[v]], {r,0,∞}, {v,0,∞ }}]/ NIntegrate[ γ[r,v,rDet]*ρ[r,v,v0], {r,0,∞}, {v,0,∞}}]; I calculate `Awfc` for some parameters `rDet` and `v0` AwfcTable = ParallelTable[ {rDet, Awfc[rDet, 0.5 v0], Awfc[rDet, v0], Awfc[rDet,2 v0], Awfc[rDet,3.7 v0]}, {rDet,0.0005,0.010,0.0005}]; using these values for the other parameters: tDet = 0.7; t = 0.230; rCloud = 0.0025; λ = 780 10 ^-9; c = λ/20; v0 = 0.00588; rDet = 0.008/2; rBeam = 0.015; All kernels throw warnings: > NIntegrate::slwcon: Numerical integration converging too slowly; suspect one > of the following: singularity, value of the integration is 0, highly > oscillatory integrand, or WorkingPrecision too small. I think that `ΔΦ` is a hightly osciallating function so I don't see a way to fix this problem. I use `TableForm[Abs[AwfcTable]]` to get the amplitude of the complex numbers in `AwfcTable`. The problem with the output is, that this amplitude is significantly larger than 1, especially for larger `v0` and small `rDet`. The function `ΔΦ` I am calculating the weighted mean of, has a maximum amplitude of 1 so I think the weighted mean should also have a maximum of 1. (The complex phase is consistent with what I expect from a different approach to the problem I did using MATLAB) Is there something wrong with my reasoning (meaning that the result _Mathematica_ gives is actually correct), or is does the fault lie with my implementation of the problem? | [
-0.0003662238013930619,
0.006845196709036827,
-0.011622808873653412,
0.008118978701531887,
-0.021054044365882874,
-0.010480034165084362,
0.005333343055099249,
-0.0013698795810341835,
-0.008996358141303062,
-0.01138960849493742,
-0.004977844655513763,
0.0036220906767994165,
-0.014467489905655... | [
0.21339815855026245,
0.0474567636847496,
0.6996511816978455,
-0.35329198837280273,
0.05457761883735657,
0.43516775965690613,
0.030226171016693115,
-0.8125925660133362,
0.18745481967926025,
-0.3919547498226166,
0.1028076782822609,
0.7754760384559631,
-0.4125571548938751,
-0.0844658017158508... |
I am currently running my site via http on a url like `domain.com`. In order to continue to make use of CloudFlare's free tier of service, I want to put my admin on a subdomain like `secure.domain.com` and access it over SSL (I would set up `secure.domain.com` to bypass CloudFlare). Having just recovered from a scary episode of losing all of my blog's Jetpack email subscribers as a result of testing this scheme on a test server with a totally different domain name, I am a little nervous about what the effect will be of changing the URL of access to the admin. Does anyone have experience in doing this kind of thing and/or can anyone explain how Jetpack associates email subscribers with a given blog URL and what to expect when changing the URL for the admin to a subdomain? | [
0.0028031880501657724,
0.001999921165406704,
-0.0028330590575933456,
0.004754490219056606,
0.014990457333624363,
0.006396063603460789,
0.007503502070903778,
0.017354816198349,
-0.012458980083465576,
0.0035003460943698883,
-0.010894905775785446,
0.004067409783601761,
-0.013799098320305347,
... | [
0.6680408716201782,
0.24768134951591492,
0.5842656493186951,
-0.1339559555053711,
-0.13997682929039001,
-0.24254217743873596,
0.5808570384979248,
0.5062954425811768,
-0.23426130414009094,
-0.7575944066047668,
0.09878858923912048,
0.23534046113491058,
0.15162639319896698,
0.6750686764717102... |
I literally have no idea how to fix this. I've looked and looked and I've found no solution. I've reinstalled it many times. I've checked my files against someone else's that has no issues with their pixelmon mod and we have the same files, same sizes, everything. Does anyone have any idea?   | [
0.0022631760220974684,
0.007555650547146797,
0.002610353520140052,
0.028342023491859436,
-0.002912284340709448,
-0.0198377575725317,
0.004022897221148014,
0.006691945716738701,
-0.022587178274989128,
0.0056700087152421474,
-0.007785875350236893,
0.015188779681921005,
0.008950833231210709,
... | [
0.21162857115268707,
-0.18645700812339783,
0.2742164731025696,
0.29387593269348145,
-0.22740672528743744,
-0.22805653512477875,
0.27606016397476196,
0.26055413484573364,
-0.2452823966741562,
-0.16956059634685516,
-0.18036387860774994,
0.1529853790998459,
-0.23993326723575592,
0.21688264608... |
I have published the data in Geoserver (Tomcat and OS Ubuntu) and I am developing a map application in my local machine (OS windows xp) using asp.net. In my application, I can get WMS feature information in IE8 with the confirmation msg "This page is accessing information that is not under its control. This poses a security risk. Do you want to continue?" given Yes. But in Chrome, i cannt get feature info. It says XMLHttpRequest cannot load http://xxx.xxx.x.xxx:8080/geoserver/cite/wms?LAYERS=cite%3Aplym_p1_examplec…e%2Fpng&INFO_FORMAT=application%2Fvnd.ogc.gml&SRS=EPSG%3A27700&X=596&Y=160. Origin http://localhost:3890 is not allowed by Access-Control-Allow-Origin. I have found similar questions Probleme with WFS protocol in OpenLayers but its for after hosting the web application in geoserver.My case is still its in under developing in my local machine. My code is: roadidPicker = new OpenLayers.Control.WMSGetFeatureInfo({ url: 'http://xxx.xxx.x.xxx:8080/geoserver/cite/wms', title: 'identify features on click', queryVisible: true }); roadidPicker.infoFormat = 'application/vnd.ogc.gml'; roadidPicker.events.register("getfeatureinfo", this, pickRoadid); osMap.addControl(roadidPicker); roadidPicker.activate(); function pickRoadid(e) { if (e.features.length > 0) { var val = e.features[0].attributes['theme']; alert(val); } } I have configured the proxy.cgi file in Geoserver (usr/lip/cgi-bin/proxy.cgi) and added my local ip address in allowedHosts = ['xxx.xxx.x.xxx', 'www.openlayers.org', 'openlayers.org', 'labs.metacarta.com', 'world.freemap.in', 'prototype.openmnnd.org', 'geo.openplans.org', 'sigma.openplans.org', 'demo.opengeo.org', 'www.openstreetmap.org', 'sample.azavea.com', 'v2.suite.opengeo.org', 'v-swe.uni-muenster.de:8080', 'vmap0.tiles.osgeo.org','www.openrouteservice.org'] and added the below line in my script. OpenLayers.ProxyHost = "/cgi-bin/proxy.cgi?url="; but it says another error in chrome console: GET localhost:3890/usr/lib/cgi-bin/….…T%3Dapplication%252Fvnd.ogc.gml%26SRS%3DEPSG%253A27700%26X%3D682%26Y%3D284 404 (Not Found) I dont know whether its right or wrong. Proxy.cgi should be copied in Geoserver or my machine? in which path? Please anyone suggest me to fix this issue | [
-0.025857850909233093,
-0.016218796372413635,
-0.009534846991300583,
0.015688326209783554,
-0.004772683139890432,
0.010379074141383171,
0.007423611357808113,
0.014827649109065533,
-0.01606418937444687,
-0.011918645352125168,
-0.004448579624295235,
0.02126089110970497,
-0.017326440662145615,
... | [
0.07472804188728333,
0.15901854634284973,
0.8949941396713257,
-0.0927555188536644,
-0.17657527327537537,
-0.2397044152021408,
0.08302364498376846,
-0.04141298681497574,
-0.3465682864189148,
-0.8944877982139587,
-0.09782254695892334,
0.369057834148407,
-0.27867087721824646,
0.03106839023530... |
Consider the regression equations below: \begin{align} Y_i &= \beta_0 + \beta_1 X_{i1} + \varepsilon_i \\\ Y_j &= \beta_0 + \beta_1 X_{j1} + \beta_2 X_{j2} + \varepsilon_j \end{align} where $Y_i,\ X_{i1},\ \varepsilon_i,\ Y_j,\ X_{j1},\ \& \ X_{j2},\ \varepsilon_j$ are vectors, and $_i$ and $_j$ index distinct sets of observations. The $_i$ respondents did not meet a qualification criterion and hence were not asked the question that corresponds to $X_2$. The dependent variable and the first independent variable is the same in both regression equations but the second regression equation has an independent variable that is not present in the first. Obviously, I can estimate the two regressions separately but that will not be efficient. Therefore, I was considering re-writing the first one as: $$ Y_i = \beta_0 + \beta_1 X_{i1} + \beta_2 X_{i2} + \varepsilon_j $$ where $X_{i2}$ is a vector of $0$s. Then I can estimate the parameter estimates by using OLS with the equation below: $\left[ \begin{array}{ccc} Y_i\\\ Y_j\end{array} \right] = \left[ \begin{array}{ccc} {\bf 1} & X_{i1} & X_{i2} \\\ {\bf 1} & X_{j1} & X_{j2}\end{array} \right] \left[ \begin{array}{ccc} \beta_0\\\ \beta_1\\\ \beta_2\end{array} \right] + \left[ \begin{array}{ccc} \epsilon_i\\\ \epsilon_j\end{array} \right]$ In the above equation, ${\bf 1}$ stands for a vector of $1$s of the appropriate dimension. Is the above a standard approach to obtaining efficient estimates? Is there a name to this way of estimation? | [
0.005940956994891167,
0.023861680179834366,
-0.00558066088706255,
-0.0006998732569627464,
-0.0005449221935123205,
-0.006661002058535814,
0.005113231483846903,
0.0049752481281757355,
-0.0035449371207505465,
0.0003059776499867439,
-0.009359425865113735,
0.009471758268773556,
-0.027294993400573... | [
-0.653083324432373,
-0.2713165879249573,
0.3605659008026123,
-0.07531975209712982,
0.23470373451709747,
0.29129141569137573,
0.12383951991796494,
-0.49565809965133667,
-0.11917705088853836,
-0.20363718271255493,
0.0990876778960228,
0.7974405288696289,
-0.3431028425693512,
0.314792335033416... |
The HUD takes up too much space and, as a result, the game _looks_ like it's running 640x480 because of the **huge default HUD** while it's actually running at 1080p. Is there a way to make the HUD smaller?  | [
-0.024231167510151863,
-0.008770543150603771,
0.0035694444086402655,
-0.002124559134244919,
-0.013812724500894547,
-0.026374006643891335,
0.005302067380398512,
-0.01300707459449768,
-0.01368718221783638,
0.01388917863368988,
-0.00874178484082222,
0.0027083938475698233,
0.012917734682559967,
... | [
-0.1713726818561554,
-0.06915325671434402,
0.4585167169570923,
0.004327188711613417,
0.36925241351127625,
0.07534744590520859,
0.1262180507183075,
0.1471886932849884,
-0.6137309670448303,
-0.34888675808906555,
-0.28556734323501587,
0.20280425250530243,
-0.31300967931747437,
0.1355218142271... |
Since normal attacks seem to be right out in this game, I'm not sure if it matters what weapons I'm using. I understand that the damage of my weapon affects the damage of my skills (for some bizarre reason) but beyond that nothing seems to matter. For example, I seem to be able to equip bows, but I have no way to shoot them that I can determine! So, if I have a bow that does 10 DPS and a dagger that does 10 DPS (assuming my other hand is empty anyway for argument's sake) aren't these identical in everything but appearance? Also, most of my skills seem to be elemental, so if I have an elemental damage weapon, how does that damage get applied? As what type? Again, should I just go for whatever is highest? Sorry if this is a dumb question—I almost feel like the help text must have explained this and I simply missed it because this seems incredibly counterintuitive to me. (Searched and couldn't find any question that addressed this directly, but apologies if I missed one.) | [
-0.010758713819086552,
0.019811592996120453,
-0.00900525413453579,
-0.006925374269485474,
-0.006698813289403915,
-0.007434115745127201,
0.00585569441318512,
-0.004964121617376804,
-0.010938944295048714,
-0.01652189902961254,
-0.009096645750105381,
0.02030118741095066,
-0.0009114621207118034,... | [
0.2924439311027527,
-0.057881783694028854,
0.20494480431079865,
0.10828126966953278,
-0.6264848709106445,
0.1380760371685028,
0.557030975818634,
-0.19943632185459137,
-0.3804979622364044,
-0.4217466711997986,
0.43528836965560913,
0.6914406418800354,
0.27493229508399963,
-0.4299851953983307... |
I have built a wordpress plugin (wp-content/plugins/MyPlugin) and I would to check some folders permission inside the MyPlugin directory. My problem is that I don't what Wordpress constant or what PHP function I should use to target on of these directories. if( !is_writable(WP_PLUGIN_URL.'MyFolder') ) { echo 'NOK'.'<hr>'; } This piece of code always returns 'NOK' whereas I am 100% sure 'MyFolder' is actually writable . Can you give me some piece of advice here to get the right file in order to target any folder that would be in my plugin directory? thanks | [
0.008192043751478195,
-0.0031274016946554184,
-0.00968998298048973,
0.02282526157796383,
-0.010221519507467747,
0.01821272075176239,
0.007375685498118401,
0.014804993756115437,
-0.012266438454389572,
0.012017801403999329,
-0.015031086280941963,
0.005908964201807976,
-0.001495924312621355,
... | [
0.23136276006698608,
0.25812485814094543,
0.556474506855011,
-0.1996813267469406,
0.0010847393423318863,
-0.03666844218969345,
0.2819175720214844,
-0.17791053652763367,
-0.05770614743232727,
-0.6152721643447876,
0.08858200162649155,
0.17356915771961212,
-0.2716132402420044,
0.3940724432468... |
Can one show that in quantum field theory _at least some example_ massive particles propagate with speed less than speed of light, while massless travel at speed of light? Well, motion is a different thing in QM than in classical mechanics, and question might be formulated differently. | [
0.017168529331684113,
0.013871897011995316,
-0.0042451792396605015,
0.007202977314591408,
-0.04223138466477394,
-0.01706387847661972,
0.015181005001068115,
-0.013423421420156956,
-0.025351688265800476,
-0.014583136886358261,
-0.015426498837769032,
0.03435828164219856,
-0.021976348012685776,
... | [
0.06258679926395416,
-0.2776115834712982,
0.20537789165973663,
0.419949471950531,
-0.06806961447000504,
-0.22166521847248077,
-0.08648499846458435,
-0.37746116518974304,
-0.41657283902168274,
-0.5821752548217773,
0.08326835930347443,
0.2934258282184601,
-0.10106626898050308,
0.389003157615... |
I am using the below to fetch the content of a page by its name and output the html $my_wp_query = new WP_Query(); $page = $my_wp_query->query( array( 'post_type' => 'page', 'name' => 'some-page-slug' ) ); $page = $page[0]; $page_data['content'] = apply_filters('the_content', $page->post_content); The field `$page->post_content` unmodified contains a string with the content I have typed in the backend as well as a reference to the gallery in the form `[gallery="1,2,3,4"]` at the end of the string. When I run it through the `apply_filters('the_content', $page->post_content)` method, it restores the HTML tags such as the `<p>` elements, however it also includes the gallery code which has its own style and quite lengthy markup as illustrated below. <p>Residential Gallery. Etiam lobortis sapien vel erat sodales et suscipit leo pretium. Pellentesque a arcu sed velit congue sagittis ac vel erat. Curabitur non felis leo, vel vestibulum purus. Donec eget velit eget magna rutrum facilisis. In hac habitasse platea dictumst.</p> <p>Quisque sit amet pellentesque risus. Aliquam erat volutpat. Ut at magna justo, tincidunt viverra odio. Pellentesque a arcu sed velit congue sagittis ac vel erat. Curabitur non felis leo, vel vestibulum purus. Donec eget velit eget magna rutrum facilisis. In hac habitasse platea dictumst. Quisque sit amet pellentesque risus. Aliquam erat volutpat.</p> <p>Ut at magna justo, tincidunt viverra odio. Curabitur non felis leo, vel vestibulum purus. Donec eget velit eget magna rutrum facilisis. In hac habitasse platea dictumst. Quisque sit amet pellentesque risus. Aliquam erat volutpat. Ut at magna justo, tincidunt viverra odio.</p> <style type='text/css'> #gallery-1 { margin: auto; } #gallery-1 .gallery-item { float: left; margin-top: 10px; text-align: center; width: 33%; } #gallery-1 img { border: 2px solid #cfcfcf; } #gallery-1 .gallery-caption { margin-left: 0; } </style> <!-- see gallery_shortcode() in wp-includes/media.php --> <div id='gallery-1' class='gallery galleryid-0 gallery-columns-3 gallery-size-thumbnail'><dl class='gallery-item'> <dt class='gallery-icon'> <a href='http://local.definitive-final2.co.uk/blog/residential/georgian-house/georgian1/' title='georgian1'><img width="150" height="150" src="http://local.definitive-final2.co.uk/blog/wp-content/uploads/2013/01/georgian1-150x150.jpg" class="attachment-thumbnail" alt="georgian1" /></a> </dt></dl><dl class='gallery-item'> <dt class='gallery-icon'> <a href='http://local.definitive-final2.co.uk/blog/residential/georgian-house/georgian2/' title='georgian2'><img width="150" height="150" src="http://local.definitive-final2.co.uk/blog/wp-content/uploads/2013/01/georgian2-150x150.jpg" class="attachment-thumbnail" alt="georgian2" /></a> </dt></dl><dl class='gallery-item'> <dt class='gallery-icon'> <a href='http://local.definitive-final2.co.uk/blog/residential/georgian-house/georgian3/' title='georgian3'><img width="150" height="150" src="http://local.definitive-final2.co.uk/blog/wp-content/uploads/2013/01/georgian3-150x150.jpg" class="attachment-thumbnail" alt="georgian3" /></a> </dt></dl><br style="clear: both" /><dl class='gallery-item'> <dt class='gallery-icon'> <a href='http://local.definitive-final2.co.uk/blog/residential/georgian-house/georgian4/' title='georgian4'><img width="150" height="150" src="http://local.definitive-final2.co.uk/blog/wp-content/uploads/2013/01/georgian4-150x150.jpg" class="attachment-thumbnail" alt="georgian4" /></a> </dt></dl><dl class='gallery-item'> <dt class='gallery-icon'> <a href='http://local.definitive-final2.co.uk/blog/residential/paris-house/paris4/' title='paris4'><img width="150" height="150" src="http://local.definitive-final2.co.uk/blog/wp-content/uploads/2013/01/paris4-150x150.jpg" class="attachment-thumbnail" alt="paris4" /></a> </dt></dl> <br style='clear: both;' /> </div> I would like to only retrieve the HTML text with `<p>`'s etc and not the gallery's `<style>` and `<div>` tags. How might I achieve this? | [
-0.024346420541405678,
0.00820275116711855,
0.006020534783601761,
0.005228972062468529,
-0.018946148455142975,
0.012026896700263023,
0.007028316613286734,
0.021639224141836166,
-0.010938763618469238,
-0.0031343826558440924,
-0.0038695470429956913,
0.010369479656219482,
-0.009131273254752159,... | [
0.13881263136863708,
0.2441437691450119,
0.8182758688926697,
-0.29059919714927673,
-0.24826112389564514,
0.024577705189585686,
0.07189596444368362,
-0.4581052362918854,
-0.0003879882860928774,
-0.694359540939331,
-0.1367579698562622,
0.1326846182346344,
-0.47999098896980286,
0.268268406391... |
I have an issue in that after a period of time, my people start approaching a mental breakdown caused by not eating. As you can see from the below, I have plenty of food resources but one of my colonists, Holt, hasn't eaten in quite a while, to the point where she's starting and about to have a mental breakdown (as indicated by the flashing !)  Why are my people refusing to eat? What can I do about it? Are there any particular actions I must do to ensure that they all eat? | [
0.004199197515845299,
0.012576811015605927,
0.002086907159537077,
0.010056868195533752,
-0.016791649162769318,
-0.023688480257987976,
0.004875635728240013,
0.00874342955648899,
-0.011102688498795033,
-0.005097112152725458,
-0.015116214752197266,
0.011002457700669765,
-0.007024518214166164,
... | [
0.6228511333465576,
0.4499622881412506,
0.268694132566452,
-0.05013759806752205,
0.17019936442375183,
0.5035850405693054,
0.10924246907234192,
0.49201279878616333,
-0.7895111441612244,
-0.4870046377182007,
-0.06239400431513786,
0.05211057886481285,
-0.15479829907417297,
0.5808574557304382,... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.