text stringlengths 23 30.4k | embeddings_A list | embeddings_B list |
|---|---|---|
I am on Debian. I configured my NIC with a static IP (192.168.1.56). When I try to connect to a network, initially with `ifconfig eth2` I get (correctly): eth2 inet addr:192.168.1.56 .... inet6 addr: fe80:221:ff:fe96:4598/64 but after a few seconds the `192.168.1.56` disappears and after some other seconds the inet6 address disappears too. When I press in the `nm-applet` it requires a password, but in the meantime it tries to connect. At my university, the connection is a DHCP one. It works for the first few seconds but after it doesn't. How do I go about fixing this? Here it is the relevant part of the syslog: (static ip configuration) Jan 5 16:38:29 patina dhcdbd: message_handler: message handler not found under /com/redhat/dhcp/eth2 for sub-path eth2.dbus.get.reason Jan 5 16:38:29 patina NetworkManager: <debug> [1294245509.629163] nm_device_802_11_wireless_get_activation_ap(): Forcing AP 'Cagei2' Jan 5 16:38:29 patina NetworkManager: <info> User Switch: /org/freedesktop/NetworkManager/Devices/eth2 / Cagei2 Jan 5 16:38:29 patina NetworkManager: <info> Deactivating device eth2. Jan 5 16:38:29 patina NetworkManager: <WARN> nm_device_802_11_wireless_set_wep_enc_key(): error setting key for device eth2: Invalid argument Jan 5 16:38:29 patina NetworkManager: <info> Device eth2 activation scheduled... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) started... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) scheduled... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) started... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) scheduled... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) complete. Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) starting... Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2/wireless): access point 'Cagei2' is encrypted, but NO valid key exists. New key needed. Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) New wireless user key requested for network 'Cagei2'. Jan 5 16:38:29 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) complete. Jan 5 16:40:15 patina dhclient: There is already a pid file /var/run/dhclient.eth0.pid with pid 2304 Jan 5 16:40:15 patina dhclient: killed old client process, removed PID file Jan 5 16:40:15 patina dhclient: Internet Systems Consortium DHCP Client V3.1.1 Jan 5 16:40:15 patina dhclient: Copyright 2004-2008 Internet Systems Consortium. Jan 5 16:40:15 patina dhclient: All rights reserved. Jan 5 16:40:15 patina dhclient: For info, please visit http://www.isc.org/sw/dhcp/ Jan 5 16:40:15 patina dhclient: Jan 5 16:40:15 patina dhclient: Listening on LPF/eth0/00:23:8b:54:96:5d Jan 5 16:40:15 patina dhclient: Sending on LPF/eth0/00:23:8b:54:96:5d Jan 5 16:40:15 patina dhclient: Sending on Socket/fallback Jan 5 16:40:15 patina dhclient: DHCPRELEASE on eth0 to 192.168.1.1 port 67 Jan 5 16:40:15 patina avahi-autoipd(eth0)[2133]: Got SIGTERM, quitting. Jan 5 16:40:15 patina avahi-autoipd(eth0)[2133]: Callout STOP, address 169.254.6.95 on interface eth0 Jan 5 16:40:48 patina dhcdbd: message_handler: message handler not found under /com/redhat/dhcp/eth2 for sub-path eth2.dbus.get.reason Jan 5 16:40:48 patina NetworkManager: <debug> [1294245648.742672] nm_device_802_11_wireless_get_activation_ap(): Forcing AP 'Cagei2' Jan 5 16:40:48 patina NetworkManager: <info> User Switch: /org/freedesktop/NetworkManager/Devices/eth2 / Cagei2 Jan 5 16:40:48 patina NetworkManager: <info> Deactivating device eth2. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2): cancelling... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) cancellation handler scheduled... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2): waiting for device to cancel activation. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) cancellation handled. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2): cancelled. Jan 5 16:40:48 patina NetworkManager: <WARN> nm_device_802_11_wireless_set_wep_enc_key(): error setting key for device eth2: Invalid argument Jan 5 16:40:48 patina NetworkManager: <info> Device eth2 activation scheduled... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) started... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) scheduled... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) started... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) scheduled... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) complete. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) starting... Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2/wireless): access point 'Cagei2' is encrypted, but NO valid key exists. New key needed. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) New wireless user key requested for network 'Cagei2'. Jan 5 16:40:48 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) complete. Jan 5 16:40:54 patina kernel: [ 1397.945334] lo: Disabled Privacy Extensions Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) New wireless user key for network 'Cagei2' received. Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) scheduled... Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) started... Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) scheduled... Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) Stage 1 of 5 (Device Prepare) complete. Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) starting... Jan 5 16:41:06 patina NetworkManager: <info> Activation (eth2/wireless): access point 'Cagei2' is encrypted, and a key exists. No new key needed. Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'INTERFACE_ADD eth2#011#011wext#011/var/run/wpa_supplicant#011' Jan 5 16:41:08 patina kernel: [ 1411.596736] ADDRCONF(NETDEV_CHANGE): eth2: link becomes ready Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'AP_SCAN 1' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'ADD_NETWORK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was '0' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'SET_NETWORK 0 ssid 436167656932' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'SET_NETWORK 0 proto WPA2' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'SET_NETWORK 0 key_mgmt WPA-PSK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'SET_NETWORK 0 psk <key>' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> SUP: sending command 'ENABLE_NETWORK 0' Jan 5 16:41:08 patina NetworkManager: <info> SUP: response was 'OK' Jan 5 16:41:08 patina NetworkManager: <info> Activation (eth2) Stage 2 of 5 (Device Configure) complete. Jan 5 16:41:08 patina NetworkManager: <info> Supplicant state changed: 1 Jan 5 16:41:08 patina NetworkManager: <info> Activation (eth2/wireless) Stage 2 of 5 (Device Configure) successful. Connected to access point 'Cagei2'. Jan 5 16:41:08 patina NetworkManager: <info> Activation (eth2) Stage 3 of 5 (IP Configure Start) scheduled. Jan 5 16:41:08 patina NetworkManager: <info> Activation (eth2) Stage 3 of 5 (IP Configure Start) started... Jan 5 16:41:09 patina NetworkManager: <info> Activation (eth2) Beginning DHCP transaction. Jan 5 16:41:09 patina NetworkManager: <info> Activation (eth2) Stage 3 of 5 (IP Configure Start) complete. Jan 5 16:41:09 patina NetworkManager: <info> DHCP daemon state is now 12 (successfully started) for interface eth2 Jan 5 16:41:09 patina dhclient: Internet Systems Consortium DHCP Client 4.1.1-P1 Jan 5 16:41:09 patina dhclient: Copyright 2004-2010 Internet Systems Consortium. Jan 5 16:41:09 patina dhclient: All rights reserved. Jan 5 16:41:09 patina dhclient: For info, please visit https://www.isc.org/software/dhcp/ Jan 5 16:41:09 patina dhclient: Jan 5 16:41:09 patina dhclient: Listening on LPF/eth2/00:21:00:96:45:98 Jan 5 16:41:09 patina dhclient: Sending on LPF/eth2/00:21:00:96:45:98 Jan 5 16:41:09 patina dhclient: Sending on Socket/fallback Jan 5 16:41:10 patina dhclient: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 7 Jan 5 16:41:17 patina dhclient: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 12 Jan 5 16:41:18 patina kernel: [ 1421.604521] eth2: no IPv6 routers present Jan 5 16:41:29 patina dhclient: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 19 Jan 5 16:41:48 patina dhclient: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 14 Jan 5 16:41:54 patina NetworkManager: <info> Device 'eth2' DHCP transaction took too long (>45s), stopping it. Jan 5 16:41:54 patina dhclient: Internet Systems Consortium DHCP Client 4.1.1-P1 Jan 5 16:41:54 patina dhclient: Copyright 2004-2010 Internet Systems Consortium. Jan 5 16:41:54 patina dhclient: All rights reserved. Jan 5 16:41:54 patina dhclient: For info, please visit https://www.isc.org/software/dhcp/ Jan 5 16:41:54 patina dhclient: Jan 5 16:41:54 patina dhclient: Listening on LPF/eth2/00:21:00:96:45:98 Jan 5 16:41:54 patina dhclient: Sending on LPF/eth2/00:21:00:96:45:98 Jan 5 16:41:54 patina dhclient: Sending on Socket/fallback Jan 5 16:41:54 patina dhclient: DHCPRELEASE on eth2 to 137.222.253.66 port 67 Jan 5 16:41:54 patina dhclient: send_packet: Network is unreachable Jan 5 16:41:54 patina dhclient: send_packet: please consult README file regarding broadcast address. Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) Stage 4 of 5 (IP Configure Timeout) scheduled... Jan 5 16:41:55 patina NetworkManager: <info> DHCP daemon state is now 14 (normal exit) for interface eth2 Jan 5 16:41:55 patina NetworkManager: <info> DHCP daemon state is now 14 (normal exit) for interface eth2 Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) Stage 4 of 5 (IP Configure Timeout) started... Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) failure scheduled... Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) Stage 4 of 5 (IP Configure Timeout) complete. Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) failed for access point (Cagei2) Jan 5 16:41:55 patina NetworkManager: <info> Activation (eth2) failed. Jan 5 16:41:55 patina NetworkManager: <info> Deactivating device eth2. Jan 5 16:41:55 patina NetworkManager: <WARN> nm_device_802_11_wireless_set_wep_enc_key(): error setting key for device eth2: Invalid argument Jan 5 16:41:56 patina kernel: [ 1460.455299] ADDRCONF(NETDEV_CHANGE): eth2: link becomes ready Jan 5 16:42:07 patina kernel: [ 1471.332129] eth2: no IPv6 routers present Jan 5 16:46:00 patina kernel: [ 1703.976186] usb 1-3: new high speed USB device using ehci_hcd and address 2 Jan 5 16:46:00 patina kernel: [ 1704.120840] usb 1-3: New USB device found, idVendor=13fe, idProduct=1e00 Jan 5 16:46:00 patina kernel: [ 1704.120851] usb 1-3: New USB device strings: Mfr=1, Product=2, SerialNumber=3 Jan 5 16:46:00 patina kernel: [ 1704.120858] usb 1-3: Product: USB DISK 2.0 Jan 5 16:46:00 patina kernel: [ 1704.120863] usb 1-3: Manufacturer: Jan 5 16:46:00 patina kernel: [ 1704.120868] usb 1-3: SerialNumber: 078203000F3C Jan 5 16:46:00 patina kernel: [ 1704.121082] usb 1-3: configuration #1 chosen from 1 choice Jan 5 16:46:00 patina NetworkManager: <debug> [1294245960.538897] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C'). Jan 5 16:46:00 patina kernel: [ 1704.328698] Initializing USB Mass Storage driver... Jan 5 16:46:00 patina kernel: [ 1704.328913] scsi6 : SCSI emulation for USB Mass Storage devices Jan 5 16:46:00 patina kernel: [ 1704.329254] usbcore: registered new interface driver usb-storage Jan 5 16:46:00 patina kernel: [ 1704.329263] USB Mass Storage support registered. Jan 5 16:46:00 patina kernel: [ 1704.331074] usb-storage: device found at 2 Jan 5 16:46:00 patina kernel: [ 1704.331080] usb-storage: waiting for device to settle before scanning Jan 5 16:46:00 patina NetworkManager: <debug> [1294245960.740962] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C_if0'). Jan 5 16:46:00 patina NetworkManager: <debug> [1294245960.742373] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C_if0_scsi_host'). Jan 5 16:46:05 patina kernel: [ 1709.329094] usb-storage: device scan complete Jan 5 16:46:05 patina kernel: [ 1709.379165] scsi 6:0:0:0: Direct-Access USB DISK 2.0 PMAP PQ: 0 ANSI: 0 CCS Jan 5 16:46:05 patina kernel: [ 1709.380922] sd 6:0:0:0: Attached scsi generic sg2 type 0 Jan 5 16:46:05 patina NetworkManager: <debug> [1294245965.792242] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C_if0_scsi_host_0'). Jan 5 16:46:05 patina NetworkManager: <debug> [1294245965.795336] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C_if0_scsi_host_0_scsi_device_lun0'). Jan 5 16:46:05 patina NetworkManager: <debug> [1294245965.799318] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/usb_device_13fe_1e00_078203000F3C_if0_scsi_host_0_scsi_device_lun0_scsi_generic'). Jan 5 16:46:06 patina kernel: [ 1710.042152] sd 6:0:0:0: [sdb] 8060928 512-byte logical blocks: (4.12 GB/3.84 GiB) Jan 5 16:46:06 patina kernel: [ 1710.043009] sd 6:0:0:0: [sdb] Write Protect is off Jan 5 16:46:06 patina kernel: [ 1710.043020] sd 6:0:0:0: [sdb] Mode Sense: 23 00 00 00 Jan 5 16:46:06 patina kernel: [ 1710.043026] sd 6:0:0:0: [sdb] Assuming drive cache: write through Jan 5 16:46:06 patina kernel: [ 1710.047632] sd 6:0:0:0: [sdb] Assuming drive cache: write through Jan 5 16:46:06 patina kernel: [ 1710.047643] sdb: sdb1 Jan 5 16:46:06 patina kernel: [ 1710.113494] sd 6:0:0:0: [sdb] Assuming drive cache: write through Jan 5 16:46:06 patina kernel: [ 1710.113504] sd 6:0:0:0: [sdb] Attached SCSI removable disk Jan 5 16:46:06 patina NetworkManager: <debug> [1294245966.835564] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/storage_serial__USB_DISK_2_0_078203000F3C_0_0'). Jan 5 16:46:07 patina NetworkManager: <debug> [1294245967.016379] nm_hal_device_added(): New device added (hal udi is '/org/freedesktop/Hal/devices/volume_uuid_BCD8_97AF'). Jan 5 16:46:07 patina kernel: [ 1710.696802] FAT: utf8 is not a recommended IO charset for FAT filesystems, filesystem will be case sensitive! | [
0.004614595323801041,
0.0008870010497048497,
-0.019200727343559265,
-0.0004356622230261564,
-0.010020794346928596,
0.01758478581905365,
0.008729035034775734,
0.010747451335191727,
-0.011079547926783562,
0.010034525766968727,
-0.01899450644850731,
0.007162227761000395,
-0.02953454479575157,
... | [
0.5738212466239929,
0.04278799891471863,
0.09362319856882095,
0.04002558812499046,
-0.07655806094408035,
-0.4354846477508545,
0.35927730798721313,
-0.06162413954734802,
0.05372868478298187,
-0.8341995477676392,
-0.041746120899915695,
0.6012669205665588,
-0.5789101719856262,
0.1438117474317... |
I just installed Linux Mint 12 using mint4win just to test it out. I have a problem with my wireless connection. When I start Mint, it sees the network I wish to join, I select it, enter the correct password, and 1 minute later I am being prompted for a password again. During that waiting time, I was not connected at all. It appears if I turn off my Wireless (by tapping on the Wireless LED on the laptop), it wont turn on again either. My Ethernet port is broken as well, so I cannot connect with a cable either. In Windows I can connect fine using Wireless. Is there any drivers required for this to work? I cant seem to find any, and they have to be downloadable from Windows so I can put them on an USB and run them in Mint. Thank you in advance EDIT: My Wireless chip is an Atheros AR9285 | [
0.006978381425142288,
-0.014830196276307106,
-0.013386270962655544,
0.0009387328173033893,
-0.026960164308547974,
-0.009315391071140766,
0.008007725700736046,
-0.0021580024622380733,
-0.010245757177472115,
0.0003405272727832198,
-0.01815127208828926,
0.008436278440058231,
-0.0263709053397178... | [
0.4746347963809967,
0.08023912459611893,
-0.08501233160495758,
-0.045173294842243195,
0.05428362637758255,
-0.11765884608030319,
0.4093452990055084,
0.32724252343177795,
-0.09324860572814941,
-0.7496551871299744,
0.10069730132818222,
0.7513637542724609,
-0.31437647342681885,
-0.00549197895... |
I have a grayscale (one channel) image in Mathematica. How can I export it as a grayscale Bitmap? Every single option I tried so far does not have an influence on the Bitmap exported using `Export`. If I export the image as PNG, Mathematica exports the grayscale image as expected. **Simple Example** Let's create a simple grayscale image: image = Image[RandomReal[{0, 1}, {100, 150}]]; Simple export via: Export["test.bmp",image]; shows exactly the problem mentioned above. Additional options like setting `ColorSpace->"Grayscale"` does not seem to have any effect. Any kind of color- conversion before doing the export did not work either, which is not surprising to me because the image already clearly carries the information that it is a grayscale image and export via Export["test.png",image]; does export a grayscale PNG. Verification can for example be done by simple reimporting the ColorSpace Element of the exported image: Import["test.bmp","ColorSpace"]; | [
0.014564210548996925,
0.007122145500034094,
-0.0009391255443915725,
0.02116531692445278,
-0.019173823297023773,
-0.014112195000052452,
0.008441025391221046,
0.006994364783167839,
-0.017887858673930168,
-0.017999103292822838,
-0.017249830067157745,
0.004181108437478542,
-0.011731991544365883,... | [
0.13417068123817444,
-0.12179506570100784,
0.5297743678092957,
0.09196358919143677,
-0.08660278469324112,
-0.011954022571444511,
-0.009020634926855564,
-0.20016616582870483,
0.02552955038845539,
-0.8060441613197327,
0.09869670867919922,
0.6132975816726685,
-0.34739774465560913,
0.018822556... |
Is there a package which allows me to render (n)curses based output (VT100, ANSI ...) directly in LaTeX (including colors if possible)? I'm currently making screenshots and include the PNGs via graphicx. I'm just wondering whether it's possible to achieve a higher quality. | [
-0.009620134718716145,
-0.005684936884790659,
-0.0036363054532557726,
0.020862746983766556,
-0.029418107122182846,
0.002619529375806451,
0.008562490344047546,
0.004793102853000164,
-0.030715640634298325,
-0.022125225514173508,
-0.0035983133129775524,
0.011588900350034237,
-0.0019402393372729... | [
0.6902300715446472,
0.07565339654684067,
-0.020522871986031532,
-0.006111534778028727,
-0.21944104135036469,
0.13925392925739288,
0.06325913965702057,
0.1389792412519455,
-0.20833180844783783,
-0.7081547975540161,
-0.048170194029808044,
0.7222471237182617,
-0.1518762707710266,
-0.112834036... |
Is it possible to use memoir without having chapters? | [
0.07955276221036911,
0.047567009925842285,
-0.013039406388998032,
0.05144967511296272,
0.006717020645737648,
0.015054832212626934,
0.024907946586608887,
-0.06221248582005501,
-0.04119635000824928,
-0.006535008549690247,
-0.06439969688653946,
0.06279655545949936,
0.017786698415875435,
0.041... | [
0.14978519082069397,
0.11671486496925354,
-0.009219560772180557,
0.35100704431533813,
0.33597704768180847,
-0.26884061098098755,
-0.0008638613508082926,
-0.1167387142777443,
-0.21330320835113525,
-0.37200456857681274,
0.35738199949264526,
0.308846652507782,
0.3245062232017517,
-0.504942536... |
I have binary map for disease suitable area and I want to calculate the percentage using pixel count in each polygon how can I do that? I use zonal histogram and do have frequency for each polygon is there is way to use this values to calculate the percentage in each polygon? | [
0.003692088183015585,
0.01510569266974926,
-0.005250644404441118,
0.020152157172560692,
-0.03493885695934296,
-0.023371899500489235,
0.012341777794063091,
-0.009891963563859463,
-0.028604378923773766,
-0.002010439755395055,
0.0034299411345273256,
0.008766097016632557,
-0.002026369096711278,
... | [
0.3714902997016907,
0.24774813652038574,
-0.25445228815078735,
0.22594080865383148,
-0.24032112956047058,
0.26309484243392944,
0.30187270045280457,
-0.03958543390035629,
-0.07486628741025925,
-0.9568845629692078,
0.5354874134063721,
0.1894945353269577,
-0.23444895446300507,
0.4381199479103... |
When using two parts in my document and resetting the counter for the second one, I get the following warning: > destination with the same identifier (name{section.1}) has been already > used, duplicate ignored A minimal example: \documentclass{scrartcl} \usepackage{hyperref} \begin{document} \tableofcontents \newpage \part{Part 1} \section{Section 1-1} \newpage \setcounter{section}{0} \part{Part 2} \section{Section 2-1} \end{document} How can I get rid of this warning? | [
0.00951449479907751,
0.014594829641282558,
-0.0013232397614046931,
0.04911632463335991,
0.03729046881198883,
0.012101918458938599,
0.0077626691199839115,
0.017918888479471207,
-0.01369217038154602,
-0.013194761238992214,
-0.0025747751351445913,
0.004326310474425554,
-0.005874754395335913,
... | [
0.11939682811498642,
0.07471839338541031,
0.42402634024620056,
0.06129630655050278,
0.12687766551971436,
-0.08062335848808289,
0.4322539269924164,
0.14411763846874237,
-0.5734384059906006,
-0.5968210101127625,
0.004305298440158367,
-0.059434130787849426,
-0.47803252935409546,
0.43522861599... |
I used the hyperref package and pdflatex to make a nice PDF form. The form works fine in that I can fill out the fields and save the contents. However, I want to make a fancy field that performs a calculation on the values of some other fields. I can do this in Acrobat without problems, by changing the properties of one of the pdflatex generated form fields to calculate a sum. So far so good! However, when I try to "distribute" my updated form (with the calculation) Acrobat claims there are no fields in the document. However, Acrobat itself lists a bunch of fields. (See the screenshots below.) What do I have to do to get Acrobat to save the document with the form calculation correctly? If I use File > Save, the calculation does not get saved.   | [
0.00304712587967515,
0.01484538335353136,
-0.010509800165891647,
0.017827818170189857,
-0.003401420312002301,
-0.005733362399041653,
0.009392429143190384,
-0.01131795160472393,
-0.017734529450535774,
-0.023608330637216568,
0.000445998040959239,
0.00783994048833847,
-0.013634840026497841,
0... | [
-0.045597147196531296,
0.20745588839054108,
0.8164398670196533,
0.17208635807037354,
-0.2840297222137451,
-0.2562265396118164,
0.08086513727903366,
-0.1586482673883438,
-0.09951043128967285,
-0.30827367305755615,
-0.08446459472179413,
0.5454779863357544,
-0.3543711006641388,
0.057468455284... |
I am trying to understand how the VNC mechanism works. The RFB protocol 3.8 specifications say : > The update protocol is demand-driven by the client. That is, an update is > only sent from the server to the client in response to an explicit request > from the client. This gives the protocol an adaptive quality. The slower the > client and the network are, the lower the rate of updates becomes. With > typical applications, changes to the same area of the framebuffer tend to > happen soon after one another. With a slow client and/or network, transient > states of the framebuffer can be ignored, resulting in less network traffic > and less drawing for the client. It seems that this implies that the server will only ever send a `FramebufferUpdate` if the client sends a `FramebufferUpdateRequest`. Then the client must be sending these packets at a regular interval. However, when I analyzed through wireshark, I observed that this was not the case. When there was no screen or pointer activity, I did not see any packets from the client to the server. When I created some screen activity on the screen without involving the client, (I ran `xclock` by setting display to that value) the first message was from the server to the client and not a request from the client. So my question is: **Is it indeed true that the server will only send updates if the client requests them and not every time there is screen activity?** What is the frequency of updates in either case? | [
-0.012530332431197166,
0.015645865350961685,
0.000054927775636315346,
0.006702304817736149,
-0.023222390562295914,
-0.04042448103427887,
0.007923533208668232,
0.008305399678647518,
-0.011012363247573376,
-0.010494416579604149,
-0.022685309872031212,
0.021431751549243927,
-0.00539159448817372... | [
0.36447423696517944,
-0.12334469705820084,
0.7236419320106506,
-0.06965065002441406,
-0.1680186241865158,
-0.1962898224592209,
0.08028507232666016,
-0.2643837034702301,
-0.16550013422966003,
-0.5431028008460999,
-0.13489270210266113,
0.8367846012115479,
-0.24574699997901917,
0.063234493136... |
I am a complete newcomer to LaTeX. Here is what I am attempting to do - combine a thesis template (made with the `report` class) with a template based on the `apa6` class. In other words, I need the following front matter from the thesis template (title page, abstract, dedication, acknowledgements, table of contents [including roman page numbering beginning on page iii]) and the following items from the `apa6` class template (APA style headings, reference styles, running heads, etc. (with arabic numbering beginning with the first page of chapter 1) for everything else. The thesis code is located here. The `apa6` template is found in the `apa6` documentation, in particular the `shortsample.tex` file that comes with `apa6.cls`. Any pointers here? What I should try? Where I should look next? | [
-0.008499068208038807,
0.018076032400131226,
0.0006007659249007702,
0.014357587322592735,
0.014166762121021748,
0.028477702289819717,
0.008748584426939487,
0.0019871348049491644,
-0.01421977486461401,
-0.009275639429688454,
-0.006561009678989649,
0.0030383251141756773,
-0.022156940773129463,... | [
0.2574175000190735,
0.523821234703064,
0.2497195452451706,
-0.204762265086174,
0.03228581324219704,
-0.07878424972295761,
0.13359522819519043,
-0.14991718530654907,
0.48388412594795227,
-0.7124046683311462,
-0.22029490768909454,
0.3025754988193512,
-0.008077822625637054,
0.2438855171203613... |
If I remember correctly, courthouses are always destroyed when a city is conquered. On the other hand, when a civ loses one of their cities and then conquers it again, it does not have to build a courthouse there. My question is, does a city become "mine" after building a courthouse, e.g. if I lose and then re-conquer it, will I have to build a courthouse again? Or do all cities remember only the original civ that built them, and only that civ doesn't need a courthouse? | [
-0.008657602593302727,
0.023696206510066986,
-0.0022301673889160156,
0.010738158598542213,
-0.018541036173701286,
-0.007173658348619938,
0.008638396859169006,
-0.0064150444231927395,
-0.0161859393119812,
-0.00921992864459753,
0.0012610048288479447,
0.02510184422135353,
-0.009342876262962818,... | [
0.3021736145019531,
0.6025519371032715,
-0.025692153722047806,
0.3276684582233429,
0.1778077781200409,
-0.0245120357722044,
0.09923412650823593,
0.006527766585350037,
-0.6441770792007446,
-0.4789188802242279,
0.025356555357575417,
0.12332949787378311,
-0.11315099149942398,
0.77043920755386... |
I want to change an image depth of bitmap for testing purposes. Right now I am trying to get a 2 bit palette image, and a 4444 Hicolor image. I have a true color bitmap. I used the below command line convert -depth 2 /media/bitmap/rule.bmp lut2bpp.bmp then when I used identify I got this Image: lut2bpp.bmp Format: BMP (Microsoft Windows bitmap image) Class: PseudoClass Geometry: 720x480 Type: Palette Endianess: Undefined Colorspace: RGB Channel depth: Red: 8-bits Green: 8-bits Blue: 8-bits It changed it to a palette which is great, how do I get to change channel depth? How about changing that true color 24 bit image to hi color 4444 image? | [
-0.01921330764889717,
0.004387203138321638,
-0.013519780710339546,
0.010905737057328224,
-0.003579190466552973,
-0.013296535238623619,
0.007121386006474495,
0.009504237212240696,
-0.01962488517165184,
-0.01829647086560726,
-0.007800256833434105,
0.0008231175015680492,
-0.011833783239126205,
... | [
0.47579631209373474,
-0.06497477740049362,
0.24809767305850983,
-0.10909998416900635,
-0.21377688646316528,
0.3666969835758209,
0.05393413081765175,
-0.24245354533195496,
-0.0957946851849556,
-0.6943382620811462,
0.05055687204003334,
0.4485017955303192,
-0.38559839129447937,
0.247924447059... |
Once, I have received feedback that using "see you when I see you" is not very polite. Do you have the same opinion? What other expression should I use in case I have no clue when I will see the other party? | [
-0.01637166552245617,
0.01966121979057789,
-0.009259778074920177,
0.01165007334202528,
0.021334629505872726,
0.024096040055155754,
0.012938079424202442,
-0.03389710932970047,
-0.02823466993868351,
-0.0324568971991539,
-0.010880701243877411,
0.023971471935510635,
0.019286973401904106,
0.009... | [
0.40589603781700134,
-0.17535383999347687,
0.17435060441493988,
-0.11283273994922638,
-0.25143590569496155,
-0.21645937860012054,
0.4677079916000366,
0.6800864338874817,
-0.2343640774488449,
-0.760074257850647,
0.18046899139881134,
0.2781420946121216,
0.099949911236763,
0.03568394482135773... |
I have a requirement where there are 10 Rules to be applied on data in excel. If Rule 1 and Rule 2 fails rest of the rules are not checked. But if Rule 1 and Rule 2 passes the rest of all the Rules should be verified and if any errors found- they should be logged. Is there any design pattern which I can use to keep this Rule Engine flexible for adding these 10 rules and Closed for any additional chains in the Current Rule. I was thinking of something like a Decorator Pattern. Will this help me achieve that? | [
0.014832050539553165,
0.026626242324709892,
-0.0028452719561755657,
0.023473622277379036,
0.013276292942464352,
0.03264852240681648,
0.00901766587048769,
0.003334307111799717,
-0.014739801175892353,
0.0031897001899778843,
-0.016622861847281456,
0.011147505603730679,
-0.015147238038480282,
... | [
0.6446697115898132,
0.29870763421058655,
0.4210692346096039,
0.1455966979265213,
0.07013747841119766,
-0.3294677734375,
0.047160107642412186,
-0.36080804467201233,
-0.3183353543281555,
-0.33141374588012695,
0.2542465925216675,
0.44921940565109253,
-0.2524939477443695,
-0.018793080002069473... |
I'm on Texlive 2011 on Linux, and I try to run this example: ftp://www.ctan.org/ctan/macros/latex2e/exptl/biblatex/doc/examples/21-indexing- advanced.tex I build, as recommended, with: latex 21-indexing-advanced.tex bibtex 21-indexing-advanced latex 21-indexing-advanced.tex On the second pass, I get this: ! Missing number, treated as zero. <to be read again> \l@german l.137 \raggedright ? Apparently, something is missing there, but I cannot tell what. Any help with this? * * * EDIT: The link may be old, but that is the same example which is included in current `biblatex` in TexLive 2011: $ MYLOCFILE=$(find /path/to/texlive/ -name 21-indexing-advanced.tex) $ echo $MYLOCFILE /path/to/texlive/2011/texmf-dist/doc/latex/biblatex/examples/21-indexing-advanced.tex $ wget ftp://www.ctan.org/ctan/macros/latex2e/exptl/biblatex/doc/examples/21-indexing-advanced.tex ... $ diff $MYLOCFILE ./21-indexing-advanced.tex $ # nothing is output => they are identical ... Logfile - first pass: $ pdflatex 21-indexing-advanced.tex This is pdfTeX, Version 3.1415926-2.3-1.40.12 (TeX Live 2011) restricted \write18 enabled. entering extended mode (./21-indexing-advanced.tex LaTeX2e <2011/06/27> Babel <v3.8m> and hyphenation patterns for english, dumylang, nohyphenation, lo aded. (/path/to/texlive/2011/texmf-dist/tex/latex/base/article.cls Document Class: article 2007/10/19 v1.4h Standard LaTeX document class (/path/to/texlive/2011/texmf-dist/tex/latex/base/size10.clo)) (/path/to/texlive/2011/texmf-dist/tex/latex/base/fontenc.sty (/path/to/texlive/2011/texmf-dist/tex/latex/base/t1enc.def)) (/path/to/texlive/2011/texmf-dist/tex/generic/babel/babel.sty (/path/to/texlive/2011/texmf-dist/tex/generic/babel/english.ldf (/path/to/texlive/2011/texmf-dist/tex/generic/babel/babel.def))) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.sty (/path/to/texlive/2011/texmf-dist/tex/latex/etoolbox/etoolbox.sty (/path/to/texlive/2011/texmf-dist/tex/latex/etex-pkg/etex.sty)) (/path/to/texlive/2011/texmf-dist/tex/latex/graphics/keyval.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.def) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.cfg)) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.sty (/path/to/texlive/2011/texmf-dist/tex/latex/logreq/logreq.sty (/path/to/texlive/2011/texmf-dist/tex/latex/logreq/logreq.def)) (/path/to/texlive/2011/texmf-dist/tex/latex/base/ifthen.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/url/url.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/blx-compat.def) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.def) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/bbx/authortitle.b bx (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/bbx/standard.bbx) ) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/cbx/authortitle.c bx) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.cfg) ) (/path/to/texlive/2011/texmf-dist/tex/latex/index/index.sty LaTeX Warning: Command \markboth has changed. Check if current package is valid. LaTeX Warning: Command \markright has changed. Check if current package is valid. ) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/lbx/american.lbx (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/lbx/english.lbx)) No file 21-indexing-advanced.aux. No file 21-indexing-advanced.bbl. (/path/to/texlive/2011/texmf-dist/tex/latex/base/t1cmtt.fd) LaTeX Warning: Citation 'piccato' on page 1 undefined on input line 127. LaTeX Warning: Citation 'gaonkar' on page 1 undefined on input line 127. LaTeX Warning: Citation 'malinowski' on page 1 undefined on input line 127. LaTeX Warning: Citation 'coleridge' on page 1 undefined on input line 127. LaTeX Warning: Citation 'gerhardt' on page 1 undefined on input line 127. LaTeX Warning: Citation 'cicero' on page 1 undefined on input line 127. [1{/path/to/texlive/2011/texmf-var/fonts/map/pdftex/updmap/pdftex.map} ] LaTeX Warning: Empty bibliography on input line 135. No file 21-indexing-advanced.ind. No file 21-indexing-advanced.nnd. No file 21-indexing-advanced.tnd. (./21-indexing-advanced.aux) LaTeX Warning: There were undefined references. Package biblatex Warning: Please (re)run BibTeX on the file(s): (biblatex) 21-indexing-advanced (biblatex) and rerun LaTeX afterwards. ) <~/.texlive2011/texmf-var/fonts/pk/ljfour/jknappen/ec/ecrm 1000.600pk> <~/.texlive2011/texmf-var/fonts/pk/ljfour/jknappe n/ec/ecbx1000.600pk> <~/.texlive2011/texmf-var/fonts/pk/ljfou r/jknappen/ec/ectt1440.600pk> <~/.texlive2011/texmf-var/fonts /pk/ljfour/jknappen/ec/ecbx1440.600pk> Output written on 21-indexing-advanced.pdf (1 page, 12739 bytes). Transcript written on 21-indexing-advanced.log. .. then `bibtex 21-indexing-advanced` -- then second pass: $ pdflatex 21-indexing-advanced.tex This is pdfTeX, Version 3.1415926-2.3-1.40.12 (TeX Live 2011) restricted \write18 enabled. entering extended mode (./21-indexing-advanced.tex LaTeX2e <2011/06/27> Babel <v3.8m> and hyphenation patterns for english, dumylang, nohyphenation, lo aded. (/path/to/texlive/2011/texmf-dist/tex/latex/base/article.cls Document Class: article 2007/10/19 v1.4h Standard LaTeX document class (/path/to/texlive/2011/texmf-dist/tex/latex/base/size10.clo)) (/path/to/texlive/2011/texmf-dist/tex/latex/base/fontenc.sty (/path/to/texlive/2011/texmf-dist/tex/latex/base/t1enc.def)) (/path/to/texlive/2011/texmf-dist/tex/generic/babel/babel.sty (/path/to/texlive/2011/texmf-dist/tex/generic/babel/english.ldf (/path/to/texlive/2011/texmf-dist/tex/generic/babel/babel.def))) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.sty (/path/to/texlive/2011/texmf-dist/tex/latex/etoolbox/etoolbox.sty (/path/to/texlive/2011/texmf-dist/tex/latex/etex-pkg/etex.sty)) (/path/to/texlive/2011/texmf-dist/tex/latex/graphics/keyval.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.def) (/path/to/texlive/2011/texmf-dist/tex/latex/csquotes/csquotes.cfg)) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.sty (/path/to/texlive/2011/texmf-dist/tex/latex/logreq/logreq.sty (/path/to/texlive/2011/texmf-dist/tex/latex/logreq/logreq.def)) (/path/to/texlive/2011/texmf-dist/tex/latex/base/ifthen.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/url/url.sty) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/blx-compat.def) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.def) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/bbx/authortitle.b bx (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/bbx/standard.bbx) ) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/cbx/authortitle.c bx) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/biblatex.cfg) ) (/path/to/texlive/2011/texmf-dist/tex/latex/index/index.sty LaTeX Warning: Command \markboth has changed. Check if current package is valid. LaTeX Warning: Command \markright has changed. Check if current package is valid. ) (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/lbx/american.lbx (/path/to/texlive/2011/texmf-dist/tex/latex/biblatex/lbx/english.lbx)) (./21-indexing-advanced.aux) (./21-indexing-advanced.bbl) Package biblatex Warning: backend=biber required by .bib file. (biblatex) Some features may not work properly. (/path/to/texlive/2011/texmf-dist/tex/latex/base/t1cmtt.fd) [1{/path/to/texlive/2011/texmf-var/fonts/map/pdftex/updmap/pdftex.map}] Overfull \hbox (18.58575pt too wide) in paragraph at lines 137--137 \T1/cmr/m/n/10 (1961), pp. 579^^U588[][]; Wein-berg, Steven. ^^P A Model of Le p-tons.^^Q In: \T1/cmr/m/it/10 Phys. Rev. Lett. Overfull \hbox (15.22781pt too wide) in paragraph at lines 137--137 \T1/cmr/m/n/10 rated Pd-N-heterocyclic car-benes in Mizoroki-Heck re-ac-tions.^ ^Q In: \T1/cmr/m/it/10 J. Organomet. ! Missing number, treated as zero. <to be read again> \l@german l.137 \raggedright ? | [
0.02077440172433853,
0.006826424039900303,
-0.020135389640927315,
0.008791593834757805,
-0.004358571022748947,
0.015505682677030563,
0.0070607541128993034,
0.004365412052720785,
-0.017600614577531815,
-0.014327502809464931,
-0.011521685868501663,
0.00526883639395237,
-0.017011038959026337,
... | [
-0.20488373935222626,
0.19450543820858002,
0.6155895590782166,
-0.4489012062549591,
-0.10555046796798706,
0.009489800781011581,
0.48582372069358826,
-0.03594997152686119,
0.08918536454439163,
-0.647832453250885,
0.16818293929100037,
0.7795615792274475,
-0.21199484169483185,
0.0480336025357... |
 I have a python script that runs 5 Spatial ETL tools that I wish to run as a scheduled task. Each of the tools will run successfully from ArcCatalog and also from the ArcCatalog Command line. If I try to run the tools from the DOS window, as would be done via python script for a scheduled task, arcpy cannot find the tools, although it does load the toolbox. In the attached screenshot: 1. I open a cmd.exe window 2. start python 3. import arcpy 4. Import my toolbox, which is verified as loaded on the next line 5. Use arcpy.ListTools to see what tools are loaded. the result is "[]", meaning no tools are recognized from this tool box 6. I use the same ListTools to show the Spatial Analyst tools that were loaded by import arcpy 7. I attempt to execute one of the tools from my toolbox and forget a " (my mistake!) 8. I attempt to execute one of the tools from my toolbox and receive the message "AttributeError: 'module' object has no attribute 'CopyFeatureClass_newgpetl'". What is preventing my tools from being recognized from the DOS CMD.EXE window while ArcCatalog and ArcCatalog Command Line can run them? | [
0.00042527157347649336,
0.009530832059681416,
-0.010317481122910976,
0.010716390796005726,
-0.016731752082705498,
-0.0004852995043620467,
0.006735607981681824,
0.01078072004020214,
-0.012028740718960762,
0.004729519598186016,
-0.022864803671836853,
0.003927823156118393,
-0.01089064497500658,... | [
0.09118444472551346,
-0.06072697415947914,
0.6213124990463257,
-0.03732188418507576,
-0.033712487667798996,
0.10544565320014954,
0.5449488759040833,
-0.4346294403076172,
-0.28622448444366455,
-0.6697785258293152,
-0.19832324981689453,
0.7247188687324524,
-0.1453695148229599,
-0.13555002212... |
I need to remove lines from a text file based on pattern but I need to keep the first n lines of that pattern. Input % 1 % 2 % 3 % 4 % 5 text1 text2 text3 output %1 %2 text1 text2 text3 I used `sed /^%/d` file but it deletes all the lines starting with %, `sed 3,/^%/d` doesn't help either. I need to keep first n lines of the pattern and delete the rest | [
0.00595973851159215,
0.008725099265575409,
-0.0166456401348114,
0.007471493445336819,
0.0060427226126194,
0.032950930297374725,
0.006775464862585068,
0.010705029591917992,
-0.01967206783592701,
0.019757721573114395,
-0.00913655012845993,
0.003071028273552656,
0.005076782777905464,
0.017609... | [
0.06908109784126282,
0.1494947224855423,
0.4659363031387329,
-0.42381522059440613,
-0.13786129653453827,
0.5618314146995544,
0.7395531535148621,
-0.3778931200504303,
-0.1306305080652237,
-0.7077368497848511,
-0.2059200555086136,
0.14926840364933014,
-0.6391003727912903,
0.09644576162099838... |
This is again one of the behaviors that are quite difficult to catch in a reconstructible (and simple) example, so one has to resort to animations. Basically, the issue is this: the tilde (`~`) is supposed to be a non-breaking space character, right? How come, then, when I use the `Fig.~\ref{fig:myfigure}` in the MWE below, it sometimes ends up breaking the line _exactly_ at that `~` character's location, as can be seen on the .gif below:  For instance, frame 41 there ends up like this (click for full res):  Isn't that what the tilde is supposed to prevent? How can I find out why this is happening - and how can I get a non-breaking space in my `Fig.~\ref{...}` construct? Below is the MWE, `test-tilde.tex`; it uses the package `repeat-build.sty` that I posted here (tested only on Linux currently) - and it needs to be built twice, with `-shell-escape`: pdflatex -shell-escape test-tilde.tex # first time, refs wrong here pdflatex -shell-escape test-tilde.tex # second time, refs ok (can enable the `convert` post action if desired) ... so that all PDFs, used in the animaton above, are generated. EDIT: I modified the code, so it will build without the `repeat-build.sty`; and it will reconstruct the situation on the still screenshot above (where the variable `\tester`=87); then if you want to use `repeat-build`, simply (un)comment the corresponding lines. The code is: \documentclass[% 12pt, journal, onecolumn, twoside, draftcls, letterpaper, ]{IEEEtran} \usepackage[demo]{graphicx} % \usepackage{xcolor} % \pagecolor \pagecolor{yellow!15} \usepackage[colorlinks=true,linkcolor=blue]{hyperref} \usepackage[nopar]{lipsum} \usepackage{xstring} % http://tex.stackexchange.com/a/26808/2595 \makeatletter \def\unpacklipsum#1#2#3{% \count@=#1\relax \advance\count@\m@ne \def#3{}% \loop\ifnum\count@<#2\relax \advance\count@\@ne \edef#3{#3\csname lipsum@\romannumeral\count@\endcsname}% \repeat} % http://tex.stackexchange.com/a/168754/2595 \def\loremnchars[#1]#2{% \unpacklipsum{#1}{#1}{\myunpacked}% \StrMid{\myunpacked}{1}{#2}% same as \StrLeft{\myunpacked}{#2} } \def\eloremnchars#1#2#3{% \unpacklipsum{#1}{#1}{\myunpacked}% % \StrMid is not an expandable macro (can't go in \def) - but % can "return" by saving in #3 \StrMid{\myunpacked}{1}{#2}[#3] } \def\getLoremParaNumChars#1#2{% \unpacklipsum{#1}{#1}{\myunpacked}% \StrLen{\myunpacked}[#2] \typeout{Lorem paragraph #1 has #2 characters}. } \makeatother \usepackage[format=plain,font=small,skip=12pt]{caption} % \usepackage{listings} % \usepackage{units} % \usepackage{floatrow} \floatsetup[figure]{capposition=bottom} \usepackage{etoolbox} \usepackage{pgfplots} \usepackage{pgfplotstable} \usepackage{tikz} \usetikzlibrary{shapes,positioning} \usetikzlibrary{fit} \usetikzlibrary{calc} \usetikzlibrary{decorations.markings} \usepackage{txfonts} % Times font in math; tlmgr install txfonts \renewcommand{\ttdefault}{pcr} \usepackage[UTF8]{ctex} % needs simsun.ttc in same directory \usepackage{setspace} \usepackage{siunitx} \usepackage[american]{babel} \usepackage{csquotes} \makeatletter \let\l@AMERICAN\l@american % http://tex.stackexchange.com/q/88861/ \makeatother \usepackage[style=ieee,isbn=true,doi=false,url=true,backend=biber]{biblatex} % here, set the \tester at a number, % so we don't hit an error if we % compile without repeat-build (-shell-escape) % http://tex.stackexchange.com/a/185976/2595 %%%% uncomment the below part if you want to use {repeat-build}: % \ifx\tester\undefined{% % \global\protected\def\tester{100} % % }\fi % % do not rm the .aux here; need them for fig label; % % since the process needs to run twice for correct % % label - comment the convert post action. % %% manual convert command: % %% convert -verbose -density 150 test-tilde-*.pdf[0] -crop 1024x580+137+773 -resize 512x +repage -delay 15 -loop 0 test-tilde.gif % %% gifsicle -O2 test-tilde.gif -o test-tilde-O2.gif # helps reduce size % \def\pacode{% % pwd; % % ls \jobname-*.log \jobname-*.aux; % % rm -vrf \jobname-*.log ;% \jobname-*.aux; % % % convert -verbose -density 75 \jobname-*.pdf[0] % % % -crop 512x290+68+386 % % % +repage % % % -delay 15 -loop 0 % -delay 5 % % \jobname.gif % % % ; % % % eog \jobname.gif ; % run viewer % } % \usepackage[% % vartoken=\tester,% % loop the variable token: \tester % varvals={15,16,...,29,41,42,...,52,58,59,...,65,77,78,...,89},% set \tester to change; (was from 0 to 100 {0,...,100}) % postactioncode={\pacode},% % ]{repeat-build} % \unprotect\tester % \makeatletter % \def\getrjname{\rpbuild@jobname} % \makeatother %%%% else this should be enough for a usual run: \def\tester{83} % comment this if using repeat-build! \begin{document} \title{Test title} \author{test} \maketitle \begin{abstract} The abstract text goes here. \end{abstract} \section{Test section} Test text, before figure: \setcounter{figure}{22} % if 10, then {figure} becomes 11 \begin{figure}[h!t] \singlespacing \includegraphics[width=\textwidth,height=100pt]{demo.whatever} %\caption[shortdesc]{The figure is being shown here. \protect\loremnchars[3]{255}} \captionof{figure}[shortdesc]{The figure is being shown here. \protect\loremnchars[3]{255}} \label{fig:myfigure} \end{figure} Current: doc: \jobname; \verb!\tester!: \tester ... \loremnchars[1]{150} \eloremnchars{2}{\tester}{\tretA} % \pgfmathtruncatemacro{\tB}{\tester/2} % \eloremnchars{4}{\tB}{\tretB} % \eloremnchars{5}{\tester}{\tretC} % \tretA\space on Fig.~\ref{fig:myfigure}. \tretB\space on Fig.~\ref{fig:myfigure}, \tretC\space on Fig.~\ref{fig:myfigure}. \loremnchars[3]{712}.. And ... \lipsum[7] %\getLoremParaNumChars{3}{\tret} %\typeout{\tret} %\eloremnchars{3}{712}{\tret} \typeout{\tret} \end{document} | [
-0.005045211408287287,
0.01887817680835724,
-0.007808437570929527,
0.009244837798178196,
-0.005911229178309441,
-0.015524988994002342,
0.006413755007088184,
0.00954801868647337,
-0.011437475681304932,
0.006581084802746773,
-0.01525137573480606,
0.005773107521235943,
-0.0023397377226501703,
... | [
0.010917817242443562,
-0.1260884404182434,
0.6236175298690796,
0.26030877232551575,
0.11762094497680664,
-0.21333558857440948,
-0.024902530014514923,
0.24856996536254883,
-0.34969231486320496,
-0.30509597063064575,
0.00517995236441493,
0.44527730345726013,
-0.007178069092333317,
0.33700302... |
I have a multipolygon layer in GeoServer. The layer represents states within a country. I want to enable a user to search for a specific state, and then only the selected state to be displayed on the base map. The problem is two-fold: First, how to retrieve the names of polygons in the layer to enable a user to select, and second, how to display only the selected polygon on the base map. I'm using OpenLayers for presentation. In addition, I'm using GeoServer and PostGIS. Any pointers on how to proceed? EDIT: From some research, I'm guessing I'll have to use GeoServer filters (CQL and ECQL) as outlined here. However, I'm still a newbie at this, and any advice will be appreciated. | [
-0.004828492179512978,
0.006726379971951246,
-0.0076527525670826435,
0.007677185349166393,
0.001223776489496231,
0.00543287442997098,
0.007948972284793854,
-0.0035125708673149347,
-0.01625414937734604,
0.004877140745520592,
-0.00748077966272831,
0.0187434870749712,
-0.0065887002274394035,
... | [
0.07923292368650436,
0.06879126280546188,
0.4198702275753021,
0.2437828928232193,
-0.13795070350170135,
0.04560260474681854,
0.07322277128696442,
0.178456112742424,
-0.279366672039032,
-1.0123794078826904,
-0.022707881405949593,
0.1252136528491974,
-0.38659361004829407,
0.22653473913669586... |
I understand probability distribution but I am having a hard time getting a grasp on probability density function, specifically difference between dexp (density of exponential distribution) and pexp (probability distribution of exponential distribution) | [
-0.021005980670452118,
0.009492021054029465,
-0.021153690293431282,
0.005697188433259726,
-0.04363543540239334,
-0.024235587567090988,
0.01948779821395874,
0.05086234584450722,
-0.033915579319000244,
-0.06395480036735535,
-0.02500304952263832,
0.003060600021854043,
-0.03156234696507454,
-0... | [
0.8123816847801208,
0.0799107626080513,
0.22023087739944458,
0.18202444911003113,
-0.16662152111530304,
-0.060747262090444565,
-0.26521968841552734,
-0.08861792087554932,
-0.17143328487873077,
-0.5116106867790222,
0.18188150227069855,
0.36976152658462524,
-0.08543877303600311,
0.1010549142... |
I do understand the advantages of standardizing regression predictors to get standardized coefficients, in order to interpret the coefficients better. However, as I was reading multiple pages online, I figured that some people do standardize both predictor AND outcome to get standardized coefficients. It doesn't make sense to me. I am OK with standardizing predictors, but when the outcome is standardized too, we are predicting another value (not the actual Y). Is that right? I also do accept the regression results when the predictors are STANDARDIZED and the outcome is CENTRED. However, not the regression results when the predictors are STANDARDIZED and the outcome is STANDARDIZED too. Is that correct? | [
0.015574995428323746,
0.03128292039036751,
-0.018292101100087166,
0.017701972275972366,
-0.007578338496387005,
0.01740552857518196,
0.009571870788931847,
-0.011549015529453754,
-0.013409601524472237,
-0.011046438477933407,
-0.005281416699290276,
0.017679624259471893,
-0.014618422836065292,
... | [
0.5982619524002075,
0.1984250843524933,
0.12405484914779663,
0.22887906432151794,
-0.528235912322998,
-0.18358291685581207,
0.0030448907054960728,
0.15216906368732452,
-0.08464495092630386,
-0.3750728964805603,
0.4459676146507263,
0.7407341003417969,
0.0840814933180809,
0.252641499042511,
... |
I want to display single random post in it's entirety in the homepage.I've tried many methods, but the main issue was that it won't display the content, only heading, or display content, but only an excerpt. Now I have this code: <?php query_posts(array('orderby' => 'rand', 'showposts' => 1)); if (have_posts()) : while (have_posts()) : the_post(); ?> <h2><a href="<?php the_permalink() ?>"><?php the_title(); ?></a></h2> <?php the_content(); ?> <?php endwhile; endif; ?> One problem is that it doesn't display full post, only excerpt.Another problem is that it displays only heading and content and I want all the things I get in single.php.The biggest problem is that my custom code: <?php if(in_category('afghanistan')) { $recent = new WP_Query("pagename=asia/afganistanas"); while($recent->have_posts()) : $recent->the_post();?> <h3><?php the_title(); ?></h3> <?php the_content(); ?> <?php endwhile; } else { echo "Klaida!"; } ?> which displays a page inside of post depending on the category of post doesn't word in index, but does work in single.php **Is there any method to use a copy of single.php in index and give it a random post to load every time the page is refreshed ?** | [
0.006611957214772701,
0.011135756969451904,
-0.007778162602335215,
0.014627797529101372,
-0.012738484889268875,
-0.012219897471368313,
0.004539555869996548,
0.027836941182613373,
-0.012405984103679657,
0.0005253460258245468,
-0.013891752809286118,
0.005993453785777092,
0.005123436450958252,
... | [
0.4055495262145996,
0.11946333199739456,
0.6292178630828857,
-0.026896538212895393,
-0.2767525017261505,
0.09127885103225708,
0.1797218769788742,
-0.36261695623397827,
-0.15981058776378632,
-0.837303638458252,
0.005626339931041002,
-0.01165397185832262,
-0.369639128446579,
0.38813856244087... |
A "faster than light" neutrino discovery was actually the result of a loose cable. A fiber-optic cable in a GPS receiver at the European Center for Particle Physics ( CERN) near Geneva Can someone show me the light here, What's Going On ? how long was fiber-optic cable? | [
-0.01951688528060913,
0.00893738865852356,
0.006023031659424305,
0.0051855482161045074,
-0.01765117235481739,
-0.030177095904946327,
0.0122172050178051,
0.002238993067294359,
-0.025874294340610504,
-0.005782075226306915,
0.0026297878939658403,
0.015485395677387714,
0.014491229318082333,
-0... | [
0.14148664474487305,
-0.36030566692352295,
0.4579741358757019,
0.58417809009552,
0.18869946897029877,
-0.3209121525287628,
-0.03314558044075966,
0.30213186144828796,
-0.5414131879806519,
-0.389901340007782,
0.07084286212921143,
0.03928067162632942,
-0.38201043009757996,
0.19589382410049438... |
I keep seeing these posts about salvage droids and the game keeps talking about it but I don't see them anywhere in the game can someone lmk how to get them and use them? | [
0.021474795415997505,
-0.004706455394625664,
-0.005579693708568811,
-0.0010454336879774928,
0.03152548521757126,
-0.0321480855345726,
0.007253195624798536,
0.023804359138011932,
-0.024004686623811722,
-0.032846540212631226,
0.007804979104548693,
0.02455161325633526,
-0.02279716730117798,
0... | [
0.7731136083602905,
0.30173712968826294,
-0.1648973524570465,
0.4535825550556183,
0.06696803867816925,
0.20949676632881165,
-0.1321180760860443,
0.41820335388183594,
-0.39344698190689087,
-0.3414374589920044,
0.16834308207035065,
0.2809939682483673,
0.13007141649723053,
0.12451480329036713... |
A problem that's been annoying me for years (http://newsgroups.derkeiler.com/Archive/Comp/comp.text.tex/2008-04/msg01387.html) has been how to create a column'ed bibliography. E.g., take a look at the bibliography in: http://www.cgd.ucar.edu/staff/trenbert/trenberth.papers/UppalaERA05.pdf  That's what I want! Ideally, I'd like four columns (Author, Year, Journal/Source, Title+Pages) rather than the three shown in the example (although that's a start). | [
-0.0013043154031038284,
0.0008036370272748172,
-0.0081412922590971,
0.010496380738914013,
-0.01576625555753708,
0.010659407824277878,
0.0030973399989306927,
0.003397177206352353,
-0.01602770760655403,
-0.0007610493339598179,
-0.00003367476165294647,
-0.0003419695422053337,
-0.009100725874304... | [
-0.10396376252174377,
0.190801739692688,
0.6177166700363159,
0.18374574184417725,
-0.03552777320146561,
-0.24286483228206635,
-0.1031658947467804,
-0.008288402110338211,
-0.6496281623840332,
-0.25931188464164734,
-0.11376997828483582,
0.5617011189460754,
0.21865196526050568,
0.130583986639... |
I just started playing Smite, so I may not be using the correct terminolgy. During the games, I used the `F` and `G` "spells" quite often. The `F` spell restored mana, and the `G` spell gives a speed boost. I compare these to the summoner spells in League of Legends where you have a handful to choose from before the game. Does Smite have spells to chose from as well, or is it currently just the mana restore and speed boost spells? I looked on the smite wiki, and all I found was the Controls page, where it labels the `F` and `G` keys as Store Active 1 and Store Active 2 respectively. | [
-0.005856376141309738,
0.007306614890694618,
-0.005304817575961351,
-0.006239690352231264,
0.019836099818348885,
-0.03286244720220566,
0.0077826823107898235,
-0.01104121096432209,
-0.013982212170958519,
0.028639821335673332,
0.0002980446442961693,
0.006891121156513691,
-0.0007059797644615173... | [
0.10442638397216797,
-0.10234556347131729,
0.4573993980884552,
0.0353303924202919,
-0.19101658463478088,
-0.09764739125967026,
0.22461223602294922,
-0.11573528498411179,
-0.27067020535469055,
-0.30503126978874207,
0.1698838770389557,
0.374317467212677,
-0.11165154725313187,
0.0007480896892... |
How can 2 or more developers code a certain feature/module of a software? Let's assume that the module is big and feature rich. How would they prevent each other from overlapping their code? Say, we have the same method but is implemented in a different way. Do you think it might be better to have one focused at a specific feature only? Is a version control system like Git would help solve the problem? Is it correct that it allows developers to have these "branches" and then merge it later on? What's your take on this? | [
0.006792937405407429,
0.012321325950324535,
-0.010318552143871784,
0.0056512667797505856,
-0.0058167134411633015,
0.005267674569040537,
0.005970005877315998,
0.001259539625607431,
-0.02070389874279499,
0.0007263112347573042,
-0.012901605106890202,
0.032479941844940186,
0.014663082547485828,
... | [
0.5162485241889954,
-0.1187686175107956,
-0.17435501515865326,
0.40854495763778687,
0.020155243575572968,
-0.4015815556049347,
0.08813310414552689,
-0.20890770852565765,
-0.4753614068031311,
-0.3593902289867401,
0.05185374245047569,
0.47046738862991333,
-0.35698622465133667,
0.092577889561... |
I am reading the statistical mechanics by Pathria in Chap 12. I have a question about the Landau free energy. What is the physical reasoning for that the free energy could be a functional of the order parameter? Normally, the free energy is used when we choose the canonical ensemble and we we know that the free energy is the characteristic function of $N,V,T$ of the system. | [
-0.0018521987367421389,
0.025109076872467995,
-0.0062035187147557735,
0.008939007297158241,
0.005649250466376543,
-0.008368173614144325,
0.010664654895663261,
-0.022680656984448433,
-0.01589169166982174,
-0.042925868183374405,
-0.026160161942243576,
0.01701875403523445,
-0.02261831983923912,... | [
0.2618353068828583,
-0.13801856338977814,
0.10515543818473816,
0.09855800122022629,
-0.33432456851005554,
0.10247410088777542,
-0.18541443347930908,
-0.39209163188934326,
-0.20499669015407562,
-0.6910363435745239,
-0.16620847582817078,
0.18893291056156158,
-0.3843287229537964,
0.7835438251... |
I need this command (sh4-linux-objcopy) in RHEL 6. I could not find the package by `yum install` , neither there is a corresponding rpm. Here is how I want to use it: sh4-linux-objdump -h vmlinux | grep .empty_zero_page | awk '{print $4}' But `objdump` is installed and works fine. So is there any difference between `objdump` and `sh4-linux-objdump` and what is `sh4-linux-` | [
0.003067540470510721,
0.007987350225448608,
-0.007364341057837009,
0.020952969789505005,
-0.025284159928560257,
0.019715558737516403,
0.010672027245163918,
0.005654100328683853,
-0.01948745921254158,
-0.017314715310931206,
-0.012337025254964828,
0.01724528893828392,
-0.004428800195455551,
... | [
0.03093435987830162,
0.0738280713558197,
0.6849619746208191,
-0.15581122040748596,
-0.07725801318883896,
-0.16856050491333008,
0.35524871945381165,
-0.024945328012108803,
-0.02668784372508526,
-0.7214678525924683,
-0.04061967879533768,
0.8661143779754639,
-0.4052700698375702,
0.18977262079... |
I once read a paper, in which: * a fluid in a container was heated from below, * after reaching temperature $T_1$, a circular motion (convection) was clearly distinguishable, in form of **cylinder** , * after reaching temperature $T_2$, the circular motion splitted into **two** circular convections, side by side two cylinders, * after reaching temperatures $T_3, T_4, \ldots$, etc. **waves** appeared on cyliders' flat-sides (where the height is measured), * the frequency of the waves doubled at temperatures $T_4, T_5, \ldots$. Ratio of every temperature pair $T_1/T_2, T_2/T_3, \ldots$ was the same as in logistic map ratios of parameter $r$, and in every other map - the Feigenbaum number, the _period-doubling route to chaos_. I cannot find the paper again.. Does anyone remember such paper? Or maybe other? | [
-0.01233585923910141,
0.013941727578639984,
-0.009941344149410725,
0.0213019922375679,
-0.01137864775955677,
-0.0060026561841368675,
0.008652194403111935,
-0.0023379228077828884,
-0.01512549165636301,
-0.0003682142123579979,
0.0002773283049464226,
0.003925482742488384,
-0.013486041687428951,... | [
0.19306249916553497,
0.11458109319210052,
0.5863221883773804,
0.30463531613349915,
-0.155950129032135,
0.0013222479028627276,
-0.6493932008743286,
0.146075040102005,
-0.45925745368003845,
-0.07601764798164368,
-0.0805475115776062,
0.08107519894838333,
-0.039835378527641296,
0.7714723944664... |
Is it possible to have the footnotes not to repeat in the same page and take the old number? In other words, even if the footnote labels are same, can the footnotes have same number. here is my MWE. \documentclass[a4paper,10pt]{book} \usepackage[para]{footmisc} % to have footnotes in a line \begin{document} This is my test document which has these four footnotes. Footnote1\footnote{one}, footnote2\footnote{two}, footnote3\footnote{three} and last footnote\footnote{one} to show that fourth one repeats. \end{document} The question is in the above example, i would like the last footnote to have 1 and not 4. | [
0.01828375831246376,
0.027533091604709625,
-0.00863536074757576,
0.03035343810915947,
-0.006795736961066723,
0.001860191929154098,
0.007926331833004951,
0.01806599088013172,
-0.01693509891629219,
-0.039711251854896545,
-0.006813815329223871,
0.014749330468475819,
-0.004374148324131966,
-0.... | [
0.05575581640005112,
0.28645753860473633,
0.48883184790611267,
0.22865931689739227,
-0.12192194163799286,
0.13087545335292816,
0.3627559542655945,
-0.4293873906135559,
-0.400008887052536,
-0.6404333114624023,
0.3579941689968109,
0.3369305729866028,
-0.035868577659130096,
-0.014732562936842... |
\documentclass{article} \usepackage{tikz} \usetikzlibrary{arrows} \begin{document} \begin{tikzpicture} \draw[step=1cm,color=gray] (0,0) grid (3,3); \draw[draw=black,solid, -triangle 90,fill=black] (0,0) -- (0,3); \end{tikzpicture} \end{document} I've tried changing this arrow with ultra thick and double, but the arrow head size changes too. I'm trying to only increase the width of the line. I came across a similar post, but `thick arc` doesn't seem to work even when I add all the libraries. | [
0.007368909660726786,
0.006462136283516884,
-0.0020365407690405846,
0.015284903347492218,
-0.011679996736347675,
-0.0012658947380259633,
0.007687984500080347,
-0.0030844162683933973,
-0.011593203991651535,
-0.023686915636062622,
-0.015086416155099869,
-0.003545759478583932,
-0.00065225991420... | [
0.0875428169965744,
-0.06637613475322723,
0.8091216683387756,
-0.2741062641143799,
-0.2972182631492615,
0.1256318837404251,
0.5179327726364136,
-0.6352458596229553,
-0.026181142777204514,
-0.4251812994480133,
0.099620021879673,
0.5425356030464172,
-0.37667661905288696,
-0.2016412764787674,... |
## Question What'd be an optimal way to redirect the home page of a WordPress site to a category archive without involving external factors (e.g. Apache's .htaccess) i.e. within WordPress? ## Why? (Only to justify the question. Please let this not turn the question into _too localized_.) It's a news site. The categories are used as _Editions_ , for example: * US: `http://example.com/main/` * UK: `http://example.com/uk/` * Japan: `http://example.com/jp/` (The category based is removed.) The visitor is to be taken to the US edition (`http://example.com/main/`) by default no matter where he/she is from, and that is why I want the redirect. ## Options Considered I know I can simply modify the main query and have the home page only include posts from the specific category (Edition). But the way the features of the site are constructed, I find the redirect to be more feasible and moreover appropriate. Aside from that, I want the URLs to make the current location clear to the visitor. As for why I can't use the simpler and better performant .htaccess-based redirection -- all the functionality needs to be either within the theme or as a separate plugin. Hence, looking for the best way to do it with WordPress/PHP. I've also considered adding this in `home.php` template file: <?php wp_redirect( 'http://example.com/main/' ); ?> And decided against it as it "almost doubles server load for home page". | [
0.0007752524688839912,
0.00749584473669529,
0.002087209839373827,
0.030640285462141037,
-0.00018721562810242176,
0.0056083919480443,
0.005420690402388573,
0.007413066923618317,
-0.010691594332456589,
-0.0018616074230521917,
-0.016226397827267647,
0.0020539206452667713,
-0.004975685849785805,... | [
0.3509480655193329,
0.1419283002614975,
0.5932813882827759,
0.03571942821145058,
-0.49676313996315,
-0.2534029185771942,
0.31523507833480835,
0.059430431574583054,
-0.32055792212486267,
-0.1823163777589798,
-0.43514665961265564,
0.14386381208896637,
-0.29906409978866577,
0.0165478587150573... |
I am reasking this question in a new thread. My previous post How Can I Use Solve/Reduce Output? seemed to have led to a an interesting dissussion of how to (better) solve the problem that produced the output rather than in how to use the output further as I had intended. [Perhaps it indirectly answered my question by telling me that if I wanted to know more I needed to reformulate the problem and the solution as qiven was as far as Mathematica could go with the original formulation.] My questions is "Can I use the following result for further Mathematica calculations?" [i.e. I am not particularly interested in this solution other than using it as an **example** expression of a conditional with restricted domain.] solQ= (x | y) \[Element] Rationals && -1 <= x <= 1 && (y == -Sqrt[1 - x^2] || y == Sqrt[1 - x^2]) I do know of one Mathematica function that will accept this as input: In[9]:= FindInstance[solQ, {x, y}, 10] During evaluation of In[9]:= FindInstance::fwsol: Warning: FindInstance found only 3 instance(s), but it was not able to prove 10 instances do not exist. >> Out[9]= {{x -> -1, y -> 0}, {x -> 100/2501, y -> 2499/2501}, {x -> 1, y -> 0}} [However (as an aside) this function has strange behavior when I ask for 50 instances it can only find 2 for sure!] In[11]:= FindInstance[solQ, {x, y}, 50] During evaluation of In[11]:= FindInstance::fwsol: Warning: FindInstance found only 2 instance(s), but it was not able to prove 50 instances do not exist. >> Out[11]= {{x -> -1, y -> 0}, {x -> 1, y -> 0}} Are there other Mathematica functions that will accept these types of solutions (like solQ) as a valid input expressions that would help me further explore the solution? Perhaps some graphics functions? Perhaps in an Assumption? ...? Perhaps the solution as given above satisfies a Mathematician (which I am not) by "proving" that a solution exits and giving a few examples? Thank you all for your patience with these rather poorly defined questions [coming from a 75 year old physicist in retirement, with no cohort for discussion, just trying to (re)learn some math and Mathematica.] | [
0.02783256769180298,
0.014443036168813705,
-0.0012591504491865635,
0.006103656720370054,
-0.03312678635120392,
0.00009373622015118599,
0.006465201266109943,
0.01523502916097641,
-0.01657859794795513,
-0.02380172722041607,
-0.0006592644494958222,
0.018565021455287933,
-0.010301318019628525,
... | [
-0.09939784556627274,
0.07995276898145676,
0.3774823546409607,
0.03685310482978821,
-0.45693156123161316,
-0.09687350690364838,
-0.13020525872707367,
-0.26283541321754456,
-0.011481410823762417,
-0.4515068233013153,
-0.01821642741560936,
0.2533184587955475,
-0.04644831269979477,
-0.1181448... |
Here's the scenario: I have a small web design business and was using a branded backlink on the bottom of all client sites. Recently this has become a bit taboo with the Google updates so I went back to a few of my sites and made it so there's only a homepage backlink. After doing this, I've had a drop in rankings, despite this apparently being a best practice. Is this likely a temporary drop that will pick back up? For any new sites, I still want to have a link on all pages of client sites as it's good advertising. I plan to have a do-follow homepage link and then no- follow every other link - is this a good idea? | [
0.015373174101114273,
0.014796698465943336,
-0.006269622128456831,
0.014761161990463734,
-0.004457863979041576,
0.009067301638424397,
0.006450565531849861,
-0.009853003546595573,
-0.01362452283501625,
-0.01368610467761755,
-0.004569857381284237,
0.025164924561977386,
-0.00957498513162136,
... | [
0.7044744491577148,
0.12830659747123718,
0.07029006630182266,
-0.005296475253999233,
-0.030767038464546204,
0.018303897231817245,
0.40610578656196594,
0.4904352128505707,
-0.49833470582962036,
-0.5349305272102356,
0.37113603949546814,
0.4176141917705536,
0.04711269587278366,
0.212291419506... |
I have a bash script which seems to lose the value of the readonly constant after the first time thru the for-in loop. For example: #!/bin/bash readonly DIR="./groups/" for output in "${array[@]}" do catstring+="$DIR$output " done printf "$catstring" cat $catstring > outputfile The array has a bunch of names in it like: file1 file2 file3, etc. The output from the printf statement is "./groups/file1 file2 file3". What I'm expecting is "./groups/file1 ./groups/file2 ./groups/file3". Why is bash losing the value of $DIR after the first time thru the for-in loop? | [
-0.0015911860391497612,
0.017037272453308105,
-0.01936749927699566,
0.02287171222269535,
0.005293859634548426,
-0.00865164678543806,
0.008002211339771748,
0.008641116321086884,
-0.019825950264930725,
0.0005648513324558735,
-0.01209127064794302,
-0.0020079491659998894,
-0.00967602338641882,
... | [
0.13472677767276764,
0.09558097273111343,
0.1540759950876236,
-0.4047970473766327,
-0.21221515536308289,
-0.15377742052078247,
0.10121473670005798,
-0.5493233799934387,
-0.15454258024692535,
-0.5203412771224976,
0.31667324900627136,
0.5949404835700989,
-0.31770193576812744,
0.0093909604474... |
On the menu, I'm talking about custom menus, I have a footer with several columns, each column will have a list of items, each column menus needs to be updatable from the menus option on wp admin, so I created the section on functions.php that calls the menus and also added each menu on the footer.php document, something like: <?php wp_nav_menu( array('menu' => 'f-1', 'menu_class' => '', 'container' => 'nav' )); ?> <?php wp_nav_menu( array('menu' => 'f-2', 'menu_class' => '', 'container' => 'nav' )); ?> <?php wp_nav_menu( array('menu' => 'f-3', 'menu_class' => '', 'container' => 'nav' )); ?> <?php wp_nav_menu( array('menu' => 'f-4', 'menu_class' => '', 'container' => 'nav' )); ?> This is how I call the menus: add_theme_support('nav-menus'); if ( function_exists('register_nav_menus')) { register_nav_menus( array( 'main' => 'Main Nav', 'footer-noticias-financieras' => 'Footer Noticias Financieras', 'footer-otras-publicaciones' => 'Footer Otras Publicaciones', 'footer-foro' => 'Footer Foro' ) ); } The thing is, when I place the menus using the theme menus area it prints automatically on my footer the column with 2 options, home and about, the "Automatically add new top-level pages" is not selected so not sure what's going on, anybody with the same problem? Thanks so much! A.- | [
0.008641132153570652,
0.013282142579555511,
-0.0029047473799437284,
0.019316771999001503,
-0.00924654956907034,
-0.009006007574498653,
0.007125432137399912,
0.0062044239602983,
-0.02332855947315693,
0.003082497278228402,
-0.013397008180618286,
0.0049549611285328865,
0.002996039343997836,
0... | [
-0.11422938108444214,
-0.038619380444288254,
0.6243508458137512,
-0.261296808719635,
-0.06894835829734802,
0.3251379430294037,
-0.06787189841270447,
-0.20609591901302338,
-0.2243727445602417,
-0.6423730850219727,
-0.08995372802019119,
0.5322661995887756,
-0.31286877393722534,
0.03911727294... |
In my native language, there is a single word to express the high energy and enthusiasm shown by many at the beginning of a new job/project/romance etc. Normally it is used with a bit of skepticism/sarcasm. I am looking for a single word or phrase in English that expresses the same | [
-0.003238044446334243,
0.020045846700668335,
0.004204523283988237,
0.023252548649907112,
-0.02899022214114666,
-0.014150572940707207,
0.009761003777384758,
-0.006335713434964418,
-0.018819659948349,
-0.024644553661346436,
0.0018668813863769174,
0.014711104333400726,
0.04849208518862724,
0.... | [
0.656943678855896,
0.33732810616493225,
-0.03642897307872772,
-0.2629806399345398,
-0.22083832323551178,
-0.05927496775984764,
0.4928860366344452,
0.7280017137527466,
-0.21923065185546875,
-0.5249868631362915,
-0.16815954446792603,
0.128298819065094,
0.2798217236995697,
-0.0763053819537162... |
Steam now has trading cards. In order to get them, you're supposed to play specific games _(currently there areonly 6 games supported, but they've said more will come)_. However, it's only possible to find **half** of the total trading cards for each game - you have to trade for the other half. **Can I find the same trading card twice in the same game** _(and if so, will that count against my allowed card-pickups)_? **Is it possible to trade for a duplicate of a card you already own?** | [
0.0070689525455236435,
0.013020215556025505,
-0.0037591527216136456,
0.00426903460174799,
0.01784731075167656,
0.008558982983231544,
0.004255867563188076,
-0.01002219133079052,
-0.018672844395041466,
-0.009755102917551994,
-0.0027976457495242357,
0.01913304068148136,
-0.01008068397641182,
... | [
0.30548784136772156,
-0.2702457904815674,
0.6438318490982056,
0.0913182869553566,
-0.3378583490848541,
-0.29239487648010254,
-0.1579601764678955,
0.20356173813343048,
-0.11146813631057739,
-0.4167404770851135,
0.2256924957036972,
0.1930581033229828,
0.14260613918304443,
0.2634054124355316,... |
I have tried to - enable `Use jQuery` & then save changes. But after saving changes, checkmark `Use jQuery` automatically unchecks itself. I am not sure whats going on. Please help me here.  | [
0.005249917972832918,
0.01371206808835268,
0.008964245207607746,
0.01891128532588482,
0.01606575772166252,
-0.022615039721131325,
0.0067609455436468124,
-0.006673669908195734,
-0.024914240464568138,
0.003617957467213273,
-0.012776648625731468,
0.004949235823005438,
-0.028197024017572403,
0... | [
0.27071624994277954,
-0.10654246062040329,
0.3703242242336273,
0.16985103487968445,
0.04085526242852211,
-0.2388642132282257,
0.3659857511520386,
0.09756667166948318,
-0.5122738480567932,
-0.8272011876106262,
-0.06944583356380463,
0.5965020060539246,
0.11225499212741852,
0.1325617134571075... |
I have some experience writing small tools in Haskell and I find it very intuitive to use, especially for writing filters (using `interact`) that process their standard input and pipe it to standard output. Recently I tried to use one such filter on a file that was about 10 times larger than usual and I got a `Stack space overflow` error. After doing some reading (e.g. here and here) I have identified two guidelines to save stack space (experienced Haskellers, please correct me if I write something that is not correct): 1. Avoid recursive function calls that are not tail-recursive (this is valid for all functional languages that support tail-call optimization). 2. Introduce `seq` to force early evaluation of sub-expressions so that expressions do not grow too large before they are reduced (this is specific to Haskell, or at least to languages using lazy evaluation). After introducing five or six `seq` calls in my code my tool runs smoothly again (also on the larger data). However, I find the original code was a bit more readable. Since I am not an experienced Haskell programmer I wanted to ask if introducing `seq` in this way is a common practice, and how often one will normally see `seq` in Haskell production code. Or are there any techniques that allow to avoid using `seq` too often and still use little stack space? | [
-0.0065130130387842655,
0.012092707678675652,
-0.006295893806964159,
0.014982903376221657,
0.012104940600693226,
-0.0026656708214432,
0.008724178187549114,
0.006715494208037853,
-0.013560539111495018,
0.0003245985135436058,
0.009621737524867058,
0.005509697832167149,
-0.0035040793009102345,
... | [
0.20421762764453888,
-0.06673246622085571,
-0.09790684282779694,
-0.14532005786895752,
-0.16746895015239716,
-0.2404395192861557,
0.5115170478820801,
-0.24371616542339325,
0.0408790186047554,
-0.5336335897445679,
0.1508755385875702,
0.6907682418823242,
-0.6453897356987,
-0.2415162622928619... |
I am trying to get the last star on the Alps Classic map of Season 2 which requires me to perform 3 barrel rolls in a single game. However, I am having difficulty finding the 3 spots required for me to perform the rolls. Where can they be found and are there any tips in completing the challenge? | [
0.0038513692561537027,
0.02026941440999508,
-0.023795822635293007,
0.006324336398392916,
-0.03154143691062927,
0.02796633541584015,
0.011677752248942852,
0.0027320922818034887,
-0.03117452748119831,
0.007144314236938953,
-0.015759283676743507,
0.0038115514907985926,
-0.0443093478679657,
0.... | [
0.1823837012052536,
0.20460738241672516,
0.5030530095100403,
0.05332280695438385,
-0.1748417764902115,
0.12892988324165344,
1.0791178941726685,
0.18011733889579773,
-0.40046927332878113,
-0.47304272651672363,
0.1966373175382614,
0.013760681264102459,
0.22804157435894012,
0.1035132780671119... |
Still learning the fundamentals of the language I would like to ask you what advantages there might be in writing something like: a[1] = 2; a[2] = 4; a[3] = "x"; It seems to me that it is always better to write a = {2, 4, "x"}; Do you know about any practical constructs where indexed variables would offer an advantage? | [
-0.009951974265277386,
0.028002183884382248,
-0.040215592831373215,
7.499705247937527e-7,
-0.03059432841837406,
0.027896109968423843,
0.010326438583433628,
-0.009063263423740864,
-0.0200674906373024,
-0.021137012168765068,
-0.00008306856761919335,
0.0052225650288164616,
-0.011734253726899624... | [
0.17747266590595245,
0.010791447944939137,
0.19902212917804718,
0.137251615524292,
0.061411548405885696,
-0.26189544796943665,
0.1335286945104599,
0.022077549248933792,
-0.09094011783599854,
-0.4181779623031616,
0.021736763417720795,
0.4383154511451721,
0.04624278098344803,
-0.099676109850... |
So in solid state classes we learned about phenomena like band structure and others arising from a periodic potential. Then we get to doing actual experiment and find out that materials being single crystals are actually fairly rare exceptions rather than the rule: more often, the material is a polycrystal or even amorphous. This kind of upsets my knowledge of what to expect, because all these things I learned were based on perfect crystals. But, then, I talk to people more knowledgeable than I about things like the band structure of an evaporated film, and they seem to give me the infuriatingly vague answer that we can still kind of expect to see the same properties of the band structure, but maybe not exactly. I don't even know where to begin theoretically -- it seems like if you have a polycrystal, it's now more like a bunch of quantum wells the width of the grain sizes mashed together, and they can tunnel into each other through their grain boundaries. I guess I could think of something using the knowledge of tunneling I know, but considering the grains are often oriented in random directions, I wouldn't know what to do. So my question is, is there a way to know what to expect of a polycrystal when you know the properties of the material in bulk? Or is it necessarily the type of thing you'd have to figure out from experiment? | [
-0.0063254339620471,
0.009840423241257668,
-0.0005084512522444129,
0.002557467669248581,
-0.00416549714282155,
-0.018116062507033348,
0.006810027174651623,
0.00030534528195858,
-0.017466586083173752,
0.012136024422943592,
-0.0013196845538914204,
0.01783464103937149,
0.014716442674398422,
0... | [
0.5966076254844666,
0.17005346715450287,
-0.2222631871700287,
0.5044283270835876,
-0.03879182040691376,
-0.4014604985713959,
0.6475081443786621,
-0.013754668645560741,
-0.31650310754776,
-0.16761843860149384,
0.007338653784245253,
-0.0032000320497900248,
-0.04648321494460106,
0.20636767148... |
I'm generating a PDF document with pdflatex (more precisely, latexmk invoked from TexMaker 3.4. My LaTeX install is TeXLive 20120719). I can read the document fine in Okular and Gmail's attachment preview, but a colleague that runs Acrobat Reader on Windows reports: `"There was a problem reading this document (131)"`. **EDIT** : The error disappears when using `\usepackage[demo]{graphicx}`, which puts black boxes instead of the real images. They are all PDFs, PNGs and JPGs. Instead of using `demo`, I added the code for a `\demoincludegraphics` command, which makes the file work in Reader when using it to replace the logo in the headers, as shown below. But enabling it (using `\includegraphics` instead of `\demoincludegraphics`) and removing everything else also works in Reader! There is some strange global interaction that makes the file appear broken to Reader when enabling all the graphics... My code compiles fine, reason why I only attach the preamble and a simple text for a Minimal Working Example: \documentclass[12pt,lettersize,twoside]{article} \usepackage[utf8]{inputenc} \usepackage[spanish,es-tabla]{babel} \usepackage[fixlanguage]{babelbib} \usepackage{amssymb,amsmath} \usepackage{graphicx} \usepackage{array} \usepackage{caption} \usepackage{subfig} \usepackage{rotating} \usepackage{booktabs} \usepackage{multirow} \usepackage{calc} \usepackage{stfloats} \usepackage{fixltx2e} \usepackage{url} \usepackage[colorlinks=true,linkcolor=blue,citecolor=red]{hyperref} \usepackage[all]{hypcap} % Margins \usepackage[top=2cm, bottom=2cm, left=3cm, right=2cm]{geometry} % Make header show up in the first page \makeatletter \renewcommand\maketitle{\par \begingroup \renewcommand\thefootnote{\@fnsymbol\c@footnote}% \def\@makefnmark{\rlap{\@textsuperscript{\normalfont\@thefnmark}}}% \long\def\@makefntext##1{\parindent 1em\noindent \hb@xt@1.8em{% \hss\@textsuperscript{\normalfont\@thefnmark}}##1}% \if@twocolumn \ifnum \col@number=\@ne \@maketitle \else \twocolumn[\@maketitle]% \fi \else \newpage \global\@topnum\z@ % Prevents figures from going at top of page. \@maketitle \fi \thispagestyle{fancy}\@thanks % was {empty} \endgroup \setcounter{footnote}{0}% \global\let\thanks\relax \global\let\maketitle\relax \global\let\@maketitle\relax \global\let\@thanks\@empty \global\let\@author\@empty \global\let\@date\@empty \global\let\@title\@empty \global\let\title\relax \global\let\author\relax \global\let\date\relax \global\let\and\relax } \makeatother % Remove date field in title; Remove preceding vertical space \makeatletter \renewcommand{\@maketitle}{ \vspace{-3cm} \begin{center}% {\LARGE \@title}\\ \vspace{2mm}% {\@author} \end{center}% \par} \makeatother \makeatletter % Demo includegraphic boxes \newcommand{\demoincludegraphics}[2][]{% \setlength{\fboxsep}{-\fboxrule}% Remove frame separator/gap \framebox{\rule{0pt}{100pt}\rule{150pt}{0pt}}% Framed box of 150pt x 100pt } % Headers Config \usepackage{fancyhdr} \pagestyle{fancy} \fancyhead{} \renewcommand{\headrulewidth}{0.4pt} \setlength{\headheight}{2.2cm} \addtolength{\textheight}{-1.62cm} \fancyhead[LO,RE]{\demoincludegraphics[height=1.6cm]{logo-slogan}} \fancyhead[LE,RO]{} % Footers Config \fancyfoot{} \renewcommand{\footrulewidth}{0pt} \fancyfoot[C]{\thepage} % Notas al pie etiquetadas por simbolos \renewcommand{\thefootnote}{\fnsymbol{footnote}} \title{Doc Title} \author{My name} \begin{document} \maketitle This is a test. It includes graphics and other floats. \end{document} **EDIT** : I found a workaround: Processing the generated pdf file through Ghostscript, as follows: gs -dSAFER -dBATCH -dNOPAUSE -sDEVICE=pdfwrite -sOutputFile=output.pdf input.pdf where `input.pdf` is the file generated by pdfLaTeX, and `output.pdf` the one that will be readable by Adobe Reader and the rest of the PDF readers. | [
0.002629039343446493,
-0.0017258310690522194,
-0.005546778440475464,
0.014044547453522682,
0.007357494905591011,
0.00732162082567811,
0.009245519526302814,
-0.003960346803069115,
-0.01552523672580719,
-0.033737994730472565,
-0.01697847619652748,
0.006405557505786419,
0.012119529768824577,
... | [
0.2994677722454071,
-0.026265593245625496,
0.4233439266681671,
-0.12210847437381744,
0.10617765039205551,
-0.008121097460389137,
0.2754463851451874,
-0.14158548414707184,
0.04988781735301018,
-0.9028754830360413,
-0.06895749270915985,
0.7310649156570435,
-0.19785527884960175,
-0.1034054160... |
Where can I find Left 4 Dead and Left 4 Dead 2 soundtracks? Do they officially exist? | [
0.020169708877801895,
-0.008554328233003616,
-0.011852635070681572,
0.01118494663387537,
0.09125108271837234,
0.060262225568294525,
0.018846165388822556,
0.03483375534415245,
-0.03011457994580269,
-0.03810443356633186,
-0.030995091423392296,
0.08906078338623047,
0.04004727676510811,
0.0487... | [
0.7865162491798401,
0.20184366405010223,
0.11832215636968613,
0.40619155764579773,
0.05954601243138313,
-0.5371876955032349,
-0.19244863092899323,
0.09180638194084167,
0.02467278577387333,
-0.2984684407711029,
0.04797099158167839,
0.6254738569259644,
0.007736721076071262,
0.104255802929401... |
I'm curious about the radiant intensity distribution of pulsars: what's the general dependence of intensity on angle, and what are typical angular beam widths? How much does the beam width vary between pulsars? (Presumably this is tied to magnetic field strength.) Even something as simple as a very sketchy plot of intensity vs. angle would be great. It's easy enough to find plots of observed intensity vs. time for individual pulsars, but it takes a bit to get from those to the distribution at the source. | [
0.01880117878317833,
0.007500276900827885,
-0.006164855789393187,
0.031211774796247482,
0.002195604145526886,
0.001986035145819187,
0.010312280617654324,
0.004417382180690765,
-0.01696740835905075,
-0.024686738848686218,
-0.007513679563999176,
0.021828947588801384,
-0.008289283141493797,
-... | [
0.3910169005393982,
-0.06963354349136353,
0.39557766914367676,
0.08989593386650085,
-0.37338289618492126,
-0.07307209074497223,
-0.19875866174697876,
0.07947232574224472,
-0.11791335046291351,
-0.563115119934082,
0.15910622477531433,
0.4184003174304962,
0.22854189574718475,
0.4326434135437... |
Firstly, I'm completely new to using `TeX`, so please bear with my ignorance. I am writing an application that will require the program to display some symbols, namely those in the `MusixTeX` library, but I don't know how to go about doing the displaying. Can anyone guide me as to how I might write code to do this? I'll be programming in either `C#` or `JavaScript`, so an answer for either (or both) of those languages would be really helpful. As I said, I'm really new to using `TeX` in general, so the more detailed of an answer you can give, the better. I'm not afraid of work, so if it's a laborious process, that's fine. I'd just like to be able to do it! | [
-0.00144656328484416,
0.01138782687485218,
-0.0020820528734475374,
-0.0019983064848929644,
0.005027436185628176,
0.0019915448501706123,
0.005311858840286732,
0.014492455869913101,
-0.016127411276102066,
-0.006775637622922659,
0.002499269787222147,
-0.0016750423237681389,
-0.00786986760795116... | [
0.636324942111969,
0.327604204416275,
0.00611101882532239,
0.010088545270264149,
-0.04140334576368332,
0.04019086807966232,
0.15264037251472473,
0.08508190512657166,
0.09198451042175293,
-0.6554260849952698,
0.3005528151988983,
0.7754477262496948,
-0.320822536945343,
-0.21251346170902252,
... |
My target is to run a file to be executed on each day whether the WordPress site runs or not. How to execute cron file? I have got stuck in this is issue. This is new task for me and i don't got any correct solution to my question. Please help me to find out the solution. | [
-0.00035702480818144977,
0.015591070987284184,
-0.010037845931947231,
0.027128295972943306,
-0.009750980883836746,
0.006292259786278009,
0.007875428535044193,
0.021284623071551323,
-0.025434527546167374,
-0.01583218201994896,
-0.019272226840257645,
0.015599483624100685,
-0.006723077967762947... | [
0.5779105424880981,
0.14726263284683228,
0.44496190547943115,
0.1259785294532776,
-0.15124347805976868,
0.2480354756116867,
0.30143171548843384,
0.2198672741651535,
-0.0867188423871994,
-0.7037872672080994,
0.18680989742279053,
0.5519323945045471,
-0.3695657253265381,
0.2730069160461426,
... |
I am trying to create a google earth like wep application. **My problem is that I can't get to render a full view of the earth.** The `zoom_all()` function generates the following error : RuntimeError: could not zoom to combined layer extents using zoom_all because proj4 could not back project any layer extents into the map srs (set map 'maximum-extent' to override layer extents) And if I define an envelope, I can only manage to have a cropped view of the earth with mapnik.Envelope(-4500000, -4500000, 4500000, 4500000) or a blank image with mapnik.Envelope(-4600000, -4600000, 4600000, 4600000) **EDIT** : I also tried specifying `maximum-extent` in the xml file but the result was the same as specifying the envelope in the python script. **What am I doing wrong ? Any hint will be appreciated :)** My configuration is : * WGS84 shapefile coastlines-split-4326.zip (http://openstreetmapdata.com/data/coastlines) * mapnik 2.2.0-pre * archlinux This is the python script that renders the image : #!/usr/bin/python2 import mapnik image = 'world-xml.png' map_config = 'world.xml' m = mapnik.Map(1000, 1000) mapnik.load_map(m, map_config) bbox = mapnik.Envelope(-4500000, -4500000, 4500000, 4500000) m.zoom_to_box(bbox) mapnik.render_to_file(m, image) And this is the xml config file : <Map background-color="#ffffff" srs="+proj=ortho +lat_0=0 +lon_0=0 +ellps=WGS84 +units=m +x_0=0 +y_0=0 +no_defs"> <Style name="projet-carte"> <Rule> <PolygonSymbolizer fill="#ffffff" /> <LineSymbolizer stroke="#000000" stroke-width="0.1" /> </Rule> </Style> <Layer name="world" srs="+proj=longlat +ellps=WGS84 +datum=WGS84 +no_defs"> <StyleName>projet-carte</StyleName> <Datasource> <Parameter name="type">shape</Parameter> <Parameter name="file">lines.shp</Parameter> </Datasource> </Layer> </Map> | [
-0.003921518102288246,
0.0021711678709834814,
-0.0056661562994122505,
0.017752094194293022,
-0.031824320554733276,
0.0066313361749053,
0.008679736405611038,
0.026529617607593536,
-0.013399593532085419,
-0.013905098661780357,
-0.0039013680070638657,
0.01794825866818428,
-0.0029908400028944016... | [
0.20693930983543396,
0.12159165740013123,
0.5557481646537781,
-0.07044408470392227,
-0.27460458874702454,
-0.08854301273822784,
0.36626631021499634,
-0.17351992428302765,
-0.10994566977024078,
-0.7403607964515686,
0.0890810489654541,
0.673792839050293,
-0.1220850944519043,
0.07592711597681... |
_**General_ relativity** is a general concept that apples to that applies to the universe at large. **_Special_ relativity** is a _"subset"_ of general relativity that applies to the subset of the universe that shares a certain characteristic. "Subset" is not the right word here - what's a word that addresses the conceptual relationship between a "general concept" and a _"special(ized)"_ application of that concept to a more strictly-defined context? | [
0.007033215370029211,
0.01447493489831686,
-0.0002963499864563346,
0.01996268890798092,
0.02737383358180523,
-0.01841890998184681,
0.008964312262833118,
-0.016708068549633026,
-0.014702469110488892,
0.009256911464035511,
-0.008194323629140854,
0.010696779936552048,
0.0243796668946743,
0.01... | [
0.10657569020986557,
0.22274941205978394,
0.19925326108932495,
-0.07766184210777283,
-0.22316928207874298,
0.18915383517742157,
-0.11998840421438217,
0.042295608669519424,
-0.2581566870212555,
-0.5423749089241028,
-0.3075253665447235,
0.16578693687915802,
-0.26452720165252686,
0.3191215395... |
I thought that PSTricks package was not possible to use in pdfLaTeX but the user Dima claims otherwise. How can I force pdfLaTeX to use PSTricks then? | [
-0.007492695935070515,
0.013704676181077957,
-0.030130360275506973,
0.04489880055189133,
0.05293881148099899,
-0.016134407371282578,
0.01703219674527645,
-0.024015052244067192,
-0.03890872001647949,
-0.03956780582666397,
-0.01891872100532055,
0.0011133315274491906,
-0.0058211591094732285,
... | [
0.4013139009475708,
0.03800894320011139,
0.27855437994003296,
0.19971442222595215,
-0.14481274783611298,
-0.591079831123352,
0.0838145911693573,
-0.26244381070137024,
0.17461787164211273,
-0.417993426322937,
0.07530134171247482,
0.47305828332901,
-0.5039349794387817,
0.14531557261943817,
... |
A client's website appears to have google adsense blocked, the code doesn't load instead it returns a 403 forbidden. We have been through the troubleshooter : https://support.google.com/adsense/troubleshooter/ The highlights: Nobody at the client's office remembers ever seeing an email to say the site was banned. They were advised by a friend of a friend of a... (etc) who works for Google [who couldn't find the site on any blocked list] to just open another account, but the new account was auto-rejected. The trouble shooter suggests address PIN verification failure, but I cannot find anywhere in the adsense app that says that the address is not verified, in fact they have had payments in the past; but they have not used adsense for a while, an alternative ad provider has recently been a victim of the recession, so are wanting to relaunch. The ads are not in a any sort of nested frame, just on a basic test page with real content and minimal design so there is nothing to break. We have gradually stripped more and more off the page to try and get them to work. Both synchronous and asynchronous codes have the same issue. So the troubleshooter points us to the form : https://support.google.com/adsense/contact/cant_see_ad?rd=1 and we get the response > According to our records, ad serving was disabled to your site because the > site was in violation of AdSense programme policies. So we check the policy appeal page and get this message > No sites have had ad serving disabled for your account. So we submit a review request, basically saying, help we're stuck and don't know what to do, and get an auto responder back > Unfortunately, we weren't able to match the Issue ID you provided with any > actions related to the publisher ID you provided. Because we don't have an issue ID to provide! What on earth do we do now? | [
-0.011779822409152985,
0.0034584589302539825,
0.004350611474364996,
0.015344647690653801,
-0.013379968702793121,
0.0048157162964344025,
0.006860497407615185,
0.00922686792910099,
-0.02239174209535122,
0.009831535629928112,
-0.01050734892487526,
0.012269813567399979,
0.0003994968719780445,
... | [
0.26788607239723206,
0.38940438628196716,
0.4861460328102112,
-0.186295285820961,
0.13562773168087006,
-0.017197608947753906,
0.3113071322441101,
-0.03972071036696434,
-0.5411213040351868,
-0.13218365609645844,
0.07970279455184937,
0.47020721435546875,
-0.4838908314704895,
-0.0834008678793... |
Trying to create 1 mosaic dataset from 2 million files. Would it be better to break the files into regions and then create regional mosaic datasets and then make a mosaic dataset of the mosaic datasets (does that make any sense?)? Or just slug it out and do one dataset of 2 million files? BTW the datatype is CIB | [
0.004571393132209778,
0.035909149795770645,
-0.021661197766661644,
0.012945525348186493,
0.004525965079665184,
0.009687403216958046,
0.012346059083938599,
0.0349322184920311,
-0.024787791073322296,
-0.012893370352685452,
-0.015552775003015995,
0.0014069030294194818,
0.009694816544651985,
0... | [
0.5103245377540588,
0.054089583456516266,
0.13558638095855713,
0.06347845494747162,
0.024546552449464798,
-0.5891969203948975,
-0.003877362236380577,
0.2520337402820587,
-0.2642592787742615,
-0.9973235130310059,
-0.17127925157546997,
-0.24713005125522614,
-0.6818865537643433,
0.57838159799... |
How can one easily found an outlier in PH model and why are they so important? Can anybody give me some advice or suggest which R function may I use? Thanks for comments and help. | [
0.026388080790638924,
0.03964337706565857,
-0.029129693284630775,
0.018796227872371674,
-0.05724022909998894,
0.01852049119770527,
0.007967024110257626,
0.039387334138154984,
-0.04521458223462105,
-0.07796190679073334,
0.009319907054305077,
0.021611761301755905,
-0.0018614835571497679,
0.0... | [
0.5423129796981812,
0.10262255370616913,
-0.05798865854740143,
0.3179592192173004,
-0.26075491309165955,
-0.024127421900629997,
0.2715807259082794,
0.19044369459152222,
-0.07009779661893845,
-0.4524080455303192,
0.3057808578014374,
0.27308011054992676,
0.23347730934619904,
0.30867460370063... |
I'm trying to do an adb backup of my Galaxy S3 (i747m), but after receiving the "Now unlock your device and confirm the backup operation" order from adb, I never receive the prompt on the device. It just sits there. Locking and unlocking it does nothing, entering the command locked or unlocked changes nothing, setting the desktop backup password in development options does nothing, disabling the lockscreen does nothing. Device is running the stock Samsung 4.1 ROM and is rooted. Computer is windows 7. How can I get this backup to go through? | [
-0.01525750569999218,
0.0027621842455118895,
-0.01130586490035057,
0.019363347440958023,
-0.03451996296644211,
0.006902916356921196,
0.009811685420572758,
0.0011631352826952934,
-0.017461195588111877,
-0.0031675808131694794,
-0.026541512459516525,
0.01190927717834711,
-0.013563611544668674,
... | [
-0.02264324203133583,
0.2925063967704773,
0.7873634696006775,
-0.015570737421512604,
0.4271501302719116,
-0.1320461928844452,
0.7026668787002563,
-0.010304552502930164,
-0.2753777503967285,
-0.2522878348827362,
-0.41762104630470276,
0.8149527311325073,
-0.15509921312332153,
0.2397488206624... |
I have two rasters of the same region and of same resolution and I want to plot in a scatter plot their values and compute the best - fit line with r-square. Any idea how can I do this on Arcmap or QGIS or extract the values of the rasters in a file so I can import it into Excel? Unfortunately I don't use R neither python. | [
0.008400867693126202,
0.02088715136051178,
-0.007740417495369911,
0.011213471181690693,
-0.02391239069402218,
0.013052459806203842,
0.009094866923987865,
0.01218140684068203,
-0.02464016154408455,
-0.041251543909311295,
0.007362257223576307,
0.01328633539378643,
-0.015003841370344162,
-0.0... | [
0.3109734356403351,
0.08624381572008133,
0.2522059977054596,
-0.06707349419593811,
-0.3008713126182556,
0.1219952329993248,
0.13183064758777618,
-0.1408691257238388,
0.11633427441120148,
-0.8696208596229553,
0.16225966811180115,
0.8886513710021973,
-0.3084658980369568,
-0.1546936333179474,... |
You flip a coin and if it is a head I pay you 1 pound but if it is a tail you pay me 2 pounds. You have 50 pounds and you stop when you spend all of your money or you flipped coin 100 times. What is the expected value of your gain in this game? | [
0.024228785187005997,
0.016222942620515823,
-0.022103019058704376,
0.002917560748755932,
0.014992493204772472,
-0.02720208466053009,
0.012298382818698883,
-0.02584332972764969,
-0.022564303129911423,
-0.012820527888834476,
-0.021031418815255165,
0.01780553162097931,
-0.005386071745306253,
... | [
0.4422440528869629,
-0.25817203521728516,
0.549291729927063,
-0.1070810854434967,
-0.20113572478294373,
0.005027650855481625,
0.11643864959478378,
-0.11769726872444153,
-0.5880380868911743,
-0.19116410613059998,
0.576886773109436,
0.2542542815208435,
-0.13139736652374268,
-0.09991977363824... |
This is related to the following Retain ESRI symbology in Mapinfo TAB . As an initial step (in arcpy) I want to access the symbology of all the datasets and create a document which has the details of the symbology as per the graphic below.  Is this possible (should I invest the time to work it out) or if it isn't possible I will just have to get cracking on doing it manually on hundreds of MXD's (:- To break down the steps -hopefully I can automate at least part of the process. 1. scrolling through mxd's and getting the datasets that are on is no problem lyrFile = arcpy.mapping.Layer if lyrFile.visible == True: 2. as it finds a dataset it should create a .lyr file (as most are just shp/gdb files) and then AttribXXX=lyrFile.xxx where XXX and xxx are BRIGHTNESS = lyrFile.brightness (above format to following) CONTRAST —contrast DATASETNAME —datasetName DATASOURCE —dataSource DESCRIPTION —description LABELCLASSES —labelClasses LONGNAME —longName NAME —name SHOWLABELS —showLabels TRANSPARENCY —transparency WORKSPACEPATH —A layer's workspace or connection file path **The issue is I can't find anything that gets the RGB, line/point/polygon type, line weights, font etc. If these can be found then the next step** 1. write each of the values above to a CSV (easy)...for example (this has to be expanded) f = open(origPath+'\\LayerAtrributes_LOG.txt', 'a') f.write(str(DATASETNAME)+","+str(DATASOURCE)+","+str(DESCRIPTION)+","+str(MXDName)+"\n") f.close() 2. If possible a jpg of the symbology should be produced as in the example and the link to it added to the csv. Am is asking to much of arcpy? | [
0.016239434480667114,
0.016404975205659866,
-0.0027876910753548145,
0.014334844425320625,
0.008599061518907547,
0.0009979724418371916,
0.006677547469735146,
0.028362877666950226,
-0.015991870313882828,
-0.012237180024385452,
-0.002068952424451709,
0.0028163082897663116,
-0.024338915944099426... | [
0.39012283086776733,
-0.14087773859500885,
0.6008058786392212,
0.24208900332450867,
-0.05339440330862999,
-0.15746164321899414,
0.16765306890010834,
-0.281389057636261,
-0.39508476853370667,
-0.7647852301597595,
0.08102277666330338,
0.5633354783058167,
-0.04858078807592392,
-0.316349208354... |
I have 2 websites, the main website located at `example.com`, and another one which is on `a.example.com`. The problem is that I have some duplicated meta description problems to the website located on the subdomain and I can't solve them right now. Will they affect my SEO rankings to my main page? | [
0.008329153060913086,
0.009934194386005402,
0.00887301005423069,
0.02937167137861252,
0.002836985047906637,
0.032665789127349854,
0.011186718009412289,
0.03932403773069382,
-0.02165813557803631,
-0.04022032767534256,
-0.023172758519649506,
0.009345576167106628,
-0.024265481159090996,
0.037... | [
0.3570735454559326,
0.26395052671432495,
0.3763677477836609,
-0.17540057003498077,
-0.5203678607940674,
0.19192323088645935,
0.7030529379844666,
0.14581124484539032,
-0.3845021426677704,
-0.5517429113388062,
0.31562089920043945,
0.32656532526016235,
-0.1627601683139801,
-0.1311051249504089... |
I’m reviewing an article, and can’t give details but here is the situation, and it’s got me puzzled Patients were divided into 4 categories (call them A B C and D), which were exhaustive and exclusive. Adjusted hazard ratios were computed for these four groups for all patients and for two subgroups of patients (call them Lung and Heart). The two subgroups were also exhaustive and exclusive. For the full sample, the hazard ratios for B C and D (compared to A) were 5.08, 4.39, and 1.81. For the lung group, they were 1.15, 1.16 and 1.07. For the heart group they were 2.55, 1.69 and 1.28. That is, they were lower for both subgroups than for the whole population. The lung group and heart group are not the same size; from the info provided, it’s not clear whether A B C and D are equally divided across heart and lung groups. So …. This seems like Simpson’s paradox. But I’ve not seen that term used for survival analysis. I don’t see why it could NOT be so applied. My feeling here is to suggest that heart and lung should ONLY be presented separately, but I’m not completely sure. (Excuse me if I am missing something obvious – dealing with some family health problems and not sleeping that well) Thanks for any insights or thoughts | [
-0.00170566001906991,
0.0127941332757473,
-0.018543053418397903,
0.02066199481487274,
0.012164800427854061,
-0.016431577503681183,
0.0076034292578697205,
-0.004005613271147013,
-0.016991687938570976,
-0.020472444593906403,
-0.002538112923502922,
0.0065846689976751804,
-0.014775261282920837,
... | [
0.33982300758361816,
0.08837902545928955,
-0.0005982499569654465,
-0.26197049021720886,
-0.04810097813606262,
0.5328280925750732,
0.43175244331359863,
-0.5958188772201538,
0.10735398530960083,
-0.5425222516059875,
-0.023312965407967567,
0.19155149161815643,
-0.16711865365505219,
0.06654331... |
Is there a representation for modern programming constructs that are becoming common in modern OOP/hybrid languages like mixins/traits/categories/partial classes ? | [
-0.014103805646300316,
0.036280788481235504,
-0.014029502868652344,
0.00803751964122057,
-0.019604157656431198,
-0.01274745725095272,
0.015022546984255314,
0.017116136848926544,
-0.027021499350667,
0.001312741544097662,
-0.013685697689652443,
0.04812924936413765,
0.023609451949596405,
0.00... | [
0.28118008375167847,
-0.2342943698167801,
-0.43388092517852783,
0.44518741965293884,
0.07717011868953705,
0.3077852129936218,
-0.18261119723320007,
0.13978581130504608,
-0.3474806547164917,
-0.49666017293930054,
-0.04485895112156868,
0.16317453980445862,
-0.1403951346874237,
-0.29737257957... |
We have a data set with two covariates and a categorical grouping variable and want to know if there are significant differences between the slope or intercept among the covariates associated with the different grouping variables. We've used anova() and lm() to compare the fits of three different models: 1) with a single slope and intercept, 2) with different intercepts for each group, and 3) with a slope and an intercept for each group. According to the anova() general linear test, the second model is the most appropriate of the three, there is a significant improvement to the model by including a separate intercept for each group. However, when we look at the 95% confidence intervals for these intercepts -- they all overlap, suggesting there aren't significant differences between the intercepts. How can these two results be reconciled? We thought another way of interpreting the results of the model- selection method was that there has to be at least one significant difference among the intercepts... but perhaps this is not correct? Below is the R code to replicate this analysis. We've used the dput() function so you can work with exactly the same data we're grappling with. # Begin R Script # > dput(data) structure(list(Head = c(1.92, 1.93, 1.79, 1.94, 1.91, 1.88, 1.91, 1.9, 1.97, 1.97, 1.95, 1.93, 1.95, 2, 1.87, 1.88, 1.97, 1.88, 1.89, 1.86, 1.86, 1.97, 2.02, 2.04, 1.9, 1.83, 1.95, 1.87, 1.93, 1.94, 1.91, 1.96, 1.89, 1.87, 1.95, 1.86, 2.03, 1.88, 1.98, 1.97, 1.86, 2.04, 1.86, 1.92, 1.98, 1.86, 1.83, 1.93, 1.9, 1.97, 1.92, 2.04, 1.92, 1.9, 1.93, 1.96, 1.91, 2.01, 1.97, 1.96, 1.76, 1.84, 1.92, 1.96, 1.87, 2.1, 2.17, 2.1, 2.11, 2.17, 2.12, 2.06, 2.06, 2.1, 2.05, 2.07, 2.2, 2.14, 2.02, 2.08, 2.16, 2.11, 2.29, 2.08, 2.04, 2.12, 2.02, 2.22, 2.22, 2.2, 2.26, 2.15, 2, 2.24, 2.18, 2.07, 2.06, 2.18, 2.14, 2.13, 2.2, 2.1, 2.13, 2.15, 2.25, 2.14, 2.07, 1.98, 2.16, 2.11, 2.21, 2.18, 2.13, 2.06, 2.21, 2.08, 1.88, 1.81, 1.87, 1.88, 1.87, 1.79, 1.99, 1.87, 1.95, 1.91, 1.99, 1.85, 2.03, 1.88, 1.88, 1.87, 1.85, 1.94, 1.98, 2.01, 1.82, 1.85, 1.75, 1.95, 1.92, 1.91, 1.98, 1.92, 1.96, 1.9, 1.86, 1.97, 2.06, 1.86, 1.91, 2.01, 1.73, 1.97, 1.94, 1.81, 1.86, 1.99, 1.96, 1.94, 1.85, 1.91, 1.96, 1.9, 1.98, 1.89, 1.88, 1.95, 1.9, 1.94, NA, 1.84, 1.83, 1.84, 1.96, 1.74, 1.91, 1.84, 1.88, 1.83, 1.93, 1.78, 1.88, 1.93, 2.15, 2.16, 2.23, 2.09, 2.36, 2.31, 2.25, 2.29, 2.3, 2.04, 2.22, 2.19, 2.25, 2.31, 2.3, 2.28, 2.25, 2.15, 2.29, 2.24, 2.34, 2.2, 2.24, 2.17, 2.26, 2.18, 2.17, 2.34, 2.23, 2.36, 2.31, 2.13, 2.2, 2.27, 2.27, 2.2, 2.34, 2.12, 2.26, 2.18, 2.31, 2.24, 2.26, 2.15, 2.29, 2.14, 2.25, 2.31, 2.13, 2.09, 2.24, 2.26, 2.26, 2.21, 2.25, 2.29, 2.15, 2.2, 2.18, 2.16, 2.14, 2.26, 2.22, 2.12, 2.12, 2.16, 2.27, 2.17, 2.27, 2.17, 2.3, 2.25, 2.17, 2.27, 2.06, 2.13, 2.11, 2.11, 1.97, 2.09, 2.06, 2.11, 2.09, 2.08, 2.17, 2.12, 2.13, 1.99, 2.08, 2.01, 1.97, 1.97, 2.09, 1.94, 2.06, 2.09, 2.04, 2, 2.14, 2.07, 1.98, 2, 2.19, 2.12, 2.06, 2, 2.02, 2.16, 2.1, 1.97, 1.97, 2.1, 2.02, 1.99, 2.13, 2.05, 2.05, 2.16, 2.02, 2.02, 2.08, 1.98, 2.04, 2.02, 2.07, 2.02, 2.02, 2.02), Site = structure(c(2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 2L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 5L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 3L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 4L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 1L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L, 6L), .Label = c("ANZ", "BC", "DV", "MC", "RB", "WW"), class = "factor"), Leg = c(2.38, 2.45, 2.22, 2.23, 2.26, 2.32, 2.28, 2.17, 2.39, 2.27, 2.42, 2.33, 2.31, 2.32, 2.25, 2.27, 2.38, 2.28, 2.33, 2.24, 2.21, 2.22, 2.42, 2.23, 2.36, 2.2, 2.28, 2.23, 2.33, 2.35, 2.36, 2.26, 2.26, 2.3, 2.23, 2.31, 2.27, 2.23, 2.37, 2.27, 2.26, 2.3, 2.33, 2.34, 2.27, 2.4, 2.22, 2.25, 2.28, 2.33, 2.26, 2.32, 2.29, 2.31, 2.37, 2.24, 2.26, 2.36, 2.32, 2.32, 2.15, 2.2, 2.29, 2.37, 2.26, 2.24, 2.23, 2.24, 2.26, 2.18, 2.11, 2.23, 2.31, 2.25, 2.15, 2.3, 2.33, 2.35, 2.21, 2.36, 2.27, 2.24, 2.35, 2.24, 2.33, 2.32, 2.24, 2.35, 2.36, 2.39, 2.28, 2.36, 2.19, 2.27, 2.39, 2.23, 2.29, 2.32, 2.3, 2.32, NA, 2.25, 2.24, 2.21, 2.37, 2.21, 2.21, 2.27, 2.27, 2.26, 2.19, 2.2, 2.25, 2.25, 2.25, NA, 2.24, 2.17, 2.2, 2.2, 2.18, 2.14, 2.17, 2.27, 2.28, 2.27, 2.29, 2.23, 2.25, 2.33, 2.22, 2.29, 2.19, 2.15, 2.24, 2.24, 2.26, 2.25, 2.09, 2.27, 2.18, 2.2, 2.25, 2.24, 2.18, 2.3, 2.26, 2.18, 2.27, 2.12, 2.18, 2.33, 2.13, 2.28, 2.23, 2.16, 2.2, 2.3, 2.31, 2.18, 2.33, 2.29, 2.26, 2.21, 2.22, 2.27, 2.32, 2.24, 2.25, 2.17, 2.2, 2.26, 2.27, 2.24, 2.25, 2.09, 2.25, 2.21, 2.24, 2.21, 2.22, 2.13, 2.24, 2.21, 2.3, 2.34, 2.35, 2.32, 2.46, 2.43, 2.42, 2.41, 2.32, 2.25, 2.33, 2.19, 2.45, 2.32, 2.4, 2.38, 2.35, 2.39, 2.29, 2.35, 2.43, 2.29, 2.33, 2.31, 2.28, 2.38, 2.32, 2.43, 2.27, 2.4, 2.37, 2.27, 2.41, 2.32, 2.38, 2.23, 2.33, 2.21, 2.34, 2.19, 2.34, 2.35, 2.35, 2.31, 2.33, 2.41, 2.53, 2.39, 2.17, 2.16, 2.38, 2.34, 2.33, 2.33, 2.29, 2.43, 2.28, 2.34, 2.38, 2.3, 2.29, 2.43, 2.36, 2.24, 2.35, 2.38, 2.4, 2.36, 2.42, 2.28, 2.45, 2.33, 2.32, 2.33, 2.31, 2.44, 2.37, 2.4, 2.35, 2.33, 2.31, 2.36, 2.43, 2.38, 2.4, 2.38, 2.46, 2.33, 2.38, 2.23, 2.24, 2.39, 2.36, 2.19, 2.32, 2.37, 2.39, 2.34, 2.39, 2.23, 2.25, 2.29, 2.39, 2.35, NA, 2.28, 2.35, 2.38, 2.34, 2.17, 2.29, NA, 2.26, NA, NA, NA, 2.24, 2.33, 2.23, 2.28, 2.29, 2.23, 2.2, 2.27, 2.31, 2.31, 2.26, 2.28)), .Names = c("Head", "Site", "Leg"), class = "data.frame", row.names = c(NA, -312L )) # plot graph library(ggplot2) qplot(Head, Leg, color=Site, data=data) + stat_smooth(method="lm", alpha=0.2) + theme_bw()  # create linear models lm.1 <- lm(Leg ~ Head, data) lm.2 <- lm(Leg ~ Head + Site, data) lm.3 <- lm(Leg ~ Head*Site, data) # evaluate linear models anova(lm.1, lm.2, lm.3) anova(lm.1, lm.2) # > anova(lm.1, lm.2) # Analysis of Variance Table # Model 1: Leg.3.1 ~ Head.W1 # Model 2: Leg.3.1 ~ Head.W1 + Site # Res.Df RSS Df Sum of Sq F Pr(>F) # 1 302 1.25589 # 2 297 0.91332 5 0.34257 22.28 < 2.2e-16 *** # examining the multiple-intercepts model (lm.2) summary(lm.2) coef(lm.2) confint(lm.2) # extracting the intercepts intercepts <- coef(lm.2)[c(1, 3:7)] intercepts.1 <- intercepts[1] intercepts <- intercepts.1 + intercepts intercepts[1] <- intercepts.1 intercepts # extracting the confidence intervals ci <- confint(lm.2)[c(1, 3:7),] ci[2:6,] <- ci[2:6,] + confint(lm.2)[1,] ci[,1] # putting everything together in a dataframe labels <- c("ANZ", "BC", "DV", "MC", "RB", "WW") ci.dataframe <- data.frame(Site=labels, Intercept=intercepts, CI.low = ci[,1], CI.high = ci[,2]) ci.dataframe # plotting intercepts and 95% CI qplot(Site, Intercept, geom=c("point", "errorbar"), ymin=CI.low, ymax=CI.high, data=ci.dataframe, ylab="Intercept & 95% CI")  Just to summarize -- the problem is that the 95% CIs for the intercepts all overlap, but the model selection method suggests that the best model is one that fits different intercepts. So I'm inclined to think either our model selection method is flawed or the 95% CIs for the intercept estimates were calculated incorrectly. Any thoughts would be greatly appreciated! | [
0.002049581613391638,
0.02025517448782921,
-0.02814338542521,
0.015393747016787529,
-0.029765110462903976,
0.011666852980852127,
0.01179877296090126,
0.004561110399663448,
-0.008200952783226967,
-0.01083296537399292,
-0.0115232914686203,
0.0058163926005363464,
-0.00023709982633590698,
0.01... | [
-0.18744874000549316,
-0.38975048065185547,
-0.036171697080135345,
0.19276465475559235,
-0.06466347724199295,
0.6055746674537659,
0.14942552149295807,
-0.5845669507980347,
0.012585180811583996,
-0.7304153442382812,
0.13766010105609894,
0.3864659070968628,
-0.11895929276943207,
0.0836522132... |
How can I get the current .tex source line number? I want to use it to generate useful error messages in a package I am writing. There are some errors that can only be detected at the end of the document, so I want to create a command that checks for them and tells at what line the error was made. To do that, I need to be able to save the source line number in the command that was misapplied. Preferably a command that can be used like `\sourcelineno` in this example: \documentclass{minimal} \newcommand \showlineno {line \sourcelineno} \begin{document} You are now at \showlineno. This paragraph starts at \showlineno, and some more text, and runs until \showlineno. \end{document} where the output would be You are now at line 6. This paragraph starts at line 8, and some more text, and runs until line 10. | [
0.02279249206185341,
0.005762661341577768,
-0.010843180119991302,
0.012304000556468964,
0.0035554352216422558,
0.011928870342671871,
0.007203157991170883,
-0.0063915313221514225,
-0.019314032047986984,
-0.004571486264467239,
-0.00012855557724833488,
-0.0027428369503468275,
0.0003785921726375... | [
0.15712137520313263,
-0.1637454777956009,
0.751380205154419,
0.0882117822766304,
0.40328487753868103,
-0.36060309410095215,
0.15530207753181458,
-0.03981729969382286,
0.04802025854587555,
-0.6949383020401001,
0.08609571307897568,
0.7870931625366211,
-0.1428183913230896,
0.12950360774993896... |
When ever I try to fastboot my kindle fire HD 7 inch. Running 7.4.3 it does not pick up the device (it is also rooted) when I try the command `fastboot -i 0x1949 getvar product`. I connect the kindle and it just reboots as it normally would the command window and it still says waiting for device. I installed the SDK correctly since it allowed me to root my kindle just fine. Do I need another tool or should my kindle not be rooted? | [
-0.014367349445819855,
-0.0006687488057650626,
-0.006652967073023319,
0.023981744423508644,
-0.0331234484910965,
-0.0027664608787745237,
0.00897974893450737,
0.004693054594099522,
-0.01851198449730873,
0.007353084161877632,
-0.01227356307208538,
0.006674395874142647,
-0.007647810038179159,
... | [
0.2800742983818054,
-0.35803237557411194,
0.7302234172821045,
0.0394720733165741,
-0.00022803974570706487,
-0.24392376840114594,
0.47871941328048706,
0.03586195036768913,
-0.1680145412683487,
-0.6748930811882019,
-0.23458322882652283,
0.5668807625770569,
0.09389955550432205,
0.003315215697... |
I want to export LD_LIBRARY_PATH to system services and all users. System services are run before login, so .bashrc is not applicable. How to achieve this? | [
0.04351429641246796,
0.028149940073490143,
-0.008531169965863228,
0.02584480121731758,
0.019337570294737816,
0.00046884064795449376,
0.015902206301689148,
0.0435737743973732,
-0.026194846257567406,
-0.04668238386511803,
-0.04402551427483559,
0.014449667185544968,
-0.07052857428789139,
0.04... | [
0.5120901465415955,
-0.06278041750192642,
0.3074609637260437,
-0.034729715436697006,
-0.15264052152633667,
-0.3622082471847534,
0.24143992364406586,
-0.1942272186279297,
0.0036805940326303244,
-0.7444062232971191,
-0.07293656468391418,
0.6121859550476074,
-0.7327788472175598,
0.08265879750... |
A lot of times when I run `yum search`, it tends first to download the metadata from a repository. How do I force it to only search the local cache? I prefer quick results to accurate data in this case, and am on a slow line. | [
-0.001423523062840104,
0.011527144350111485,
-0.019876496866345406,
0.016708852723240852,
-0.01755341701209545,
0.013750143349170685,
0.010101786814630032,
-0.002575248247012496,
-0.03545429930090904,
-0.00857069343328476,
-0.013718362897634506,
0.01007560919970274,
-0.029532859101891518,
... | [
0.37849512696266174,
0.24131865799427032,
-0.006433975882828236,
0.3325234353542328,
0.0739966481924057,
-0.4870545566082001,
0.20448708534240723,
0.23525100946426392,
-0.013347533531486988,
-0.19274388253688812,
0.1247430145740509,
0.49100571870803833,
0.18812918663024902,
0.5147562623023... |
I am doing project on adapting Linux for scalable multi core architecture. I am testing Linux's performance with benchmarks. When I am running the benchmarks in the system then other programs like background processes were running in the system. I think system load due to background processes is not fixed, as a result execution time of benchmarks may vary due to system's varying load. Then how to execute benchmarks uniformly. I already searched in net and asked researchers about it but not yet able to find solution of this problem. If possible please try to help me. | [
0.02085787057876587,
0.011177510023117065,
-0.01139415055513382,
0.018264396116137505,
-0.007248255889862776,
-0.0223856121301651,
0.0073197707533836365,
0.01346365362405777,
-0.012802272103726864,
0.0046555884182453156,
-0.00942384172230959,
0.017650237306952477,
-0.002070345915853977,
0.... | [
0.5477976202964783,
0.0006069373339414597,
-0.2070906162261963,
0.08000808209180832,
0.00308805494569242,
-0.10519705712795258,
0.02736142836511135,
0.3614388406276703,
-0.37872761487960815,
-0.6980903148651123,
0.2936888337135315,
0.5135465860366821,
-0.11006496846675873,
0.27429810166358... |
I ask this question so that I can use it to my advantage. At what point does ping become an issue for a player? This would help me know (as a jungler for instance) who to help (gank bot more) or who to attack (gank an opposing player more) Thanks | [
-0.014866421930491924,
0.03180783987045288,
0.004790503531694412,
0.0005025806021876633,
-0.014327285811305046,
0.020881105214357376,
0.008346298709511757,
-0.0008521098061464727,
-0.032581835985183716,
-0.0033704363740980625,
0.004677423741668463,
0.03189116343855858,
-0.012567613273859024,... | [
0.19274750351905823,
-0.08943197131156921,
0.4250892102718353,
0.03979624807834625,
-0.3545830547809601,
-0.10364477336406708,
-0.15411515533924103,
-0.2163144052028656,
-0.19723433256149292,
-0.7982645630836487,
0.36893391609191895,
0.45015886425971985,
-0.2872708737850189,
-0.36446312069... |
as far as I know backlinks are good for SEO when they come from external or websites or domains, what about subdomain? do they have the same value of external backlinks or they are just like internal ones? especially when I have the possibility to put each subdomain on a different IP/country Thanks. | [
0.018974287435412407,
0.008405880071222782,
-0.0010744528844952583,
0.027747102081775665,
0.01055651530623436,
0.03109699860215187,
0.012163451872766018,
0.019019778817892075,
-0.0407227948307991,
-0.02384885959327221,
-0.012105333618819714,
0.02178417518734932,
0.011912125162780285,
0.029... | [
0.5519413352012634,
-0.17139776051044464,
0.4243619740009308,
0.39659732580184937,
-0.4505641758441925,
-0.2613579034805298,
0.5207087993621826,
0.42518261075019836,
-0.4779569208621979,
-0.3824266195297241,
-0.020237546414136887,
0.23178336024284363,
-0.17472849786281586,
0.06938555836677... |
I need to get the title of the attached USB device. I can do that with `lsusb`. udev allows some substitutions when I write rules: say, we can use `$kernel` to get name of the device, or `$devpath` to get path to the device. But problem is that `lsusb` returns string like that: Bus 005 Device 032: ID 0403:6001 Future Technology Devices International, Ltd FT232 USB-Serial (UART) IC But udev's devpath is: /devices/pci0000:00/0000:00:1d.0/usb5/5-2 Bus number is the same (`5`), but numbers are different: `Device 032` seems to be some logical number (when I reattach the device, this number increases), and `2` seems to be physical device number. So udev returns physical number, and I need to get logical number. Then, i can retrieve data like this: `lsusb -D /dev/bus/usb/005/032` So, how can I get logical device number `032` by physical path like `/devices/pci0000:00/0000:00:1d.0/usb5/5-2` ? | [
-0.0046156891621649265,
0.003579826094210148,
-0.014593574218451977,
0.006456366740167141,
-0.0056851147674024105,
0.0241602323949337,
0.00783240981400013,
0.010217978619039059,
-0.012543534860014915,
-0.014545056968927383,
-0.0049052126705646515,
-0.005224329419434071,
0.002006664639338851,... | [
0.02119814231991768,
-0.004007373936474323,
0.5516214966773987,
0.2844984233379364,
0.141730397939682,
-0.017182864248752594,
-0.08571436256170273,
-0.14058420062065125,
-0.19601263105869293,
-0.6853869557380676,
-0.19797103106975555,
0.5644848942756653,
-0.2794525623321533,
0.497176170349... |
I've been through the Act many times and have not seen this drop. Where is it?  | [
-0.014818012714385986,
0.006028107367455959,
0.011775804683566093,
0.030026504769921303,
0.028884408995509148,
-0.010492580011487007,
0.0059247566387057304,
0.020759303122758865,
-0.03336154669523239,
0.027869384735822678,
-0.01769101433455944,
0.01692996546626091,
-0.02662206068634987,
0.... | [
0.6670195460319519,
0.1236928179860115,
0.2691449820995331,
0.22353625297546387,
-0.18086956441402435,
0.09283167868852615,
0.35018643736839294,
0.3130786418914795,
-0.27011099457740784,
-0.2502162456512451,
0.2759633958339691,
-0.24193647503852844,
0.022400138899683952,
-0.059444297105073... |
I am trying to import the print plugin within an application built using the Opengeo Suite sdk and I have run into issues as is described in the following thread https://getsatisfaction.com/opengeo/topics/doubts_about_the_use_of_print_from_openlayers_ready_gxp After having followed all the instructions provided, I still cannot get the print plugin the work. The print button appears but I get the message Uncaught TypeError: undefined is not a function This seems to be on the following line: var scaleLine = new OpenLayers.Control.ScaleLine({ ... I have also tried using an alternate approach wherein I have a handler for the print button as is shown below: var app = new gxp.Viewer({ portalConfig: { layout: "border", region: "center", ... ... ... // layer sources sources: { local: { ptype: "gxp_wmscsource", url: "/geoserver/wms", version: "1.1.1" }, osm: { ptype: "gxp_osmsource" }, geonetwork: { ptype: "gxp_cataloguesource", url: "/geonetwork/srv/en/csw", title: "Geonetwork" }, ol: { ptype: "gxp_olsource" } }, // map and layers map: { id: "mymap", // id needed to reference map in portalConfig above title: "Map", //projection: "EPSG:900913", //center: [-10764594.758211, 4523072.3184791], projection: "EPSG:4326", center: [-97, 38], zoom: 3, layers: [{ source: "ol", type: "OpenLayers.Layer.WMS", args: ["World map", "http://vmap0.tiles.osgeo.org/wms/vmap0", {layers: 'basic'}], group: "background" }, { source: "ol", type: "OpenLayers.Layer", args: ["Blank"], visibility: false, group: "background" //source: "osm", //name: "mapnik", //group: "background" }//, { //source: "local", //name: "usa:states", //selected: true //}//, { //source: "google", //name: "ROADMAP", //group: "background" //} ], bbar: [{ text: "Print...", handler: showPrintWindow }], items: [{ xtype: "gx_zoomslider", vertical: true, height: 100 }] } }); function showPrintWindow() { var printWindow = new Ext.Window({ title: "Print", modal: true, border: false, resizable: false, width: 360, autoHeight: true, items: new GeoExt.ux.PrintPreview({ autoHeight: true, printMapPanel: { // limit scales to those that can be previewed limitScales: true, // no zooming on the map map: {controls: [ new OpenLayers.Control.Navigation({ zoomBoxEnabled: false, zoomWheelEnabled: false }), new OpenLayers.Control.PanPanel() ]} }, printProvider: { // using get for remote service access without same origin // restriction. For async requests, we would set method to "POST". method: "GET", //method: "POST", // capabilities from script tag in Printing.html. capabilities: printCapabilities, listeners: { "print": function() {printWindow.close();} } }, includeLegend: true, mapTitle: "PrintMapPanel Demo", sourceMap: app.map }) }).show().center(); } From within the showPrintWindow function the sourceMap is set to app.map. However, when I click on the print button now, I get the error message: Uncaught TypeError: Object #<Object> has no method 'getProjection' PrintMapPanel.js:176 GeoExt.PrintMapPanel.Ext.extend.initComponent PrintMapPanel.js:176 Ext.Component ext-all.js:7 L ext-base.js:7 L ext-base.js:7 L ext-base.js:7 L ext-base.js:7 L ext-base.js:7 GeoExt.ux.PrintPreview.Ext.extend.initComponent PrintPreview.js:156 Ext.Component ext-all.js:7 L ext-base.js:7 L ext-base.js:7 L ext-base.js:7 showPrintWindow app.js:185 Ext.Button.Ext.extend.onClick ext-all.js:7 I I am not entirely sure how I can get this working and any help is greatly appreciated. | [
-0.01839008554816246,
0.000634449883364141,
-0.0006156867602840066,
0.008932041935622692,
-0.009568355977535248,
-0.00833586510270834,
0.008114397525787354,
0.02828814834356308,
-0.015545537695288658,
-0.012932859361171722,
-0.01027311198413372,
0.01196129247546196,
-0.00877828523516655,
0... | [
0.41011732816696167,
0.24435313045978546,
0.5831423401832581,
-0.40345898270606995,
-0.19244223833084106,
-0.2280687540769577,
0.3788098394870758,
-0.10276586562395096,
-0.12844665348529816,
-0.7197045683860779,
0.13265757262706757,
0.5847570300102234,
-0.47810041904449463,
0.2459146678447... |
I have a document that is to be compiled in PDF (using `pdflatex`) and XHTML (using `tex4ht`). For the XHTML version, I wish to have the images linking to their source in a new window (`target="_blank"`). Now, this is perfectly possible by some fiddling with `\HCode`, but I prefer not having to write this _every time_ when I include a graphic in my TeX source. Is there a `\Configure` command I can use in my `tex4ht` config file, so that an a-tag is wrapped around the resulting img-tag automatically? Note: I noticed `\Configure{IMG}`, but I assume that's for generated images, + there is not clear "image source" variable I can use. | [
-0.014673980884253979,
0.007974484004080296,
-0.004302645102143288,
0.003150791395455599,
0.006026377901434898,
-0.017108358442783356,
0.0062028635293245316,
-0.0022879517637193203,
-0.012352967634797096,
-0.0060469359159469604,
-0.009678136557340622,
-0.0038587101735174656,
-0.0064547327347... | [
0.30101892352104187,
0.21868674457073212,
0.7185680866241455,
0.08779222518205643,
0.022359900176525116,
-0.10276252031326294,
-0.08296291530132294,
-0.057562291622161865,
-0.03563214838504791,
-1.0190248489379883,
0.002472597174346447,
0.8085283637046814,
-0.2689497768878937,
-0.041338875... |
I want to buy a domain that is in a status pendingDelete. who.is shows: Registrar: NAMEKING.COM, INC. Whois Server: whois.nameking.com Referral URL: http://www.nameking.com Status: pendingDelete Expiration Date: 2011-10-08 Creation Date: 2003-10-08 Last Update Date: 2011-12-08 Name Servers: ns1.dsredirection.com ns2.dsredirection.com Information Updated: Wed, 14 Dec 2011 18:14:18 UTC Can I know why this domain isn't expired yet (considering its expiration date)? Consider that I have actived a backorder from godaddy too | [
-0.0036723222583532333,
0.006134297698736191,
0.011437427252531052,
0.013700651004910469,
0.016842296347022057,
0.017192646861076355,
0.007736247032880783,
0.01619523949921131,
-0.010838414542376995,
0.029434137046337128,
-0.009222624823451042,
0.015307801775634289,
0.007160150911659002,
0... | [
0.5944961905479431,
0.024191347882151604,
1.1974130868911743,
0.07171623408794403,
0.25970205664634705,
-0.43653762340545654,
0.36900267004966736,
-0.11506950110197067,
-0.10686620324850082,
-0.2744041383266449,
0.2661677896976471,
0.4404743015766144,
0.18891005218029022,
0.815670728683471... |
The best I can achieve so far is to use AutoShare (https://play.google.com/store/apps/details?id=com.dngames.autoshareg). I take the photo, share to autoshare, and it's done. It would be great to be able to snap the picture and have it automatically sent via email, avoiding the share step. Is there any way to do this? | [
-0.01961706578731537,
-0.011514576151967049,
-0.00573016656562686,
0.009826414287090302,
0.023469438776373863,
-0.008236776106059551,
0.007020187098532915,
0.01486791018396616,
-0.019860733300447464,
0.011788835749030113,
-0.01498762983828783,
0.012708798982203007,
0.006085017696022987,
0.... | [
0.743316113948822,
0.15147627890110016,
0.5898788571357727,
-0.11143796145915985,
0.1224789246916771,
-0.0352640263736248,
0.20988884568214417,
-0.26036861538887024,
-0.09325142204761505,
-0.6026455163955688,
0.3072092533111572,
0.48106566071510315,
0.1332167237997055,
-0.13034142553806305... |
I have a serious problem. When I write a new post in the edit window, I try to upload a picture via the media upload button, and it just doesn't add them. I just cant upload any type on media at all to my posts. How can I fix it? | [
0.004694958683103323,
-0.004682266619056463,
-0.011900254525244236,
0.024490807205438614,
-0.0018471883377060294,
-0.027603259310126305,
0.007342152763158083,
0.03262011706829071,
-0.02333945780992508,
0.0045755947940051556,
-0.01746450923383236,
0.010217135772109032,
0.0008282496128231287,
... | [
0.48644912242889404,
0.46011534333229065,
0.4331429898738861,
0.13317276537418365,
-0.17433442175388336,
-0.19719085097312927,
0.5124287009239197,
0.10863927006721497,
-0.29044458270072937,
-0.6199986338615417,
0.34439101815223694,
0.23312915861606598,
-0.3588886260986328,
0.26090738177299... |
I am using `Cylinder` to produce wide flat disks (in _Mathematica_ 8). This works just fine except that the circular base of such a cylinder turns out to be really just a 40-gon which is simply too coarse an approximation to a circle for what I have in mind. Is there a way to convince _Mathematica_ to use say a 200-gon as a circular base for a cylinder? Here is an example of the kind of picture that I am trying to create. Zoom in to see how coarse the cylinders' curved surfaces pan out. Graphics3D[{Cylinder[{{0, 0, 0}, 0.0011 {0.`, 0.`, -0.9510565162951536`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, 0.`, 0.9510565162951536`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.85065080835204`, 0.`, -0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.85065080835204`, 0.`, 0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.6881909602355868`, -0.5`, -0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.6881909602355868`, 0.5`, -0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.6881909602355868`, -0.5`, 0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.6881909602355868`, 0.5`, 0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.2628655560595668`, -0.8090169943749475`, \ -0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.2628655560595668`, 0.8090169943749475`, -0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.2628655560595668`, -0.8090169943749475`, 0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.2628655560595668`, 0.8090169943749475`, 0.42532540417601994`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, 0.`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, 0.`, 1.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.42532540417602`, 0.3090169943749474`, 0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.7236067977499789`, 0.5257311121191336`, 0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.16245984811645317`, 0.5`, 0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.2628655560595668`, 0.8090169943749473`, 0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.27639320225002106`, 0.8506508083520399`, 0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.42532540417602`, -0.3090169943749474`, 0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.7236067977499789`, -0.5257311121191336`, 0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.85065080835204`, 0.`, 0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.16245984811645317`, -0.5`, 0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.27639320225002106`, -0.8506508083520399`, 0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.2628655560595668`, -0.8090169943749473`, 0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.5257311121191336`, 0.`, 0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.8944271909999159`, 0.`, 0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.6881909602355868`, -0.5`, 0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.6881909602355868`, 0.5`, 0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.7236067977499789`, -0.5257311121191336`, \ -0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.42532540417602`, -0.3090169943749474`, \ -0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, 0.`, -1.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.2628655560595668`, -0.8090169943749473`, \ -0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.16245984811645317`, -0.5`, -0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.27639320225002106`, -0.8506508083520399`, \ -0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.7236067977499789`, 0.5257311121191336`, -0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.42532540417602`, 0.3090169943749474`, -0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.85065080835204`, 0.`, -0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.27639320225002106`, 0.8506508083520399`, -0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.16245984811645317`, 0.5`, -0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.2628655560595668`, 0.8090169943749473`, -0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.8944271909999159`, 0.`, -0.4472135954999579`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.5257311121191336`, 0.`, -0.8506508083520399`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.6881909602355868`, 0.5`, -0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.6881909602355868`, -0.5`, -0.5257311121191336`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, 1.`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.5877852522924731`, 0.8090169943749473`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.9510565162951535`, 0.3090169943749474`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.9510565162951535`, -0.3090169943749474`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {-0.5877852522924731`, -0.8090169943749473`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.`, -1.`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.5877852522924731`, -0.8090169943749473`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.9510565162951535`, -0.3090169943749474`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.9510565162951535`, 0.3090169943749474`, 0.`}}, .1], Cylinder[{{0, 0, 0}, 0.0011 {0.5877852522924731`, 0.8090169943749473`, 0.`}}, .1]}] | [
-0.009120611473917961,
0.008116846904158592,
-0.01698419451713562,
0.02096133679151535,
-0.01806204579770565,
-0.010937271639704704,
0.006871497258543968,
-0.0107490923255682,
-0.012856467626988888,
-0.04092903435230255,
-0.0018267820123583078,
-0.0003587103565223515,
0.004697652067989111,
... | [
0.1010114923119545,
0.36293795704841614,
0.32544630765914917,
0.299956738948822,
0.013617601245641708,
-0.011095903813838959,
-0.4343453347682953,
-0.011353629641234875,
-0.32758304476737976,
-0.39763203263282776,
0.34626898169517517,
0.8438317179679871,
-0.1264040470123291,
0.106780551373... |
I want to run my entire WordPress site through a CDN to improve performance (HTML, not just static CSS/JS/media). Lets say the origin server is example.com and www.example.com is the URL where traffic will go through the CDN set-up as an origin-pull. The problem is an infinite loop will be created when visiting www.example.com which passes the request to WordPress which will look at the request and say "Actually this should be www.example.com according to the home_url and site_url defined in the Settings" Sooo... 1) Is there an easy way make WordPress stop redirecting to www under a certain condition (like checking for the x-forwarded-for headers which will be added to requests coming through the CDN)? 2) Am I going to have to set a constant to change the site URL for requests from the CDN and then set-up some sort of output buffering callback to replace http://example.com to http://www.example.com ? I know there are plugins out there that handle this. I'm looking to roll my own so I can figure out how this all works. Any guidance would be much appreciated. | [
-0.014873525127768517,
0.013670742511749268,
-0.005250401794910431,
0.031114734709262848,
-0.01675717532634735,
-0.01424499973654747,
0.008803103119134903,
-0.0007174412603490055,
-0.01639891415834427,
0.0078653609380126,
-0.00675031915307045,
0.008475825190544128,
0.0023945942521095276,
0... | [
0.5896646976470947,
0.11542104929685593,
0.3613930940628052,
-0.1348380744457245,
0.1662965565919876,
-0.028016995638608932,
0.2072760909795761,
0.4279184639453888,
-0.06898751109838486,
-0.5183291435241699,
-0.023373102769255638,
0.07935382425785065,
-0.5467166304588318,
0.412248611450195... |
I am trying to do field data collection using QGIS. One of the features that I regard as necessary for data management is timestamps. For example, a timestamp can indicate that an inspection record has been updated. I have figured out how to add triggers using Spatialite to add a timestamp on update, however the timestamp is added when the edited records are saved and not when the identify dialog box is closed. This means that timestamps can be out by many minutes. One solution would be to force the updated record to be saved by using a custom form with an OK button (Can do the form with Qt- Designer but don't know how to add the save command to the button yet.) However my preferred solution would be to take a timestamp for new records, when the point or line is created. The reason for this is if I am surveying a road, I can stand on the road, collect the point and them move to the side to complete data collection. I have been unable to work out how to get Spatialite to run a trigger that puts in the start of feature editing. Perhaps this is also something that needs to be done by code in a custom form. Is there an On-dirty event or signal in QGIS that indicates when someone starts editing a record and that can be used to capture the time? Could I also ask what the Python code (and what to hook it to) that would put such a timestamp in the timestamp field. AndrewM | [
0.006961206439882517,
0.004526819568127394,
-0.010452977381646633,
0.02383560873568058,
0.01926380954682827,
0.018165817484259605,
0.01034469809383154,
0.026948601007461548,
-0.01752854883670807,
0.031005417928099632,
-0.005234123673290014,
0.012656611390411854,
-0.013367816805839539,
0.01... | [
0.40276268124580383,
-0.09627573937177658,
0.5532802939414978,
-0.015598930418491364,
-0.025472193956375122,
-0.2571678161621094,
0.45187637209892273,
-0.25367265939712524,
-0.06305379420518875,
-0.7045204043388367,
0.13903489708900452,
0.11798328906297684,
-0.11103393882513046,
0.13060699... |
I'm working through K&R and I notice that their code is extremely tightly spaced. I haven't developed C since at University, but professionally most source from other languages I've worked in has used blank lines to aid readability. Is C generally more tightly packed or is it a historical matter related to display size in the 70s? Exercise 1-16 seems to hinder readability by, for example, an `if` immediately after `for` loop that doesn't have braces. /* mygetline: read a line into s, return length */ int mygetline(char s[], int lim) { int c,i; for (i=0; i<lim-1 && (c=getchar()) != EOF && c!='\n'; ++i) s[i] = c; if (c == '\n') { s[i] = c; ++i; } s[i] = '\0'; return i; } | [
-0.009438160806894302,
0.00465014111250639,
-0.015090649016201496,
0.01907046139240265,
-0.0037983949296176434,
0.004043016117066145,
0.007616193499416113,
-0.02447107993066311,
-0.013432338833808899,
-0.007356069982051849,
-0.012368280440568924,
0.004044584929943085,
0.004626012872904539,
... | [
0.2742713391780853,
-0.05463629215955734,
0.19211789965629578,
-0.17375870048999786,
-0.23552446067333221,
0.2767193019390106,
0.3084360361099243,
0.06272341310977936,
-0.2867477238178253,
-0.5550789833068848,
-0.11972054839134216,
0.016759468242526054,
-0.22810930013656616,
-0.06899034231... |
Could you please tell me whether I use _that of_ in an appropriate way or not? > Here are the results of the calculations. **That of** calculation number > four is pretty difficult to get. | [
-0.015991870313882828,
0.016653010621666908,
-0.05079318583011627,
0.034620486199855804,
-0.01729920320212841,
-0.01590343378484249,
0.015198701061308384,
-0.004647830035537481,
-0.041499678045511246,
-0.030287964269518852,
0.010165396146476269,
0.008729377761483192,
-0.01139730866998434,
... | [
0.3648253083229065,
0.3344752788543701,
0.16950954496860504,
-0.023242762312293053,
-0.11661087721586227,
0.2156316339969635,
0.15615807473659515,
-0.30535200238227844,
-0.12126606702804565,
-0.40964794158935547,
0.2458232045173645,
0.6982439756393433,
0.011232203803956509,
-0.179761156439... |
I code in Python and I'm starting to use the Mock library. Although I've read through all of the documentation, something that I'm finding hard to understand is 'What makes a good mock'. For example, say I've got a function that makes a HTTP request: def fetch_url(url, method=urlfetch.GET, data=''): """Send a HTTP request""" result = urlfetch.fetch(url=url, method=method, payload=data, headers={'Access-Control-Allow-Origin': '*'}) return result.content What would a good mock for this function look like? Furthermore, what would then be tested? From my understanding, the characteristics of a good test are that the test checks: 1. That the code isn't broken 2. That the code acts as expected 3. That the code returns the expected output if it returns something That said, what would be the characteristics of a good mock? | [
0.01058940589427948,
0.0004889813717454672,
-0.014352510683238506,
-0.0030557464342564344,
-0.026145819574594498,
-0.0033101935405284166,
0.005221965257078409,
0.009095117449760437,
-0.013744819909334183,
-0.013038057833909988,
-0.00033141812309622765,
0.004081808030605316,
0.005338330287486... | [
0.41643908619880676,
0.32654881477355957,
0.551643967628479,
-0.08152259141206741,
-0.15758861601352692,
0.04432595521211624,
0.27489590644836426,
-0.2548491954803467,
0.13440261781215668,
-0.7609049677848816,
0.13799816370010376,
0.3455493152141571,
-0.12799832224845886,
0.173907831311225... |
Suppose you shake water inside a container, then at first the waves goes up and down strongly but they gradually dissipate. What makes them dissipate? | [
-0.021305004134774208,
0.03676751255989075,
0.029115237295627594,
0.03997962549328804,
0.02272319793701172,
-0.029022540897130966,
0.015805630013346672,
-0.01844002678990364,
-0.018165115267038345,
0.04389415681362152,
-0.0070014288648962975,
0.019895492121577263,
-0.01374145969748497,
0.0... | [
0.32922232151031494,
0.01595163159072399,
0.11101572215557098,
0.16748186945915222,
0.05980491638183594,
0.05105717480182648,
0.11990194022655487,
-0.35594236850738525,
-0.3796101212501526,
0.021142950281500816,
0.16274434328079224,
0.05314581096172333,
-0.06113610044121742,
0.522279679775... |
So I'm reading this text on Quantum Mechanics, and it goes through a few chapters that I understand fairly well including probability. But then it says that all quantities, like position and energy of an object, are represented in a matrix, and that quantities have associated probability distributions. I kind of get this, although I'm a little unclear about whether we're talking about full m-by-n matrices or just vectors. If it's vectors, yeah, I'm sort of familiar with that. But if not, how do you use a full m-by-n matrix to represent a quantity? And then further along, it says that is the mean of a matrix, but doesn't say what that is. Is it the average of all the coordinates in the matrix, so it's $\displaystyle \frac{1}{mn}\sum_{(i,j)\in \ulcorner m \urcorner \times \ulcorner n \urcorner}a_{ij}$? Or are each of the columns supposed to represent separate quantities and then I guess the mean is a vector of the means of the columns? The only guidance the text gives in this regard is "Some of the basic rules of quantum mechanics involve simple relations between quantities, expressed in terms of matrices, and corresponding relations between mean values. Consider a quantity represented by a matrix $M$. Let $<M>$ denote its mean value. For any number $z$, the matrix $zM$ represents the original quantity multiplied by $z$. Its mean value is $<zM>=z<M>$." And so on. But nowhere does it define the mean of a matrix, it just jumps into this notation. Some quick websearching showed that there doesn't seem to be any consensus on what is meant by the mean of a matrix representing a quantity. | [
0.014216195791959763,
0.008216758258640766,
0.00927172414958477,
0.003988220356404781,
0.013852018862962723,
-0.009511185809969902,
0.004095891490578651,
0.019042305648326874,
-0.012442718259990215,
-0.012281511910259724,
-0.007916577160358429,
0.01414412446320057,
0.009620501659810543,
-0... | [
0.5632452964782715,
0.048355571925640106,
0.01768333837389946,
0.18613889813423157,
-0.11881180852651596,
0.04517342150211334,
-0.3746986985206604,
0.034134041517972946,
-0.5407202839851379,
-0.3782218098640442,
0.07103231549263,
0.1948363035917282,
-0.1448453664779663,
0.41924092173576355... |
I'm pretty sure I've read "Ad question one..." in the sense of "Addressing question one..." before, but the only usage on the web I find is by someone with a Hispanic name and m-w.com doesn't seem to know it. I've always thought it was derived from a Latin expression and (therefore) sophisticated. | [
-0.004875154234468937,
0.01337088830769062,
-0.014731258153915405,
0.009873797185719013,
-0.002870051423087716,
0.008238199166953564,
0.00878980290144682,
0.025300946086645126,
-0.020612867549061775,
0.02185639925301075,
-0.00922616571187973,
0.008059140294790268,
0.010652046650648117,
0.0... | [
0.3703400194644928,
0.3306809663772583,
0.20567508041858673,
-0.0976567417383194,
-0.4617552161216736,
-0.09886053949594498,
0.052003826946020126,
0.4684236943721771,
-0.078120656311512,
-0.1420273780822754,
-0.06594474613666534,
0.2522035241127014,
-0.07509928941726685,
0.2250491529703140... |
 This is the equation about zeta. `NSolve[eq]` could not solve this equation. (I think it is too complex to solve by `NSolve`.) I want to know how can I solve this equation, which has such a complicated form. This should be the code for the equation: eq = (0. - 30.1996 I) ζ^6 + (0. + 9.81486 I) ζ^5 Sqrt[(0. - 640.701 I) + ζ^2] + (0. + 20.3847 I) ζ^4 ((0. - 640.701 I) + ζ^2) + (0. + 20.3847 I) ζ^5 Sqrt[(0. - 1127.53 I) + ζ^2] - (0. + 20.3847 I) ζ^3 ((0. - 640.701 I) + ζ^2) Sqrt[(0. - 1127.53 I) + ζ^2] + (0. + 9.81486 I) ζ^4 ((0. - 1127.53 I) + ζ^2) - (0. + 9.81486 I) ζ^3 Sqrt[(0. - 640.701 I) + ζ^2] ((0. - 1127.53 I) + ζ^2) + 408.549 ( (0. - 0.630868 I) ζ^3 - (0. + 14.0909 I) ζ^2 Sqrt[(0. - 640.701 I) + ζ^2] - (0. + 1.21667 I) ζ ((0. - 640.701 I) + ζ^2) - (0. + 6.45738 I) ζ^2 Sqrt[(0. - 1127.53 I) + ζ^2] + (0. + 14.7217 I) ((0. - 640.701 I) + ζ^2) Sqrt[(0. - 1127.53 I) + ζ^2] + (0. + 0.585806 I) ζ ((0. - 1127.53 I) + ζ^2) + (0. + 7.08825 I) Sqrt[(0. - 640.701 I) + ζ^2] ((0. - 1127.53 I) + ζ^2)) == 0; | [
0.009078082628548145,
-0.003116664942353964,
-0.027060333639383316,
0.0010320654837414622,
-0.010553155094385147,
0.0008391784504055977,
0.006170577369630337,
-0.00892688985913992,
-0.017024792730808258,
-0.008148716762661934,
-0.0019684431608766317,
0.006506573408842087,
-0.0251947604119777... | [
-0.14455056190490723,
0.24292504787445068,
0.8616710305213928,
-0.20897023379802704,
-0.15553466975688934,
-0.11414714902639389,
0.3228152096271515,
-0.40823787450790405,
-0.1910533607006073,
-0.33264413475990295,
-0.05013477802276611,
0.6985030174255371,
-0.15204869210720062,
0.4709564447... |
  I'd like to calculate x value in this equation. Basically, I tried to 2 types of method which are FindRoot and NSolve. But, I have failed the calculation caused by these errors up to now. If there is anyone who knows this problems, plz let me know what I should do first! Thank you for your cooperation. | [
0.008232960477471352,
0.0037451665848493576,
-0.0010107093257829547,
0.01065768115222454,
-0.019616026431322098,
-0.008871166966855526,
0.004271218553185463,
0.0066635021939873695,
-0.019772207364439964,
0.0018850727938115597,
-0.006739161908626556,
-0.005619712173938751,
-0.0244715735316276... | [
0.12223974615335464,
0.13572710752487183,
0.5069092512130737,
0.09706411510705948,
-0.3211570382118225,
-0.15013031661510468,
0.5745123028755188,
-0.030456479638814926,
-0.3731299936771393,
-0.4948665201663971,
0.04082879796624184,
0.6587976217269897,
0.0477592907845974,
0.0344655700027942... |
What are the origins and differences between these two? Same for Granddad/Grandpa? Why was there the need for the two different names? | [
-0.015333710238337517,
0.06120071932673454,
0.010866057127714157,
0.03462231159210205,
-0.048471368849277496,
0.10893131792545319,
0.019072089344263077,
0.08444236218929291,
-0.03472878783941269,
-0.012651043012738228,
0.011407347396016121,
0.031777527183294296,
0.053083084523677826,
-0.04... | [
0.20294158160686493,
0.3563189208507538,
-0.03221496194601059,
0.07883454114198685,
0.25608882308006287,
-0.20696237683296204,
0.3440788984298706,
0.012914793565869331,
-0.5018880367279053,
-0.40815824270248413,
0.32413962483406067,
-0.23642398416996002,
0.055833712220191956,
0.93663614988... |
I am trying to simply run a command: htlatex texfilename "xhtml,charset=utf-8" In a terminal command line, it runs well. If I create a shell script file that looks like: #!/bin/bash htlatex texfilename "xhtml,charset=utf-8" It produces the following: This is pdfTeX, Version 3.1415926-2.4-1.40.13 (TeX Live 2012) restricted \write18 enabled. entering extended mode LaTeX2e <2011/06/27> Babel <v3.8m> and hyphenation patterns for english, dumylang, nohyphenation, ge rman-x-2012-05-30, ngerman-x-2012-05-30, afrikaans, ancientgreek, ibycus, arabi c, armenian, basque, bulgarian, catalan, pinyin, coptic, croatian, czech, danis h, dutch, ukenglish, usenglishmax, esperanto, estonian, ethiopic, farsi, finnis h, french, friulan, galician, german, ngerman, swissgerman, monogreek, greek, h ungarian, icelandic, assamese, bengali, gujarati, hindi, kannada, malayalam, ma rathi, oriya, panjabi, tamil, telugu, indonesian, interlingua, irish, italian, kurmanji, latin, latvian, lithuanian, mongolian, mongolianlmc, bokmal, nynorsk, polish, portuguese, romanian, romansh, russian, sanskrit, serbian, serbianc, s lovak, slovenian, spanish, swedish, turkish, turkmen, ukrainian, uppersorbian, welsh, loaded. * and stops. How can I make this work? Should I give some options to `htlatex`? I am using a Mac. | [
-0.01422665175050497,
0.02089821547269821,
-0.005409782286733389,
0.006911979988217354,
-0.00099729816429317,
-0.008883706294000149,
0.009261916391551495,
0.0019366078777238727,
-0.018501147627830505,
-0.04901831969618797,
-0.015072774142026901,
-0.007466227747499943,
-0.01210317574441433,
... | [
0.3174861669540405,
0.18565118312835693,
0.6034483909606934,
-0.4138801395893097,
0.022331004962325096,
-0.18512828648090363,
0.45033738017082214,
-0.09205672144889832,
0.15316730737686157,
-0.4286462664604187,
-0.07190553843975067,
0.6597101092338562,
-0.3886275887489319,
0.01434574369341... |
A basketball is thrown straight upwards so that it reaches the bottom of the rim (2.00 m above the launch point) 0.50 s after it is thrown. What is the maximum height above the launch point which the ball reaches? Now, I use the motion equations but the answer is not as desired. It results for me a negative value for displacement. What I did was assumed initial velocity = 0, got acceleration as 16 m/s^2. And then put these in the third motion equation. But I didn't quite get a right displacement value. Any help appreciated. | [
-0.0022006432991474867,
0.007735809776932001,
-0.009010848589241505,
0.0038186905439943075,
-0.014956298284232616,
-0.02764030173420906,
0.008710436522960663,
-0.014160188846290112,
-0.018169403076171875,
-0.031645189970731735,
-0.007958403788506985,
0.012114424258470535,
-0.0127048026770353... | [
0.09374717622995377,
0.33616387844085693,
0.2283530980348587,
-0.022637225687503815,
-0.44275835156440735,
0.16685999929904938,
0.32375097274780273,
-0.109755739569664,
-0.6816841959953308,
-0.4182620644569397,
0.2725447714328766,
0.4249912202358246,
0.1591360718011856,
-0.0033636491280049... |
Does anyone know how to turn off the auto-complete feature when texting? Sometimes I just want to text and not have my device try and tell me what I'm thinking. Other times this is OK, but I just want to turn off the autocomplete sometimes. I looked in the phone settings (not very hard I guess...) and I didn't really see anything to allow this. Anyone know how this is done? **EDIT:** I am using a Droid X. | [
0.02374524623155594,
-0.01604202203452587,
-0.02393665909767151,
0.0009610580746084452,
-0.005530527792870998,
0.003796411445364356,
0.006577021908015013,
0.02344195544719696,
-0.029903411865234375,
0.01734231784939766,
-0.006435451563447714,
0.007599099073559046,
0.002209323225542903,
0.0... | [
0.6091210842132568,
0.3775733709335327,
0.27066367864608765,
0.14350907504558563,
0.035853225737810135,
-0.372448593378067,
0.4213106334209442,
0.6116981506347656,
-0.027755554765462875,
-0.5093690752983093,
0.11664304882287979,
0.5116207599639893,
-0.2673681676387787,
-0.00585581967607140... |
My issue is that I have a horizontal menu, with a few hashed - '#' - custom menu links to prevent users from going to empty pages (the main categories were just made pages for the link structure). Now the issue further is that in my attempt to create a select dropdown menu for mobile, I am using logic to see if a nav item is a child to another nav item, but my custom nav items are not being registered as parents. Is there a work around to this? To be specific: I need to find out which nav items are children and add dashes to them for a better user experience: ex: parent - child -- 2 child | [
0.0056177289225161076,
0.0025418391451239586,
0.005832277704030275,
0.017255151644349098,
-0.005143588408827782,
-0.01406959630548954,
0.006552300415933132,
0.015703372657299042,
-0.015617826953530312,
0.0011970531195402145,
-0.011302840895950794,
0.008769195526838303,
0.00534479133784771,
... | [
-0.062006741762161255,
0.21405453979969025,
0.1091599240899086,
0.11241655051708221,
0.08726409822702408,
0.24029667675495148,
0.12949472665786743,
0.27200785279273987,
-0.3631092607975006,
-0.4024180769920349,
-0.0643693059682846,
0.1788475662469864,
-0.2753708064556122,
0.341757595539093... |
I've been looking into getting a specific type for Hidden Power, but the best I can find is the (now outdated) Smogon Hidden Power Guide. The Sixth Generation of Pokemon has introduced the Fairy type, throwing the value-table out of whack. So, what are the values needed to get particular types? | [
-0.0011702969204634428,
0.01510987151414156,
-0.0018078488064929843,
0.01651042141020298,
0.01114971749484539,
-0.015510614030063152,
0.009686564095318317,
0.029047563672065735,
-0.04466388747096062,
0.013011869974434376,
-0.0029737299773842096,
0.018006926402449608,
-0.02452869899570942,
... | [
0.5777920484542847,
-0.36617711186408997,
0.2058672159910202,
0.19522127509117126,
-0.3212651014328003,
-0.2965475022792816,
0.4438439607620239,
0.1058938279747963,
0.036231059581041336,
-0.04801076650619507,
0.08848626911640167,
0.12977327406406403,
-0.08509226888418198,
0.302798062562942... |
Suppose that a cord is wrapped around the rim a disk of radius $R$. The disk is allowed to rotate around its central axis (the line passing through the center and perpendicular to the disk surface). The force from the cord is $F$. Then I am told that the magnitude of torque on the disk is $RF$. _I could not understand how $RF$ follows from the definition net torque $T= \sum \vec r_i \times \vec F_i$_ when the sum is taken over **all** particle. Things become more confusing as I notice that the force $F_i$ on any single particle of the object must not be zero, because each particle is rotating together with the rigid object. Any help is appreciated. Additional Info: The fact that net $T=\sum \vec r_i \times \vec F_i$ is used, for example, in the proof of Newton's second law for rotation $T = I \varepsilon$. The proof (as far as I know) proceeds from the case of a single particle and then generalizes to rigid objects by considering an object as being a combination of many particles. | [
-0.007286508101969957,
-0.0005186977796256542,
-0.004254491068422794,
0.011434740386903286,
-0.01224227249622345,
-0.008878080174326897,
0.007100195623934269,
-0.012847905978560448,
-0.010878115892410278,
-0.0005679531604982913,
-0.008073161356151104,
0.00923813134431839,
-0.0159790124744176... | [
0.08397237956523895,
-0.18977580964565277,
0.9408321380615234,
0.11859408766031265,
-0.08023473620414734,
0.17495283484458923,
-0.2702111005783081,
-0.44739535450935364,
-0.2107720822095871,
-0.25386542081832886,
-0.12806077301502228,
0.8185940384864807,
0.12812680006027222,
0.571811616420... |
In continuation of this question, I tried to make a list of items each consisting of unbreakable text. This can be done manually using minipage: \documentclass[10pt,a4paper]{article} \usepackage{lipsum} %for example \setlength{\textheight}{7cm} %about one item fits per page \begin{document} \begin{itemize} \item\begin{minipage}[t]{\linewidth} \lipsum[1] \end{minipage} \item\begin{minipage}[t]{\linewidth} \lipsum[2] \end{minipage} \end{itemize} \end{document} What I would like it to to define a list environment that is just like itemize (or whatever) only that the items are automatically inserted into a minipage as in my manual example. The solution should be robust for nested list environments, so if `\item` is to be redifined, it has to be done so in a very gentle way...not sure how. After redefining the environment itemize-block (for example) the following code should compile and not break the text in each item across pages: \begin{itemize-block} \item \lipsum[1] \item \begin{itemize} \item let us make sure it compiles even with \item a nested list... \end{itemize} \item \lipsum[2] \end{itemize-block} This can surely be done...only I have no idea how. | [
0.030788984149694443,
0.0028260969556868076,
-0.0028361412696540356,
0.01055485662072897,
0.0016917928587645292,
0.016270240768790245,
0.00810585729777813,
0.029204197227954865,
-0.01346602663397789,
0.0030850814655423164,
-0.009982507675886154,
-0.0037528821267187595,
-0.0071303704753518105... | [
-0.125202476978302,
-0.233184814453125,
0.5156087875366211,
0.20791412889957428,
0.162796288728714,
0.11656274646520615,
0.5442104339599609,
-0.2178041636943817,
-0.2749226689338684,
-0.6931063532829285,
-0.36042535305023193,
0.26852577924728394,
-0.11856406927108765,
0.10803605616092682,
... |
I have a fairly large dataset with a few fields containing time-related data. This data comes in various shapes and sizes, but most of it can be parsed and rephrased in more appropriate formats for human-reading. Formats range from `2010-01-01` to less obvious like `2008-52` (the second number is the number of editions a magazine had -- in this case, 1 for each week of the year). Now I trying to transform it features for crunching. I first I thought it make sense to calculate the "age", i.e., the time difference between `now()` and the timestamp I have in my data, and later normalize it. I know this works (computationally speaking), but I don't know if is the best approach or produce anything useful. Is this approach useful? Is there a better way to handle "date" features? | [
-0.002179389353841543,
0.01299512479454279,
-0.019768379628658295,
0.00973595678806305,
0.012351571582257748,
0.020746631547808647,
0.006481743883341551,
0.00823751837015152,
-0.011552048847079277,
-0.004813046660274267,
-0.001341142226010561,
0.00022564362734556198,
0.005010704509913921,
... | [
0.3290095627307892,
0.39416536688804626,
0.19244544208049774,
0.05921578407287598,
0.17074716091156006,
0.05789684131741524,
-0.15906628966331482,
0.24397075176239014,
-0.3323182165622711,
-0.452417254447937,
0.10552900284528732,
0.045093782246112823,
0.5615193843841553,
0.5596086978912354... |
I have created my thesis using `fncychap` to customize the chapters. There are also chapters starred created using `chapter*{}`. I would add an `hrule` in those chapters. As the user _mafp_ suggested me, it is possible using `titlesec` package. Unfortunately those two packages enter in conflict. Here it is the previosly topic Add a hrule to only chapter not numbered Here it is **MWE** : \documentclass[12pt,a4paper,oneside]{book} \usepackage[utf8]{inputenc} \usepackage{lipsum} \usepackage[margin=3cm,bindingoffset=0.5cm]{geometry} \usepackage[Lenny]{fncychap} \begin{document} \chapter*{Intro} \lipsum[1-3] \chapter{My Chapter} \lipsum[1-3] \end{document} I would add hrule to the starred chapter Intro without affecting My chapter | [
-0.0003877475392073393,
0.01861545816063881,
-0.005974219646304846,
0.02477961778640747,
0.025216611102223396,
-0.0012587769888341427,
0.009619692340493202,
0.0066681113094091415,
-0.017117472365498543,
-0.010478844866156578,
-0.019298002123832703,
0.0070319101214408875,
-0.02681171521544456... | [
0.10299740731716156,
0.17061690986156464,
0.7414000034332275,
-0.0788009986281395,
0.10077017545700073,
0.04856041818857193,
0.020692972466349602,
-0.3010950982570648,
-0.07723889499902725,
-0.5401055812835693,
-0.03218190744519234,
0.3055735230445862,
0.15083958208560944,
0.38071686029434... |
I have a directory containing about 7,000 music files. I used lame to recursively re-encode all files in it to a separate directory, outputting all files with the same relative path and file name. The output files have a .mp3 extension, but some of the input files had different extensions (.wma, .aac, etc). I can see that there is a file count difference of ~100 files missing in the output directory. What I want to do is run a compare of the two directories and obtain a list of the files that exist in the source, but not in the destination. This would be simple enough except I need to ignore differences in file extension. I've tried using rsync with dry-run turned on but I couldn't figure out a way to ignore file extensions. I've also tried diff but was unable to find an option to only check by name but ignore file extensions. I started thinking I could just do a recursive ls on both directories, remove the file extensions, and then compare the outputs, but I really have no idea on where to start with modifying the ls output using sed or awk since I'm a compete noob with those. | [
0.010911139659583569,
0.018698517233133316,
-0.008021293208003044,
0.009103124961256981,
0.00016706599853932858,
-0.00876513123512268,
0.008363369852304459,
0.008886313997209072,
-0.014604938216507435,
-0.018614593893289566,
-0.006325472146272659,
-0.004709350876510143,
0.012861525639891624,... | [
0.3270878493785858,
0.3350118398666382,
0.1946251392364502,
0.21556983888149261,
0.11133091896772385,
0.14455176889896393,
-0.02492932230234146,
-0.3592674732208252,
-0.37884384393692017,
-0.40350085496902466,
0.129526287317276,
0.5831588506698608,
-0.07501398772001266,
0.3792295753955841,... |
I would like to start/stop GrooveIP when I connect/disconnect from my home wifi. I searched the manifest file (using DexPlorer from the play store) and found these intents. > > <receiver name=".utilities.PhoneStart" > > <intent-filter > > <action name="android.intent.action.BOOT_COMPLETED" /> > <category name="android.intent.category.HOME" /> > </intent-filter /> > <intent-filter > > <action name="com.gvoip.START_GROOVE_IP" /> > <action name="com.gvoip.STOP_GROOVE_IP" /> > <action name="com.gvoip.CHECK_START_GROOVE_IP" /> > <action name="com.gvoip.CHECK_STOP_GROOVE_IP" /> > </intent-filter> > I set up a new task with a "Send Intent" and insert the action name for "Action" but I leave everything else default. None of the tasks seem to start or stop GrooveIP. And how would CHECK_START_GROOVE_IP work? Wouldn't that have to return a value? As for now, I am starting GrooveIP just by launching it, then another task that returns to the home screen, but that isn't as clean as I would like. I've tried doing the "Kill App" task to stop GrooveIP, but that doesn't work. Any suggestions? Thanks | [
0.004493539221584797,
0.003784989705309272,
0.01240126695483923,
0.01414896547794342,
-0.03373216465115547,
0.006511767860502005,
0.008367784321308136,
0.017936933785676956,
-0.01611400581896305,
0.026463612914085388,
-0.012053142301738262,
0.015990089625120163,
-0.0005839706864207983,
0.0... | [
-0.07279942184686661,
-0.14877629280090332,
0.6581634879112244,
-0.2593785524368286,
-0.08873847126960754,
-0.1433260142803192,
0.32060664892196655,
0.16176119446754456,
-0.08329026401042938,
-0.7519609928131104,
0.12683846056461334,
0.5225143432617188,
-0.5202032327651978,
-0.175758406519... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.