qid int64 1 74.7M | question stringlengths 12 33.8k | date stringlengths 10 10 | metadata list | response_j stringlengths 0 115k | response_k stringlengths 2 98.3k |
|---|---|---|---|---|---|
403,997 | I sometimes encounter explanations of phenomena where only entropy is quoted as a driver, for example, osmosis, diffusion, hydrophobic effect... But can entropy ever be the sole driver of a process? I encountered 'entropic force', is it a real force? Aren't energy changes always present in these processes?
If this the case, when it comes to concentration gradiënts, these gradients 'contain' potential energy but I can't figure out where it originates from. For example, a water molecule that is a part of the cage around a lipophilic molecule is missing out on potential hydrogen bonding and thus feels a net force which is opposed. This missing out on stable interactions, feeling a force which is opposed...is what 'creates' potential energy. I can't see these things in diffusion and osmosis for neutral solutes (when the solute has charge, the electrostatic force comes into play - electrochemical potential - this feels more natural to consider as potential energy) which seem just to be random mixing... but in the case of osmosis the gradiënt is able to do work and create a hydrostatic pressure difference which proves the potential energy there since it can be converted into another subtype? The semi-permeable membrane seems essential in these phenomena for work to be done (osmosis, transmembrane potential generation in biology)? Free diffusion in bulk solution without membranes could never do work and all potential energy would be dissipated as heat?
The role of charge in all of this...
Osmosis is a colligative property, so the charge of the solute doesn't actually matter.
But when the membrane is only permeable to the solute, a chemical gradiënt can store electric potential energy over the membrane. The chemical gradiënt does work on the charged particles by pushing them against their electric potential. But what causes this push against the electric field, the difference in concentration? But this seems more of a probabilistic factor.The electric field over de membrane does negative work and the concentration gradiënt positive work?
Coming back to osmosis, the solutes gradiënt does positive work and the hydrostatic pressure performs negative work. So charge doesn't play a part after all?
What happens when equilibrium is reached in all of these cases? Zero work?
Thank you | 2018/05/05 | [
"https://physics.stackexchange.com/questions/403997",
"https://physics.stackexchange.com",
"https://physics.stackexchange.com/users/194664/"
] | As I see it, the driver behind all the system processes you describe involves disequilibrium, i.e., pressure, temperature, concentration etc. differentials or gradients; and the achievement of equilibrium as a result of the process results in entropy production, that is, an overall increase in entropy (system plus surroundings) as occurs with all real (irreversible) processes. So in that sense, one can think of entropy production as a driver of processes. | All these processes are due to random events. The atoms do not "feel" a concentration gradient, but random walks will cause diffusion to even out differences in concentration. |
3,822,647 | here is my scenario:
* DHCPD machine (I cannot edit DHCPD settings)
* Machine 1
+ Boots a Live GNU/Linux and offers net-boot with TFTPD
* Machine 2
+ Tries to net-boot, but ends with PXE-E53 error
If I run DHCPD on Machine 1, everything is fine, 'cos I can setup all I need.
How could I setup a PXE environment without affecting DHCPD Machine settings?
Thanks,
hamen | 2010/09/29 | [
"https://Stackoverflow.com/questions/3822647",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/108742/"
] | I think you can as you need a DHCP server to redirect the "call" to the proper TFTP server and thus to the install or OS files you want.
You need a specific DHCP server where *filename "linux-install/pxelinux.0";* is properly defined. Otherwise your Machine 2 will not be able to find the necessary files and thus your PXE boot will fail.
Have a look a this (http://idoitonamac.blogspot.com/2012/03/os-x-lion-as-pxe-server.html). I now it's Mac, but you'll see that I managed to have a PXE server on one machine. | For the case where there is already in place a DHCP startegy PXE relies on the services of an extra "proxyDHCP server" that does not assign IP addresses but only provides the PXE parameters (bootstrap file name, and TFTP IP address) to booting stations that identify themself as PXE clients. |
85,873 | Inkscape has the simplify tool, which removes some nodes along the path and makes it smooth. Is there something like simplify, but which instead makes the path more edgy? The path is traced bitmap, so doing so by hand is impossible.
For comparison, this is a bitmap traced and post-processed as vector (probably in Photoshop and then Illustrator), which keeps the edges of the color patches and gives it some energy:
[](https://i.stack.imgur.com/8P5VV.png)
while this is my attempt (with a different image) to produce similar effect in Inkscape, where, however, every smoothing from the trace leads to rounded corners and theresult feels blunt:
[](https://i.stack.imgur.com/wa25j.png)
Ideas? | 2017/02/27 | [
"https://graphicdesign.stackexchange.com/questions/85873",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/63147/"
] | Select all nodes within a path and change node type to Corner. That will make the path into a polygon like in your first image.[](https://i.stack.imgur.com/Alchm.png) | Maybe the Fractalize extension might do what you want... or it might be way too strong of an effect! It's on the Extensions menu under Modify Path. |
48,977 | I tried for awhile to use the SOFLAM to laser target air vehicles, but no one else on the team was hitting them with javelins or laser guided tanks. (Same problem when I jump in a tank in the CITV station).
Is there any way to use the SOFLAM individually to destroy enemy vehicles? | 2012/01/30 | [
"https://gaming.stackexchange.com/questions/48977",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/7/"
] | No not really but you could switch kits. A lot of people forget how useful switching kits can really be. I sometimes set up a SOFLAM as Recon and leave it to designate targets automatically while I find myself a javalin. | No, there really isn't. If no one is using your targetting, you're better off doing something else with your time. The piddling amount of points you get for laser designation really isn't worth it.
I do agree it is a bit deflating. I especially hate being in the CITV station because it's so hard to coordinate with a regular tank driver to have guided missiles. |
48,977 | I tried for awhile to use the SOFLAM to laser target air vehicles, but no one else on the team was hitting them with javelins or laser guided tanks. (Same problem when I jump in a tank in the CITV station).
Is there any way to use the SOFLAM individually to destroy enemy vehicles? | 2012/01/30 | [
"https://gaming.stackexchange.com/questions/48977",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/7/"
] | Yes, but it isn't too useful:
Set up your SOFLAM, aim it in a useful direction, and then disconnect. Now jump into a vehicle with the Guided Shell or Guided Missile unlock. Wait until an enemy vehicle blunders through the area the SOFLAM is sitting and watching.
If you want help from teammates, you should always **announce** that you are making laser-designations, because *most players can't see them*. Only players who are already carrying laser-related equipment can see the diamond-indicator. Your engineers who have SMAWs probably don't even know that there's a SOFLAM, so they won't bother to change to Javelin. | No, there really isn't. If no one is using your targetting, you're better off doing something else with your time. The piddling amount of points you get for laser designation really isn't worth it.
I do agree it is a bit deflating. I especially hate being in the CITV station because it's so hard to coordinate with a regular tank driver to have guided missiles. |
48,977 | I tried for awhile to use the SOFLAM to laser target air vehicles, but no one else on the team was hitting them with javelins or laser guided tanks. (Same problem when I jump in a tank in the CITV station).
Is there any way to use the SOFLAM individually to destroy enemy vehicles? | 2012/01/30 | [
"https://gaming.stackexchange.com/questions/48977",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/7/"
] | No not really but you could switch kits. A lot of people forget how useful switching kits can really be. I sometimes set up a SOFLAM as Recon and leave it to designate targets automatically while I find myself a javalin. | Using the SOFLAM isn't really worth all that many points anyway, but if you can find a good squad they will be glad to spawn with a javalin and make use of your targeting. The SOFLAM is very powerful and can basically take out any target with ease with two + coordinated players so DICE made it a little tedious to use properly. If you could just drop a SOFLAM die and then spawn as an engineer then there would be no helicopters planes or tanks alive in the entire game. It is very satisfying to designate a pesky chopper and watch it have no chance of evading your missle haha. |
48,977 | I tried for awhile to use the SOFLAM to laser target air vehicles, but no one else on the team was hitting them with javelins or laser guided tanks. (Same problem when I jump in a tank in the CITV station).
Is there any way to use the SOFLAM individually to destroy enemy vehicles? | 2012/01/30 | [
"https://gaming.stackexchange.com/questions/48977",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/7/"
] | No not really but you could switch kits. A lot of people forget how useful switching kits can really be. I sometimes set up a SOFLAM as Recon and leave it to designate targets automatically while I find myself a javalin. | Yes, but it isn't too useful:
Set up your SOFLAM, aim it in a useful direction, and then disconnect. Now jump into a vehicle with the Guided Shell or Guided Missile unlock. Wait until an enemy vehicle blunders through the area the SOFLAM is sitting and watching.
If you want help from teammates, you should always **announce** that you are making laser-designations, because *most players can't see them*. Only players who are already carrying laser-related equipment can see the diamond-indicator. Your engineers who have SMAWs probably don't even know that there's a SOFLAM, so they won't bother to change to Javelin. |
48,977 | I tried for awhile to use the SOFLAM to laser target air vehicles, but no one else on the team was hitting them with javelins or laser guided tanks. (Same problem when I jump in a tank in the CITV station).
Is there any way to use the SOFLAM individually to destroy enemy vehicles? | 2012/01/30 | [
"https://gaming.stackexchange.com/questions/48977",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/7/"
] | Yes, but it isn't too useful:
Set up your SOFLAM, aim it in a useful direction, and then disconnect. Now jump into a vehicle with the Guided Shell or Guided Missile unlock. Wait until an enemy vehicle blunders through the area the SOFLAM is sitting and watching.
If you want help from teammates, you should always **announce** that you are making laser-designations, because *most players can't see them*. Only players who are already carrying laser-related equipment can see the diamond-indicator. Your engineers who have SMAWs probably don't even know that there's a SOFLAM, so they won't bother to change to Javelin. | Using the SOFLAM isn't really worth all that many points anyway, but if you can find a good squad they will be glad to spawn with a javalin and make use of your targeting. The SOFLAM is very powerful and can basically take out any target with ease with two + coordinated players so DICE made it a little tedious to use properly. If you could just drop a SOFLAM die and then spawn as an engineer then there would be no helicopters planes or tanks alive in the entire game. It is very satisfying to designate a pesky chopper and watch it have no chance of evading your missle haha. |
22,487,388 | I'm working on something that sends data from one program over UDP to another program at a known IP and port. The program at the known IP and port receives the message from the originating IP but thanks to the NAT the port is obscured (to something like 30129). The program at the known IP and port wants to send an acknowledgement and/or info to the querying program. It can send it back to the original IP and the obscured port #. But how will the querying program know what port to monitor to get it back on? Or, is there a way (this is Python) to say "send this out over port 3200 to known IP (1.2.3.4) on port 7000? That way the known IP/port can respond to port 30129, but it'll get redirected to 3200, which the querying program knows to monitor. Any help appreciated. And no, TCP is not an option. | 2014/03/18 | [
"https://Stackoverflow.com/questions/22487388",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/867731/"
] | Okay, I figured it out - the trick is to use the same sock object to receive that you used to send. At least in initial experiments, that seems to do the trick. Thanks for your help. | The simple answer is you don't care what the "real" (ie: pre-natted) port is. Just reply to the nat query and allow the nat to handling delivering the result. If you ABSOLUTELY have to know the source UDP port, include the information in your UDP packet -- but I strongly recommend against this. |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | If you click on the list of invitees, there should be an "Add Invitees" field which appears below the list.
 | I use an Exchange account and iCloud, and I notice my iCloud events remain editable while my Exchange events get "locked" after creation. There is no way to click the list. After the event is created, all the editing functions (changing time, changing notes, adding invitees, etc) are not clickable to edit. |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | If you click on the list of invitees, there should be an "Add Invitees" field which appears below the list.
 | I did eventually find a solution: I logged into the external calendar via browser (in my case, exchange/Outlook) and could edit there. Apple 'Calendar' (formerly iCal) would create the event for exchange, but after a step or two could not edit, or even cancel with notifications. (Again, I would reply to the earlier solution not working, but I can't add comment there.) |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I solved the issue of not being able to add an invitee by exporting all my appointments and creating a new iCloud Calendar.
* go to **File** > **New Calendar** > **iCloud**
* create a new calendar
* go to **File** > **Export** > **Export...**
* export the old calendar
* go to **File** > **Import...**
* select the file to which you exported the old calendar
* click **Import**
* select the newly added calendar
* click **Ok**
* delete the old calendar | If you click on the list of invitees, there should be an "Add Invitees" field which appears below the list.
 |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I think I figured out the reason. I found out this statement in Apple support page (<https://support.apple.com/guide/calendar/if-you-cant-change-a-calendar-or-event-on-mac-icl1033/mac>):
>
> If you can’t change an event you created, or you can’t change your
> status for an event you were invited to, it might be because you’re
> using an email address in Calendar that isn’t on your card in
> Contacts. Make sure all your email addresses are listed on your
> Contacts card. See Edit contacts.
>
>
>
I decided to completely remove my Exchange account from the Internet Accounts on my mac. Then when signing again, I used as an email address my full email address: firstname.lastname@company.com and when signing to my account (via Microsoft web login) we need to use the shorter version of the email (in our company): username@company.com. It seems the problem lied in that I used short version of the email for the default email address in my calendar, which caused some permission issues etc. Hope this helps someone. | I use an Exchange account and iCloud, and I notice my iCloud events remain editable while my Exchange events get "locked" after creation. There is no way to click the list. After the event is created, all the editing functions (changing time, changing notes, adding invitees, etc) are not clickable to edit. |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I solved the issue of not being able to add an invitee by exporting all my appointments and creating a new iCloud Calendar.
* go to **File** > **New Calendar** > **iCloud**
* create a new calendar
* go to **File** > **Export** > **Export...**
* export the old calendar
* go to **File** > **Import...**
* select the file to which you exported the old calendar
* click **Import**
* select the newly added calendar
* click **Ok**
* delete the old calendar | I use an Exchange account and iCloud, and I notice my iCloud events remain editable while my Exchange events get "locked" after creation. There is no way to click the list. After the event is created, all the editing functions (changing time, changing notes, adding invitees, etc) are not clickable to edit. |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I think I figured out the reason. I found out this statement in Apple support page (<https://support.apple.com/guide/calendar/if-you-cant-change-a-calendar-or-event-on-mac-icl1033/mac>):
>
> If you can’t change an event you created, or you can’t change your
> status for an event you were invited to, it might be because you’re
> using an email address in Calendar that isn’t on your card in
> Contacts. Make sure all your email addresses are listed on your
> Contacts card. See Edit contacts.
>
>
>
I decided to completely remove my Exchange account from the Internet Accounts on my mac. Then when signing again, I used as an email address my full email address: firstname.lastname@company.com and when signing to my account (via Microsoft web login) we need to use the shorter version of the email (in our company): username@company.com. It seems the problem lied in that I used short version of the email for the default email address in my calendar, which caused some permission issues etc. Hope this helps someone. | I did eventually find a solution: I logged into the external calendar via browser (in my case, exchange/Outlook) and could edit there. Apple 'Calendar' (formerly iCal) would create the event for exchange, but after a step or two could not edit, or even cancel with notifications. (Again, I would reply to the earlier solution not working, but I can't add comment there.) |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I solved the issue of not being able to add an invitee by exporting all my appointments and creating a new iCloud Calendar.
* go to **File** > **New Calendar** > **iCloud**
* create a new calendar
* go to **File** > **Export** > **Export...**
* export the old calendar
* go to **File** > **Import...**
* select the file to which you exported the old calendar
* click **Import**
* select the newly added calendar
* click **Ok**
* delete the old calendar | I did eventually find a solution: I logged into the external calendar via browser (in my case, exchange/Outlook) and could edit there. Apple 'Calendar' (formerly iCal) would create the event for exchange, but after a step or two could not edit, or even cancel with notifications. (Again, I would reply to the earlier solution not working, but I can't add comment there.) |
266,326 | I can not find a way to add an invitee after the event has been created and the invite emails sent. Is it possible?
Below is the already created event, where I do not see the field to "add invitees" any more:

Here is an event that has not yet been created, thus possible to add invitees:

EDIT: I use Exchange as the backend for the calendar... though had same issues with Google Calendar as backend as well.
Using Calendar Version 9.0 (2155.15) | 2016/12/27 | [
"https://apple.stackexchange.com/questions/266326",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/81319/"
] | I solved the issue of not being able to add an invitee by exporting all my appointments and creating a new iCloud Calendar.
* go to **File** > **New Calendar** > **iCloud**
* create a new calendar
* go to **File** > **Export** > **Export...**
* export the old calendar
* go to **File** > **Import...**
* select the file to which you exported the old calendar
* click **Import**
* select the newly added calendar
* click **Ok**
* delete the old calendar | I think I figured out the reason. I found out this statement in Apple support page (<https://support.apple.com/guide/calendar/if-you-cant-change-a-calendar-or-event-on-mac-icl1033/mac>):
>
> If you can’t change an event you created, or you can’t change your
> status for an event you were invited to, it might be because you’re
> using an email address in Calendar that isn’t on your card in
> Contacts. Make sure all your email addresses are listed on your
> Contacts card. See Edit contacts.
>
>
>
I decided to completely remove my Exchange account from the Internet Accounts on my mac. Then when signing again, I used as an email address my full email address: firstname.lastname@company.com and when signing to my account (via Microsoft web login) we need to use the shorter version of the email (in our company): username@company.com. It seems the problem lied in that I used short version of the email for the default email address in my calendar, which caused some permission issues etc. Hope this helps someone. |
17,621,342 | For example, right now I have a cell like this:

But I want it like this.

Is it possible in excel VBA?
Thank you!! | 2013/07/12 | [
"https://Stackoverflow.com/questions/17621342",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2495069/"
] | Use the Text To Columns feature under the Data tab to break this into 2 columns. Then, if you really need a line, select the last column and add a left border. | Is this data all in a single cell?
You will struggle to get it to align correctly, with or without VBA - each character occupies a slightly different width. **Edited** Well, it may behave better with a fixed-width font, as suggested by Tim Williams, but I still doubt that it will work consistently. (You could only insert spaces and the pipe '|' character, but it won't line up correctly.)
This data belongs in two columns, and multiple rows, with a border between them. |
32,034 | My question is pretty much summed up in the title. My story includes a lot of narration. Narrating events, narrating character's thoughts. There are several intervals in each chapter where the characters engage dialogue, but most of the story telling is done through narration.
How viable is this for writing an enjoyable story?
Edit: just to clarify what "too much" means, there is generally more narration than actual dialogue. | 2017/12/16 | [
"https://writers.stackexchange.com/questions/32034",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/27734/"
] | I don't think this balance matters terribly much.
The more critical measure is whether there is tension due to conflict. If your narration is describing a battle, for example, it can go on for pages without any dialogue. If your narration is describing people at high risk (trying to infiltrate a lair, for example) it can go on for pages.
It is tension that keeps readers turning pages to find out what happens on the next page. The tension is usually caused by conflict, but can also be caused by novelty. Like a character seeing something for the first time, that the reader also finds captivating. A giant alien space station or something. A living dinosaur.
It is easier to create tension in dialogue than in narration, just because the speakers can disagree, misunderstand, get confused or angry or resistant.
Thus a story that is mostly narration is more difficult for the writer to keep interesting, and you risk people getting bored or fatigued by the amount of information they are given without anything ***happening*** in the story with the characters. But if you can craft your narrations to engage people's imagination and give them a simulated imaginary "experience" then your story can be fine. | I like good narration and am annoyed by too much pointless dialog.
Actually, the dialog does not even need to be pointless. I become annoyed by dialog.
So, as a reader, to answer your question, there is **no such thing** as "too much narration." But there is, to this reader, such a thing as "too much dialog."
In my critique groups, I see some writers relying super - heavily on dialog. This is presumably because it is in some ways easier to write. (It is also harder to write well, in other ways.)
I believe many find it straightforward to type out a conversation, whereas describing an imaginary scene/setting requires, well, imagination and language. So, I have seen some writers relying heavily on dialog, and what ends up happening is the conversation is not well - anchored within a setting, and also the nuances of the conversation (the shading of words or stances of the people or contextual thoughts or subtext) are lost. I think some writers feel that the 'quick pacing' that they achieve from a crisp dialog and the 'all important white space' that one sees touted, are the goal.
The goal is a good story! Not speed and white space.
**Edit:**
I had actually read statistics on this ratio for fantasy books some months ago.
>
> Density: One way in which these books differ significantly from one another is in the proportion of narration to dialogue. The texts of the majority of titles are **less than 50% dialogue**, ranging from a low, narrative-heavy score of **13% dialogue** for The Wizard of Earthsea, to a much chattier 3**7% dialogue in The Final Empire (Mistborn #1)**. But the real odd-balls are Santiago, which is 59% dialogue, and **The Last Unicorn, which scores a whopping 63% talky-talk.** These two outliers seem so at odds with the rest of the group that I had to go into the text and examine it myself, to be sure that there wasn’t some kind of bug in my analysis tool, but my visual inspection did reveal an awful lot of dialogue in these two books.
>
>
>
<http://creativityhacker.ca/2013/07/05/analyzing-dialogue-lengths-in-fantasy-fiction/>
**Second Edit:**
Are you actually asking about whether it is OK to have info-dumps? because that is a separate question but some conflate it with narration. |
15,411,408 | I got confused about some issues:
1: The one month duration of auto-renewable subscriptions is 30 days or does it depend on a natural month?
Because I can only test in sandbox mode,so the duration is just several minutes...
Maybe Apple just simply calculates it like this: 2013\01\15 -> 2013\02\15 -> 2013\03\15. If so,the second issue comes up
2: For example: I buy a monthly auto-renewable subscriptions at 2013\03\31 ,because 2013\04 only has 30 days, then what is the expires\_date of my subscriptions? 2013\04\30 or 2013\05\01 or other date ? | 2013/03/14 | [
"https://Stackoverflow.com/questions/15411408",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1530528/"
] | It's adding 1 month, not 30 days. The number of days in 1 month varies. So purchasing a subscription on 3/31 would end on 4/30.
You can use NSDate, NSCalendar, and NSDateComponents to add a month to a date and see how long it will last. More info here: [Modifying NSDate to represent 1 month from today](https://stackoverflow.com/questions/185780/modifying-nsdate-to-represent-1-month-from-today) | I can confirm this from my own app's data on both iOS and Android.
If someone purchases a monthly renewing subscription on, for example, the 15th of one month, then it will renew on the 15th of the next month, irrespective of how many days were in the month.
If someone purchases the subscription on the 31st of March, then it will renew on May 1st (because April has no 31st day, the renewal date will jump to the next available day in the calendar year - this also applies to February in a leap year, etc).
Given this, and assuming your app's primary income is from subscriptions, over time you should expect to see lower than average sales on the 31st of any given month because there won't be as many renewals. But it is no cause for concern as it is counteracted by the fact that you should expect to see higher than average sales on the 1st of any month that follows a long month anyway. Don't be dissapointed if you check your stats on the 31st of any month. |
14,839 | I have done a three months independent research at a national lab in US with guidance from a scientist who was working there, but he is very busy and not much time is left to ask him to write a recommendation letter (RL) since the application deadline is coming.
I want to include this experience in my SOP, and wonder, in general, how is the reaction of graduate admission committees to research experience which is not backed up by a RL? | 2013/12/17 | [
"https://academia.stackexchange.com/questions/14839",
"https://academia.stackexchange.com",
"https://academia.stackexchange.com/users/8289/"
] | I believe the research experience with a national lab would be very helpful to your graduate school application. You should do your best to ask that senior scientist to write the recommendation letter for you. And he should understand it's part of his job to write recommendation letters.
In the worst case he will never have time to write the letter, my suggestion is to ask the human resource department of the lab to write a letter to **certify that you had worked at that lab**. This is, of course, not the recommendation letter. But, at least the certification letter proves that you did work there. How the admission commitee will react is another story. You have no control over it. You just need to do your best! | You need a letter from the senior person under whom you did the work! If you don't have one, this is like getting a *bad* letter. People know that they have this responsibility to junior people, so, although, yes, it involves some work on their part, it would be irresponsible to shirk it. You need that letter, I think, or people will wonder... |
14,839 | I have done a three months independent research at a national lab in US with guidance from a scientist who was working there, but he is very busy and not much time is left to ask him to write a recommendation letter (RL) since the application deadline is coming.
I want to include this experience in my SOP, and wonder, in general, how is the reaction of graduate admission committees to research experience which is not backed up by a RL? | 2013/12/17 | [
"https://academia.stackexchange.com/questions/14839",
"https://academia.stackexchange.com",
"https://academia.stackexchange.com/users/8289/"
] | I believe the research experience with a national lab would be very helpful to your graduate school application. You should do your best to ask that senior scientist to write the recommendation letter for you. And he should understand it's part of his job to write recommendation letters.
In the worst case he will never have time to write the letter, my suggestion is to ask the human resource department of the lab to write a letter to **certify that you had worked at that lab**. This is, of course, not the recommendation letter. But, at least the certification letter proves that you did work there. How the admission commitee will react is another story. You have no control over it. You just need to do your best! | One of the duties of a researcher who takes on a mentor role is to write letters of recommendation for his students. Your advisor will understand this responsibility. You may want to read the answers to [this question regarding writing your own recommendation letter](https://academia.stackexchange.com/questions/1452/points-to-remember-when-having-to-write-recommendation-letter-yourself?rq=1), as this may be relevant to your situation, but you should *always* be willing to ask for a letter. |
63,253 | The History Channel program *Life After People* focuses on how civilization would fare if humans instantly vanished. One episode says that one hour after people, unattended oil refineries blow up in flames. Another episode says that simple gas leaks are enough to turn residential neighborhoods like Levittown, New York, into raging infernos.
Yet the show does not cover the fate of humanity's primary fossil fuel--coal. When humans instantly disappear (how they did is not relevant to the show or the question), what fate will befall the coal factories? | 2016/12/03 | [
"https://worldbuilding.stackexchange.com/questions/63253",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/10274/"
] | Automated loss-of-fuel shutdown - not very exciting
===================================================
For coal powered plants, pulverized coal is fed from hoppers into the furnace. Human operator action is required to switch between hopper feed sources, or to refill the currently on-use hopper.
With no humans, the hopper which is being fed will eventually empty. I would assume that the plant has an automated loss-of-fuel shutdown procedure. I have experience with oil fired systems, which will perform an auto-shutdown when fuel runs out; coal plants probably do something similar.
Mostly, the auto-shutdown consists of ensuring that the output electrical generators are disconnected from the grid, so as their frequency drops they don't try to turn into motors. In oil plants, the feed fuel tanks all automatically shut their safety vales, to make sure there is no accidental feed into a not-running burner; again I would assume coal does something similar. | Even the most automated plants will run out of fuel within a few hours, coal plants require trainloads of coal per day and there are no people to run the trains, trucks, and tractors. |
63,253 | The History Channel program *Life After People* focuses on how civilization would fare if humans instantly vanished. One episode says that one hour after people, unattended oil refineries blow up in flames. Another episode says that simple gas leaks are enough to turn residential neighborhoods like Levittown, New York, into raging infernos.
Yet the show does not cover the fate of humanity's primary fossil fuel--coal. When humans instantly disappear (how they did is not relevant to the show or the question), what fate will befall the coal factories? | 2016/12/03 | [
"https://worldbuilding.stackexchange.com/questions/63253",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/10274/"
] | Even the most automated plants will run out of fuel within a few hours, coal plants require trainloads of coal per day and there are no people to run the trains, trucks, and tractors. | I'm not clear on the meaning of "coal factories", but coal fired thermal energy plants will shut down after the fuel in the ready bins runs out without anyone to refill them.
Coke plants (burning off the impurities in coal to make a pure carbon fuel for steel mills) will also shut down without operator supervision, although if they were suddenly unmanned during the coking process, there is a chance they would catch fire and all the coke/coal in and around the plant would be consumed in the fire.
Coal mines will generally fill up with water without running pumps, and this includes underground plants and pit mines. Exposed seams of coal from open air mining could potentially catch fire if a =proper heat source was induced (lightning strike, or a forest fire), creating underground coal seam fires which can burn for decades, |
63,253 | The History Channel program *Life After People* focuses on how civilization would fare if humans instantly vanished. One episode says that one hour after people, unattended oil refineries blow up in flames. Another episode says that simple gas leaks are enough to turn residential neighborhoods like Levittown, New York, into raging infernos.
Yet the show does not cover the fate of humanity's primary fossil fuel--coal. When humans instantly disappear (how they did is not relevant to the show or the question), what fate will befall the coal factories? | 2016/12/03 | [
"https://worldbuilding.stackexchange.com/questions/63253",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/10274/"
] | Automated loss-of-fuel shutdown - not very exciting
===================================================
For coal powered plants, pulverized coal is fed from hoppers into the furnace. Human operator action is required to switch between hopper feed sources, or to refill the currently on-use hopper.
With no humans, the hopper which is being fed will eventually empty. I would assume that the plant has an automated loss-of-fuel shutdown procedure. I have experience with oil fired systems, which will perform an auto-shutdown when fuel runs out; coal plants probably do something similar.
Mostly, the auto-shutdown consists of ensuring that the output electrical generators are disconnected from the grid, so as their frequency drops they don't try to turn into motors. In oil plants, the feed fuel tanks all automatically shut their safety vales, to make sure there is no accidental feed into a not-running burner; again I would assume coal does something similar. | I'm not clear on the meaning of "coal factories", but coal fired thermal energy plants will shut down after the fuel in the ready bins runs out without anyone to refill them.
Coke plants (burning off the impurities in coal to make a pure carbon fuel for steel mills) will also shut down without operator supervision, although if they were suddenly unmanned during the coking process, there is a chance they would catch fire and all the coke/coal in and around the plant would be consumed in the fire.
Coal mines will generally fill up with water without running pumps, and this includes underground plants and pit mines. Exposed seams of coal from open air mining could potentially catch fire if a =proper heat source was induced (lightning strike, or a forest fire), creating underground coal seam fires which can burn for decades, |
228,131 | What I mean by this question is that the ship should follow the basic anatomy of the protagonist Enterprises such as the Constitution and Galaxy classes. I.e. they should have two nacelles connected by distinct pylons, primary and secondary hulls, some sort of saucer section, etc.
Is there something smaller than a full-blown starship such as something the size of the USS Defiant or a shuttlecraft that follows this plan? Or are the NX-class and the Freedom-class the limit? | 2020/03/02 | [
"https://scifi.stackexchange.com/questions/228131",
"https://scifi.stackexchange.com",
"https://scifi.stackexchange.com/users/3823/"
] | "two nacelles connected by distinct pylons, primary and secondary hulls, some sort of saucer section"
It's crew module is more of a tea-cup than a saucer, but with two nacelles on pylons, a secondary hull made from a surplus nuclear missile and a primary hull made of scavenged titanium, Cochrane's 'The Phoenix' seems to have been the first and smallest of that formula.
<https://memory-alpha.fandom.com/wiki/Phoenix?file=Phoenix_top_view.jpg> shows clearly the crew module is attached above the narrowing of the missile's nose cone rather than being a single part of that hull. | It turns out there's not an easy way to answer this, because canon is sloppy or missing.
Oberth Class (120M? 150M?)
==========================
The [Oberth class](http://www.ex-astris-scientia.org/articles/oberth-size.htm) meets all the requirements at 120M (per Star Trek III production notes). But, as the link above notes, it's got massive logistical problems with that number
>
> If the Oberth were only 120m long, there would be only two complete decks of only 2m clear height inside a saucer of 39m diameter for the lower deck and 32m for the upper deck, respectively. The small saucer would have to include the standard-sized bridge, a computer core, quarters for 80 crew members, three cargo bays and, of course, several science labs as we expect it on a dedicated science ship. The window arrangement with the useless "skylights" for the lower deck, at the expense of some area in the upper deck, seems idiotic, unless it's something else. No need to mention that the Lilliputian decks of the 120m ship would not match what we have seen as interior sets in "Star Trek III" and four TNG episodes.
>
>
>
As such, the Oberth class is probably closer to 150M at its smallest, and possibly more than 200M by the time TNG rolled around.
Daedalus Class(140M?)
=====================
The [Daedalus class](https://memory-alpha.fandom.com/wiki/Daedalus_class) seems like a better candidate. They were 105M or 140M long and meet all the requirements of having distinct nacelles with pylons and a split hull design (with a giant ball instead of a saucer, but close enough). The catch is we have no official canon specs, so subsequent mentions vary, but some [official-ish sources](https://memory-alpha.fandom.com/wiki/Star_Trek:_The_Official_Starships_Collection) peg it in at 140M [which seems workable](http://www.ex-astris-scientia.org/articles/daedalus-problems.htm)
>
> [Maybe the model's windows are wrong] in order to allow the ship to be composed of 15 decks of 2.8m each, giving us an overall length of 140m. This is just a compromise, but it keeps the ship reasonably small compared to the Constitution class, while it alleviates all the problems of the 105m Daedalus. The size comparison with the NX class and the Constitution class demonstrates that the Daedalus class works well at a length of 140m.
>
>
>
NX Class (225M)
===============
If we're limiting ourselves to pure canon, this is probably the only acceptable winner. At around 225M, this is the smallest we can be certain meets all the qualifications. With *ST: ENT* revolving around a single ship, we know how this class was laid out and how big she was. There's no TOS or TNG era ships that were this small and it's hard to find many below 300M, let alone far below. |
228,131 | What I mean by this question is that the ship should follow the basic anatomy of the protagonist Enterprises such as the Constitution and Galaxy classes. I.e. they should have two nacelles connected by distinct pylons, primary and secondary hulls, some sort of saucer section, etc.
Is there something smaller than a full-blown starship such as something the size of the USS Defiant or a shuttlecraft that follows this plan? Or are the NX-class and the Freedom-class the limit? | 2020/03/02 | [
"https://scifi.stackexchange.com/questions/228131",
"https://scifi.stackexchange.com",
"https://scifi.stackexchange.com/users/3823/"
] | It turns out there's not an easy way to answer this, because canon is sloppy or missing.
Oberth Class (120M? 150M?)
==========================
The [Oberth class](http://www.ex-astris-scientia.org/articles/oberth-size.htm) meets all the requirements at 120M (per Star Trek III production notes). But, as the link above notes, it's got massive logistical problems with that number
>
> If the Oberth were only 120m long, there would be only two complete decks of only 2m clear height inside a saucer of 39m diameter for the lower deck and 32m for the upper deck, respectively. The small saucer would have to include the standard-sized bridge, a computer core, quarters for 80 crew members, three cargo bays and, of course, several science labs as we expect it on a dedicated science ship. The window arrangement with the useless "skylights" for the lower deck, at the expense of some area in the upper deck, seems idiotic, unless it's something else. No need to mention that the Lilliputian decks of the 120m ship would not match what we have seen as interior sets in "Star Trek III" and four TNG episodes.
>
>
>
As such, the Oberth class is probably closer to 150M at its smallest, and possibly more than 200M by the time TNG rolled around.
Daedalus Class(140M?)
=====================
The [Daedalus class](https://memory-alpha.fandom.com/wiki/Daedalus_class) seems like a better candidate. They were 105M or 140M long and meet all the requirements of having distinct nacelles with pylons and a split hull design (with a giant ball instead of a saucer, but close enough). The catch is we have no official canon specs, so subsequent mentions vary, but some [official-ish sources](https://memory-alpha.fandom.com/wiki/Star_Trek:_The_Official_Starships_Collection) peg it in at 140M [which seems workable](http://www.ex-astris-scientia.org/articles/daedalus-problems.htm)
>
> [Maybe the model's windows are wrong] in order to allow the ship to be composed of 15 decks of 2.8m each, giving us an overall length of 140m. This is just a compromise, but it keeps the ship reasonably small compared to the Constitution class, while it alleviates all the problems of the 105m Daedalus. The size comparison with the NX class and the Constitution class demonstrates that the Daedalus class works well at a length of 140m.
>
>
>
NX Class (225M)
===============
If we're limiting ourselves to pure canon, this is probably the only acceptable winner. At around 225M, this is the smallest we can be certain meets all the qualifications. With *ST: ENT* revolving around a single ship, we know how this class was laid out and how big she was. There's no TOS or TNG era ships that were this small and it's hard to find many below 300M, let alone far below. | "Classic design" I assume you mean TOS? or the saucer/body/nacelle design.
The smallest is the "Nova class" the ship comes in at 221 meters in length and follows the classic design philosophy.
"Size according to main website"
<https://www.startrek.com/article/inside-the-u-s-s-equinox-ncc-72381>
Never the less, approximate size is difficult to determine without placing against Voyager.
The ships Master system display extrapolated against pixels, show it to be 198 meters in length.
[](https://i.stack.imgur.com/GJLLD.jpg) |
228,131 | What I mean by this question is that the ship should follow the basic anatomy of the protagonist Enterprises such as the Constitution and Galaxy classes. I.e. they should have two nacelles connected by distinct pylons, primary and secondary hulls, some sort of saucer section, etc.
Is there something smaller than a full-blown starship such as something the size of the USS Defiant or a shuttlecraft that follows this plan? Or are the NX-class and the Freedom-class the limit? | 2020/03/02 | [
"https://scifi.stackexchange.com/questions/228131",
"https://scifi.stackexchange.com",
"https://scifi.stackexchange.com/users/3823/"
] | "two nacelles connected by distinct pylons, primary and secondary hulls, some sort of saucer section"
It's crew module is more of a tea-cup than a saucer, but with two nacelles on pylons, a secondary hull made from a surplus nuclear missile and a primary hull made of scavenged titanium, Cochrane's 'The Phoenix' seems to have been the first and smallest of that formula.
<https://memory-alpha.fandom.com/wiki/Phoenix?file=Phoenix_top_view.jpg> shows clearly the crew module is attached above the narrowing of the missile's nose cone rather than being a single part of that hull. | "Classic design" I assume you mean TOS? or the saucer/body/nacelle design.
The smallest is the "Nova class" the ship comes in at 221 meters in length and follows the classic design philosophy.
"Size according to main website"
<https://www.startrek.com/article/inside-the-u-s-s-equinox-ncc-72381>
Never the less, approximate size is difficult to determine without placing against Voyager.
The ships Master system display extrapolated against pixels, show it to be 198 meters in length.
[](https://i.stack.imgur.com/GJLLD.jpg) |
232,253 | So, I am developing a fictitious world that is very similar to our earth. Generally, most of the continents stay the same way they are (Maybe North America is split in two smaller continents), with slight alteration in shape and size. However, these are the major changes I am considering to add.
* First, there is a landmass the size of Australia slap into where Arctic Ocean is, right at the north pole.
* Second is Antarctica's landmass at the south pole is reduced to half of its size (Or remove as much as the landmass that Arctic gains, so it doesn't affect sea level significantly).
How would this affect climate overall? I heard that with landmass at the pole would make ice cap being able to form larger, so would an "Arctic continent" cause North America and Eurasia to become colder at the top from the bigger, more firmed ice cap?
And with smaller Antarctica, would that make southern hemisphere warmer due to more water surface area and less ice cap? | 2022/07/02 | [
"https://worldbuilding.stackexchange.com/questions/232253",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/97031/"
] | It depends a lot on what the geologic history of your world is.
On the one hand, the climatic history of the Cenozoic is almost entirely dependent on the history of Antarctica. You have a large landmass positioned directly over the South Pole where ice can build up on it, *then* you have the continent lose its connections to Australia and South America so there is a current circling around it (the circum-Antarctic current) to produce a natural cooling effect.
On the one hand, a smaller Antarctica by itself would mean that the ice age would not be as severe because there is less space for ice to build up on Antarctica. It also depends on how it separates from Australia/South America. If it's 50% the size, does it separate earlier, or does this mean it's positioned more northernly so the separation still happens in the Eocene and thus the continent may not be on top of the South Pole (and thus no circum-Antarctic current).
On the other hand, if there's a landmass at the *North* Pole, it means the same thing that happened IRL to Antarctica might happen there instead. It depends on if there is enough space for a circum-Arctic current to form around this new landmass, which would potentially cause an early ice age, which in turn would really change the evolutionary history of Earth. | **The Northern hemisphere landmasses around the Arctic would likely be a bit warmer. Antarctica would still be largely unchanged, possibly slightly colder.**
The climate is primarily driven by the sun and the atmosphere. The amount of energy being received by the planet isn't changing and neither is the composition of the atmosphere or the planet, thus the outcome is really that not much changes. There might be a slight increase in the planet's albedo (that's how reflective it is) if snow is more permanent in the north which would result in a slight increase in heat loss to space, but it's unlikely to make any notable difference as the region is reflective most of the time currently.
However, Antarctica is particularly cold because:
* It's high up and high altitude air is cold
* It's surrounded by water which stores a lot of thermal energy
* The angle of the incoming sunlight is very shallow
If none of these change, Antartica's cold climate also stays the same. It may get a little colder if it's surrounded by even more thermal storing water.
Meanwhile in the north, a continent the size of Australia would punch a pretty big chunk out of the Arctic ocean. The Arctic is currently cold, but not that cold, because:
* It's basically at sea level
* The angle of the incoming sunlight is very shallow
* It's surrounded by land, which will instead be providing heat rather than trying to take it
If you remove the thermal "battery" that is the arctic ocean, more energy that would've been stored up in that water will instead be in the land resulting in slightly higher temperatures there. |
375,276 | This question keeps coming up every here at meta and it gets positive responses.
Yet it has not been implemented yet.
Why?!
Using a phone and copy pasting a code block is a true nightmare!
Sometimes it can take minutes to make sure you select the correct text.
Try to copy paste an array with multiple dimensions and several layers, it's just not user friendly at all.
I understand some people think "this will encourage copy pasting answers".
Maybe. Yes it will be easier, but if you really want to copy paste an answer is this the "thing" that makes them give up? No, I don't think so.
It's hard to copy the code blocks from any platform in my opinion, but especially the smaller the screen is.
Because of that, if it at least can be implement to mobile devices then I will be happy.
Other threads from the past:
2009 [Shortcut or button for copying posted code from Stack Overflow](https://meta.stackexchange.com/questions/32625/shortcut-or-button-for-copying-posted-code-from-stack-overflow)
2014 [Select All / Copy All Button for Code](https://meta.stackoverflow.com/questions/251883/select-all-copy-all-button-for-code)
2017 [Copy button for code blocks](https://meta.stackoverflow.com/questions/345667/copy-button-for-code-blocks)
Video of it:
Go to 1:30
<https://youtu.be/U6SfRPwTKqo>
With phones being used more and more this problem is "worse" than back then (maybe not 2017).
*This question was written on a phone* | 2018/10/13 | [
"https://meta.stackoverflow.com/questions/375276",
"https://meta.stackoverflow.com",
"https://meta.stackoverflow.com/users/5159168/"
] | This feature should have been implemented *much earlier*, as we can see from the linked meta discussion above from 2014 or even 2009.
It's not only about phone. Even on a computer, I always need to drag the mouse multiple times to correctly select the code area, it's very cumbersome. This also affects users who merely browse the site, since user often needs to copy the code in answer to verify on their own machine.
Manually selecting code area is way too inconvenient. This simple feature would be a much better and useful UI improvement than some fancy layout change that happened recently. | If I were to wager a guess...this feature's not implemented because the people who would primarily benefit from it (e.g. users who post answers on phones) are a much smaller demographic in general.
The larger demographic has an existing work around in that they can simply select all of the text they want and copy that using their keyboard.
I could at least see rationale for wanting to do this, but I don't personally see it being worth the dev time. |
375,276 | This question keeps coming up every here at meta and it gets positive responses.
Yet it has not been implemented yet.
Why?!
Using a phone and copy pasting a code block is a true nightmare!
Sometimes it can take minutes to make sure you select the correct text.
Try to copy paste an array with multiple dimensions and several layers, it's just not user friendly at all.
I understand some people think "this will encourage copy pasting answers".
Maybe. Yes it will be easier, but if you really want to copy paste an answer is this the "thing" that makes them give up? No, I don't think so.
It's hard to copy the code blocks from any platform in my opinion, but especially the smaller the screen is.
Because of that, if it at least can be implement to mobile devices then I will be happy.
Other threads from the past:
2009 [Shortcut or button for copying posted code from Stack Overflow](https://meta.stackexchange.com/questions/32625/shortcut-or-button-for-copying-posted-code-from-stack-overflow)
2014 [Select All / Copy All Button for Code](https://meta.stackoverflow.com/questions/251883/select-all-copy-all-button-for-code)
2017 [Copy button for code blocks](https://meta.stackoverflow.com/questions/345667/copy-button-for-code-blocks)
Video of it:
Go to 1:30
<https://youtu.be/U6SfRPwTKqo>
With phones being used more and more this problem is "worse" than back then (maybe not 2017).
*This question was written on a phone* | 2018/10/13 | [
"https://meta.stackoverflow.com/questions/375276",
"https://meta.stackoverflow.com",
"https://meta.stackoverflow.com/users/5159168/"
] | This feature should have been implemented *much earlier*, as we can see from the linked meta discussion above from 2014 or even 2009.
It's not only about phone. Even on a computer, I always need to drag the mouse multiple times to correctly select the code area, it's very cumbersome. This also affects users who merely browse the site, since user often needs to copy the code in answer to verify on their own machine.
Manually selecting code area is way too inconvenient. This simple feature would be a much better and useful UI improvement than some fancy layout change that happened recently. | I don't get the scenario you have in mind. You said you feel the urge of this feature because you want:
>
> To write answers. I post most of my answers from my phone
>
>
>
Odds are that you want to quote a snippet from the whole code and you have to select it anyways before copying; if you want a button to copy the entire code i don't see how it can be any useful while answering since you will paste the whole block and then modify it through answer window which is as well time consuming. I can't see honestly any scenario where this feature can be useful on phone. Even on desktop it's not worth implementing it since questions contain 99% of the times custom lines of code and also 99% of the times you have custom code to add. So at the end it's faster to write it down maybe copying unusually long lines with the help of keyboard shortcuts. |
5,372,340 | I need to save several documents to the cloud and need to save the documents, document metadata, and words/phrases for searching.
My plan is to use a symmetric cypher for encrypting the whole document, but I'm unsure of the right way to hash each word. I would like something secure, but I don't want to increase the count of characters in each word unnecessarily.
What implementation is most suitable for doing a symmetric encryption against a document, and what is the best way to hash a word or phrase without making it many times larger than it needs to be? | 2011/03/20 | [
"https://Stackoverflow.com/questions/5372340",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/328397/"
] | First, I suggest different tags. It sounds like you're really interested in offloading searching to a server in a cryptographically secure way (such that the server doesn't have access to the plaintext and such that the client need not transfer the entire index).
Issues:
* An attacker being able to figure out which words are in the index (and which are not) could be an issue for you. You should state whether it is as a part of your requirements.
* An attacker being able to figure out which items in the index occur more frequently could be an issue for you. You should state whether it is as a part of your requirements.
* An attacker being able to associate words with a document could be an issue for you. You should state whether it is as a part of your requirements.
* An attacker may be able to subvert the server entirely and observe queries / retrievals. You should state security needs in this circumstance as well.
* Probably others I haven't thought of.
I'm assuming that you're designing your own, but there is probably some prior art, research, etc. that would be smarter than I am below:
For the first, I suggest that you should hash the words, combining the plaintext with a secret (not shared with the index server) before hashing, and truncating the hash to the point where it is likely to be non-unique in the index. This costs you hash efficiency, but helps prevent an attacker from using the hash as a plaintext equivalent or experimentally determining the secret
For the second and third, you should encrypt any indexed data (such as counts or document+position) and decrypt it on the client. This may cost you latency.
For the fourth, you'd want to consider concealing real requests inside groups of unrelated requests, things like that, but you'd want a lot of math to make sure you weren't still vulnerable to statistical analysis.
For the fifth, do some web research. I'm confident there will be stuff out there, and this is a pretty specific (and less common) need, so you'll want someone who put more thought into it than I just have. | Your requirements are mutually exclusive. That kind of metadata will leak a huge amount of information about the document content, to the point it can't be called secure.
Furthermore, encrypting individual words is futile. The difficulty of breaking encryption is usually said to be as difficult as breaking the key, but this assumes the information content in the plaintext is greater than that in the key. For single words, that certainly isn't true. |
124,938 | >
> Can you guess how old I am?
>
>
>
What tense should I use with *guess*?
>
> I **guess** you are around 30.
>
>
> I **am guessing** you are around 30.
>
>
> | 2017/04/07 | [
"https://ell.stackexchange.com/questions/124938",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/47521/"
] | One of the keepers of the International System of Units is the International Bureau of Weights and Measures (*Bureau international des poids et mesures*). [Its website](http://www.bipm.org/en/publications/si-brochure/section5-3.html) says, among many other things:
>
> ... a space is always used to separate the unit from the number ...
> Even when the value of a quantity is used as an adjective, a space is
> left between the numerical value and the unit symbol. Only when the
> name of the unit is spelled out would the ordinary rules of grammar
> apply, so that in English a hyphen would be used to separate the
> number from the unit.
>
>
>
It gives the example of 'a 35-millimetre film'. I assume that 'a 35 mm film' would be interchangeable.
'Litre' is a 'Non-SI unit accepted for use with the International System of Units' and has the alternative abbreviations L and l ([link](http://www.bipm.org/en/publications/si-brochure/table6.html)), with a footnote:
>
> (f) The litre, and the symbol lower-case l, were adopted ... in 1879
> ... The alternative symbol, capital L, was adopted [in] 1979 ..., in
> order to avoid the risk of confusion between the letter l (el) and the
> numeral 1 (one).
>
>
>
(Generally speaking, only units based on people's names are in upper case.)
So, your original example would be either:
>
> a 1000-litre reactor
>
>
>
or
>
> a 1000 L reactor
>
>
>
(With the the possibility of kilolitre or kL.)
That said, if you are formally or informally using a published or house style guide, then you would go with that. And it might make a difference whether you are writing for a specialist or general reader. | Numeric metric expressions with the abbreviated metric measurement name should be written after the number with no space.
>
> The cultivation was performed in a 1000L bioreactor.
>
>
> I bought 250kg of X, 250ml of Y, 250cm of Z, etc.
>
>
> 3cm x 4cm strip.
>
>
> |
136,677 | **[SOURCE](https://www.eduzip.com/ssc/ssc-cgl-tier-1-9th-august-2015-morning/english-comprehension.html)** (Indian civil service exam; one of dozens)
>
> The second pigeon flew just as the first **pigeon had flown.**
>
>
>
A) No improvement
B) one had done
C) one had flown away
D) had done
Which part, among the parts given above, is appropriate for the bold part in the above given sentence ?
I tried to find answer on google but got contradictory answers, somewhere its B as answer somewhere C. To my ears options B sounds best among all options. Perhaps it would have been better with *one had flown* but that's not in any of the given alternative. | 2017/07/18 | [
"https://ell.stackexchange.com/questions/136677",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/46809/"
] | >
> 1. The second pigeon **flew**…
>
>
>
This is the simple past.
We also learn a previous pigeon had performed the same action earlier.
>
> 2. …just as the first pigeon **had flown**.
>
>
>
This is the past perfect tense.
The original sentence could remain as it is, but a good writer will probably sense that writing the term *pigeon*, or repeating the same verb *fly* twice is redundant. In order to overcome this, a pronoun is needed to substitute “pigeon”.
>
> 2. …just as the first **one** had flown.
>
>
>
Now, I quite like this version but it is not included in the multiple choice. The closest equivalent is C) *…one had flown away*. The adverb *[away](https://www.macmillandictionary.com/dictionary/british/away_1)*, in this context, means “further from a place, thing or person”, and “fly away” is a very common collocation. So, the OP could choose C).
>
> 2. C) …just as the first **one** had flown **away**
>
>
>
If the writer wanted to use a pronoun, *and* avoid repeating the same verb, the auxiliary verb, *do*, is used.
>
> 3. B) The second pigeon **flew** just as the first one **had done**
>
>
>
Here, the adverb *away* is not mentioned at all, if had been added to the first clause, then the second clause would fit perfectly,e.g. “The second pigeon flew away just as the first one had done.”
There is the construction **do + so**, where different forms of ***do so*** substitutes the verb, and its complement. e.g. a) *The second pigeon **returned to its coop** just as the first one had **done so***. b) *They asked me to revise the essay and **I did so** (= I revised the essay.)* c) Dangerous currents. Anyone who swims here [**does so**](https://www.collinsdictionary.com/dictionary/english/at-ones-peril) at their own peril. In the OP's example there is no complement in *The first pigeon flew* and therefore "so" is not required.
For more information about ***Do* as a substitute verb**, visit the [*Cambridge Dictionary Grammar*](https://dictionary.cambridge.org/grammar/british-grammar/common-verbs/do) website.
D) is incorrect because a noun or pronoun is missing:
>
> 2. D)…just as the first had done
>
>
>
The first *what*? It might be a white dove, or a duck for all we know.
So all this boils down to personal preference, and ***style***. There is nothing grammatically incorrect with clauses B) or C), either one, in my opinion, is appropriate. | All are bad. The sentence is a muddled jumble. D is probably best.
To express similarity of flight...
>
> The second pigeon flew just like the first.
>
>
> The second pigeon flew just like the first pigeon flew.
>
>
>
To express concurrency of flight or as it's worded, flying.
>
> The second pigeon flew while the first pigeon flew.
>
>
> The second pigeon flew while the first pigeon was flying.
>
>
>
To express concurrency of taking to flight (i.e. going from perch to flight) use similar constructs. The use of "also" further reinforces concurrency over similarity.
>
> The first pigeon took flight (or more commonly "took off") just as the second was [also] taking flight (taking off).
>
>
> One pigeon took flight just as the other took flight.
>
>
>
The words "just as" create some confusion in your sentence since "just as" as you used it is more likely to express concurrency over similarity. Additionally, there is an issue with "flew". I think you mean "took flight".
Your sentence as it's worded, however, tends to express similarity while using the concurrent form. |
136,677 | **[SOURCE](https://www.eduzip.com/ssc/ssc-cgl-tier-1-9th-august-2015-morning/english-comprehension.html)** (Indian civil service exam; one of dozens)
>
> The second pigeon flew just as the first **pigeon had flown.**
>
>
>
A) No improvement
B) one had done
C) one had flown away
D) had done
Which part, among the parts given above, is appropriate for the bold part in the above given sentence ?
I tried to find answer on google but got contradictory answers, somewhere its B as answer somewhere C. To my ears options B sounds best among all options. Perhaps it would have been better with *one had flown* but that's not in any of the given alternative. | 2017/07/18 | [
"https://ell.stackexchange.com/questions/136677",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/46809/"
] | >
> 1. The second pigeon **flew**…
>
>
>
This is the simple past.
We also learn a previous pigeon had performed the same action earlier.
>
> 2. …just as the first pigeon **had flown**.
>
>
>
This is the past perfect tense.
The original sentence could remain as it is, but a good writer will probably sense that writing the term *pigeon*, or repeating the same verb *fly* twice is redundant. In order to overcome this, a pronoun is needed to substitute “pigeon”.
>
> 2. …just as the first **one** had flown.
>
>
>
Now, I quite like this version but it is not included in the multiple choice. The closest equivalent is C) *…one had flown away*. The adverb *[away](https://www.macmillandictionary.com/dictionary/british/away_1)*, in this context, means “further from a place, thing or person”, and “fly away” is a very common collocation. So, the OP could choose C).
>
> 2. C) …just as the first **one** had flown **away**
>
>
>
If the writer wanted to use a pronoun, *and* avoid repeating the same verb, the auxiliary verb, *do*, is used.
>
> 3. B) The second pigeon **flew** just as the first one **had done**
>
>
>
Here, the adverb *away* is not mentioned at all, if had been added to the first clause, then the second clause would fit perfectly,e.g. “The second pigeon flew away just as the first one had done.”
There is the construction **do + so**, where different forms of ***do so*** substitutes the verb, and its complement. e.g. a) *The second pigeon **returned to its coop** just as the first one had **done so***. b) *They asked me to revise the essay and **I did so** (= I revised the essay.)* c) Dangerous currents. Anyone who swims here [**does so**](https://www.collinsdictionary.com/dictionary/english/at-ones-peril) at their own peril. In the OP's example there is no complement in *The first pigeon flew* and therefore "so" is not required.
For more information about ***Do* as a substitute verb**, visit the [*Cambridge Dictionary Grammar*](https://dictionary.cambridge.org/grammar/british-grammar/common-verbs/do) website.
D) is incorrect because a noun or pronoun is missing:
>
> 2. D)…just as the first had done
>
>
>
The first *what*? It might be a white dove, or a duck for all we know.
So all this boils down to personal preference, and ***style***. There is nothing grammatically incorrect with clauses B) or C), either one, in my opinion, is appropriate. | I disagree with both answers. All the four option are possible. The **D** version being rather informal. The rest is only a matter of style and emphasis. However both **B** and **D** improve the sentence.
Of all the options, clearly "**C**" isn't the best one (it doesn't improve anything). I would go with "**B**", which is a very brief and suitable option. |
75,263 | I read a recipe that said don't preheat the oven when baking a pound cake. Why wouldn't you preheat the oven? Every recipe I've ever read said preheat the oven. This is really confusing to me. | 2016/11/04 | [
"https://cooking.stackexchange.com/questions/75263",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/51728/"
] | I have seen (yeast) bread recipes that rely on an un-preheated oven and seem to take advantage of the slow heating process (give the leavener some time to act before incapacitating it by overheating). While pound cake is unlikely to use yeast, there was likely a similar idea on the recipe writers mind. Maybe getting a somewhat denser texture by firming the outer layer before activating the leavener in the core is the goal here. Such recipes tend to be unreliable because they make many assumptions about the bakeware and oven used.
EDIT: Another reason for somebody to develop a recipe that does not use preheating would be simply saving active time, especially for people having slow-to-preheat ovens - mix, put in cold oven, start oven, set a timer and step away. | Without seeing the recipe and likely experimenting, it is hard to know for sure, but a few thoughts on pre-heating:
The shorter a cooking time, typically the more important it is to preheat. If you are doing a 5 minute cookie bake, you are unlikely to get your desired results if you fail to pre-heat. If you are doing a 6 hour roast, well, you are probably just wasting energy preheating.
The general idea is, if your recipe calls for a temperature for 30 minutes, then it needs to be up to that temp and stable before you start. A counter argument is how can it be when you open the oven to put the item in and radically lower the temperature when you do that.
Part of the tradition of most recipes saying to pre-heat was wood stoves especially, but older models in general which heated slowly and one person's might take longer to come to temperature than someone else's. So, baking for 30 minutes in my stove that comes to temperature in 10 minutes would result in half cooked in someone else's over that took 20 minutes to come to temp, and burnt for someone else who pre-heated, so the pre-heat command put everyone on an even field. Today, most ovens heat up fairly quickly so it is less of an issue other than for shorter cooking times and things that need precision such as delicate pastries. BUT, and this is a big but, with many modern ovens which heat to temp quickly, they do so with higher power. So, if you put things, especially short cooking time items, rather than taking longer to cook, they will burn, but may also be half raw. While you have you oven set at say 350F, it may actually be pouring 500F heat at your dinner rolls until the full oven temp stabilizes, and you end up burning the surface while not getting the middle cooked.
Now, one reason I can think that a recipe might specifically say not to pre-heat would be to take advantage of this. That initial higher power while the oven is heating and stabilizing could help form a crisper crust and then allow the bulk of the cooking to occur at an effectively slightly lower temp, similar starting an item at a high temp and then allowing the temperature to fall as is sometimes done for a bread with a crisp shell like crust but delicate interior. The effect would be a lessor example of this. |
75,263 | I read a recipe that said don't preheat the oven when baking a pound cake. Why wouldn't you preheat the oven? Every recipe I've ever read said preheat the oven. This is really confusing to me. | 2016/11/04 | [
"https://cooking.stackexchange.com/questions/75263",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/51728/"
] | One obvious reason is to avoid wasting energy. A lot of people tend to like recipes which avoid preheating for this reason alone.
But, assuming that there's more than that here, my guess is that the recipe is trying to use the slower heating as a way to achieve a greater rise. Whether your pound cake recipe relies on leavening (such as baking powder/soda) or whether it's a traditional pound cake recipe that relies only on the air bubbles beaten into the batter while mixing, a gradual temperature rise can in some circumstances result in more expansion.
Basically, baking is a race against time in crust formation vs. internal expansion. You want to allow enough time for the internal gas bubbles to expand and make the cake (or bread or whatever) lighter, because once the structure starts to harden (with egg coagulation, and then starch gelatinization), you can't expand any more. On the other hand, as a cake or loaf rises, it becomes increasingly unstable. If you wait too long to form a crust and then stabilize the internal structure, your whole cake could collapse.
Frequently, the inside portion of a cake or loaf continues to expand even after the crust hardens, resulting in cracking on the top surface (common in pound cakes, perhaps even desirable according to some). In extreme cases with things like yeast breads, it can even result in a "blow-out" where the pressure from internal expansion literally blows a big hole in the crust.
With pound cakes, this balance can be a particular problem. As already mentioned, hardening the crust too early can result in excessive cracking. But hardening the crust later can also be a problem if the internal structure is unevenly heated (and thus unevenly stable). If the crust hardens while the cake is rising very high (e.g., from rapid expansion under high oven heat), but the internal structure is unstable underneath and collapses a bit, it can result in the "crust separation" where there's an air gap between the top crust and the rest of the cake. It's possible that heating more slowly can make that less likely by heating the whole batter more evenly and ensuring the internal parts of the cake are closer to setting along with the top crust.
Anyhow, to generalize a bit more, one of the reasons to slowly heat a baked good is to try to allow more time for a gradual expansion. On the other hand, internal air within batter/dough will eventually "leak out." It's actually a continuous process, since batter/dough is permeable to air. With cake batter, you can often actually see bubbles rising and popping on the surface of a cake early in the baking process. So you need *some* structure and crust to form to prevent all that air from escaping and collapsing your batter too. The trick, again, is to balance the time given for bubbles to form/expand vs. how fast the cake sets (i.e., structure hardens). That's actually one of the primary reasons for different baking temperatures in different leavened baked goods.
While allowing expansion during preheating can sometimes be helpful, it's also hard to manage. Ovens don't preheat at the same rate. If your oven preheats too slowly, it could actually cause your batter to rise too much before it stabilizes, thus resulting in collapse. As pointed out in another answer, some ovens with radiant heat elements (especially electric models) can actually introduce a "broiling" effect on the outer surface during preheating, which could counteract the positive effects if the "broiled" crust hardens early and prevents further expansion.
Since it's less predictable, most recipes tend to recommend preheating for things like cakes and pastry. And no matter what, you'll have to monitor doneness more frequently the first few times you do a recipe like this in a particular oven, since the time window for being "done" will likely vary much more significantly for recipes that don't preheat. | Without seeing the recipe and likely experimenting, it is hard to know for sure, but a few thoughts on pre-heating:
The shorter a cooking time, typically the more important it is to preheat. If you are doing a 5 minute cookie bake, you are unlikely to get your desired results if you fail to pre-heat. If you are doing a 6 hour roast, well, you are probably just wasting energy preheating.
The general idea is, if your recipe calls for a temperature for 30 minutes, then it needs to be up to that temp and stable before you start. A counter argument is how can it be when you open the oven to put the item in and radically lower the temperature when you do that.
Part of the tradition of most recipes saying to pre-heat was wood stoves especially, but older models in general which heated slowly and one person's might take longer to come to temperature than someone else's. So, baking for 30 minutes in my stove that comes to temperature in 10 minutes would result in half cooked in someone else's over that took 20 minutes to come to temp, and burnt for someone else who pre-heated, so the pre-heat command put everyone on an even field. Today, most ovens heat up fairly quickly so it is less of an issue other than for shorter cooking times and things that need precision such as delicate pastries. BUT, and this is a big but, with many modern ovens which heat to temp quickly, they do so with higher power. So, if you put things, especially short cooking time items, rather than taking longer to cook, they will burn, but may also be half raw. While you have you oven set at say 350F, it may actually be pouring 500F heat at your dinner rolls until the full oven temp stabilizes, and you end up burning the surface while not getting the middle cooked.
Now, one reason I can think that a recipe might specifically say not to pre-heat would be to take advantage of this. That initial higher power while the oven is heating and stabilizing could help form a crisper crust and then allow the bulk of the cooking to occur at an effectively slightly lower temp, similar starting an item at a high temp and then allowing the temperature to fall as is sometimes done for a bread with a crisp shell like crust but delicate interior. The effect would be a lessor example of this. |
38,798 | On one of my characters, the guy in the wagon at Whiterun will take me to almost any city in Skyrim.
On my slightly newer character, he only lets me go to four cities - Markarth, Riften, Solitude and Windhelm.
My first character has all those plus places like Morthal, Winterhold, and Falkreath.
What's the reason behind this? It's not based on which cities I've visited (as I've only visited Riften out of the four on my second character). | 2011/11/27 | [
"https://gaming.stackexchange.com/questions/38798",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/6139/"
] | Locations only show up if you have enough gold to travel to them, it is likely that your newer character did not have sufficiant gold to travel along the more expensive routes. | Try to scroll down. May be the issue |
38,798 | On one of my characters, the guy in the wagon at Whiterun will take me to almost any city in Skyrim.
On my slightly newer character, he only lets me go to four cities - Markarth, Riften, Solitude and Windhelm.
My first character has all those plus places like Morthal, Winterhold, and Falkreath.
What's the reason behind this? It's not based on which cities I've visited (as I've only visited Riften out of the four on my second character). | 2011/11/27 | [
"https://gaming.stackexchange.com/questions/38798",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/6139/"
] | Locations only show up if you have enough gold to travel to them, it is likely that your newer character did not have sufficiant gold to travel along the more expensive routes. | I just had the same issue except only the 20 gold options were available. I traveled to them all and back to whiterun and that seemed to fix the problem. |
38,798 | On one of my characters, the guy in the wagon at Whiterun will take me to almost any city in Skyrim.
On my slightly newer character, he only lets me go to four cities - Markarth, Riften, Solitude and Windhelm.
My first character has all those plus places like Morthal, Winterhold, and Falkreath.
What's the reason behind this? It's not based on which cities I've visited (as I've only visited Riften out of the four on my second character). | 2011/11/27 | [
"https://gaming.stackexchange.com/questions/38798",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/6139/"
] | I just had the same issue except only the 20 gold options were available. I traveled to them all and back to whiterun and that seemed to fix the problem. | Try to scroll down. May be the issue |
3,793,311 | how can I get a list of all C API functions with Lua?
For example, there are plenty C functions in a game(Crysis) which
can be called with
Lua:
g\_gameRules.game:FunctionInC()
There are many functions I know but how can I get all available functions? | 2010/09/25 | [
"https://Stackoverflow.com/questions/3793311",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/458007/"
] | If you are talking about window popups then sure they are considered bad usabiltiy practice. A user feels that he is about to view an advertisement and closes it all sudden. Also it doesnt falls into seamless view of site. Popups are sure a distraction we have currently been working on a project where we use Lightbox modalbox from jQuery to avoid usage of popups which have been very well recieved by users. See
<http://colorpowered.com/colorbox/> for examples of Lightboxes.
I always like to read <http://www.useit.com/alertbox/990530.html> for usability concerns. | You may not count this as the answer you wanted, but I have used CSS dialog popups with great effect in some of my web apps.
See here: [CSS Message Boxes with Different Types](http://www.jankoatwarpspeed.com/post/2008/05/22/CSS-Message-Boxes-for-different-message-types.aspx) |
3,793,311 | how can I get a list of all C API functions with Lua?
For example, there are plenty C functions in a game(Crysis) which
can be called with
Lua:
g\_gameRules.game:FunctionInC()
There are many functions I know but how can I get all available functions? | 2010/09/25 | [
"https://Stackoverflow.com/questions/3793311",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/458007/"
] | If you are talking about window popups then sure they are considered bad usabiltiy practice. A user feels that he is about to view an advertisement and closes it all sudden. Also it doesnt falls into seamless view of site. Popups are sure a distraction we have currently been working on a project where we use Lightbox modalbox from jQuery to avoid usage of popups which have been very well recieved by users. See
<http://colorpowered.com/colorbox/> for examples of Lightboxes.
I always like to read <http://www.useit.com/alertbox/990530.html> for usability concerns. | With the arrival of Web 2.0, it is better to use dialog boxes like [jQuery UI](http://jqueryui.com/demos/dialog/) |
139,990 | Regarding this question
[Is it okay to repost a question from another website that was never answered there?](https://meta.stackexchange.com/questions/68543/is-it-okay-to-repost-a-question-from-another-website-that-was-never-answered-the),
is it okay to repost a question which I have asked at another website (such as the Eclipse forums). It shouldn't be plagiarism, because it's my own question, but is this an "etiquette" violation? | 2012/07/16 | [
"https://meta.stackexchange.com/questions/139990",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/190828/"
] | It may also be helpful to consider improving the question in a way that ensures you get a better answer than those you received (or didn't receive) on the other site. If you didn't get a satisfactory answer there, you may want to indicate why (e.g. including details about solutions you were provided at the other site, and why those were unacceptable), or look more closely at your question and see if there are details you left out that made it harder for folks to help solve your problem. Especially if you received follow-up questions there from potential answerers, whether or not you answered them, etc.
As long as you form a complete question on SO, including a link to the "other" question to give people context might be a good idea, or adding it as a comment to your question at least. | Since you're the author and plagiarism isn't an issue, the only factor you should consider is quality. Stack Exchange welcomes all good questions, regardless of whether you've asked them elsewhere.
In this context, "good" depends on a number of things, including whether your question is on-topic, how well your question is formatted, the amount of background/supporting information you provide.... |
139,990 | Regarding this question
[Is it okay to repost a question from another website that was never answered there?](https://meta.stackexchange.com/questions/68543/is-it-okay-to-repost-a-question-from-another-website-that-was-never-answered-the),
is it okay to repost a question which I have asked at another website (such as the Eclipse forums). It shouldn't be plagiarism, because it's my own question, but is this an "etiquette" violation? | 2012/07/16 | [
"https://meta.stackexchange.com/questions/139990",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/190828/"
] | Some people get upset if they find the same user posting the same question to multiple forum sites *at the same time.* It can be annoying if I spend ten minutes answering your question on site A, and then find your question, already answered, on site B. It's even more annoying if I see if on site C, as well!
But if a question isn't answered on site A after a reasonable period of time, it seems fine to ask it on another site. And you could make the case that since SE isn't actually a forum, but rather a repository of knowledge, slightly different rules apply. Even if the question is answered elsewhere, it may belong here as well, to serve as a permanent record. | Since you're the author and plagiarism isn't an issue, the only factor you should consider is quality. Stack Exchange welcomes all good questions, regardless of whether you've asked them elsewhere.
In this context, "good" depends on a number of things, including whether your question is on-topic, how well your question is formatted, the amount of background/supporting information you provide.... |
20,822 | In the debt based fiat system of currency currently being monopolized, the amount of money added into circulation can cause it to deflate or inflate in value. How does the network calculate the proper rate at which to add more bitcoins? | 2014/01/20 | [
"https://bitcoin.stackexchange.com/questions/20822",
"https://bitcoin.stackexchange.com",
"https://bitcoin.stackexchange.com/users/12432/"
] | The answer for @jtorba is not strictly correct.
Difficulty is the main control mechanism used to determine the amount of Bitcoins released over time. The amount of bitcoins released is according to the bitcoin era as defined [here](https://en.bitcoin.it/wiki/Controlled_supply)
Currently the system tries to maintain a release of bitcoins every 10 minutes - which is part of the config settings that all miners adhere to. To do this the code takes into account the average block completion time every 2016 blocks.
It is important to note that this is NOT affected by the number of miners, but instead affected by the computing power.
A block is discovered by a process of hashing - which is a repetitive computing process - until a result is found that matches a rule. Obviously the FASTER you can repeat this process the better chance you have of finding the answer that fits the rule before anyone else. It is specifically designed not to matter how many computers you have doing this task.
The rule becomes increasingly difficult - hence the use of the word difficulty - and therefore more computing effort is required to cycle through the possibilities becomes harder.
This self adjusting process is the main mechanism that maintains the release of bitcoins to the schedule in the link above. | The simple answer is that it tries to keep bitcoin creation as steady as possible, but halving the rate every 4 years in order to sustain inflation in value.
The network makes mining difficult or easier by increasing/decreasing difficulty in order to achieve a 10 minute block time. If the amount of miners increases, that means a block will be solved less than 10 minutes and that becomes the que for the network to make it harder to mine. And if it on average took longer than 10 mins to mine a block, that the que to ease up mining. To put it simply, if the amount of miners doubled over night, it becomes twice as hard to mine bitcoins.
The network adjusts to the amount of people using the currency. |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | I also made complete demo for 9-bit UART emulation (based on even/odd parity). You can find it <http://bohdan-danishevsky.blogspot.com/2016/10/9-bit-serial-communication-in-linux.html>.
All sources available on git.
You can easily adapt it for your device. Hope you like it. | The accepted answer, which used an Atmel processor with true 9 bit data URT communicating with a Pi looks to have been either abandoned, or taken commerical.
So I am going with <https://www.vendingtools.ro/en> for Eur 70, and that will interface my Pi to the MDB 9 data bit bus.
---
[Update]
See also
<https://www.qibixx.com/en/products/mdb-interface/>
<https://blog.abrantix.com/webshop/product/mdb-to-raspberry-pi/> |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | My [pigpio](http://abyz.me.uk/rpi/pigpio/) library supports reading and writing 9-bit serial data. It uses bit banging so you can use any available GPIO.
If I remember correctly any speeds of 19.2 kbps or slower were pretty stable.
What bits per second do you need?
Reading ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioSerialReadOpen), [Python](http://abyz.me.uk/rpi/pigpio/python.html#bb_serial_read_open)) is slightly easier than writing ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioWaveAddSerial), [Python](http://abyz.me.uk/rpi/pigpio/python.html#wave_add_serial)). | You can't *RELIABLE* connect RPi serial directly to MDB bus due to 9-bit format and strict MDB timings. Messages between MDB peripheral and RPi need to be converted on-the-fly and in real-time.
Check this link it will help: [DIY MDB-UART converter](https://github.com/perdidor/Arduino-MDB-UART) |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | My [pigpio](http://abyz.me.uk/rpi/pigpio/) library supports reading and writing 9-bit serial data. It uses bit banging so you can use any available GPIO.
If I remember correctly any speeds of 19.2 kbps or slower were pretty stable.
What bits per second do you need?
Reading ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioSerialReadOpen), [Python](http://abyz.me.uk/rpi/pigpio/python.html#bb_serial_read_open)) is slightly easier than writing ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioWaveAddSerial), [Python](http://abyz.me.uk/rpi/pigpio/python.html#wave_add_serial)). | I was assigned a project to run a snack vending machine that uses MDB protocol for payment and I have completed the project using Pi Zero (Orange).
I have tried 9 bit hardware serial and software serial libraries and had timing problems on Pi Zero. MDB's 9bit serial communication became a pain. MDB protocol says peripherals should have %5 tolerance for serial communication timing however different peripheral vendors have different tolerances, not compliant with MDB protocol. When you think that you have accomplished serial communication but try a different vendor's payment peripheral, it just don't work. So don't rely on MDB protocol datasheet. I got sick and tired of implementing MDB controller for buggy vendors. Also some peripherals can drain excessive amonts of current from uart pins during their internal boot process and might damage your serial communication layer. So, you better use an abstraction. Optocouplers are fine but still I wouldn't recommend handling MDB serial communication using Pi Zero. Better way is to use a middle layer approach using an AVR.
Rather using Uart on Pi Zero for MDB communication, I used an Atmega328 AVR for MDB handling, polling etc. Atmega328 controls the MDB peripherals using Software Serial library and sends human readable data to Pi Zero on hardware serial. All electronics scheme, sources and Pi Zero Armbian image, Python code for Vending operations are available here:
<http://eliverse.com/content/vendiverse/>
You can check the wiki page for further details on controlling motors, product delivery sensors, coolers and character LCD displays. It is a complete vending machine controller project and it is being used by couple of vending machine producers. |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | My [pigpio](http://abyz.me.uk/rpi/pigpio/) library supports reading and writing 9-bit serial data. It uses bit banging so you can use any available GPIO.
If I remember correctly any speeds of 19.2 kbps or slower were pretty stable.
What bits per second do you need?
Reading ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioSerialReadOpen), [Python](http://abyz.me.uk/rpi/pigpio/python.html#bb_serial_read_open)) is slightly easier than writing ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioWaveAddSerial), [Python](http://abyz.me.uk/rpi/pigpio/python.html#wave_add_serial)). | All serial data is by definition, 1 bit. It is up to the interfaces reading *and* writing that data how to agree on how they interpret the bits to and from meaningful data.
If you want 9 bits of data, and a parity bit, and a stop and a start bit. Then it is up to you to convert your data into that format, and to interpret data you read in that format. The pigpio module mentioned in another answer will give you the hardware interfacing you need, or you can write your own. If you are developing in python, I suggest having a look at the bitString.py module by Scott Griffiths as a library that makes it fairly easy to manipulate bit based data. |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | My [pigpio](http://abyz.me.uk/rpi/pigpio/) library supports reading and writing 9-bit serial data. It uses bit banging so you can use any available GPIO.
If I remember correctly any speeds of 19.2 kbps or slower were pretty stable.
What bits per second do you need?
Reading ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioSerialReadOpen), [Python](http://abyz.me.uk/rpi/pigpio/python.html#bb_serial_read_open)) is slightly easier than writing ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioWaveAddSerial), [Python](http://abyz.me.uk/rpi/pigpio/python.html#wave_add_serial)). | The accepted answer, which used an Atmel processor with true 9 bit data URT communicating with a Pi looks to have been either abandoned, or taken commerical.
So I am going with <https://www.vendingtools.ro/en> for Eur 70, and that will interface my Pi to the MDB 9 data bit bus.
---
[Update]
See also
<https://www.qibixx.com/en/products/mdb-interface/>
<https://blog.abrantix.com/webshop/product/mdb-to-raspberry-pi/> |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | I also made complete demo for 9-bit UART emulation (based on even/odd parity). You can find it <http://bohdan-danishevsky.blogspot.com/2016/10/9-bit-serial-communication-in-linux.html>.
All sources available on git.
You can easily adapt it for your device. Hope you like it. | All serial data is by definition, 1 bit. It is up to the interfaces reading *and* writing that data how to agree on how they interpret the bits to and from meaningful data.
If you want 9 bits of data, and a parity bit, and a stop and a start bit. Then it is up to you to convert your data into that format, and to interpret data you read in that format. The pigpio module mentioned in another answer will give you the hardware interfacing you need, or you can write your own. If you are developing in python, I suggest having a look at the bitString.py module by Scott Griffiths as a library that makes it fairly easy to manipulate bit based data. |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | I was assigned a project to run a snack vending machine that uses MDB protocol for payment and I have completed the project using Pi Zero (Orange).
I have tried 9 bit hardware serial and software serial libraries and had timing problems on Pi Zero. MDB's 9bit serial communication became a pain. MDB protocol says peripherals should have %5 tolerance for serial communication timing however different peripheral vendors have different tolerances, not compliant with MDB protocol. When you think that you have accomplished serial communication but try a different vendor's payment peripheral, it just don't work. So don't rely on MDB protocol datasheet. I got sick and tired of implementing MDB controller for buggy vendors. Also some peripherals can drain excessive amonts of current from uart pins during their internal boot process and might damage your serial communication layer. So, you better use an abstraction. Optocouplers are fine but still I wouldn't recommend handling MDB serial communication using Pi Zero. Better way is to use a middle layer approach using an AVR.
Rather using Uart on Pi Zero for MDB communication, I used an Atmega328 AVR for MDB handling, polling etc. Atmega328 controls the MDB peripherals using Software Serial library and sends human readable data to Pi Zero on hardware serial. All electronics scheme, sources and Pi Zero Armbian image, Python code for Vending operations are available here:
<http://eliverse.com/content/vendiverse/>
You can check the wiki page for further details on controlling motors, product delivery sensors, coolers and character LCD displays. It is a complete vending machine controller project and it is being used by couple of vending machine producers. | The accepted answer, which used an Atmel processor with true 9 bit data URT communicating with a Pi looks to have been either abandoned, or taken commerical.
So I am going with <https://www.vendingtools.ro/en> for Eur 70, and that will interface my Pi to the MDB 9 data bit bus.
---
[Update]
See also
<https://www.qibixx.com/en/products/mdb-interface/>
<https://blog.abrantix.com/webshop/product/mdb-to-raspberry-pi/> |
38,642 | I have a powered USB hub. Can I use it to power the Raspberry Pi Zero using up just one Micro USB port?
Or does the power have to come from one particular port that is only used to provide power?
Edit (2021-10): What i'm trying to ask is, can i use the micro USB port for both power and data simultaneously? This is still unclear. | 2015/11/27 | [
"https://raspberrypi.stackexchange.com/questions/38642",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/37313/"
] | My [pigpio](http://abyz.me.uk/rpi/pigpio/) library supports reading and writing 9-bit serial data. It uses bit banging so you can use any available GPIO.
If I remember correctly any speeds of 19.2 kbps or slower were pretty stable.
What bits per second do you need?
Reading ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioSerialReadOpen), [Python](http://abyz.me.uk/rpi/pigpio/python.html#bb_serial_read_open)) is slightly easier than writing ([C](http://abyz.me.uk/rpi/pigpio/cif.html#gpioWaveAddSerial), [Python](http://abyz.me.uk/rpi/pigpio/python.html#wave_add_serial)). | I also made complete demo for 9-bit UART emulation (based on even/odd parity). You can find it <http://bohdan-danishevsky.blogspot.com/2016/10/9-bit-serial-communication-in-linux.html>.
All sources available on git.
You can easily adapt it for your device. Hope you like it. |
4,349 | Accepted answers on controversial topics on Meta pose a very real problem.
* They give an advantage to whatever position the asker of the question supports, as he can mark that as accepted and it will bubble to the top, whether or not it's the answer that the most people agree with, regardless of how well it is actually supported.
* To newer users, accepted answers on meta appear to be the correct answer, and they might believe that it marks the policy that the community has decided to go with.
* Accepted answers give the appearance that the matter is closed and no longer up for discussion.
I'm a relative newbie, but have been around and very active for the last 4 months. I very much doubt I was alone in [not understanding that accepted answers are generally meaningless on Meta](https://gaming.meta.stackexchange.com/questions/4269/what-does-an-accepted-answer-mean-in-the-context-of-meta-discussions).
For one example of this, in the topic of [What's the difference between FortressCraft and Minecraft? Who cares?](https://gaming.meta.stackexchange.com/questions/2853/whats-the-difference-between-fortresscraft-and-minecraft-who-cares), the community seems to have yet to reach much of a consensus, but an answer is already marked as accepted. As mentioned above, I think this makes that position seem "official" to newer users who don't know better, and discourages them from participating in a discussion that already seems to be closed.
My original request was that I would like to see the ability to accept an answer removed from Meta, but I was very quickly convinced that this was too harsh an approach. If, [as indicated](https://gaming.meta.stackexchange.com/a/4270/3062), it accepted answers truly mean nothing, I'd like to see them not bubble to the top of the sort order. All it does is serve create confusion and bias.
Update: My example question was kind of bad, since it was originally opened in September 2011 and just now came up for additional discussion. So I don't mean to imply that the asker is doing anything tricky to support his position. But the principle holds. | 2012/03/26 | [
"https://gaming.meta.stackexchange.com/questions/4349",
"https://gaming.meta.stackexchange.com",
"https://gaming.meta.stackexchange.com/users/3062/"
] | I think they mean something on questions asked by moderators looking for consensus on a topic.
For example, in the last ITG debate, where agent86 made the process through meta questions, the accepted solution was then accepted as official.
If we have a problem with a particular accepted answer on a meta discussion (a problem that's disruptive) that does not represent an official solution, we can always delete it. | A part of Meta's system is Eating our own Dogfood even if it isn't 100% perfectly suited to unanswerable discussions. The Q/A format isn't perfect here but it's familiar and it works well when people play nice.
The top answer is obviously most important on the main site and casual users will just skim for the accepted/top answer, and we want that. But meta's not like that; if you're here, you're probably an advanced user who's enganged and willing/wanting to read the whole conversation.
I don't think Meta needs to be optimized for scanning like the main site's questions; we do that for correctness and anonymous visitors. Meta doesn't really get anonymous visitors and often we don't have "correctness" either. I don't think it's a problem worth messing with the system for. |
4,349 | Accepted answers on controversial topics on Meta pose a very real problem.
* They give an advantage to whatever position the asker of the question supports, as he can mark that as accepted and it will bubble to the top, whether or not it's the answer that the most people agree with, regardless of how well it is actually supported.
* To newer users, accepted answers on meta appear to be the correct answer, and they might believe that it marks the policy that the community has decided to go with.
* Accepted answers give the appearance that the matter is closed and no longer up for discussion.
I'm a relative newbie, but have been around and very active for the last 4 months. I very much doubt I was alone in [not understanding that accepted answers are generally meaningless on Meta](https://gaming.meta.stackexchange.com/questions/4269/what-does-an-accepted-answer-mean-in-the-context-of-meta-discussions).
For one example of this, in the topic of [What's the difference between FortressCraft and Minecraft? Who cares?](https://gaming.meta.stackexchange.com/questions/2853/whats-the-difference-between-fortresscraft-and-minecraft-who-cares), the community seems to have yet to reach much of a consensus, but an answer is already marked as accepted. As mentioned above, I think this makes that position seem "official" to newer users who don't know better, and discourages them from participating in a discussion that already seems to be closed.
My original request was that I would like to see the ability to accept an answer removed from Meta, but I was very quickly convinced that this was too harsh an approach. If, [as indicated](https://gaming.meta.stackexchange.com/a/4270/3062), it accepted answers truly mean nothing, I'd like to see them not bubble to the top of the sort order. All it does is serve create confusion and bias.
Update: My example question was kind of bad, since it was originally opened in September 2011 and just now came up for additional discussion. So I don't mean to imply that the asker is doing anything tricky to support his position. But the principle holds. | 2012/03/26 | [
"https://gaming.meta.stackexchange.com/questions/4349",
"https://gaming.meta.stackexchange.com",
"https://gaming.meta.stackexchange.com/users/3062/"
] | To echo [Ben's answer](https://gaming.meta.stackexchange.com/a/4351/3389) somewhat, you sorta need to be a little more discerning when reading meta discussions. Meta should be inviting, but it's not the place to get quick rules of thumb or policies: that's what [the FAQ](https://gaming.stackexchange.com/faq) is for.
The ideal meta question—where the top answer exactly matches unassailable policy and should be treated like the word of god—doesn't exist. Every decision requires some interpretation and asking questions of more experienced users. Yes, meta is messy, but I think that's by design: this is where the sausage is made. Discussions are messy. Guidelines and policies made by humans are messy.
When trying to understand a meta discussion, sort order, scores, and accepted answers are less important than what the general arguments are. In most situations, however, the accepted answer is pretty close to the summary of consensus and the scores reflect what position should be (or has been) implemented.
Sometimes it's not, and many times there are good tells: the accepted answer has a bunch of negative comments, or is down voted, or is just inferior to the other arguments presented in other answers. Sometimes, however, it's not clear what the consensus is without reading everything and coming up with a determination for oneself.
But in trying to solve that edge case by removing accepted answers from being shown first because they might be contrary to what's been implemented, you introduce a new one: if something gets implemented long after the discussion happens (so by default it doesn't have as many votes), or if something gets implemented in a way that's unpopular, the canonical answer gets buried amongst all the noise of the surrounding discussion.
I'd rather keep the tool we have—and is used in good faith by most people to push the canonical or consensus answer to the top—rather than remove it in favor of sorting everything by votes. Stack Exchange isn't a popularity contest, and there's no reason why the messy discussions we have should be treated like one either. | I think they mean something on questions asked by moderators looking for consensus on a topic.
For example, in the last ITG debate, where agent86 made the process through meta questions, the accepted solution was then accepted as official.
If we have a problem with a particular accepted answer on a meta discussion (a problem that's disruptive) that does not represent an official solution, we can always delete it. |
4,349 | Accepted answers on controversial topics on Meta pose a very real problem.
* They give an advantage to whatever position the asker of the question supports, as he can mark that as accepted and it will bubble to the top, whether or not it's the answer that the most people agree with, regardless of how well it is actually supported.
* To newer users, accepted answers on meta appear to be the correct answer, and they might believe that it marks the policy that the community has decided to go with.
* Accepted answers give the appearance that the matter is closed and no longer up for discussion.
I'm a relative newbie, but have been around and very active for the last 4 months. I very much doubt I was alone in [not understanding that accepted answers are generally meaningless on Meta](https://gaming.meta.stackexchange.com/questions/4269/what-does-an-accepted-answer-mean-in-the-context-of-meta-discussions).
For one example of this, in the topic of [What's the difference between FortressCraft and Minecraft? Who cares?](https://gaming.meta.stackexchange.com/questions/2853/whats-the-difference-between-fortresscraft-and-minecraft-who-cares), the community seems to have yet to reach much of a consensus, but an answer is already marked as accepted. As mentioned above, I think this makes that position seem "official" to newer users who don't know better, and discourages them from participating in a discussion that already seems to be closed.
My original request was that I would like to see the ability to accept an answer removed from Meta, but I was very quickly convinced that this was too harsh an approach. If, [as indicated](https://gaming.meta.stackexchange.com/a/4270/3062), it accepted answers truly mean nothing, I'd like to see them not bubble to the top of the sort order. All it does is serve create confusion and bias.
Update: My example question was kind of bad, since it was originally opened in September 2011 and just now came up for additional discussion. So I don't mean to imply that the asker is doing anything tricky to support his position. But the principle holds. | 2012/03/26 | [
"https://gaming.meta.stackexchange.com/questions/4349",
"https://gaming.meta.stackexchange.com",
"https://gaming.meta.stackexchange.com/users/3062/"
] | To echo [Ben's answer](https://gaming.meta.stackexchange.com/a/4351/3389) somewhat, you sorta need to be a little more discerning when reading meta discussions. Meta should be inviting, but it's not the place to get quick rules of thumb or policies: that's what [the FAQ](https://gaming.stackexchange.com/faq) is for.
The ideal meta question—where the top answer exactly matches unassailable policy and should be treated like the word of god—doesn't exist. Every decision requires some interpretation and asking questions of more experienced users. Yes, meta is messy, but I think that's by design: this is where the sausage is made. Discussions are messy. Guidelines and policies made by humans are messy.
When trying to understand a meta discussion, sort order, scores, and accepted answers are less important than what the general arguments are. In most situations, however, the accepted answer is pretty close to the summary of consensus and the scores reflect what position should be (or has been) implemented.
Sometimes it's not, and many times there are good tells: the accepted answer has a bunch of negative comments, or is down voted, or is just inferior to the other arguments presented in other answers. Sometimes, however, it's not clear what the consensus is without reading everything and coming up with a determination for oneself.
But in trying to solve that edge case by removing accepted answers from being shown first because they might be contrary to what's been implemented, you introduce a new one: if something gets implemented long after the discussion happens (so by default it doesn't have as many votes), or if something gets implemented in a way that's unpopular, the canonical answer gets buried amongst all the noise of the surrounding discussion.
I'd rather keep the tool we have—and is used in good faith by most people to push the canonical or consensus answer to the top—rather than remove it in favor of sorting everything by votes. Stack Exchange isn't a popularity contest, and there's no reason why the messy discussions we have should be treated like one either. | A part of Meta's system is Eating our own Dogfood even if it isn't 100% perfectly suited to unanswerable discussions. The Q/A format isn't perfect here but it's familiar and it works well when people play nice.
The top answer is obviously most important on the main site and casual users will just skim for the accepted/top answer, and we want that. But meta's not like that; if you're here, you're probably an advanced user who's enganged and willing/wanting to read the whole conversation.
I don't think Meta needs to be optimized for scanning like the main site's questions; we do that for correctness and anonymous visitors. Meta doesn't really get anonymous visitors and often we don't have "correctness" either. I don't think it's a problem worth messing with the system for. |
56,659 | Downloading the zip for Raspbian from <https://www.raspberrypi.org/downloads/raspbian/> is taking hours. There is also a small torrent file that can be downloaded. How can this be used to install Raspbian? | 2016/10/22 | [
"https://raspberrypi.stackexchange.com/questions/56659",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/35590/"
] | This works on Pi3
Create 3 shell scripts, one for each service.
[](https://i.stack.imgur.com/CYiXh.jpg)
I run this from the GUI (open in terminal) and works fine.
You still need to enter auth credentials though.
You will need to change paths - my files are all in Home/pi
hope that helps | In order to do this headless, kicking off three terminals I'd look into [crontab](https://unix.stackexchange.com/questions/57852/crontab-job-start-1-min-after-reboot). Otherwise you should be able to run the script as a [daemon](https://stackoverflow.com/questions/19233529/run-bash-script-as-daemon). Remember that Raspbian is just Linux, and that you'll have luck in finding answers when googling by replacing Raspberry Pi with Linux. |
82,147 | Well, I just upgraded one of my machines to windows 7, and it turns out that my usb wireless adapter just isn't working. There is no driver for it in windows 7 (yes, I've looked everywhere). So I'm willing to buy a new one, but want to make sure it's gonna work perfect. So it either needs to have a windows 7 driver or have a vista driver that works under windows 7.
I checked my local hardware store, and there are no adapters there with drivers for windows 7, but there are some that have drivers for vista. I don't mind using a vista driver, as long as it works OK. I'd hate to have a buggy internet connection. These are the ones I found that have vista drivers:
Encore Eletronics: [ENUWI-G](http://www.encore-usa.com/product_item.php?region=us&bid=2&pgid=4&pid=18)
Encore Eletronics: [ENUWI-N](http://www.encore-usa.com/product_item.php?region=us&bid=2&pgid=2&pid=277)
Linksys: [WUSB300N](http://www.linksysbycisco.com/US/en/support/WUSB300N)
Linksys: [WUSB54GC v1.0](http://www.linksysbycisco.com/US/en/support/WUSB54GC)
So I ask thee: **Has anyone here tested one of these Wifi USB adapters in windows 7 64bit? Did it work? If it did, did you have to do anything special?**
Thanks in advance. I tried asking the actual salespeople at the store, but they don't know.
EDIT: Forgot to say, it's 64bit. | 2009/12/12 | [
"https://superuser.com/questions/82147",
"https://superuser.com",
"https://superuser.com/users/6371/"
] | i have the ENUWI-N USB WLAN adapter and it works fine with Windows 7 x64 | Right in the linksys pages that you linked is this Windows 7 compatibility list for their adapters...
<http://linksys.custhelp.com/cgi-bin/linksys.cfg/php/enduser/std_adp.php?p_faqid=20872> |
82,147 | Well, I just upgraded one of my machines to windows 7, and it turns out that my usb wireless adapter just isn't working. There is no driver for it in windows 7 (yes, I've looked everywhere). So I'm willing to buy a new one, but want to make sure it's gonna work perfect. So it either needs to have a windows 7 driver or have a vista driver that works under windows 7.
I checked my local hardware store, and there are no adapters there with drivers for windows 7, but there are some that have drivers for vista. I don't mind using a vista driver, as long as it works OK. I'd hate to have a buggy internet connection. These are the ones I found that have vista drivers:
Encore Eletronics: [ENUWI-G](http://www.encore-usa.com/product_item.php?region=us&bid=2&pgid=4&pid=18)
Encore Eletronics: [ENUWI-N](http://www.encore-usa.com/product_item.php?region=us&bid=2&pgid=2&pid=277)
Linksys: [WUSB300N](http://www.linksysbycisco.com/US/en/support/WUSB300N)
Linksys: [WUSB54GC v1.0](http://www.linksysbycisco.com/US/en/support/WUSB54GC)
So I ask thee: **Has anyone here tested one of these Wifi USB adapters in windows 7 64bit? Did it work? If it did, did you have to do anything special?**
Thanks in advance. I tried asking the actual salespeople at the store, but they don't know.
EDIT: Forgot to say, it's 64bit. | 2009/12/12 | [
"https://superuser.com/questions/82147",
"https://superuser.com",
"https://superuser.com/users/6371/"
] | i have the ENUWI-N USB WLAN adapter and it works fine with Windows 7 x64 | At first, the WUSB54GC v1.0 did work with Windows 7 64 bit. But later it went into a code 10. Microsoft do have a list of wireless adapters that will work with Windows 7. Also, they are working on drivers for the WUSB54GC to work on the Windows 7 64 bit. |
151,387 | I Accepted two conditional job offers from two large investment banks. Both of these companies use the same background check company (First Advantage)
Is there a risk the background check company tells one company that I have accepted an offer from another company? | 2020/01/16 | [
"https://workplace.stackexchange.com/questions/151387",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/113652/"
] | Go ahead and accept conditional job offers, until you have an unconditional and confirmed offer in hand.
A conditional job offer is **not** a guaranteed one - there is no reason to consider it as one, most likely the other party (the company offering the job) is not.
However, once you have accepted an unconditional offer and have it signed and sealed, you should not be accepting any more offers and communicate cancellation to any outstanding conditional offers.
That said, background checks are considered very sensitive process, and unless it is actually related to the process, the identity of the person undergoing the check is not revealed to anyone in any matter. Only the results are shared with the organization requesting for the check.
And moreover, any conditional job offer that you have today is usually not part of a background check anyways, it's about the previous engagements, if at all. | >
> I there a risk the background check company tells one company that I have accepted an offer from another company?
>
>
>
There is always a risk depending on how extensive of a background check either of the companies have asked the background check company to perform. To prevent such incidents in the future, you should not simultaneously accept multiple job offers. You could potentially lose both offers if the companies are aware of what you have done. |
18,147 | Tempest devices could via intercepting from far away the electro-magnetic waves coming out from one's computer detect what one does on the computer. Employing the Faraday cage principle, one could prevent such detections. A. Stanoyevitch in his book Discrete Structures with Contemporary Applications, CRC Press, 2011, p.301 writes:
"Buildings can be fitted using a special insulation procedure that protects against tempest devices, but any company or individual in the US who has this insulation must first obtain a license from the federal Government."
Could this be true? If yes, which laws are involved in this context? | 2017/04/03 | [
"https://law.stackexchange.com/questions/18147",
"https://law.stackexchange.com",
"https://law.stackexchange.com/users/4642/"
] | The only applicable law for the US would seem to be FCC regulations against interfering with "any radio communications of any station licensed or authorized". [Marriott Hotels](https://apps.fcc.gov/edocs_public/attachmatch/DA-14-1444A1_Rcd.pdf) was fined substantially for blocking people's hotspots within the hotel. As I understand "tempest device", this is a method of obtaining information from stray EMR from computers and the like, which picks up on radiation emitted from a device. Stray radiation is not licensed communication. The problem would be if a room were insulated to block all EM radiation coming in or going out, including licensed transmissions. The main enforcement interest of FCC is with devices that actively jam signals (by transmitting interference). But the law, [47 USC 333](https://www.law.cornell.edu/uscode/text/47/333), states that "No person shall willfully or maliciously interfere with or cause interference to any radio communications of any station licensed or authorized by or under this chapter or operated by the United States Government". Insulating a room from incoming licensed radio signals is thus not allowed, under a "what the law says" reading of the law. However, laws are frequently interpreted in terms of beliefs about an intent, so it's possible that a court could eventually decide that the law was not intended to prevent construction of a Faraday cage room. The issue would then be whether an intentional action to do something allowed which has an (unintended) disallowed but non-harmful consequence would be deemed a violation (the user would be okay with blocking licensed transmissions into the room). I don't know of clearly applicable case law on that point. | A Faraday cage can prevent EM emissions that can (in theory) be used to eavesdrop on electronic devices. But no, there is no U.S. law against putting metallic screens around a room or within a building. Just as there is no U.S. law against putting shades in windows that would prevent visual surveillance of anything inside. |
44,204,878 | Can someone explain how I can configure Azure cosmosDB to not be so expensive? 30€ for 3 days??
It's a small test environment with very few users.
In the backend, I only see that it will cost some cents ...
Here is my Cost overview:
[](https://i.stack.imgur.com/agu6V.png)
I have around 18 collections but all are only kbits big.
It the backend database for the js cms : <http://keystonejs.com/>
Example scaling for one collection in azure:
[](https://i.stack.imgur.com/hmTPf.png) | 2017/05/26 | [
"https://Stackoverflow.com/questions/44204878",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/6506286/"
] | The problem is that you are billed by collection. So even if you use only a little storage space and you make only a couple of requests, you have to pay the minimum RU per document collection. | You created 18 collections, meaning you have 18 billable database constructs. And with the lowest-cost collection starting at roughly 20 Euro per month (about 5 Euro per 100 RU per month, minimum 400 RU), the math works out: You basically created an environment with 400\*18=7200 RU. Perhaps just use fewer collections to reduce your cost footprint (note: collections have no rules around documents being homogeneous). You could use a single collection, since you have such a low amount of data.
**EDIT FEB 2018** - note: with database-level RUs (a feature added a few months after the OPs question was posted), the cost model would be very different. 18 collections would be able to share an 1800RU database-level allocation (the minimum is 400RU, up to 4 collections, then 100RU per additional collection). With database-level RU, the cost of an 18-collection configuration would start at 25% of the original cost, based on 400RU per collection. |
98,534 | In our current adventures my PCs are in a specific very large City. By searching, they found a specific black market shop, the route leading to which was a bit complicated.
Can they roll an intelligence check to determine if they remember the complex route, or do I have to describe the roads one by one until they find it again (which is not a problem)?
I try to run the campaign as realistic as possible and this situation is like a common real life problem. We tend to use marks along the way (shops, signs) to remember specific routes, but a lot of times we fail because we don't remember minor details (which is described as high check). | 2017/04/21 | [
"https://rpg.stackexchange.com/questions/98534",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/33564/"
] | The DMG (p.237) says, that **Intelligence checks are used for memory and reason**, so this would be a fitting check.
>
> [...] or do I have to describe the roads one by one until they find it again (which is not a problem)
>
>
>
This would be a *"test the characters, not the players"* situation. I would strongly discourage you from doing so, except you aim for exactly that.
>
> We tend to use marks along the way (shops, signs) to remember specific routes, but a lot of times we fail because we don't remember minor details (which is described as high check).
>
>
>
Maybe give them advantage on the Intelligence check if they describe you how they remember the way or let them use **Investigation checks to find their way back on track if they get lost.**
All in all, if there are no consequences for failed checks, I would like to refer to the **DMG (p.237)** which recommends that you assume, if they have got the time, they succeed automatically taking ten times the normal amount of time needed to get there. | What is the consequence of failure?
===================================
Let's assume that they fail. What happens now? I'm guessing that they will just use whatever method caused them to find it in the first place to find it in the second place. If so: **why are we wasting valuable playing time on this?**
Only if you decide that there *is* a consequence for failure should you bother making any type of check at all.
[Old school or new school?](http://www.lulu.com/items/volume_63/3019000/3019374/1/print/3019374.pdf)
----------------------------------------------------------------------------------------------------
Do you want to test the players (old school) or the characters (new school) or a bit of both [see Zen Moment 2]?
**Neither method is better than the other: it is purely a matter of *your* group's preference!**
If you want to test the players then describe what they see and they have to rely on their own memories. If you want to test the characters then an Intelligence check is appropriate with whatever proficiency the player can make a case for - Perception and Investigation spring to mind. |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | One method that has worked for me if you desperately need to use that cork again is to shave the end with a knife to create a taper so that it will slide into the opening of the bottle and then with the pressure of your hand you can squeeze it down in.
I always keep a set of reusable rubber corks around though for this very problem. They're cheap, come in sets of 2-4 and will fit pretty much any bottle. | I can sometimes get them back in by flipping them over and using the corkscrew end.
But if you can, you're better off using one of those rubber stoppers with a vacuum pump (e.g. "vacu-vin"). It will remove a lot of the air, preventing the wine from oxidizing as much. |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | One method that has worked for me if you desperately need to use that cork again is to shave the end with a knife to create a taper so that it will slide into the opening of the bottle and then with the pressure of your hand you can squeeze it down in.
I always keep a set of reusable rubber corks around though for this very problem. They're cheap, come in sets of 2-4 and will fit pretty much any bottle. | Flip it over and insert the dry end back in the bottle. I do it all the time! |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | Saw a bit of diameter off the cork with a bread knife and happily plug it back in (it may not seal it completely but it'll do overnight) | Flip it over and insert the dry end back in the bottle. I do it all the time! |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | There is no trick, it just won't work. Synthetic corks are popular as replacements to cork not only because they are cheaper, but more effective at preserving wine as they don't dry out, and they expand more in the neck keeping a tighter seal. This makes it more difficult, if not impossible, to get them back in.
The simple and easy solution is to buy re-usable bottle stoppers. There are many different types, I prefer the ones which have a lever or button you push down to expand the stopper as they are best at preventing spills and leaks. There are vacuum sealers as well, however IMHO they're gimmicks and don't improve the storage of wine. | I just tried it and got it back in. Put the cork in at a 30-45 degree angle and keep applying pressure. This works best if you can sit down with your feet under you butt and keep the bottle tight between your thighs. Apply pressure and take a butter knife and keep working the edges in. Once in just put your weight on the cork and ta da! |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | One method that has worked for me if you desperately need to use that cork again is to shave the end with a knife to create a taper so that it will slide into the opening of the bottle and then with the pressure of your hand you can squeeze it down in.
I always keep a set of reusable rubber corks around though for this very problem. They're cheap, come in sets of 2-4 and will fit pretty much any bottle. | When I have tried and am alone with no muscle, I have tried every trick internet search gives.... not even warm/hot water worked for me with plastic corks in large champagne bottles... what worked at end of day even after going to buy a whole other bottle and its cork also not giving way.... butter knife!
Just keep prying under cork with a butter knife (it may bend- don't use one that may be missed if you bend it!). Just continue to pry each side you can get it up into the cork and sure enough pried way to which I could then pull off the cork! |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | I can sometimes get them back in by flipping them over and using the corkscrew end.
But if you can, you're better off using one of those rubber stoppers with a vacuum pump (e.g. "vacu-vin"). It will remove a lot of the air, preventing the wine from oxidizing as much. | I shaved the plastic rind off with a scissors, squeezed it in as far as I could with my hands, then put the hard plastic bottom of a bottle of ibuprofen on top of the cork and pressed down. I imagine any kind of grippable, flat and hard thing would work. The ibuprofen bottle made it much easier to apply force once the cork was in a workable position. |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | Saw a bit of diameter off the cork with a bread knife and happily plug it back in (it may not seal it completely but it'll do overnight) | Why not expand the wine bottle head using heat then fit the cork back in maybe that should work |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | I can sometimes get them back in by flipping them over and using the corkscrew end.
But if you can, you're better off using one of those rubber stoppers with a vacuum pump (e.g. "vacu-vin"). It will remove a lot of the air, preventing the wine from oxidizing as much. | I just tried it and got it back in. Put the cork in at a 30-45 degree angle and keep applying pressure. This works best if you can sit down with your feet under you butt and keep the bottle tight between your thighs. Apply pressure and take a butter knife and keep working the edges in. Once in just put your weight on the cork and ta da! |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | I shaved the plastic rind off with a scissors, squeezed it in as far as I could with my hands, then put the hard plastic bottom of a bottle of ibuprofen on top of the cork and pressed down. I imagine any kind of grippable, flat and hard thing would work. The ibuprofen bottle made it much easier to apply force once the cork was in a workable position. | Why not expand the wine bottle head using heat then fit the cork back in maybe that should work |
33,242 | The past couple times I've bought wine stopped with a [synthetic cork](http://en.wikipedia.org/wiki/File%3aVinova_synthetic_wine_closure.jpg), I've had a very difficult time reinserting the cork after opening the bottle. It seems the cork expands after leaving the bottle, and it's made of such a rigid material that sometimes I can't squeeze it back in. (And no, inserting the back end doesn't work, as it sometimes does with real cork.)
Is there a trick to getting an expanded synthetic cork back in a wine bottle? | 2013/04/05 | [
"https://cooking.stackexchange.com/questions/33242",
"https://cooking.stackexchange.com",
"https://cooking.stackexchange.com/users/7107/"
] | I just tried it and got it back in. Put the cork in at a 30-45 degree angle and keep applying pressure. This works best if you can sit down with your feet under you butt and keep the bottle tight between your thighs. Apply pressure and take a butter knife and keep working the edges in. Once in just put your weight on the cork and ta da! | Why not expand the wine bottle head using heat then fit the cork back in maybe that should work |
54,160,011 | Fairly new to capybara and cucumber (and testing in general) here.
I am doing some black box acceptance testing and I need to test a page that can only be reached by typing in the url.
That page requires authentication so when the page is reached through a call to visit() I am redirected to the login page.
The problem is after logging in the app does not redirect me back to that page but goes to the homepage instead so I can't just fill in the login form and move on.
I have tried logging in first and then performing a call to that page but it seems the session is cleared after any call to visit().
How can I keep the session after a call to visit, or solve this problem in any other way?
Ideally I would like to log in once and visit a bunch of pages only reachable by typing the uri, without having to log in again and again. Is that possible?
I am only using the chrome selenium driver but at some point will use a headless driver.
Thanks! | 2019/01/12 | [
"https://Stackoverflow.com/questions/54160011",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1242439/"
] | There might be a signing issue. Is your app signed? The default signing is the debug signing. When you just build and deploy or run it on a connected device, it is signed by default as a debug sign.
If it is **not** uniquely signed, then running or installing on other devices might not be possible. Look at the app signing instruction and guide by Google, [here](https://developer.android.com/studio/publish/app-signing), if you wish to distribute your app on other devices.
Another problem might be in the security authorization of other devices. Your device on which you are building has the developer options turned on. To run and install unsigned apps you'll need to activate the developer options in the device you want to run it on. Look at this [link](https://developer.android.com/studio/debug/dev-options) about Developer Options and how to turn it on. | by default the Android Studio packages just the needed files and installs the app in your mobile. If you share to other mobiles, the app will not install in few devices due to a few reasons like OS version, files mismatch, SHA keys mismatch, etc.
However, if you build the APK and share, the APK is equipped to be installed in any device starting from the minimum SDK version to the target SDK version. |
646,806 | So I'm trying to [make a pulse sensor using TCRT5000](http://www.ee.iitb.ac.in/%7Estallur/wp-content/uploads/2017/02/Heart-Rate-Measurement-using-PPG1.pdf) IR sensor and arduino UNO (giving out 4.8V really from 5V pin) (Baud Rate 19200), and I'm supposed to get a pulse plot like this
[](https://i.stack.imgur.com/iileQ.png)
But I'm getting it like this
[](https://i.stack.imgur.com/m3N1I.png)
As you can see, absolutely no pulse is noticeable.
Below is the schematic
[](https://i.stack.imgur.com/H5vVo.png)
Below are some essential points:
1. In the phototransistor stage, I don't understand the symbol with the Arrow on the extreme left, I thought it was a Pot set to 1K Ohm so I used a 1K resistor between the 150 Ohm resistor and +5V.
2. My 1uF capacitor in the second stage is tantalum. My second capacitor, the 4.7uF, is electrolytic.
3. The schematic says the pin 11 of the LM324 should be connected to -5V, since I don't know how to get -5V out of Arduino I just connected it to GND on Arduino.
4. I did not have a IN4004 diode so I used a IN4007.
Where am I going wrong? | 2022/12/18 | [
"https://electronics.stackexchange.com/questions/646806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/323019/"
] | If you need a negative voltage, you can use this (only V2- part of schematic, D2-D3-C10-C11) or use outputs of Arduino with software ...
[](https://i.stack.imgur.com/IKgqv.png) | Antonio51 has given the right solution, but I found to my application the use of a TC7660 IC more suitable with 10uF Capacitor (as advised on another forum). But since Antonio's answer is correct and the source for my further inquiry, that stays as the solution. |
646,806 | So I'm trying to [make a pulse sensor using TCRT5000](http://www.ee.iitb.ac.in/%7Estallur/wp-content/uploads/2017/02/Heart-Rate-Measurement-using-PPG1.pdf) IR sensor and arduino UNO (giving out 4.8V really from 5V pin) (Baud Rate 19200), and I'm supposed to get a pulse plot like this
[](https://i.stack.imgur.com/iileQ.png)
But I'm getting it like this
[](https://i.stack.imgur.com/m3N1I.png)
As you can see, absolutely no pulse is noticeable.
Below is the schematic
[](https://i.stack.imgur.com/H5vVo.png)
Below are some essential points:
1. In the phototransistor stage, I don't understand the symbol with the Arrow on the extreme left, I thought it was a Pot set to 1K Ohm so I used a 1K resistor between the 150 Ohm resistor and +5V.
2. My 1uF capacitor in the second stage is tantalum. My second capacitor, the 4.7uF, is electrolytic.
3. The schematic says the pin 11 of the LM324 should be connected to -5V, since I don't know how to get -5V out of Arduino I just connected it to GND on Arduino.
4. I did not have a IN4004 diode so I used a IN4007.
Where am I going wrong? | 2022/12/18 | [
"https://electronics.stackexchange.com/questions/646806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/323019/"
] | If you need a negative voltage, you can use this (only V2- part of schematic, D2-D3-C10-C11) or use outputs of Arduino with software ...
[](https://i.stack.imgur.com/IKgqv.png) | Regarding the question first part, the ideal method needs to be to check the output of each stage progressively and make alterations till that expected output is achieved. It was discovered that 1K pot was not set at 1K and needed to be altered to receive the expected output at stage 1. In this case the 1k pot needed to be turned to zero. |
646,806 | So I'm trying to [make a pulse sensor using TCRT5000](http://www.ee.iitb.ac.in/%7Estallur/wp-content/uploads/2017/02/Heart-Rate-Measurement-using-PPG1.pdf) IR sensor and arduino UNO (giving out 4.8V really from 5V pin) (Baud Rate 19200), and I'm supposed to get a pulse plot like this
[](https://i.stack.imgur.com/iileQ.png)
But I'm getting it like this
[](https://i.stack.imgur.com/m3N1I.png)
As you can see, absolutely no pulse is noticeable.
Below is the schematic
[](https://i.stack.imgur.com/H5vVo.png)
Below are some essential points:
1. In the phototransistor stage, I don't understand the symbol with the Arrow on the extreme left, I thought it was a Pot set to 1K Ohm so I used a 1K resistor between the 150 Ohm resistor and +5V.
2. My 1uF capacitor in the second stage is tantalum. My second capacitor, the 4.7uF, is electrolytic.
3. The schematic says the pin 11 of the LM324 should be connected to -5V, since I don't know how to get -5V out of Arduino I just connected it to GND on Arduino.
4. I did not have a IN4004 diode so I used a IN4007.
Where am I going wrong? | 2022/12/18 | [
"https://electronics.stackexchange.com/questions/646806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/323019/"
] | Antonio51 has given the right solution, but I found to my application the use of a TC7660 IC more suitable with 10uF Capacitor (as advised on another forum). But since Antonio's answer is correct and the source for my further inquiry, that stays as the solution. | Regarding the question first part, the ideal method needs to be to check the output of each stage progressively and make alterations till that expected output is achieved. It was discovered that 1K pot was not set at 1K and needed to be altered to receive the expected output at stage 1. In this case the 1k pot needed to be turned to zero. |
47,431 | I use hangouts on Android and on the desktop (browser extension) and these messages periodically appear in my Gmail inbox. So every day I'm having to manually archive all these hangout messages that have appeared in my inbox.
(In my case it is really just 1 hangout with 1 person. Opening any of these hangout messages gives me access to the entire hangout, just at a different point in the conversation.)
How can I auto-archive these hangout messages in Gmail? There does not seem to be a "hangouts" filter?
I've noticed [this question](https://webapps.stackexchange.com/questions/47370/gmail-problem-with-hangout-and-filters "Gmail - problem with Hangout and filters") (asked yesterday) which suggests that the archiving of hangout messages is normal behaviour, but they have set up a filter that has apparently broken this!? ~~I don't have any filters set for the user in question.~~ | 2013/08/02 | [
"https://webapps.stackexchange.com/questions/47431",
"https://webapps.stackexchange.com",
"https://webapps.stackexchange.com/users/5434/"
] | It appears that chat/hangout messages are auto-archived in Gmail afterall - providing you don't have *any* filters set on the user you are chatting with! (This sounds like a bug if you ask me!)
I previously had a filter set on this particular user that simply prevented any emails from that user being seen as spam - but this was apparently enough to prevent all hangout messages from that user being archived. I had, however, already deleted this filter - but it appears to have taken some time for the system to catch up, since I was still getting hangout messages in my inbox after I had initially deleted the filter.
Thanks to [Grzegorz](https://webapps.stackexchange.com/questions/47370/gmail-problem-with-hangout-and-filters) for help in diagnosing this. | I had the same problem but your solution did not apply to my case. Instead, the following filter works for me:
>
> Matches: **label:chats label:inbox**
>
>
> Do this: Skip Inbox
>
>
>
I had originally matched against "*in*:chats" and "*in*:inbox", but Gmail automatically converts those. Don't get concerned when a warning pops up when creating this filter.
>
> Filter searches containing "label:", "in:", "is:", or stars criteria (i.e. "has:yellow-star") are not recommended as they will never match incoming mail. Do you still wish to continue to the next step?
>
>
>
This is perfectly fine for me since chat history is not what I would consider "incoming mail". |
47,431 | I use hangouts on Android and on the desktop (browser extension) and these messages periodically appear in my Gmail inbox. So every day I'm having to manually archive all these hangout messages that have appeared in my inbox.
(In my case it is really just 1 hangout with 1 person. Opening any of these hangout messages gives me access to the entire hangout, just at a different point in the conversation.)
How can I auto-archive these hangout messages in Gmail? There does not seem to be a "hangouts" filter?
I've noticed [this question](https://webapps.stackexchange.com/questions/47370/gmail-problem-with-hangout-and-filters "Gmail - problem with Hangout and filters") (asked yesterday) which suggests that the archiving of hangout messages is normal behaviour, but they have set up a filter that has apparently broken this!? ~~I don't have any filters set for the user in question.~~ | 2013/08/02 | [
"https://webapps.stackexchange.com/questions/47431",
"https://webapps.stackexchange.com",
"https://webapps.stackexchange.com/users/5434/"
] | It appears that chat/hangout messages are auto-archived in Gmail afterall - providing you don't have *any* filters set on the user you are chatting with! (This sounds like a bug if you ask me!)
I previously had a filter set on this particular user that simply prevented any emails from that user being seen as spam - but this was apparently enough to prevent all hangout messages from that user being archived. I had, however, already deleted this filter - but it appears to have taken some time for the system to catch up, since I was still getting hangout messages in my inbox after I had initially deleted the filter.
Thanks to [Grzegorz](https://webapps.stackexchange.com/questions/47370/gmail-problem-with-hangout-and-filters) for help in diagnosing this. | I set up a new filter that looks for @profiles.google.com as part of the sender and immediately archives those. It seems to be working well. |
47,431 | I use hangouts on Android and on the desktop (browser extension) and these messages periodically appear in my Gmail inbox. So every day I'm having to manually archive all these hangout messages that have appeared in my inbox.
(In my case it is really just 1 hangout with 1 person. Opening any of these hangout messages gives me access to the entire hangout, just at a different point in the conversation.)
How can I auto-archive these hangout messages in Gmail? There does not seem to be a "hangouts" filter?
I've noticed [this question](https://webapps.stackexchange.com/questions/47370/gmail-problem-with-hangout-and-filters "Gmail - problem with Hangout and filters") (asked yesterday) which suggests that the archiving of hangout messages is normal behaviour, but they have set up a filter that has apparently broken this!? ~~I don't have any filters set for the user in question.~~ | 2013/08/02 | [
"https://webapps.stackexchange.com/questions/47431",
"https://webapps.stackexchange.com",
"https://webapps.stackexchange.com/users/5434/"
] | I had the same problem but your solution did not apply to my case. Instead, the following filter works for me:
>
> Matches: **label:chats label:inbox**
>
>
> Do this: Skip Inbox
>
>
>
I had originally matched against "*in*:chats" and "*in*:inbox", but Gmail automatically converts those. Don't get concerned when a warning pops up when creating this filter.
>
> Filter searches containing "label:", "in:", "is:", or stars criteria (i.e. "has:yellow-star") are not recommended as they will never match incoming mail. Do you still wish to continue to the next step?
>
>
>
This is perfectly fine for me since chat history is not what I would consider "incoming mail". | I set up a new filter that looks for @profiles.google.com as part of the sender and immediately archives those. It seems to be working well. |
130,691 | This is a question from a study guide for a military officer exam:
I have fluid flowing through an elbow in a pipe (the elbow is in the shape of an upside down and backwards "J") and the flow goes from south to north to east. If holes are drilled at the upper left outside part of the bend and the lower right inside part of the bend, at which point will pressure be lower? Please see attached image for a picture of the question.
When I have thought about this problem, I keep coming back to pressure being transmitted equally to all parts of a fluid, so I don't understand how the pressure can be different at one of the holes. The book gives the answer as the inside bend which I can force myself to accept since I can picture the fluid attempting to go straight to the outside bend hole and thus hitting that part harder, but I don't really understand it. | 2014/08/12 | [
"https://physics.stackexchange.com/questions/130691",
"https://physics.stackexchange.com",
"https://physics.stackexchange.com/users/57155/"
] | When a pipe is bent the outside curve - what would be the longest path through the curve - has the highest pressure and the lowest speed. The inside curve - the shortest path through the curve - has the lowest pressure and the highest speed. In short, when the path of a fluid in steady-state flow bends, the pressure on the outside of the bend is always higher than the pressure on the inside of the bend.
The pressure would be lower on the lower right inside point of the bend.
I could no view that attached image; however, I hope that helped. | The same answer as above, but I also happened to have analysis results for this question. Hope this helps,
[](https://i.stack.imgur.com/YHsxU.png)
The above image is velocity contour of the cross section of the pipe.
Below Image is for Pressure
[](https://i.stack.imgur.com/5FHiq.png) |
130,691 | This is a question from a study guide for a military officer exam:
I have fluid flowing through an elbow in a pipe (the elbow is in the shape of an upside down and backwards "J") and the flow goes from south to north to east. If holes are drilled at the upper left outside part of the bend and the lower right inside part of the bend, at which point will pressure be lower? Please see attached image for a picture of the question.
When I have thought about this problem, I keep coming back to pressure being transmitted equally to all parts of a fluid, so I don't understand how the pressure can be different at one of the holes. The book gives the answer as the inside bend which I can force myself to accept since I can picture the fluid attempting to go straight to the outside bend hole and thus hitting that part harder, but I don't really understand it. | 2014/08/12 | [
"https://physics.stackexchange.com/questions/130691",
"https://physics.stackexchange.com",
"https://physics.stackexchange.com/users/57155/"
] | If you imagine the bend portion of the pipe to be a part of a circle, it becomes apparent that there is centripetal force present. As the water travels through the bend, there is a centripetal force that acts inwards towards the centre of the "circle". There must be a reactionary force to this which acts in the opposite direction, in other words the outer wall of the pipe. This reactionary force upon the area of the outer wall causes a higher pressure. | The same answer as above, but I also happened to have analysis results for this question. Hope this helps,
[](https://i.stack.imgur.com/YHsxU.png)
The above image is velocity contour of the cross section of the pipe.
Below Image is for Pressure
[](https://i.stack.imgur.com/5FHiq.png) |
5,537,890 | i am newb in using SQLite in mobile development having experience in sql server, MySQL on web based applications.
After googling around I found sql browser and navicut to use for sqlite manager. While working on both of them, realising they are very buggy.Now i am trying Firefox SQLite plugin.Can someone please suggest good SQLite client for mac. Till your answers I am trying Firefox sqlite add-on for mac.
Thanks. | 2011/04/04 | [
"https://Stackoverflow.com/questions/5537890",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/490953/"
] | Try [Base](http://menial.co.uk/software/base/).
>
> Base is an application for creating, designing, editing and browsing SQLite 3 database files. It's a proper Mac OS X application. Fast to launch, quick to get in to and get the data you need.
>
>
> | [SQLVue](http://itunes.apple.com/us/app/sqlvue/id426397771?mt=12), which my company just released on the Mac App Store this week, is a lightweight SQLite development tool for Mac OS X. |
5,537,890 | i am newb in using SQLite in mobile development having experience in sql server, MySQL on web based applications.
After googling around I found sql browser and navicut to use for sqlite manager. While working on both of them, realising they are very buggy.Now i am trying Firefox SQLite plugin.Can someone please suggest good SQLite client for mac. Till your answers I am trying Firefox sqlite add-on for mac.
Thanks. | 2011/04/04 | [
"https://Stackoverflow.com/questions/5537890",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/490953/"
] | Try [Base](http://menial.co.uk/software/base/).
>
> Base is an application for creating, designing, editing and browsing SQLite 3 database files. It's a proper Mac OS X application. Fast to launch, quick to get in to and get the data you need.
>
>
> | [SQLiteFlow](http://www.sqliteflow.com/)
Since you're using SQLite in iOS app development, I recently found SQLiteFlow is nice. And one of its outstanding feature I think is it can handle database directory changes, this really mean something when you debug your iOS app in an iOS simulator which your SQLite database directory will change every time your app launch through Xcode. |
5,537,890 | i am newb in using SQLite in mobile development having experience in sql server, MySQL on web based applications.
After googling around I found sql browser and navicut to use for sqlite manager. While working on both of them, realising they are very buggy.Now i am trying Firefox SQLite plugin.Can someone please suggest good SQLite client for mac. Till your answers I am trying Firefox sqlite add-on for mac.
Thanks. | 2011/04/04 | [
"https://Stackoverflow.com/questions/5537890",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/490953/"
] | I'm not sure which SQLite Manager you were mentioning, there's a Firefox add-on and a native Mac app. I've used the Mac app, [SQLiteManager](http://www.sqlabs.net/sqlitemanager.php) for a while and found that it's quick and easy to use, even if it has a few quirks. I don't think it's buggy at all. It is pricey though.
I couldn't stand the Firefox plugin (which is called SQLite Manager, with a space). | [SQLVue](http://itunes.apple.com/us/app/sqlvue/id426397771?mt=12), which my company just released on the Mac App Store this week, is a lightweight SQLite development tool for Mac OS X. |
5,537,890 | i am newb in using SQLite in mobile development having experience in sql server, MySQL on web based applications.
After googling around I found sql browser and navicut to use for sqlite manager. While working on both of them, realising they are very buggy.Now i am trying Firefox SQLite plugin.Can someone please suggest good SQLite client for mac. Till your answers I am trying Firefox sqlite add-on for mac.
Thanks. | 2011/04/04 | [
"https://Stackoverflow.com/questions/5537890",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/490953/"
] | I'm not sure which SQLite Manager you were mentioning, there's a Firefox add-on and a native Mac app. I've used the Mac app, [SQLiteManager](http://www.sqlabs.net/sqlitemanager.php) for a while and found that it's quick and easy to use, even if it has a few quirks. I don't think it's buggy at all. It is pricey though.
I couldn't stand the Firefox plugin (which is called SQLite Manager, with a space). | [SQLiteFlow](http://www.sqliteflow.com/)
Since you're using SQLite in iOS app development, I recently found SQLiteFlow is nice. And one of its outstanding feature I think is it can handle database directory changes, this really mean something when you debug your iOS app in an iOS simulator which your SQLite database directory will change every time your app launch through Xcode. |
61,228 | Why would my furnace run it's cycle, shut down and then start up again a few minutes later. Continuing this cycle all night?? Thanks | 2015/02/27 | [
"https://diy.stackexchange.com/questions/61228",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/34166/"
] | If the home is maintaining the temperature set on the thermostat (at least AT the thermostat) then it's probable that you are near, but not quite at, "design load" for the furnace. At design load, the furnace runs 100% of the time to keep up with heat loss in the house and maintain temperature - if it gets colder than that outside (below the "design temperature"), the house gets colder than set. At anything less than design load, the furnace runs a fraction of the time, and is off a fraction of the time. The closer the outside temperature is to "design temperature" the longer the run time and shorter the off time. As it gets warmer outside, the off time increases and the run time gets shorter.
Anything that makes the house lose heat faster raises the effective design temperature (doesn't change what temperature was assumed to design it, but alters the actual one.)
Anything that makes the house lose heat slower lowers the effective design temperature (doesn't change what temperature was assumed to design it, but alters the actual one.)
If the house is NOT maintaining the set temperature AND the furnace is shutting off regularly, something may be in need of maintenance/adjustment. | As Ecnerwal said, your furnace cannot keep up with heating your home. This could be caused by a number of things.
1. Your furnace itself is inadequate and cannot keep up with the load. The rating in BTUs is not high enough.
2. There are inefficiencies in your home which is causing the furnace to work harder than it was designed to.
* The furnace requires service.
* Insulation in your home may be inadequate.
* There could be air leaks and drafts caused by old windows, doors, etc.
* Too many leaks in the duct work are causing heat loss.
3. You are trying to maintain a temperature that is too high for the furnace to handle.
There are a number of services available that will come into your home and do an energy audit. They will be able to figure out what issues you have in your home. They will do tests like the blower door test to determine how many air leaks you have. They also have special tools such as thermal cameras that can see where the heat loss is. They will also inspect your insulation to see if it meets code. I would also recommend that you have a technician come in and service your heating equipment. |
61,228 | Why would my furnace run it's cycle, shut down and then start up again a few minutes later. Continuing this cycle all night?? Thanks | 2015/02/27 | [
"https://diy.stackexchange.com/questions/61228",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/34166/"
] | If the home is maintaining the temperature set on the thermostat (at least AT the thermostat) then it's probable that you are near, but not quite at, "design load" for the furnace. At design load, the furnace runs 100% of the time to keep up with heat loss in the house and maintain temperature - if it gets colder than that outside (below the "design temperature"), the house gets colder than set. At anything less than design load, the furnace runs a fraction of the time, and is off a fraction of the time. The closer the outside temperature is to "design temperature" the longer the run time and shorter the off time. As it gets warmer outside, the off time increases and the run time gets shorter.
Anything that makes the house lose heat faster raises the effective design temperature (doesn't change what temperature was assumed to design it, but alters the actual one.)
Anything that makes the house lose heat slower lowers the effective design temperature (doesn't change what temperature was assumed to design it, but alters the actual one.)
If the house is NOT maintaining the set temperature AND the furnace is shutting off regularly, something may be in need of maintenance/adjustment. | Thank you all for your replies. The issue was simple. My filters were clogged. First thing I should have looked at. Obviously with the dust build up, not enough cool air was coming in to transfer |
23,344,251 | Are there any classes in Windows RunTime APIs to find the current device information(OS info, RAM size, power source) ?
The equivalent class in Windows Phone 8 is Microsoft.Phone.Info.DeviceStatus (<http://bit.ly/deviceStatus>)
Any Help? | 2014/04/28 | [
"https://Stackoverflow.com/questions/23344251",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2140743/"
] | According to my knowledge and what I've searched currently there is no method that will tell you that. Similar problem according to WinRT was raised [here](https://stackoverflow.com/a/19924303/2681948).
As I've also searched looking at [MSDN](http://msdn.microsoft.com/en-us/library/windows/apps/hh464945.aspx), there is also no alternative to System.Power.
You can check [the battery status](http://msdn.microsoft.com/en-us/library/windowsphone/develop/windows.phone.devices.power(v=vs.105).aspx) thought, but that won't help you to check if it is powered or not.
I'm also curious how this will be solved, hence IMO there should be such an ability when developing for phone to check if it's powered. Under WP8.0 for example BackgroundTransfers for large files required external power supply, and no I'm not able to check it - maybe it will be solved in final release. | As mentioned earlier, there is currently no straightforward way to get the device info on a windows phone 8.1 runtime app although its possible to get some infor using the:
Windows.Security.ExchangeActiveSyncProvisioning.EasClientDeviceInformation class <http://msdn.microsoft.com/en-us/library/windows/apps/windows.security.exchangeactivesyncprovisioning.easclientdeviceinformation.aspx>
Hope this helps |
587,573 | Is there any English dialect that distinguishes the stressed /oʊ/ as in *goat, throat, slope, broke, stroke*, etc. from the final, unstressed /oʊ/ as in *sparrow, arrow, tomorrow, yellow, window*, etc?
What does the unstressed /oʊ/ sound like in those dialects? | 2022/04/14 | [
"https://english.stackexchange.com/questions/587573",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/250563/"
] | ...the lyrics were written to scan
Scan means 'fit the line'.
Definition:
scan - verb (POEM)
If a poem or part of a poem scans, it follows a pattern of regular beats: This line doesn't scan - it has too many syllables.
<https://dictionary.cambridge.org/dictionary/english/scan>
There are several definitions for scan, scroll down to see the one I mention above. | What about **metrical**?
>
> pertaining to meter or poetic measure
>
> [source](https://www.dictionary.com/browse/metrical)
>
>
>
"*Don't take too much literal meaning from this song. The lyrics were written metrically*."
Another example (my emphasis):
>
> for he knew that **if he had written metrically in Latin** as the other poets of past times had done, he would only have done service to men of letters [..]
>
> From: *Vita di Dante*, by Bocaccio, transl. by Philip H. Wicksteed. [source](https://books.google.fi/books?id=haKfDwAAQBAJ)
>
>
> |
587,573 | Is there any English dialect that distinguishes the stressed /oʊ/ as in *goat, throat, slope, broke, stroke*, etc. from the final, unstressed /oʊ/ as in *sparrow, arrow, tomorrow, yellow, window*, etc?
What does the unstressed /oʊ/ sound like in those dialects? | 2022/04/14 | [
"https://english.stackexchange.com/questions/587573",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/250563/"
] | ...the lyrics were written to scan
Scan means 'fit the line'.
Definition:
scan - verb (POEM)
If a poem or part of a poem scans, it follows a pattern of regular beats: This line doesn't scan - it has too many syllables.
<https://dictionary.cambridge.org/dictionary/english/scan>
There are several definitions for scan, scroll down to see the one I mention above. | This only captures part of your intended meaning, but you could say that **the lyrics are just filler.** [Cambridge defines "filler" in this sense as:](https://dictionary.cambridge.org/us/dictionary/english/filler)
>
> something of lower quality included in a film, broadcast, speech, record, etc. in order to fill all the time or space
>
>
> |
1,116,283 | I need SFTP functionality on my Windows 2012 VPS server.
According to [this page](https://learn.microsoft.com/en-us/windows-server/administration/openssh/openssh_install_firstuse?tabs=gui), one of the prerequisites is that I'm running at least Windows Server 2019 or Windows 10 (build 1809).
My hosting company says they can migrate everything to Windows Server 2022, but there will be costs and my sites will obviously be down for a little while.
1. Does anyone know if a SFTP server can be installed on Windows Server 2012?
2. Does anyone have experience doing this? Any recommendations?
3. Would this be greatly simplified if I used plain ol' FTP? I don't think our data is particularly sensitive. | 2022/11/21 | [
"https://serverfault.com/questions/1116283",
"https://serverfault.com",
"https://serverfault.com/users/317530/"
] | Domain Admins will be able to read and write, and Authenticated Users will be able to read. | >
> If I set write permissions for domain admins, but then read-only for 'authenticated users', which takes precedence?
>
>
> Does the domain admins write permission trump the authenticated users read-only permission? Or will the domain admins be unable to write because domain admins are included in authenticated users?
>
>
>
In the Windows ACL model, neither has higher precedence than the other – instead the **sum** of all matching "Allow" permissions is used. So if you grant X to Authenticated Users but Y+Z to Domain Admins, the user is effectively granted X+Y+Z.
However, *Deny* entries have higher precedence than any "Allow" entry. (Again, the sum of all matching "Deny" entries will be denied.)
This applies equally to NTFS files, AD entries, and most other securable objects. For AD specifically, the algorithm is documented at [MS-ADTS](https://learn.microsoft.com/en-us/openspecs/windows_protocols/ms-adts/c55e113e-25a2-45eb-abba-bf5a9e521a8f). |
28,764 | I've enabled WiFi sync for my 4S and my first gen iPad, both running iOS5. When my phone is plugged in to my computer with a USB cable, it shows up in the sidebar, but as soon as it's unplugged it goes away. The iPad stays in the sidebar until I manually eject it.
It seems that my iPad is behaving 'correctly', at least based on [This kbase article](http://support.apple.com/kb/PH2555).
Toggling WiFi sync when the phone is connected doesn't seem to solve it.
I've also tried the steps outlined [in this article](http://support.apple.com/kb/TS4062) to no avail.
This was a fresh install, I didn't import a backup from a previous device. I've restored and set up as new, and the same behavior happens.
I had originally thought it was because my computer had switched to an 802.11n 5Ghz network while my phone remained on the 802.11n 2.4Ghz network, but even when they're on the same exact network I see this behavior. | 2011/10/22 | [
"https://apple.stackexchange.com/questions/28764",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/8308/"
] | I had this problem with an iPad. I did the initial setup over USB, enabled Wi-Fi sync, but then when I removed the iPad, it disappeared from the iTunes sidebar, but my iPhone didn't have that problem.
Turned out that I had been disconnecting the USB end of the cable from my Mac first; disconnecting the dock connector end from the iPad (then removing the USB end) made it stay in the sidebar. Completely bonkers, really doesn't seem like it should be that way (or even that iTunes could discern what was disconnected when), but that's what fixed it for me. | It disappears because the "show" selection on DEVICES needs to be activated and then it wil come back |
28,764 | I've enabled WiFi sync for my 4S and my first gen iPad, both running iOS5. When my phone is plugged in to my computer with a USB cable, it shows up in the sidebar, but as soon as it's unplugged it goes away. The iPad stays in the sidebar until I manually eject it.
It seems that my iPad is behaving 'correctly', at least based on [This kbase article](http://support.apple.com/kb/PH2555).
Toggling WiFi sync when the phone is connected doesn't seem to solve it.
I've also tried the steps outlined [in this article](http://support.apple.com/kb/TS4062) to no avail.
This was a fresh install, I didn't import a backup from a previous device. I've restored and set up as new, and the same behavior happens.
I had originally thought it was because my computer had switched to an 802.11n 5Ghz network while my phone remained on the 802.11n 2.4Ghz network, but even when they're on the same exact network I see this behavior. | 2011/10/22 | [
"https://apple.stackexchange.com/questions/28764",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/8308/"
] | I have tried reinstalling iTunes, restoring the software, rebooting the router and pretty much every combination of checkboxes checked and unchecked and you know what fixed it for me?
Rebooting my MacBook.
I've been struggling with this for several hours and it never came to me to reboot my laptop.
So my guess is that it has to do with the computer's wifi connection. Flush that cache, reconnect that wifi, or just reboot the thing.
I hope this helps someone. I am very relieved. | It disappears because the "show" selection on DEVICES needs to be activated and then it wil come back |
240,556 | After a hard drive died on a RAID5 setup on a legacy system, the Vmware Virtual Machine on it kept on running for a period of 6 months. Needless to say, finally it crashed bad. Now, the dead hard drive has been replaced and RAID5 restoration tools have been used to recover the lost data.
Yet, the logical drive is still giving IO errors. Is this normal? Should we be worried about further data loss?
Thanks in advance | 2011/02/26 | [
"https://serverfault.com/questions/240556",
"https://serverfault.com",
"https://serverfault.com/users/49376/"
] | You aren't telling us enough:
Are you using a hardware RAID controller? Which one? If not, software RAID on what operating system? Did the recovery have finished without reporting any errors? Did the RAID management tools for the RAID solution you are using report the volume as working fine? What kind of I/O errors do you get? Are they reported by the host OS or the guest machine?
Also, I can't help but to remark that you should have started to worry six months ago when the drive died. If a drive fails in a RAID setup you have to replace it immediately, especially with a RAID5. Any error on one of the remaining disks will fail the recovery, so there is no time to loose. | Yep, you should for sure but that's not 100 % warranty, you know. I'd use Linux Software RAID as a tool to check how many stripes have had bad CRCs. That would give overall damage picture. |
8,317 | I have noticed that interior of the ISS appears to be very disorganized.
Let me clarify what I mean by that. I grew up watching Star Trek, and the Enterprise was *the* reference of clean space-faring vehicle. Things were in drawers, hidden. Or structured in a clean way. Computers were built into larger panels. There were touch screens. Have you noticed they always needed to uncover a panel to access some occasionally accessed system? It was minimalistic design.
When I look at the ISS, everything is everywhere, and it seems like nothing is attached firmly to the main frame. Things are just loosely attached to the wall and there are so many cables, devices, objects, ... that one wonders
* how do they not cut themselves all day with some sharp edge
* how do they not break or make a cable loose
* how come those objects, machines, devices did not break already by bumping into each other a long time ago
given there is no stabilizing force in a form of gravity.
Another aspect is psychological. How can those astronauts feel comfortable and focus on the job, when there is constant visual noise in their peripheral area? | 2015/03/02 | [
"https://space.stackexchange.com/questions/8317",
"https://space.stackexchange.com",
"https://space.stackexchange.com/users/8610/"
] | In Star Trek there is artificial gravity, and lots of staff. On the ISS there is neither. The crew spends a lot of time maintaining the station. They barely get time for science operations, with all the time they spend on maintenance. Going to a crew of 7, which Commercial Crew vehicles will enable (bringing 4 at a time, acting as a lifeboat for 4, instead of Soyuz with only capacity for 3), is expected to greatly increase available time for science. Just by adding one more body, on top of all the time the current staff of 6, will more than double available science time.
There is not a lot of storage space in the station, all things considered, and when cargo comes up on a vehicle, it comes in bulk, all at once. They track where everything is, else it gets lost in the scrum. The Dragon and Cygnus vehicles are mostly volume limited, but the ATV and HTV bring up a fairly larger amount all at once. It all has to go somewhere, and you need to keep track of where everything is.
There was a [great video of Sunnita Williams](https://www.youtube.com/watch?v=doN4t5NKW-k) hunting for something. She looked up the location on the computer, went to the specified location, opened a cabinet, and pulled out three layers deep of boxes. The item she wanted was in the box at the back.
Thus there is a great deal of order, hiding behind the seeming disorder. Looks can be deceiving.
The inner storage areas of the station are really well organized. The visible surfaces definitely look like a hodge-podge, but they are taking advantage of available surface area to store things they use a lot. | At 4,000 to 14,000 dollar per kg in launch costs, pretty panels are a huge waste of money. It needs to be fixed solidly enough to withstand a crewmember bumping into it, and easily detachable enough to be replaced when needed. Beyond that, the crewmembers are professionals who watch where they're going and don't need a kiddieproofed environment. |
28,089,123 | I want to know the right concept about it. If I have a MVC application with Repository Pattern, where the BL should be?
* Should it be inside the Model? Model should have all the business
logic before call the unitofwork to insert or not the data into
database?
* Should it be in the controller? Before call the model?
* Should I have a service layer to do the business logic and decide if
I should call the Model to call the UnitOfWork to save the data?
A good explanation will help a lot too. | 2015/01/22 | [
"https://Stackoverflow.com/questions/28089123",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2274562/"
] | The short answer - it depends. If it's a fairly complex or sizable application, I like to create a service layer project with the repositories as dependencies. If it's a small application, I'll put the logic in the controller. In my opinion, if it takes more time and effort to create the service layer than it would be to create the application (i.e. one or two controllers), then it doesn't make sense to me to go that route. You also need to consider the likelihood that the application will grow. What might start small could grow into something much bigger and in that case, again, it might be more beneficial to create the separate service layer. | The third one... and then some.
Your application structure could look like this (each in different projects):
* Data storage layer (e.g. SQL database)
* ORM (e.g. NHibernate or Entity Framework)
* Domain (including abstract repositories and entities)
* Service layer (and optionally business)
* MVC application (which has it's own models relating to the entities)
but there are many ways to go about this depending on the complexity and size of your application. |
28,089,123 | I want to know the right concept about it. If I have a MVC application with Repository Pattern, where the BL should be?
* Should it be inside the Model? Model should have all the business
logic before call the unitofwork to insert or not the data into
database?
* Should it be in the controller? Before call the model?
* Should I have a service layer to do the business logic and decide if
I should call the Model to call the UnitOfWork to save the data?
A good explanation will help a lot too. | 2015/01/22 | [
"https://Stackoverflow.com/questions/28089123",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2274562/"
] | The short answer - it depends. If it's a fairly complex or sizable application, I like to create a service layer project with the repositories as dependencies. If it's a small application, I'll put the logic in the controller. In my opinion, if it takes more time and effort to create the service layer than it would be to create the application (i.e. one or two controllers), then it doesn't make sense to me to go that route. You also need to consider the likelihood that the application will grow. What might start small could grow into something much bigger and in that case, again, it might be more beneficial to create the separate service layer. | There is no "correct" answer to this question, it is primarily opinion-based. You can read about my opinion in the following project wiki:
<https://github.com/danludwig/tripod/wiki/Why-Tripod%3F>
<https://github.com/danludwig/tripod/wiki/Dependency-and-Control-Inversion>
<https://github.com/danludwig/tripod/wiki/Tripod-101>
<https://github.com/danludwig/tripod/wiki/Testing,-Testing,-1-2-3>
<https://github.com/danludwig/tripod/wiki/Command-Query-Responsibility-Segregation-(CQRS)>
Another piece of advice I would like to offer is never put any business logic in viewmodels or entities. These classes should not have methods, only properties to contain data. Separate your data from behavior. Use models for data, and other types for behavior (methods). |
148,354 | In games where you have to use an account to sign in or otherwise authenticate yourself (e.g MMO games), it is often prohibited to share your account with other people.
For example in the [World of Warcraft Official Terms of Use Agreement](http://eu.blizzard.com/en-gb/company/legal/wow_tou.html):
>
> [...] You may not share the Account with anyone, except that if you are a parent or guardian, you may permit one (1) minor child to use the Account when not in use by you. You are liable for all uses of the Account that has been enabled by you […]
>
>
>
And the [League of Legends Terms of Use](http://eune.leagueoflegends.com/en/legal/termsofuse):
>
> [...] You can’t share your account or Login Credentials with anyone. You can’t sell, transfer or allow any other person to access your account or Login Credentials, or offer to do so. You’re entirely responsible for maintaining the confidentiality of your Login Credentials. [...]
>
>
>
What is the reason for this? Is it because of legal reasons and / or security reasons? I imagine account sharing would not be a problem if there weren't some real-world consequences for the *company*, not just the end-user. | 2017/09/13 | [
"https://gamedev.stackexchange.com/questions/148354",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/77195/"
] | It's important to have a firm grasp of precisely who the legal entities are in any contractual agreement, so you know who you have to sue or blame or whatever if it ever comes to that.
Less seriously, if account-sharing were permissible, then blaming "somebody else who was using the account at the time" would be a reasonable response to any punishments for violating any other rules. By prohibiting sharing that way, a company eliminates the hassle of having to deal with that argument (which is usually impossible to disprove) in response to bans. Obviously people still try to use that defense, but a company can simply point to the rule against account sharing and say "well that's not allowed either." It makes life way simpler for the GM team.
It also has a small side-benefit: it means that if you and your friend want to play, you have to buy two copies of the game, you can't buy one and share it. This is obviously also better for the company. | From both a legal & security standpoint, the single biggest factor I'm aware of is accountability. If you cannot determine who is responsible for the account, you cannot reasonably hold anyone accountable for actions related to the account. For instance, money laundering is easier if you cannot trace account ownership. Forbidding shared accounts won't prevent laundering in all forms, but it is choosing to not make it easier & demonstrates an effort on the part of the developer to not facilitate such things.
You didn't ask specifically about this aspect, but I feel it's equally important - there are also design implications to shared accounts. For instance, as a designer, you might not want players to pay others to accomplish objectives due to the impact it has on the game experience. You might not be able to full prevent such things, but you can take steps to discourage them & forbidding it via the terms of use may be part of such a solution. |
148,354 | In games where you have to use an account to sign in or otherwise authenticate yourself (e.g MMO games), it is often prohibited to share your account with other people.
For example in the [World of Warcraft Official Terms of Use Agreement](http://eu.blizzard.com/en-gb/company/legal/wow_tou.html):
>
> [...] You may not share the Account with anyone, except that if you are a parent or guardian, you may permit one (1) minor child to use the Account when not in use by you. You are liable for all uses of the Account that has been enabled by you […]
>
>
>
And the [League of Legends Terms of Use](http://eune.leagueoflegends.com/en/legal/termsofuse):
>
> [...] You can’t share your account or Login Credentials with anyone. You can’t sell, transfer or allow any other person to access your account or Login Credentials, or offer to do so. You’re entirely responsible for maintaining the confidentiality of your Login Credentials. [...]
>
>
>
What is the reason for this? Is it because of legal reasons and / or security reasons? I imagine account sharing would not be a problem if there weren't some real-world consequences for the *company*, not just the end-user. | 2017/09/13 | [
"https://gamedev.stackexchange.com/questions/148354",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/77195/"
] | In addition to all the other answers, there is another reason: **Ensuring leaderboard integrity by prohibiting account sharing and multi-account usage.** A famous example is the rhythm game [osu!](https://osu.ppy.sh), which states as its **first** rule:
[](https://osu.ppy.sh/wiki/Rules)
Due to the game's ranking system ("[performance points](https://osu.ppy.sh/wiki/Performance_Points)") absence of this rule would destroy the leaderboards, as a top player could easily dominate their rank range by just repeating their top scores. | I run Wildfire, a multiplayer network backend for Crysis Wars.
Like other networks, we don't allow account sharing- our ToS never actually goes into detail as to why this is, but I made this decision because:
* One copy of the game per account- allowing users to share their accounts would *dramatically* skew the statistics of that account. For commercial networks, this would simply mean that they miss out on the revenue from multiple accounts purchasing a unique licence for the product.
* Accountability- if a user of the account violates the ToS, it has an impact for all users using the said user account. Separate accounts mean that only one person is banned/suspended/warned, affecting just that one person when compared to multiple people. In addition to this, allowing multiple people to use a single account would make it *very* difficult to pin the blame on a single person.
* Leaderboards- multiple people using one account would *seriously* inflate the statistics of the account (similar to my first point). By this, I mean that for Crysis Wars (an FPS) the amount of kills/deaths/points attributed to an account will be abnormally higher, and this is unfair for legitimate users.
* Security- multiplayer backend systems, like my own, log *everything* (and I mean everything). There are several security systems that I use that these logs support- one of them is that the system automatically locks-out a user if it detects that it is using multiple IP addresses at the same time. Similar systems do the same. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.