qid int64 1 74.7M | question stringlengths 12 33.8k | date stringlengths 10 10 | metadata list | response_j stringlengths 0 115k | response_k stringlengths 2 98.3k |
|---|---|---|---|---|---|
164,605 | I just bought an ASA 5505 that my data center is setting up for me. They have told me that setting the ASA to routed with NAT will break Web/DNS server on inside network.
For example: WAN IP address 66.xxx.47.x - translated to LAN IP address 192.168.0.1 on inside network - will not provide WAN IP to the inside Web/DNS server, which apparently will break DNS -- of course DC provides no other details.
Are most people providing web services behind an ASA set in Transparent mode?? Looks like transient mode has some drawbacks, not terminating VPN traffic being one of them.
Seems like routed/NAT is the most secure/versatile, but maybe I'm not seeing the benefits of transient mode, on the surface, seems like a quick & dirty way to get up & running, hopefully more than that.
Feedback appreciated, I have to make the call on this soon. | 2010/07/27 | [
"https://serverfault.com/questions/164605",
"https://serverfault.com",
"https://serverfault.com/users/46210/"
] | ASA static entries let you use a "dns" manipulator which will rewrite the DNS response on the fly. If they are saying this is not possible they do not know what they are doing with the ASA. | It depends on what kind of DNS setup you have.
We have our DNS servers and web servers behind an ASA 5510 using NAT.
On the DNS server (BIND), we provide different information depending on which IP address the request comes from. This is [Split-horizon DNS](http://en.wikipedia.org/wiki/Split-horizon_DNS).
If the request comes from inside, we reply with an internal IP.
If the request comes from outside, we reply with an external IP.
For example, if the request comes from a normal outside host, we reply with the 66.xxx.47.x IP of the webserver. However, if an internal host requests the IP, we reply with the 192.168.0.x IP of the webserver.
Basically, it's completely possible to put your DNS and webservers behind an ASA using NAT as long as you configure your DNS server properly. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | 1. No. GPS is not autonomous. Errors in satellite vehicles can and have propagated to ground stations (I have one on record from a working GPS receiver placing me kilometers from my actual location that coincided with a published failed software update in the satellites). Additionally, selective availability, while not likely, can be turned on by DoD any time it likes. Because there are so many thing to go wrong, it's not possible to use GPS in systems that have a hard performance requirement.
2. Iphone (like all cell phones) has a small GPS antenna and the GPS is akin to "would you like fries with that". i.e. GPS is an add on rather than its reason for existing, and is inferior in nearly every way to a dedicated GPS unit. .
3. Can't say. GPS results are statistical, not hard and fast. Study standard deviations and such like to get an idea of how to specify required GPS performance. After that, I am sure money comes into it, and for the right amount, you can get what you are after, but only as long as you can specify it.
You may be interested in studying up on "Receiver autonomous integrity monitoring" (RAIM) as well.
If you can't be bothered with all the above and want a simple answer - buy a dedicated GPS with the smallest TTFF (Cold start, non assisted) you can afford. This is as good a guide to sensitivity and quality as you will get without becoming a rocket scientist. | Over a small area you may be able to do something. I don't know if you have a small area to deal with, and 10ft might be pushing it even then. Basically you fix the position of a known point by some means (it might not even have to be fixed absolutely depending on your goal). *Long term* GPS averaging might do that for you. Then you use this known value to calibrate an offset. You'd need a laptop connected to a dedicated GPS unit and probably custom code. If you could also access differential GPS signals you'd need a beacon receiver for that, but wouldn't need the known point or your own code.
But over a sufficiently small area compass triangulation can get surprisingly good *relative* accuracy; once again you need well-surveyed references (and line of sight to them). |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | 1. No. GPS is not autonomous. Errors in satellite vehicles can and have propagated to ground stations (I have one on record from a working GPS receiver placing me kilometers from my actual location that coincided with a published failed software update in the satellites). Additionally, selective availability, while not likely, can be turned on by DoD any time it likes. Because there are so many thing to go wrong, it's not possible to use GPS in systems that have a hard performance requirement.
2. Iphone (like all cell phones) has a small GPS antenna and the GPS is akin to "would you like fries with that". i.e. GPS is an add on rather than its reason for existing, and is inferior in nearly every way to a dedicated GPS unit. .
3. Can't say. GPS results are statistical, not hard and fast. Study standard deviations and such like to get an idea of how to specify required GPS performance. After that, I am sure money comes into it, and for the right amount, you can get what you are after, but only as long as you can specify it.
You may be interested in studying up on "Receiver autonomous integrity monitoring" (RAIM) as well.
If you can't be bothered with all the above and want a simple answer - buy a dedicated GPS with the smallest TTFF (Cold start, non assisted) you can afford. This is as good a guide to sensitivity and quality as you will get without becoming a rocket scientist. | This answer is based strictly upon my personal experience.
I am not trying to sell you anything.
There are consumer grade GPS receivers which will get you approximately 3 meters of repeatable accuracy if they are used properly, and there is little or no overhead canopy.
I have had amazing results using a Garmin GLO which receives single frequency GPS, and Glonass data, along with WAAS corrections. ( About $110.00 )
It connected via bluetooth, and worked with multiple GIS, and mapping software applications, as well as various versions of the Windows operating system.
I have also used it with several models of semi-rugged, and rugged tablets without issue.
I have read that it is compatible with Android, Linux, IOS, and just about everything else out there. I only have Windows experience with it.
There were the occasional flyers that were in the five meter range, but in most cases, I was locating previously surveyed locations (Data collected with Survey Grade, Dual Frequency, RTK corrected, GPS/Glonass receivers) within a sub two meter radius.
One of the things I have found is, the units tend to work better when you are moving slightly. Just standing still, the position tends to walk around creating about a four meter circle.
The Garmin I was using had a 10 HZ update rate which created a very smooth line if I was mapping access routes, or roads.
The mapping data I collected, overlaid extremely well on previously located features.
There are other consumer grade external receivers that are probably capable of collecting data with this type of accuracy, I am just not experienced with them.
I have yet to see an embedded, discreet GPS chip within a mobile phone that approaches the results I have gotten with the Garmin.
Selective Availability has been turned off for a long time now, but it can be turned on at the whim of the Department of Defense.
It would really serve no purpose to do so at this point in time, or in any foreseeable scenario with regards to national security. Everyone knows where everything is. (Sorry, bad generalization.)
With regards to locating drainage tiles, I have done a great deal of this using a tool called a Magnetomatic. It is basically a telescopic car antenna attached to a swivel that is mounted through a handle. When you cross something that is buried, the antenna will swing in the direction of the object or line that is buried.
It works on plastic, poly, tile, transite, and metal lines, as well as fiber optic, and twisted pair copper lines. At least it does for me, and most of the people I work with.
Some people are unable to get it to work for them. Most of them are left handed, and the others are people who are unable to wear a movement style watch. (I believe these people have reversed polarity, but I cannot prove it.)
The process is similar to witching (divining) for water.
You can also bend two pinflags at a 90 degree angle to make short handles for each, put one in each hand, with the long ends parallel to each other, and parallel to the ground, about waist high, keeping them a shoulder's width apart. They will cross when you encountered something buried.
If you are using the GPS to measure a distance from a known location to find other unknown locations, say starting at a pump in the corner of a field, wanting to find the grid of drain pipes every 100 feet, the Garmin will get you close, but you will still want to verify the locations somehow, rather than blindly trusting the locations provided by the Garmin. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | 1. No. GPS is not autonomous. Errors in satellite vehicles can and have propagated to ground stations (I have one on record from a working GPS receiver placing me kilometers from my actual location that coincided with a published failed software update in the satellites). Additionally, selective availability, while not likely, can be turned on by DoD any time it likes. Because there are so many thing to go wrong, it's not possible to use GPS in systems that have a hard performance requirement.
2. Iphone (like all cell phones) has a small GPS antenna and the GPS is akin to "would you like fries with that". i.e. GPS is an add on rather than its reason for existing, and is inferior in nearly every way to a dedicated GPS unit. .
3. Can't say. GPS results are statistical, not hard and fast. Study standard deviations and such like to get an idea of how to specify required GPS performance. After that, I am sure money comes into it, and for the right amount, you can get what you are after, but only as long as you can specify it.
You may be interested in studying up on "Receiver autonomous integrity monitoring" (RAIM) as well.
If you can't be bothered with all the above and want a simple answer - buy a dedicated GPS with the smallest TTFF (Cold start, non assisted) you can afford. This is as good a guide to sensitivity and quality as you will get without becoming a rocket scientist. | There are several ways to use GPS signals that would give you the desired accuracy.
You can use either DGPS or a SBAS (WAAS in the US) for an accuracy of meters.
For a centimeter level accuracy you would need an RTK o PPP device.
I doubt you can get the desired accuracy in less than two minutes since the startup of the device unless you have a data connection to obtain the GPS almanac (I guess you could use a satellite Internet service).
Once the device has located the GPS satellites, it should be pretty fast.
In a farm setting a SBAS system is usually the most cost-effective option. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | 1. No. GPS is not autonomous. Errors in satellite vehicles can and have propagated to ground stations (I have one on record from a working GPS receiver placing me kilometers from my actual location that coincided with a published failed software update in the satellites). Additionally, selective availability, while not likely, can be turned on by DoD any time it likes. Because there are so many thing to go wrong, it's not possible to use GPS in systems that have a hard performance requirement.
2. Iphone (like all cell phones) has a small GPS antenna and the GPS is akin to "would you like fries with that". i.e. GPS is an add on rather than its reason for existing, and is inferior in nearly every way to a dedicated GPS unit. .
3. Can't say. GPS results are statistical, not hard and fast. Study standard deviations and such like to get an idea of how to specify required GPS performance. After that, I am sure money comes into it, and for the right amount, you can get what you are after, but only as long as you can specify it.
You may be interested in studying up on "Receiver autonomous integrity monitoring" (RAIM) as well.
If you can't be bothered with all the above and want a simple answer - buy a dedicated GPS with the smallest TTFF (Cold start, non assisted) you can afford. This is as good a guide to sensitivity and quality as you will get without becoming a rocket scientist. | Ten feet isn't unreasonable, although it's right at the edge of day to day accuracy. I used a Garmin 60c along with aerial photographs to create an orienteering map. I had an overall accuracy of about 5 meters, coupled with a 8 meter differential between photo and GPS caused by the 40 meter difference in elevation between floodplain and highlands.
Ten feet from a cold start in under two minutes isn't reasonable.
A: Enable WAAS. Some of the error in GPS is due to ionization state of the upper atmosphere. (More ions, slower radio) This tends to be a large scale phenomena so by comparing the calculated GPS position, to a real location, you can figure out a correction. This correction is broadcast. This routinely gets errors down to under 3 meters -- 10 feet.
B: I found when using a GPS for making trail and cross country maps that it would take about half an hour to really lock in. The longer it had been since I last used the GPS the longer it took. If it had been off only a day, it took only a few minutes. I also found in woods that sometimes moving just a few feet would improve my accuracy. I suspect line of sight to the optimum satellite constellation was blocked by a tree.
C: The constellation of satellites in the sky makes a difference. Bunched satellites or too close to the horizon are worse than having a wide spread set over 30 degrees above the horizon.
D: Some of the errors vary over long periods (~1 hour) So the relative error of two fixes made close together in time can be quite small. I used this to good effect by having locations where I could get an accurate map fix (Benchmarks, road intersections, power pylons) Go to the location, take a waypoint, go to my new spot, take a waypoint. The difference was repeatable to 1-2 meters, if the time difference was under 10 minutes. (For orienteering having small local relative errors is more important than absolute accuracy. You want to know which side of the road a landmark is.)
E: Remember that the error is statistical, and has a Poisson distribution (long tail) When the GPS error is 3 meters, that means that half of the time, the difference between measure and actual is 3 meters or less. The other half of the time it can be more.
F: In mountains and urban locations, you can get reflections. These can move your fix by a couple of kilometers. Conditions for this in the mountains aren't common, but if your fix makes no sense, wait a few minutes.
G: A caution: Most of the US Topo maps and Canadian topo maps have contour information derived from stereo aerial photographs. The Canadian maps can have a horizontal absolute error in contour line placement of up to 90 meters, north/south and half that east west. This can result in the GPS telling you you are at the top of the cliff when you plot your location, while the reality places you at the bottom. Benchmark placement, however is very accurate.
H: Forget the phone. Fine for finding street routes, not much else.
See this thread on Robotics.SE: It talks about various differential GPS techniques that can get sub 1 meter accuracy.
<https://robotics.stackexchange.com/questions/4312/low-cost-centimeter-accurate-satellite-positioning-gnss-gps>
More info:
<https://en.wikipedia.org/wiki/Global_Positioning_System> |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | Over a small area you may be able to do something. I don't know if you have a small area to deal with, and 10ft might be pushing it even then. Basically you fix the position of a known point by some means (it might not even have to be fixed absolutely depending on your goal). *Long term* GPS averaging might do that for you. Then you use this known value to calibrate an offset. You'd need a laptop connected to a dedicated GPS unit and probably custom code. If you could also access differential GPS signals you'd need a beacon receiver for that, but wouldn't need the known point or your own code.
But over a sufficiently small area compass triangulation can get surprisingly good *relative* accuracy; once again you need well-surveyed references (and line of sight to them). | This answer is based strictly upon my personal experience.
I am not trying to sell you anything.
There are consumer grade GPS receivers which will get you approximately 3 meters of repeatable accuracy if they are used properly, and there is little or no overhead canopy.
I have had amazing results using a Garmin GLO which receives single frequency GPS, and Glonass data, along with WAAS corrections. ( About $110.00 )
It connected via bluetooth, and worked with multiple GIS, and mapping software applications, as well as various versions of the Windows operating system.
I have also used it with several models of semi-rugged, and rugged tablets without issue.
I have read that it is compatible with Android, Linux, IOS, and just about everything else out there. I only have Windows experience with it.
There were the occasional flyers that were in the five meter range, but in most cases, I was locating previously surveyed locations (Data collected with Survey Grade, Dual Frequency, RTK corrected, GPS/Glonass receivers) within a sub two meter radius.
One of the things I have found is, the units tend to work better when you are moving slightly. Just standing still, the position tends to walk around creating about a four meter circle.
The Garmin I was using had a 10 HZ update rate which created a very smooth line if I was mapping access routes, or roads.
The mapping data I collected, overlaid extremely well on previously located features.
There are other consumer grade external receivers that are probably capable of collecting data with this type of accuracy, I am just not experienced with them.
I have yet to see an embedded, discreet GPS chip within a mobile phone that approaches the results I have gotten with the Garmin.
Selective Availability has been turned off for a long time now, but it can be turned on at the whim of the Department of Defense.
It would really serve no purpose to do so at this point in time, or in any foreseeable scenario with regards to national security. Everyone knows where everything is. (Sorry, bad generalization.)
With regards to locating drainage tiles, I have done a great deal of this using a tool called a Magnetomatic. It is basically a telescopic car antenna attached to a swivel that is mounted through a handle. When you cross something that is buried, the antenna will swing in the direction of the object or line that is buried.
It works on plastic, poly, tile, transite, and metal lines, as well as fiber optic, and twisted pair copper lines. At least it does for me, and most of the people I work with.
Some people are unable to get it to work for them. Most of them are left handed, and the others are people who are unable to wear a movement style watch. (I believe these people have reversed polarity, but I cannot prove it.)
The process is similar to witching (divining) for water.
You can also bend two pinflags at a 90 degree angle to make short handles for each, put one in each hand, with the long ends parallel to each other, and parallel to the ground, about waist high, keeping them a shoulder's width apart. They will cross when you encountered something buried.
If you are using the GPS to measure a distance from a known location to find other unknown locations, say starting at a pump in the corner of a field, wanting to find the grid of drain pipes every 100 feet, the Garmin will get you close, but you will still want to verify the locations somehow, rather than blindly trusting the locations provided by the Garmin. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | Over a small area you may be able to do something. I don't know if you have a small area to deal with, and 10ft might be pushing it even then. Basically you fix the position of a known point by some means (it might not even have to be fixed absolutely depending on your goal). *Long term* GPS averaging might do that for you. Then you use this known value to calibrate an offset. You'd need a laptop connected to a dedicated GPS unit and probably custom code. If you could also access differential GPS signals you'd need a beacon receiver for that, but wouldn't need the known point or your own code.
But over a sufficiently small area compass triangulation can get surprisingly good *relative* accuracy; once again you need well-surveyed references (and line of sight to them). | Ten feet isn't unreasonable, although it's right at the edge of day to day accuracy. I used a Garmin 60c along with aerial photographs to create an orienteering map. I had an overall accuracy of about 5 meters, coupled with a 8 meter differential between photo and GPS caused by the 40 meter difference in elevation between floodplain and highlands.
Ten feet from a cold start in under two minutes isn't reasonable.
A: Enable WAAS. Some of the error in GPS is due to ionization state of the upper atmosphere. (More ions, slower radio) This tends to be a large scale phenomena so by comparing the calculated GPS position, to a real location, you can figure out a correction. This correction is broadcast. This routinely gets errors down to under 3 meters -- 10 feet.
B: I found when using a GPS for making trail and cross country maps that it would take about half an hour to really lock in. The longer it had been since I last used the GPS the longer it took. If it had been off only a day, it took only a few minutes. I also found in woods that sometimes moving just a few feet would improve my accuracy. I suspect line of sight to the optimum satellite constellation was blocked by a tree.
C: The constellation of satellites in the sky makes a difference. Bunched satellites or too close to the horizon are worse than having a wide spread set over 30 degrees above the horizon.
D: Some of the errors vary over long periods (~1 hour) So the relative error of two fixes made close together in time can be quite small. I used this to good effect by having locations where I could get an accurate map fix (Benchmarks, road intersections, power pylons) Go to the location, take a waypoint, go to my new spot, take a waypoint. The difference was repeatable to 1-2 meters, if the time difference was under 10 minutes. (For orienteering having small local relative errors is more important than absolute accuracy. You want to know which side of the road a landmark is.)
E: Remember that the error is statistical, and has a Poisson distribution (long tail) When the GPS error is 3 meters, that means that half of the time, the difference between measure and actual is 3 meters or less. The other half of the time it can be more.
F: In mountains and urban locations, you can get reflections. These can move your fix by a couple of kilometers. Conditions for this in the mountains aren't common, but if your fix makes no sense, wait a few minutes.
G: A caution: Most of the US Topo maps and Canadian topo maps have contour information derived from stereo aerial photographs. The Canadian maps can have a horizontal absolute error in contour line placement of up to 90 meters, north/south and half that east west. This can result in the GPS telling you you are at the top of the cliff when you plot your location, while the reality places you at the bottom. Benchmark placement, however is very accurate.
H: Forget the phone. Fine for finding street routes, not much else.
See this thread on Robotics.SE: It talks about various differential GPS techniques that can get sub 1 meter accuracy.
<https://robotics.stackexchange.com/questions/4312/low-cost-centimeter-accurate-satellite-positioning-gnss-gps>
More info:
<https://en.wikipedia.org/wiki/Global_Positioning_System> |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | There are several ways to use GPS signals that would give you the desired accuracy.
You can use either DGPS or a SBAS (WAAS in the US) for an accuracy of meters.
For a centimeter level accuracy you would need an RTK o PPP device.
I doubt you can get the desired accuracy in less than two minutes since the startup of the device unless you have a data connection to obtain the GPS almanac (I guess you could use a satellite Internet service).
Once the device has located the GPS satellites, it should be pretty fast.
In a farm setting a SBAS system is usually the most cost-effective option. | This answer is based strictly upon my personal experience.
I am not trying to sell you anything.
There are consumer grade GPS receivers which will get you approximately 3 meters of repeatable accuracy if they are used properly, and there is little or no overhead canopy.
I have had amazing results using a Garmin GLO which receives single frequency GPS, and Glonass data, along with WAAS corrections. ( About $110.00 )
It connected via bluetooth, and worked with multiple GIS, and mapping software applications, as well as various versions of the Windows operating system.
I have also used it with several models of semi-rugged, and rugged tablets without issue.
I have read that it is compatible with Android, Linux, IOS, and just about everything else out there. I only have Windows experience with it.
There were the occasional flyers that were in the five meter range, but in most cases, I was locating previously surveyed locations (Data collected with Survey Grade, Dual Frequency, RTK corrected, GPS/Glonass receivers) within a sub two meter radius.
One of the things I have found is, the units tend to work better when you are moving slightly. Just standing still, the position tends to walk around creating about a four meter circle.
The Garmin I was using had a 10 HZ update rate which created a very smooth line if I was mapping access routes, or roads.
The mapping data I collected, overlaid extremely well on previously located features.
There are other consumer grade external receivers that are probably capable of collecting data with this type of accuracy, I am just not experienced with them.
I have yet to see an embedded, discreet GPS chip within a mobile phone that approaches the results I have gotten with the Garmin.
Selective Availability has been turned off for a long time now, but it can be turned on at the whim of the Department of Defense.
It would really serve no purpose to do so at this point in time, or in any foreseeable scenario with regards to national security. Everyone knows where everything is. (Sorry, bad generalization.)
With regards to locating drainage tiles, I have done a great deal of this using a tool called a Magnetomatic. It is basically a telescopic car antenna attached to a swivel that is mounted through a handle. When you cross something that is buried, the antenna will swing in the direction of the object or line that is buried.
It works on plastic, poly, tile, transite, and metal lines, as well as fiber optic, and twisted pair copper lines. At least it does for me, and most of the people I work with.
Some people are unable to get it to work for them. Most of them are left handed, and the others are people who are unable to wear a movement style watch. (I believe these people have reversed polarity, but I cannot prove it.)
The process is similar to witching (divining) for water.
You can also bend two pinflags at a 90 degree angle to make short handles for each, put one in each hand, with the long ends parallel to each other, and parallel to the ground, about waist high, keeping them a shoulder's width apart. They will cross when you encountered something buried.
If you are using the GPS to measure a distance from a known location to find other unknown locations, say starting at a pump in the corner of a field, wanting to find the grid of drain pipes every 100 feet, the Garmin will get you close, but you will still want to verify the locations somehow, rather than blindly trusting the locations provided by the Garmin. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | Ten feet isn't unreasonable, although it's right at the edge of day to day accuracy. I used a Garmin 60c along with aerial photographs to create an orienteering map. I had an overall accuracy of about 5 meters, coupled with a 8 meter differential between photo and GPS caused by the 40 meter difference in elevation between floodplain and highlands.
Ten feet from a cold start in under two minutes isn't reasonable.
A: Enable WAAS. Some of the error in GPS is due to ionization state of the upper atmosphere. (More ions, slower radio) This tends to be a large scale phenomena so by comparing the calculated GPS position, to a real location, you can figure out a correction. This correction is broadcast. This routinely gets errors down to under 3 meters -- 10 feet.
B: I found when using a GPS for making trail and cross country maps that it would take about half an hour to really lock in. The longer it had been since I last used the GPS the longer it took. If it had been off only a day, it took only a few minutes. I also found in woods that sometimes moving just a few feet would improve my accuracy. I suspect line of sight to the optimum satellite constellation was blocked by a tree.
C: The constellation of satellites in the sky makes a difference. Bunched satellites or too close to the horizon are worse than having a wide spread set over 30 degrees above the horizon.
D: Some of the errors vary over long periods (~1 hour) So the relative error of two fixes made close together in time can be quite small. I used this to good effect by having locations where I could get an accurate map fix (Benchmarks, road intersections, power pylons) Go to the location, take a waypoint, go to my new spot, take a waypoint. The difference was repeatable to 1-2 meters, if the time difference was under 10 minutes. (For orienteering having small local relative errors is more important than absolute accuracy. You want to know which side of the road a landmark is.)
E: Remember that the error is statistical, and has a Poisson distribution (long tail) When the GPS error is 3 meters, that means that half of the time, the difference between measure and actual is 3 meters or less. The other half of the time it can be more.
F: In mountains and urban locations, you can get reflections. These can move your fix by a couple of kilometers. Conditions for this in the mountains aren't common, but if your fix makes no sense, wait a few minutes.
G: A caution: Most of the US Topo maps and Canadian topo maps have contour information derived from stereo aerial photographs. The Canadian maps can have a horizontal absolute error in contour line placement of up to 90 meters, north/south and half that east west. This can result in the GPS telling you you are at the top of the cliff when you plot your location, while the reality places you at the bottom. Benchmark placement, however is very accurate.
H: Forget the phone. Fine for finding street routes, not much else.
See this thread on Robotics.SE: It talks about various differential GPS techniques that can get sub 1 meter accuracy.
<https://robotics.stackexchange.com/questions/4312/low-cost-centimeter-accurate-satellite-positioning-gnss-gps>
More info:
<https://en.wikipedia.org/wiki/Global_Positioning_System> | This answer is based strictly upon my personal experience.
I am not trying to sell you anything.
There are consumer grade GPS receivers which will get you approximately 3 meters of repeatable accuracy if they are used properly, and there is little or no overhead canopy.
I have had amazing results using a Garmin GLO which receives single frequency GPS, and Glonass data, along with WAAS corrections. ( About $110.00 )
It connected via bluetooth, and worked with multiple GIS, and mapping software applications, as well as various versions of the Windows operating system.
I have also used it with several models of semi-rugged, and rugged tablets without issue.
I have read that it is compatible with Android, Linux, IOS, and just about everything else out there. I only have Windows experience with it.
There were the occasional flyers that were in the five meter range, but in most cases, I was locating previously surveyed locations (Data collected with Survey Grade, Dual Frequency, RTK corrected, GPS/Glonass receivers) within a sub two meter radius.
One of the things I have found is, the units tend to work better when you are moving slightly. Just standing still, the position tends to walk around creating about a four meter circle.
The Garmin I was using had a 10 HZ update rate which created a very smooth line if I was mapping access routes, or roads.
The mapping data I collected, overlaid extremely well on previously located features.
There are other consumer grade external receivers that are probably capable of collecting data with this type of accuracy, I am just not experienced with them.
I have yet to see an embedded, discreet GPS chip within a mobile phone that approaches the results I have gotten with the Garmin.
Selective Availability has been turned off for a long time now, but it can be turned on at the whim of the Department of Defense.
It would really serve no purpose to do so at this point in time, or in any foreseeable scenario with regards to national security. Everyone knows where everything is. (Sorry, bad generalization.)
With regards to locating drainage tiles, I have done a great deal of this using a tool called a Magnetomatic. It is basically a telescopic car antenna attached to a swivel that is mounted through a handle. When you cross something that is buried, the antenna will swing in the direction of the object or line that is buried.
It works on plastic, poly, tile, transite, and metal lines, as well as fiber optic, and twisted pair copper lines. At least it does for me, and most of the people I work with.
Some people are unable to get it to work for them. Most of them are left handed, and the others are people who are unable to wear a movement style watch. (I believe these people have reversed polarity, but I cannot prove it.)
The process is similar to witching (divining) for water.
You can also bend two pinflags at a 90 degree angle to make short handles for each, put one in each hand, with the long ends parallel to each other, and parallel to the ground, about waist high, keeping them a shoulder's width apart. They will cross when you encountered something buried.
If you are using the GPS to measure a distance from a known location to find other unknown locations, say starting at a pump in the corner of a field, wanting to find the grid of drain pipes every 100 feet, the Garmin will get you close, but you will still want to verify the locations somehow, rather than blindly trusting the locations provided by the Garmin. |
16,143 | I've already read [this](https://outdoors.stackexchange.com/questions/5446/what-specific-features-should-i-take-into-account-when-buying-gps-device-for-geo) and [this](https://outdoors.stackexchange.com/questions/4045/how-to-select-a-good-gps-receiver) to get some background knowledge on GPS tools, but I'm still left with some questions:
1. Is there any way I can get a device that consistently provides accuracy to within 10 feet in less than a minute or two?
2. I've messed around with several iPhone apps, but haven't had very good luck thus far. What's the most accurate a phone can be?
3. If I can't have accuracy to less than 10 feet within two minutes, what's the closest I can get to that goal? (Both in terms of time and accuracy)
I only need to be able to accurately determine the coordinates of specific locations, so I don't need any additional features--even poor battery life isn't a big deal. | 2017/05/14 | [
"https://outdoors.stackexchange.com/questions/16143",
"https://outdoors.stackexchange.com",
"https://outdoors.stackexchange.com/users/13109/"
] | There are several ways to use GPS signals that would give you the desired accuracy.
You can use either DGPS or a SBAS (WAAS in the US) for an accuracy of meters.
For a centimeter level accuracy you would need an RTK o PPP device.
I doubt you can get the desired accuracy in less than two minutes since the startup of the device unless you have a data connection to obtain the GPS almanac (I guess you could use a satellite Internet service).
Once the device has located the GPS satellites, it should be pretty fast.
In a farm setting a SBAS system is usually the most cost-effective option. | Ten feet isn't unreasonable, although it's right at the edge of day to day accuracy. I used a Garmin 60c along with aerial photographs to create an orienteering map. I had an overall accuracy of about 5 meters, coupled with a 8 meter differential between photo and GPS caused by the 40 meter difference in elevation between floodplain and highlands.
Ten feet from a cold start in under two minutes isn't reasonable.
A: Enable WAAS. Some of the error in GPS is due to ionization state of the upper atmosphere. (More ions, slower radio) This tends to be a large scale phenomena so by comparing the calculated GPS position, to a real location, you can figure out a correction. This correction is broadcast. This routinely gets errors down to under 3 meters -- 10 feet.
B: I found when using a GPS for making trail and cross country maps that it would take about half an hour to really lock in. The longer it had been since I last used the GPS the longer it took. If it had been off only a day, it took only a few minutes. I also found in woods that sometimes moving just a few feet would improve my accuracy. I suspect line of sight to the optimum satellite constellation was blocked by a tree.
C: The constellation of satellites in the sky makes a difference. Bunched satellites or too close to the horizon are worse than having a wide spread set over 30 degrees above the horizon.
D: Some of the errors vary over long periods (~1 hour) So the relative error of two fixes made close together in time can be quite small. I used this to good effect by having locations where I could get an accurate map fix (Benchmarks, road intersections, power pylons) Go to the location, take a waypoint, go to my new spot, take a waypoint. The difference was repeatable to 1-2 meters, if the time difference was under 10 minutes. (For orienteering having small local relative errors is more important than absolute accuracy. You want to know which side of the road a landmark is.)
E: Remember that the error is statistical, and has a Poisson distribution (long tail) When the GPS error is 3 meters, that means that half of the time, the difference between measure and actual is 3 meters or less. The other half of the time it can be more.
F: In mountains and urban locations, you can get reflections. These can move your fix by a couple of kilometers. Conditions for this in the mountains aren't common, but if your fix makes no sense, wait a few minutes.
G: A caution: Most of the US Topo maps and Canadian topo maps have contour information derived from stereo aerial photographs. The Canadian maps can have a horizontal absolute error in contour line placement of up to 90 meters, north/south and half that east west. This can result in the GPS telling you you are at the top of the cliff when you plot your location, while the reality places you at the bottom. Benchmark placement, however is very accurate.
H: Forget the phone. Fine for finding street routes, not much else.
See this thread on Robotics.SE: It talks about various differential GPS techniques that can get sub 1 meter accuracy.
<https://robotics.stackexchange.com/questions/4312/low-cost-centimeter-accurate-satellite-positioning-gnss-gps>
More info:
<https://en.wikipedia.org/wiki/Global_Positioning_System> |
1,535,635 | I hope someone can help with this. I am a statistician looking for a good book on high dimensional probability and data analysis. Basically I am looking for the equivalent of Terry Tao's 2 volume set on Analysis, but for high dimensional probability. Let me qualify what I am looking for.
Now there are a bunch of books out there with these very words in the title. I will list some below. But most of these are geared towards just pure machine learning folks or computer science. So often books on high dimensional data focus on techniques like Principle Components Analysis or Lasso, etc., to analyze high dimensional data. In developing these models, the authors start off with strong parametric assumptions about exponential family distributions or independence, etc. These book lack any sort of organic development of a theory behind adding dimensions to a data set or changes in the patterns of symmetry as a data set grows larger--both in dimensions and in the number of observations.
A basic probability text book will begin with a definition of random variables and work its way towards the Central Limit Theorem. While that is good for an intro stats course, there are a lot of problems with assuming normality even in high dimensional situations.
An example of such a book is:
[Geometric Structure of High-Dimensional Data and Dimensionality Reduction](http://rads.stackoverflow.com/amzn/click/364227496X)
[Statistics for High-Dimensional Data: Methods, Theory and Applications](http://rads.stackoverflow.com/amzn/click/3642201911)
(Please note that I am not critizing any of the books mentioned. I am just saying that these books don't fit my particular need.)
So what I am looking for is a more analytic look at how probability varies as dimensions get rather high. I use Terry Tao's book as an example of a wonderful development of analysis from basic foundations. I am looking for the same treatment for high dimensional data. I am not sure if I should be looking at a book on measure theory, or calculus on manifolds, or where?
Any suggestions would really be appreciated. | 2015/11/18 | [
"https://math.stackexchange.com/questions/1535635",
"https://math.stackexchange.com",
"https://math.stackexchange.com/users/38239/"
] | It has been some time since the question was asked and I believe the recent text "High Dimensional Probability" by Roman Vershynin might be just what you are looking for. It covers a variety of topics, focusing on the concentration phenomenon on higher dimensions. It is both theoretically heavy and intuitionally approachable.
Its current draft is also available for free, in the author's webpage.
<https://www.math.uci.edu/~rvershyn/papers/HDP-book/HDP-book.html#> | A recently published book by Wainwright, on high dimensional data, which seems useful
[High-dimensional statistics : a non-asymptotic viewpoint](https://books.google.com/books/about/High_Dimensional_Statistics.html?id=8C8nuQEACAAJ) |
41,240,745 | I am looking for a version control system which I can use for the Code of VBA possibly in a form of *add-in* but similar to **SVN** or **Git** or even a commercial software. I know there is a discussion [here](https://stackoverflow.com/questions/131605/best-way-to-do-version-control-for-ms-excel) and [there](https://stackoverflow.com/questions/2996995/how-to-use-version-control-with-vba-code) however, they are older than 6 years and that is not exactly what I want to have. And I know one can use *Tortoise SVN* just for the code but not for the whole project (all modules of the code+tables) and I am aware of it that one can do it manually. | 2016/12/20 | [
"https://Stackoverflow.com/questions/41240745",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/3305448/"
] | >
> **NOTE: This feature has since been removed.** Rubberduck still allows exporting all code files from a VBA project to the file system, but as of 2019/02/08 while it allows importing all code files at once, it will not overwrite existing modules on import, which makes this answer pretty much invalid.
>
>
>
Disclaimer: I'm a contributor on the Rubberduck VBA project.
The Rubberduck VBA add-in for the VBE, adds a number of features, including Source Control for all VBA hosts. It's actively developed, open source and free, and works in all versions of Excel (and Access, and Word, and PowerPoint, etc) from Excel 2000 onwards, and in 32 and 64-bit versions of Office
Rubberduck uses the LibGit2Sharp library, so you can integrate directly with GitHub and other cloud based services, or you can simply export modules and forms to a folder.
You can find the project on GitHub here: <https://github.com/rubberduck-vba/Rubberduck> | I don't think there are any compelling source control systems targeting Excel VBA, you'll just have to pick one of the names you already have and adapt it for Excel VBA.
So, I most recently used (and would still choose) GIT but I had/have to export the modules out to a directory. Other Enterprise Level Excel systems I've worked on have similarly export modules and versioned those separately.
As for the sheets they typically were housed as a binary file; only on one occasion did I work on a system that stored the sheet logic that lent itself to source control version comparison, it was a custom Xml format.
Interestingly, the latest Excel file formats format of xlsm and xslx are in fact zip directories of sheet logic atomised. I've looked into this file format and it does not lend itself very well to being version by source control. I think Microsoft missed a trick here. |
59,500 | How can I take a screenshot of my PC's BIOS? | 2009/10/23 | [
"https://superuser.com/questions/59500",
"https://superuser.com",
"https://superuser.com/users/15114/"
] | Take a camera, point it at the screen and take a picture.
Unless you have some kind of very fancy KVM switch that works over network (no, the cheap home user ones don't do that). Then you can do a direct screen capture on the receiving end, I think. Requires another computer, though.
Yet another option would involve a graphics card with TV out, a video capture card and another computer too. And some hope that video out works in text mode out of the box. I think you can guess where that goes.
Seriously, take a camera. It's by far the easiest and quickest way of getting a screen grab of something that happens way before you control the computer. | You can use an IP KVM switch with 'BIOS level access' (not exactly cheap though) and access the BIOS 'remotely' from another machine and take 'proper' screenshots there.
 |
59,500 | How can I take a screenshot of my PC's BIOS? | 2009/10/23 | [
"https://superuser.com/questions/59500",
"https://superuser.com",
"https://superuser.com/users/15114/"
] | Take a camera, point it at the screen and take a picture.
Unless you have some kind of very fancy KVM switch that works over network (no, the cheap home user ones don't do that). Then you can do a direct screen capture on the receiving end, I think. Requires another computer, though.
Yet another option would involve a graphics card with TV out, a video capture card and another computer too. And some hope that video out works in text mode out of the box. I think you can guess where that goes.
Seriously, take a camera. It's by far the easiest and quickest way of getting a screen grab of something that happens way before you control the computer. | Some motherboards can be set up to push the boot-up and BIOS interface out the serial port. If your's can, then you could connect it to another machine with a terminal emulator running, and take a screen capture of the terminal emulator window. |
2,427 | Should we allow genre meta tags, and if so, should they be applicable to any series within that genre so that fans of that genre can more easily find those questions in order to answer them/learn more, or only to questions inquiring about the genre itself?
For example, we currently have genre questions such as the following, which have a smattering of varied tags attached to them, such as [anime-production](https://anime.stackexchange.com/questions/tagged/anime-production "show questions tagged 'anime-production'") [japanese-language](https://anime.stackexchange.com/questions/tagged/japanese-language "show questions tagged 'japanese-language'") [terminology](https://anime.stackexchange.com/questions/tagged/terminology "show questions tagged 'terminology'") [tropes](https://anime.stackexchange.com/questions/tagged/tropes "show questions tagged 'tropes'") and others, but some of the tags chosen weren't applicable to the question (the OP might not have been able to select a more appropriate tag).
* [Where did the mecha genre originate from in Japan?](https://anime.stackexchange.com/questions/3629/where-did-the-mecha-genre-originate-from-in-japan)
* [Who is the main target audience of yuri titles?](https://anime.stackexchange.com/questions/13482/who-is-the-main-target-audience-of-yuri-titles)
* [What is the opposite of “seme” according to otakus?](https://anime.stackexchange.com/questions/14364/what-is-the-opposite-of-seme-according-to-otakus)
* [In yaoi anime, why does the uke always say 'senpai'?](https://anime.stackexchange.com/questions/13325/in-yaoi-anime-why-does-the-uke-always-say-senpai)
* [Why are the male love-interests of shoujo often portrayed to be top scorers?](https://anime.stackexchange.com/questions/2420/why-are-the-male-love-interests-of-shoujo-often-portrayed-to-be-top-scorers)
* [Have the creators of Tokyo Mew Mew or Sailor Moon commented on the similarities?](https://anime.stackexchange.com/questions/3646/have-the-creators-of-tokyo-mew-mew-or-sailor-moon-commented-on-the-similarities)
* [What does 'moe' mean?](https://anime.stackexchange.com/questions/894/what-does-moe-mean)
* [What would the list of manga genres in Japan be, based on bookstore labeling schemes?](https://anime.stackexchange.com/questions/19480/what-would-the-list-of-manga-genres-in-japan-be-based-on-bookstore-labeling-sch)
* [What is the difference between yuri and shoujo-ai, yaoi and shounen-ai and ecchi and hentai anime genre?](http://%20https://anime.stackexchange.com/questions/2681/what-is-the-difference-between-yuri-and-shoujo-ai-yaoi-and-shounen-ai-and-ecchi)
* [Is there a difference in drawing style between different anime for different demographics?](https://anime.stackexchange.com/questions/3477/is-there-a-difference-in-drawing-style-between-different-anime-for-different-dem/3483#3483)
* [What is the target demographic of Yuki Yuna is a Hero?](https://anime.stackexchange.com/questions/15533/what-is-the-target-demographic-of-yuki-yuna-is-a-hero)
I would propose that
1. A "genre" tag could be useful for people to use when asking a question about **differences between or about similarities between genres and demographics**. It would be nice if questions like those above could be located in an easy-to-find location for users interested in answering and reading questions about genre.
2. Anime/manga-specific genre tags (such as "shoujo," "mahou shoujo", "yuri," yaoi," "moe," "nichijou," "seinen," etc.) could be useful so that, for example, someone who is **an expert in "josei" could easily find questions about their area of speciality to answer, and to find answers of interest to read**, without having to wade through many questions searching by series title. I, for one, would use such tags to periodically check for questions I could likely answer.
3. Although [previous discussion](https://anime.meta.stackexchange.com/a/2239/8134) posited that "a genre-based tag *could* work, but it would need to be used on questions about the genre itself (e.g. production, tropes, etc.) and not questions about individual series," and that "Tagging series questions with genre tags would make the series tags redundant and the genre tags bloated," I would suggest that the **series tags would not become redundant**, because a user who can generally answer "shoujo" questions and is interested in reading their answers is **not likely to search title by title just to check for them**, especially since the OP might not know the title (for [identification-request](https://anime.stackexchange.com/questions/tagged/identification-request "show questions tagged 'identification-request'")) or the user might not remember the title off-hand in order to search for it but would recognize it if seen within that tag.
4. Some of the [previous reasons](https://anime.meta.stackexchange.com/a/2238/8134) given for not creating a "hentai" tag were variations on being age-appropriate for a 13+ age, which is **not applicable** regarding genre tags other than "hentai" and possibly "yaoi" and "yuri."
5. Genre tags that are not specific to anime/manga (such as "sci-fi," "romance," "comedy") would be less useful than those for genres which have originated in anime/manga. I would make a motion that **only genres which have developed through anime/manga mediums** get tags. I don't think there are so many anime/manga-specific genres that the number of genre tags would be too unwieldy, nor that frequent users would try to pile a ton of genre tags all onto one question.
6. Meta tags which were [previously voted against](https://anime.meta.stackexchange.com/a/402/8134) were "characters," "character-abilities," "variations," and "plot-explanation," which I heartily agree with as 1) not helping users to find questions and answers of interest via searching by tag, 2) "don't add much to the question," and 3) "hurt search engine visibility." Anime/manga-specific genre tags would contribute to **search engine optimization** rather than detract from it. | 2015/06/11 | [
"https://anime.meta.stackexchange.com/questions/2427",
"https://anime.meta.stackexchange.com",
"https://anime.meta.stackexchange.com/users/8134/"
] | >
> 2. Anime/manga-specific genre tags (such as "shoujo," "mahou shoujo", "yuri," yaoi," "moe," "nichijou," "seinen," etc.) could be useful so that, for example, someone who is **an expert in "josei" could easily find questions about their area of speciality to answer, and to find answers of interest to read,** without having to wade through many questions searching by series title. I, for one, would use such tags to periodically check for questions I could likely answer.
>
>
>
>
> 6. Meta tags which were previously voted against were "characters," "character-abilities," "variations," and "plot-explanation," which I heartily agree with as 1) not helping users to find questions and answers of interest via searching by tag, 2) "don't add much to the question," and 3) "hurt search engine visibility." Anime/manga-specific genre tags would contribute to **search engine optimization** rather than detract from it.
>
>
>
I disagree with **tagging a question with genre tag in addition to the series title** for the following reasons:
* Without exception, the genre tags will end up having more question than the rest of the series tag. For the same reason as why the rest of the meta tags are frowned upon, it will hurt search engine visibility. Note that this refers to ranking of our pages on external search engines -- like Google, Bing, Yahoo, etc. -- not the search engine on SE.
Since the genre tag, which has the highest count of questions, will be placed in the title of the HTML page, as the search engine gives more weight to matches in the page title over the matches in the page body, questions from our site will rank lower.
Do note that it's natural to search for the name of the series rather than the genre of the series on search engines.
* Even if the problem above is resolved, as the number of genre tags expands, so is the amount of effort to keep the tags consistent over all questions of a series, for existing questions and incoming questions.
This is not yet mentioning all the potential meta issues that come with the introduction of genre tags (e.g. Which genre tags should series X have? How should we add genre tags for questions with 2 series tags? Question ABC will exceed the 5-tag limit - which genre tag should be removed?).
* When a question is tagged with series tag, the genre is mostly not relevant to the subject-matter in the question. It only serves to describe the series tag.
While I agree that the current system is not ideal for users to look for and answer questions by genre, I don't think tagging every question with its genres is a good solution.
Maybe a better solution is a tagging system for tags, where series tags are tagged with their genres (since the main purpose here is to look for series by genre), though the chances that it's implemented here at SE is small, since it will introduce significant changes to the existing system. | I don't know if they're a good thing. I feel that they could keep expanding until they're unmaintainably large:
What is this anime? [romance](https://anime.stackexchange.com/questions/tagged/romance "show questions tagged 'romance'") [yuri](https://anime.stackexchange.com/questions/tagged/yuri "show questions tagged 'yuri'") [shoujo-ai](https://anime.stackexchange.com/questions/tagged/shoujo-ai "show questions tagged 'shoujo-ai'") [shounen](https://anime.stackexchange.com/questions/tagged/shounen "show questions tagged 'shounen'") [seinen](https://anime.stackexchange.com/questions/tagged/seinen "show questions tagged 'seinen'") [identification-request](https://anime.stackexchange.com/questions/tagged/identification-request "show questions tagged 'identification-request'")
I read a lot of shoujo, and I do see that having the tag would be useful if you wanted to only answer shoujo questions (or similar) - but a search for the word *shoujo* or finding particular series seems sufficient to me. **Tags seem more to indicate what the post is *about* rather than a description of the contents.**
Example: The question is about looking for an id for an anime. The question is not about shoujo anime. (rather it is a description of the id-request)
One thing I would recommend regardless is setting up your favourite tags so that those questions are hilighted for you.
---
I might expand on this later, but I'm currently at work - so I'm making it community wiki if other people want to tag onto this side of the argument |
18,751,027 | Would concerns help me share logic for partials? I see Rails 4 has the "concerns" folder for controllers, but it's empty on app generation.
DHH wrote a blog on concerns (but I think they are for the model): <http://37signals.com/svn/posts/3372-put-chubby-models-on-a-diet-with-concerns>
I would appreciate a concise example. | 2013/09/11 | [
"https://Stackoverflow.com/questions/18751027",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1155375/"
] | Concerns are modules that get mixed into controller or model classes for instance. DHH's post focused on models, but the same applies to controllers. It helps slim down the model or controller classes, and makes it easier to reuse common code across multiple classes.
Helpers are for short presentation-oriented methods that you want to call from your views, or very occasionally from your controllers, maybe. | Simply put:
Helpers contain methods to be used in views.
Concerns are modules to be included in controllers/models(depending which concern is it) etc.
It's just a folder with code that you could otherwise put in another place if you wanted. We use fancy terms like concerns to get paid more. |
4,828 | Defensive backs often get them from interceptions but it's possible for other players have forced fumbles, strips and blocked kicks. | 2014/06/02 | [
"https://sports.stackexchange.com/questions/4828",
"https://sports.stackexchange.com",
"https://sports.stackexchange.com/users/3704/"
] | I also believe that this could be very subjective, however, I will try to apply some logic to find a best available answer.
To take a 2014 approach, and looking at the statistics through the first three weeks of the season, it appears defensive backs may have the most influence on turnovers, at least during this season.
To explain, and according to ESPN.com statistics regarding forced fumbles and interceptions, there are 8 players with two or more recorded interceptions this season and 16 players with two or more forced fumbles this season. Of those 24 players that influenced two or more turnovers, 16, or nearly 67%, are defensive backs. | Looking at stats from the 2013...
Richard Sherman had [8 interceptions](http://www.nfl.com/stats/player?seasonId=2013&seasonType=REG&Submit=Go).
The leader in fumbles recovered was [Mychal Kendricks](http://www.teamrankings.com/nfl/player-stat/fumbles-recovered-lost-by-opposition) with 4.
The leader in forced fumbles was [Robert Mathis](http://www.teamrankings.com/nfl/player-stat/defense-forced-fumbles) with 10.
It is tough to track because interceptions are a turnover, where forced fumbles **may not** be a turnover. I cannot find stats on individual turnovers created. There are team turnover differentials, but not stats for an individual (or position).
I would conclude that defensive backs, and more specifically corner backs create the most turnovers based on the stats available. Especially in today's NFL which is a passing league. |
4,828 | Defensive backs often get them from interceptions but it's possible for other players have forced fumbles, strips and blocked kicks. | 2014/06/02 | [
"https://sports.stackexchange.com/questions/4828",
"https://sports.stackexchange.com",
"https://sports.stackexchange.com/users/3704/"
] | I also believe that this could be very subjective, however, I will try to apply some logic to find a best available answer.
To take a 2014 approach, and looking at the statistics through the first three weeks of the season, it appears defensive backs may have the most influence on turnovers, at least during this season.
To explain, and according to ESPN.com statistics regarding forced fumbles and interceptions, there are 8 players with two or more recorded interceptions this season and 16 players with two or more forced fumbles this season. Of those 24 players that influenced two or more turnovers, 16, or nearly 67%, are defensive backs. | I do not have hard research to support this, however in my experience Quarterbacks have the most *fumbles* on any given team – generally speaking. Not the least factor in that is they touch the ball more than any other player.
Given that, I believe **Defensive Ends** are *”most likely to create turnovers in football”.* [My college philosophy/logic professor is doing a facepalm.]
Probably by far and away; if someone were to ferret out the proper statistics. However – and of note – big cornerbacks (like the almighty Ronnie Lott) often have the most dramatic turnovers through interceptions and big hits across the middle; and those are the ones that imprint the largest in our minds – but I believe the numbers will flush out that the **Defensive End** is the correct answer [...unsung workhorse and most beautiful specimen of all Football players.]
(Note: It would be prudent to make sure the 2 *corners* and 2 *safety* positions are not combined into “backs” for comparison - as 4 players on a team just might collect more numbers than 2 defensive ends – and **diggers3**'s excellently made point about forced fumbles not all being turnovers notwithstanding.) |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | There are three important sources of energy in planetary bodies (or their moons) that may be important in determining their suitability for life. All of these are evident in various bodies in the Solar System.
The first is tidal heating. When one object orbits another then there will be gradient in the gravitational force across their finite sizes. If this is combined with any orbital non-circularity or difference between rotational and orbital periods, along with some fluidity in the structure of one of the bodies, then it will be heated by frictional forces as the fluid moves in response to the changing gravitational gradient. This heating is most evident on Io in our own Solar System, the closest large moon of Jupiter, and is generally more important for objects orbiting close to more massive stars or planets.
The second is heating by radioactive decay. When planetary systems are born, they can incorporate significant amounts of radioactive material. Some of this has a short half-life and may be responsible for a short period of intense heating, but other isotopes are much longer lived. According to [this article](https://blogs.scientificamerican.com/observations/nuclear-fission-confirmed-as-source-of-more-than-half-of-earths-heat/), about 20 TW of power is generated inside the Earth from the radioactive decay of uranium, thorium and potassium. This sounds a lot, but is about a factor $10^4$ less power than we receive from the Sun.
The third source is just the "heat of formation". Gathering together the raw materials of a planet/moon and compressing it into a gravitational potential well, inevitably leads to the generation of considerable heat and it can take billions of years for this heat to leak out into space. Indeed, for the Earth, the heat of formation is probably comparable with radioactive decay in terms of supplying energy to the surface from the interior. | Stars function by nuclear fusion. There is energy that isn't released by nuclear fusion, nor from the nuclear decay of elements produced by fusion.
There is the cosmic microwave background. This is radiation, but it "cold", at about -268 \*C. There is a lot of energy, but because it is so "cold" it is difficult to use it to power anything.
There is a lot of gravitational potential energy. This can be released when things fall together. So when an asteroid hits the Earth, the energy released is not from a star. However its not a dependable source of energy for life.
There are proto-stars that are powered by gravitational energy, and there are accretion discs around neutron stars and black holes that are hot due to the gravitational energy released as matter falls into them. The hard X-rays from accretion discs isn't much good for life.
There's also a lot of potential energy in a cloud of interstellar gas that could be released if it all fused to heavier elements. However the only effective way to release this energy is to make stars: Lots of energy, not easy to get at.
In principle, all mass is energy, by E=mc². However, using this energy to power something isn't easy.
Really you are asking the wrong question. There's lots of energy. The trouble is converting it to do useful work, otherwise all this energy is about as useful as a piece of coal to a hungry person. |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | There are three important sources of energy in planetary bodies (or their moons) that may be important in determining their suitability for life. All of these are evident in various bodies in the Solar System.
The first is tidal heating. When one object orbits another then there will be gradient in the gravitational force across their finite sizes. If this is combined with any orbital non-circularity or difference between rotational and orbital periods, along with some fluidity in the structure of one of the bodies, then it will be heated by frictional forces as the fluid moves in response to the changing gravitational gradient. This heating is most evident on Io in our own Solar System, the closest large moon of Jupiter, and is generally more important for objects orbiting close to more massive stars or planets.
The second is heating by radioactive decay. When planetary systems are born, they can incorporate significant amounts of radioactive material. Some of this has a short half-life and may be responsible for a short period of intense heating, but other isotopes are much longer lived. According to [this article](https://blogs.scientificamerican.com/observations/nuclear-fission-confirmed-as-source-of-more-than-half-of-earths-heat/), about 20 TW of power is generated inside the Earth from the radioactive decay of uranium, thorium and potassium. This sounds a lot, but is about a factor $10^4$ less power than we receive from the Sun.
The third source is just the "heat of formation". Gathering together the raw materials of a planet/moon and compressing it into a gravitational potential well, inevitably leads to the generation of considerable heat and it can take billions of years for this heat to leak out into space. Indeed, for the Earth, the heat of formation is probably comparable with radioactive decay in terms of supplying energy to the surface from the interior. | The answer by mentioned two internal sources of energy in planets and moons, etc.:
one) Tidal heating. In some situations tidal heating hasno significant effect, in ohters it can make a cold moon warm enough for life, in others it could cause a runaway greenhouse effect.
Two) Energy produced by the decay of radioactive isotropes.
But there is at least one other sourceo of internal energy in planets and moons, etc.
Three) The left over heat from the formation of the world as grains of dust, molecules of gas, and larger objects fell thousands or millions of miles toward each other and eventually impacted with great force. |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | Stars function by nuclear fusion. There is energy that isn't released by nuclear fusion, nor from the nuclear decay of elements produced by fusion.
There is the cosmic microwave background. This is radiation, but it "cold", at about -268 \*C. There is a lot of energy, but because it is so "cold" it is difficult to use it to power anything.
There is a lot of gravitational potential energy. This can be released when things fall together. So when an asteroid hits the Earth, the energy released is not from a star. However its not a dependable source of energy for life.
There are proto-stars that are powered by gravitational energy, and there are accretion discs around neutron stars and black holes that are hot due to the gravitational energy released as matter falls into them. The hard X-rays from accretion discs isn't much good for life.
There's also a lot of potential energy in a cloud of interstellar gas that could be released if it all fused to heavier elements. However the only effective way to release this energy is to make stars: Lots of energy, not easy to get at.
In principle, all mass is energy, by E=mc². However, using this energy to power something isn't easy.
Really you are asking the wrong question. There's lots of energy. The trouble is converting it to do useful work, otherwise all this energy is about as useful as a piece of coal to a hungry person. | There are also [electric bacteria](https://en.wikipedia.org/wiki/Electric_bacteria). The BBC's [There are microbes that eat and poo nothing but electricity](http://www.bbc.com/earth/story/20160613-there-are-microbes-that-eat-and-poo-nothing-but-electricity) coverts the topic extensively.
At first electrical forms of energy absorption and excretion seemed exotic, but this now is seen as a widespread phenomenon on Earth, including colonies that include multiple species working together to literally *complete the circuit*.
These organisms don't derive their energy directly from Sunlight or heat, but *the ones currently found on Earth* do require enough warmth to have liquid water environments in which to ply their trade.
So perhaps as long as self gravitation leftover heat and radioactivity mentioned in other answers can provide enough warmth for a liquid water environment, bacteria-like life can feed directly off of the mixtures of primordial chemicals (organics, metals, oxides, sulfides, etc.) without necessarily [eating the chemicals themselves](https://en.wikipedia.org/wiki/Chemosynthesis), but instead forming electrical pathways through which electrochemical reactions can take place.
From the BBC article:
>
> While most organisms get their electron fix from carbohydrates, some bacteria can harvest electrons in their purest form. They can effectively "eat" electrons from minerals and rocks. In a way, they are getting their electrical energy straight from the socket.
>
>
> Annette Rowe, a graduate student of Nealson, has found [six new bacterial species on the ocean floor that can live off electricity alone](http://dx.doi.org/10.3389/fmicb.2014.00784). All are very different to one another, and none of them is anything like Shewanella or Geobacter.
>
>
> She found that, when no other food source was available, the bacteria would happily take electrons directly from the electrodes. In their natural habitat, the bacteria likely take their electrons directly from iron and sulphur in the seabed.
>
>
> Examples of electron-eating bacteria found by Rowe include *Halomonas, Idiomarina, Marinobacter, and Pseudomonas* of the *Gammaproteobacteria*, and *Thalassospira and Thioclava* from the *Alphaproteobacteria*.
> Many more electron-loving bacteria have now been found. In fact all you have to do is stick an electrode in the ground and pass electrons down it, and soon the electrode will be coated with feeding bacteria. Experiments show that these bacteria essentially eat or excrete electricity.
>
>
>
>
> Some species of Geobacter, he says, can both directly transfer electrons to electrodes and also directly accept electrons from them.
> In 2015, we learned that electron-eating and electron-excreting microbes can actually team up and pass electrons between each other, wiring themselves into a common electrical grid.
>
>
>
It turns out multiple types of bacteria can work together:
>
> Different species of bacteria and archaea – ancient single-celled microbes similar to bacteria in many ways – team up to degrade the methane before it can get the surface.
>
>
> Gunter Wegener from the Max Planck Institute for Marine Microbiology in Bremen wondered how the process worked. He [collected samples of the microbes](http://dx.doi.org/10.1038/nature15733), which live at temperatures of 60C on the ocean floor, and put them under a scanning electron microscope.
>
>
> The microscope revealed thin wire-like structures protruding from the bacterial cells. Although only a few nanometres wide, the wires were several micrometres long, which is much longer than the cells themselves.
> It seems that the bacteria use these nanowires to hook up with the archaea.
>
>
> **The archaea feed on electrons from methane, oxidising the gas to generate carbonate. They then pass the electrons on to their partner bacteria along the nanowires, which act like power cables. Finally the bacteria deposit the electrons onto sulphate, producing energy that the cell can use in the process.**
>
>
> |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | You can also have planets with powerful magnetic fields, a conductive body passing through them [generates electric currents](https://www.mtnet.info/division/papers/EMWKSHP_ReviewVolumes/1972Edinburgh/DyalParkin_1972EdinburghReview_PEPI_1973.pdf), that besides simply heating can also power chemical reactions directly (like charging up a chemical battery).
There is also chemical energy. If there is a constant supply or accumulated store of reactive chemicals, that can act as fuel or food or battery electrolytes. For example, there are [clouds of organic molecules in space](https://phys.org/news/2014-09-alcohol-clouds-space.html) that could rain down on to a planet and be used as 'food', or which could collide with other clouds and react. Or planets made of different chemicals could collide and then react with one another.
As rotating dust clouds collapse under gravity, they can form [accretion disks](https://www.universetoday.com/87597/zooming-in-on-proto-planetary-disks/) in which stars and planets form. The continual collisions generate heat, causing them to glow in the infrared. This likely includes cases too small to form a star (brown dwarf systems).
[Shock waves](https://en.wikipedia.org/wiki/Shock_waves_in_astrophysics) passing through interstellar gas clouds can produce intense heating. These shock waves are usually produced by stellar explosions (supernovas) but could in principle be produced by anything that causes high speeds, like the bow shock of a planet passing through a gas cloud. | There are also [electric bacteria](https://en.wikipedia.org/wiki/Electric_bacteria). The BBC's [There are microbes that eat and poo nothing but electricity](http://www.bbc.com/earth/story/20160613-there-are-microbes-that-eat-and-poo-nothing-but-electricity) coverts the topic extensively.
At first electrical forms of energy absorption and excretion seemed exotic, but this now is seen as a widespread phenomenon on Earth, including colonies that include multiple species working together to literally *complete the circuit*.
These organisms don't derive their energy directly from Sunlight or heat, but *the ones currently found on Earth* do require enough warmth to have liquid water environments in which to ply their trade.
So perhaps as long as self gravitation leftover heat and radioactivity mentioned in other answers can provide enough warmth for a liquid water environment, bacteria-like life can feed directly off of the mixtures of primordial chemicals (organics, metals, oxides, sulfides, etc.) without necessarily [eating the chemicals themselves](https://en.wikipedia.org/wiki/Chemosynthesis), but instead forming electrical pathways through which electrochemical reactions can take place.
From the BBC article:
>
> While most organisms get their electron fix from carbohydrates, some bacteria can harvest electrons in their purest form. They can effectively "eat" electrons from minerals and rocks. In a way, they are getting their electrical energy straight from the socket.
>
>
> Annette Rowe, a graduate student of Nealson, has found [six new bacterial species on the ocean floor that can live off electricity alone](http://dx.doi.org/10.3389/fmicb.2014.00784). All are very different to one another, and none of them is anything like Shewanella or Geobacter.
>
>
> She found that, when no other food source was available, the bacteria would happily take electrons directly from the electrodes. In their natural habitat, the bacteria likely take their electrons directly from iron and sulphur in the seabed.
>
>
> Examples of electron-eating bacteria found by Rowe include *Halomonas, Idiomarina, Marinobacter, and Pseudomonas* of the *Gammaproteobacteria*, and *Thalassospira and Thioclava* from the *Alphaproteobacteria*.
> Many more electron-loving bacteria have now been found. In fact all you have to do is stick an electrode in the ground and pass electrons down it, and soon the electrode will be coated with feeding bacteria. Experiments show that these bacteria essentially eat or excrete electricity.
>
>
>
>
> Some species of Geobacter, he says, can both directly transfer electrons to electrodes and also directly accept electrons from them.
> In 2015, we learned that electron-eating and electron-excreting microbes can actually team up and pass electrons between each other, wiring themselves into a common electrical grid.
>
>
>
It turns out multiple types of bacteria can work together:
>
> Different species of bacteria and archaea – ancient single-celled microbes similar to bacteria in many ways – team up to degrade the methane before it can get the surface.
>
>
> Gunter Wegener from the Max Planck Institute for Marine Microbiology in Bremen wondered how the process worked. He [collected samples of the microbes](http://dx.doi.org/10.1038/nature15733), which live at temperatures of 60C on the ocean floor, and put them under a scanning electron microscope.
>
>
> The microscope revealed thin wire-like structures protruding from the bacterial cells. Although only a few nanometres wide, the wires were several micrometres long, which is much longer than the cells themselves.
> It seems that the bacteria use these nanowires to hook up with the archaea.
>
>
> **The archaea feed on electrons from methane, oxidising the gas to generate carbonate. They then pass the electrons on to their partner bacteria along the nanowires, which act like power cables. Finally the bacteria deposit the electrons onto sulphate, producing energy that the cell can use in the process.**
>
>
> |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | Stars function by nuclear fusion. There is energy that isn't released by nuclear fusion, nor from the nuclear decay of elements produced by fusion.
There is the cosmic microwave background. This is radiation, but it "cold", at about -268 \*C. There is a lot of energy, but because it is so "cold" it is difficult to use it to power anything.
There is a lot of gravitational potential energy. This can be released when things fall together. So when an asteroid hits the Earth, the energy released is not from a star. However its not a dependable source of energy for life.
There are proto-stars that are powered by gravitational energy, and there are accretion discs around neutron stars and black holes that are hot due to the gravitational energy released as matter falls into them. The hard X-rays from accretion discs isn't much good for life.
There's also a lot of potential energy in a cloud of interstellar gas that could be released if it all fused to heavier elements. However the only effective way to release this energy is to make stars: Lots of energy, not easy to get at.
In principle, all mass is energy, by E=mc². However, using this energy to power something isn't easy.
Really you are asking the wrong question. There's lots of energy. The trouble is converting it to do useful work, otherwise all this energy is about as useful as a piece of coal to a hungry person. | The answer by mentioned two internal sources of energy in planets and moons, etc.:
one) Tidal heating. In some situations tidal heating hasno significant effect, in ohters it can make a cold moon warm enough for life, in others it could cause a runaway greenhouse effect.
Two) Energy produced by the decay of radioactive isotropes.
But there is at least one other sourceo of internal energy in planets and moons, etc.
Three) The left over heat from the formation of the world as grains of dust, molecules of gas, and larger objects fell thousands or millions of miles toward each other and eventually impacted with great force. |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | The answer by mentioned two internal sources of energy in planets and moons, etc.:
one) Tidal heating. In some situations tidal heating hasno significant effect, in ohters it can make a cold moon warm enough for life, in others it could cause a runaway greenhouse effect.
Two) Energy produced by the decay of radioactive isotropes.
But there is at least one other sourceo of internal energy in planets and moons, etc.
Three) The left over heat from the formation of the world as grains of dust, molecules of gas, and larger objects fell thousands or millions of miles toward each other and eventually impacted with great force. | There are also [electric bacteria](https://en.wikipedia.org/wiki/Electric_bacteria). The BBC's [There are microbes that eat and poo nothing but electricity](http://www.bbc.com/earth/story/20160613-there-are-microbes-that-eat-and-poo-nothing-but-electricity) coverts the topic extensively.
At first electrical forms of energy absorption and excretion seemed exotic, but this now is seen as a widespread phenomenon on Earth, including colonies that include multiple species working together to literally *complete the circuit*.
These organisms don't derive their energy directly from Sunlight or heat, but *the ones currently found on Earth* do require enough warmth to have liquid water environments in which to ply their trade.
So perhaps as long as self gravitation leftover heat and radioactivity mentioned in other answers can provide enough warmth for a liquid water environment, bacteria-like life can feed directly off of the mixtures of primordial chemicals (organics, metals, oxides, sulfides, etc.) without necessarily [eating the chemicals themselves](https://en.wikipedia.org/wiki/Chemosynthesis), but instead forming electrical pathways through which electrochemical reactions can take place.
From the BBC article:
>
> While most organisms get their electron fix from carbohydrates, some bacteria can harvest electrons in their purest form. They can effectively "eat" electrons from minerals and rocks. In a way, they are getting their electrical energy straight from the socket.
>
>
> Annette Rowe, a graduate student of Nealson, has found [six new bacterial species on the ocean floor that can live off electricity alone](http://dx.doi.org/10.3389/fmicb.2014.00784). All are very different to one another, and none of them is anything like Shewanella or Geobacter.
>
>
> She found that, when no other food source was available, the bacteria would happily take electrons directly from the electrodes. In their natural habitat, the bacteria likely take their electrons directly from iron and sulphur in the seabed.
>
>
> Examples of electron-eating bacteria found by Rowe include *Halomonas, Idiomarina, Marinobacter, and Pseudomonas* of the *Gammaproteobacteria*, and *Thalassospira and Thioclava* from the *Alphaproteobacteria*.
> Many more electron-loving bacteria have now been found. In fact all you have to do is stick an electrode in the ground and pass electrons down it, and soon the electrode will be coated with feeding bacteria. Experiments show that these bacteria essentially eat or excrete electricity.
>
>
>
>
> Some species of Geobacter, he says, can both directly transfer electrons to electrodes and also directly accept electrons from them.
> In 2015, we learned that electron-eating and electron-excreting microbes can actually team up and pass electrons between each other, wiring themselves into a common electrical grid.
>
>
>
It turns out multiple types of bacteria can work together:
>
> Different species of bacteria and archaea – ancient single-celled microbes similar to bacteria in many ways – team up to degrade the methane before it can get the surface.
>
>
> Gunter Wegener from the Max Planck Institute for Marine Microbiology in Bremen wondered how the process worked. He [collected samples of the microbes](http://dx.doi.org/10.1038/nature15733), which live at temperatures of 60C on the ocean floor, and put them under a scanning electron microscope.
>
>
> The microscope revealed thin wire-like structures protruding from the bacterial cells. Although only a few nanometres wide, the wires were several micrometres long, which is much longer than the cells themselves.
> It seems that the bacteria use these nanowires to hook up with the archaea.
>
>
> **The archaea feed on electrons from methane, oxidising the gas to generate carbonate. They then pass the electrons on to their partner bacteria along the nanowires, which act like power cables. Finally the bacteria deposit the electrons onto sulphate, producing energy that the cell can use in the process.**
>
>
> |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | There are three important sources of energy in planetary bodies (or their moons) that may be important in determining their suitability for life. All of these are evident in various bodies in the Solar System.
The first is tidal heating. When one object orbits another then there will be gradient in the gravitational force across their finite sizes. If this is combined with any orbital non-circularity or difference between rotational and orbital periods, along with some fluidity in the structure of one of the bodies, then it will be heated by frictional forces as the fluid moves in response to the changing gravitational gradient. This heating is most evident on Io in our own Solar System, the closest large moon of Jupiter, and is generally more important for objects orbiting close to more massive stars or planets.
The second is heating by radioactive decay. When planetary systems are born, they can incorporate significant amounts of radioactive material. Some of this has a short half-life and may be responsible for a short period of intense heating, but other isotopes are much longer lived. According to [this article](https://blogs.scientificamerican.com/observations/nuclear-fission-confirmed-as-source-of-more-than-half-of-earths-heat/), about 20 TW of power is generated inside the Earth from the radioactive decay of uranium, thorium and potassium. This sounds a lot, but is about a factor $10^4$ less power than we receive from the Sun.
The third source is just the "heat of formation". Gathering together the raw materials of a planet/moon and compressing it into a gravitational potential well, inevitably leads to the generation of considerable heat and it can take billions of years for this heat to leak out into space. Indeed, for the Earth, the heat of formation is probably comparable with radioactive decay in terms of supplying energy to the surface from the interior. | ProfRob makes a good point about tidal heating, but you can have internal heat sufficient for life even without this. The heat I'm referring to is heat trapped in a planet's core from its formation, which slowly and reliably leaks out through geothermal vents.
We observe this at the bottoms of Earth's oceans, where thriving colonies of deep sea organisms live in total darkness, feeding off the heat and chemicals that come from these vents. No sunlight supports these organisms, only geothermal energy.
Note that how long this is useful depends on your planet's size. Mars used to have a liquid core, but has since cooled down enough to have its core freeze since it's smaller and therefore cools faster. The larger your planet, the longer it'd be able to provide this trickle of heat.
---
Edit: it has been made aware to me that Earth's internal heat is mostly driven by radioactive decay nowadays, which I didn't know. The point still stands however that a planet's original heat could be sufficient to jumpstart life. I'm not sure how long it could provide for this before something else would need to kick in though. |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | ProfRob makes a good point about tidal heating, but you can have internal heat sufficient for life even without this. The heat I'm referring to is heat trapped in a planet's core from its formation, which slowly and reliably leaks out through geothermal vents.
We observe this at the bottoms of Earth's oceans, where thriving colonies of deep sea organisms live in total darkness, feeding off the heat and chemicals that come from these vents. No sunlight supports these organisms, only geothermal energy.
Note that how long this is useful depends on your planet's size. Mars used to have a liquid core, but has since cooled down enough to have its core freeze since it's smaller and therefore cools faster. The larger your planet, the longer it'd be able to provide this trickle of heat.
---
Edit: it has been made aware to me that Earth's internal heat is mostly driven by radioactive decay nowadays, which I didn't know. The point still stands however that a planet's original heat could be sufficient to jumpstart life. I'm not sure how long it could provide for this before something else would need to kick in though. | There are also [electric bacteria](https://en.wikipedia.org/wiki/Electric_bacteria). The BBC's [There are microbes that eat and poo nothing but electricity](http://www.bbc.com/earth/story/20160613-there-are-microbes-that-eat-and-poo-nothing-but-electricity) coverts the topic extensively.
At first electrical forms of energy absorption and excretion seemed exotic, but this now is seen as a widespread phenomenon on Earth, including colonies that include multiple species working together to literally *complete the circuit*.
These organisms don't derive their energy directly from Sunlight or heat, but *the ones currently found on Earth* do require enough warmth to have liquid water environments in which to ply their trade.
So perhaps as long as self gravitation leftover heat and radioactivity mentioned in other answers can provide enough warmth for a liquid water environment, bacteria-like life can feed directly off of the mixtures of primordial chemicals (organics, metals, oxides, sulfides, etc.) without necessarily [eating the chemicals themselves](https://en.wikipedia.org/wiki/Chemosynthesis), but instead forming electrical pathways through which electrochemical reactions can take place.
From the BBC article:
>
> While most organisms get their electron fix from carbohydrates, some bacteria can harvest electrons in their purest form. They can effectively "eat" electrons from minerals and rocks. In a way, they are getting their electrical energy straight from the socket.
>
>
> Annette Rowe, a graduate student of Nealson, has found [six new bacterial species on the ocean floor that can live off electricity alone](http://dx.doi.org/10.3389/fmicb.2014.00784). All are very different to one another, and none of them is anything like Shewanella or Geobacter.
>
>
> She found that, when no other food source was available, the bacteria would happily take electrons directly from the electrodes. In their natural habitat, the bacteria likely take their electrons directly from iron and sulphur in the seabed.
>
>
> Examples of electron-eating bacteria found by Rowe include *Halomonas, Idiomarina, Marinobacter, and Pseudomonas* of the *Gammaproteobacteria*, and *Thalassospira and Thioclava* from the *Alphaproteobacteria*.
> Many more electron-loving bacteria have now been found. In fact all you have to do is stick an electrode in the ground and pass electrons down it, and soon the electrode will be coated with feeding bacteria. Experiments show that these bacteria essentially eat or excrete electricity.
>
>
>
>
> Some species of Geobacter, he says, can both directly transfer electrons to electrodes and also directly accept electrons from them.
> In 2015, we learned that electron-eating and electron-excreting microbes can actually team up and pass electrons between each other, wiring themselves into a common electrical grid.
>
>
>
It turns out multiple types of bacteria can work together:
>
> Different species of bacteria and archaea – ancient single-celled microbes similar to bacteria in many ways – team up to degrade the methane before it can get the surface.
>
>
> Gunter Wegener from the Max Planck Institute for Marine Microbiology in Bremen wondered how the process worked. He [collected samples of the microbes](http://dx.doi.org/10.1038/nature15733), which live at temperatures of 60C on the ocean floor, and put them under a scanning electron microscope.
>
>
> The microscope revealed thin wire-like structures protruding from the bacterial cells. Although only a few nanometres wide, the wires were several micrometres long, which is much longer than the cells themselves.
> It seems that the bacteria use these nanowires to hook up with the archaea.
>
>
> **The archaea feed on electrons from methane, oxidising the gas to generate carbonate. They then pass the electrons on to their partner bacteria along the nanowires, which act like power cables. Finally the bacteria deposit the electrons onto sulphate, producing energy that the cell can use in the process.**
>
>
> |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | There are three important sources of energy in planetary bodies (or their moons) that may be important in determining their suitability for life. All of these are evident in various bodies in the Solar System.
The first is tidal heating. When one object orbits another then there will be gradient in the gravitational force across their finite sizes. If this is combined with any orbital non-circularity or difference between rotational and orbital periods, along with some fluidity in the structure of one of the bodies, then it will be heated by frictional forces as the fluid moves in response to the changing gravitational gradient. This heating is most evident on Io in our own Solar System, the closest large moon of Jupiter, and is generally more important for objects orbiting close to more massive stars or planets.
The second is heating by radioactive decay. When planetary systems are born, they can incorporate significant amounts of radioactive material. Some of this has a short half-life and may be responsible for a short period of intense heating, but other isotopes are much longer lived. According to [this article](https://blogs.scientificamerican.com/observations/nuclear-fission-confirmed-as-source-of-more-than-half-of-earths-heat/), about 20 TW of power is generated inside the Earth from the radioactive decay of uranium, thorium and potassium. This sounds a lot, but is about a factor $10^4$ less power than we receive from the Sun.
The third source is just the "heat of formation". Gathering together the raw materials of a planet/moon and compressing it into a gravitational potential well, inevitably leads to the generation of considerable heat and it can take billions of years for this heat to leak out into space. Indeed, for the Earth, the heat of formation is probably comparable with radioactive decay in terms of supplying energy to the surface from the interior. | You can also have planets with powerful magnetic fields, a conductive body passing through them [generates electric currents](https://www.mtnet.info/division/papers/EMWKSHP_ReviewVolumes/1972Edinburgh/DyalParkin_1972EdinburghReview_PEPI_1973.pdf), that besides simply heating can also power chemical reactions directly (like charging up a chemical battery).
There is also chemical energy. If there is a constant supply or accumulated store of reactive chemicals, that can act as fuel or food or battery electrolytes. For example, there are [clouds of organic molecules in space](https://phys.org/news/2014-09-alcohol-clouds-space.html) that could rain down on to a planet and be used as 'food', or which could collide with other clouds and react. Or planets made of different chemicals could collide and then react with one another.
As rotating dust clouds collapse under gravity, they can form [accretion disks](https://www.universetoday.com/87597/zooming-in-on-proto-planetary-disks/) in which stars and planets form. The continual collisions generate heat, causing them to glow in the infrared. This likely includes cases too small to form a star (brown dwarf systems).
[Shock waves](https://en.wikipedia.org/wiki/Shock_waves_in_astrophysics) passing through interstellar gas clouds can produce intense heating. These shock waves are usually produced by stellar explosions (supernovas) but could in principle be produced by anything that causes high speeds, like the bow shock of a planet passing through a gas cloud. |
45,491 | Are there other life giving sources of energy in space apart from stars? (like nebulae, radiation, etc.)? Or are all possible life giving potential sources some variation of a star? | 2021/08/28 | [
"https://astronomy.stackexchange.com/questions/45491",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/42626/"
] | ProfRob makes a good point about tidal heating, but you can have internal heat sufficient for life even without this. The heat I'm referring to is heat trapped in a planet's core from its formation, which slowly and reliably leaks out through geothermal vents.
We observe this at the bottoms of Earth's oceans, where thriving colonies of deep sea organisms live in total darkness, feeding off the heat and chemicals that come from these vents. No sunlight supports these organisms, only geothermal energy.
Note that how long this is useful depends on your planet's size. Mars used to have a liquid core, but has since cooled down enough to have its core freeze since it's smaller and therefore cools faster. The larger your planet, the longer it'd be able to provide this trickle of heat.
---
Edit: it has been made aware to me that Earth's internal heat is mostly driven by radioactive decay nowadays, which I didn't know. The point still stands however that a planet's original heat could be sufficient to jumpstart life. I'm not sure how long it could provide for this before something else would need to kick in though. | You can also have planets with powerful magnetic fields, a conductive body passing through them [generates electric currents](https://www.mtnet.info/division/papers/EMWKSHP_ReviewVolumes/1972Edinburgh/DyalParkin_1972EdinburghReview_PEPI_1973.pdf), that besides simply heating can also power chemical reactions directly (like charging up a chemical battery).
There is also chemical energy. If there is a constant supply or accumulated store of reactive chemicals, that can act as fuel or food or battery electrolytes. For example, there are [clouds of organic molecules in space](https://phys.org/news/2014-09-alcohol-clouds-space.html) that could rain down on to a planet and be used as 'food', or which could collide with other clouds and react. Or planets made of different chemicals could collide and then react with one another.
As rotating dust clouds collapse under gravity, they can form [accretion disks](https://www.universetoday.com/87597/zooming-in-on-proto-planetary-disks/) in which stars and planets form. The continual collisions generate heat, causing them to glow in the infrared. This likely includes cases too small to form a star (brown dwarf systems).
[Shock waves](https://en.wikipedia.org/wiki/Shock_waves_in_astrophysics) passing through interstellar gas clouds can produce intense heating. These shock waves are usually produced by stellar explosions (supernovas) but could in principle be produced by anything that causes high speeds, like the bow shock of a planet passing through a gas cloud. |
12,914 | I've got a Windows API Window. I've managed to make it windowed, and "fullscreen windowed". But I can't seem to make the window just full-screen- for example, even if I specify `WS_POPUP`, I can still freely move the cursor to my second screen. However, if I launch full-screen games, they don't permit it, and exist only exclusively or minimized. How can I make a window that behaves this way?
Edit: Excuse me. Limiting the mouse movement was just one example of how a full-screen window behaves differently to a non-fullscreen window, I'm not looking for how to limit the mouse movement.
Or is this *set* in the D3DPRESENT\_PARAMETERS, rather than just communicated? | 2011/05/30 | [
"https://gamedev.stackexchange.com/questions/12914",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/4662/"
] | My approach to this is two-fold
* First, I abstract the source of control for every game object. That lets me implement a "user controlled entity", which takes input from gamepad/keyboard/mouse/etc for the player entity, and implement AI routines for NPCs.
* Second, each entity can only express an "intent" ... so the output from the abstracted control object in point 1 only says, "I want to move in that direction". That intent is then implemented by some other routine which checks against map details, does collision, etc. | you can implement your object control in map calss, but that would create some god object and it'll be too difficult to debug or add features your program in future. there are two approches for this problem :
1- one is to handle all the inputs in the player class, and then ask map to move your player as player calculated, this can also be applied for any other dynamic or static objects in your scene. note that in this apprich map class only check for collision or other similar things , and doesn't care how or why is the object moving.
2- you can also share all map data with other dynamic classes (using physics engine and creating an object for every child of map is a way of doing this) and check all the thing you have to consider while moving objects inside their own class.
the first approch is much more simpler and take less time to implement, because you consider all your object the same and move them in a same manner, but in the second aproch you have much more flexibility with your obejects moving (for example you can make some object to ignore collsion) |
12,914 | I've got a Windows API Window. I've managed to make it windowed, and "fullscreen windowed". But I can't seem to make the window just full-screen- for example, even if I specify `WS_POPUP`, I can still freely move the cursor to my second screen. However, if I launch full-screen games, they don't permit it, and exist only exclusively or minimized. How can I make a window that behaves this way?
Edit: Excuse me. Limiting the mouse movement was just one example of how a full-screen window behaves differently to a non-fullscreen window, I'm not looking for how to limit the mouse movement.
Or is this *set* in the D3DPRESENT\_PARAMETERS, rather than just communicated? | 2011/05/30 | [
"https://gamedev.stackexchange.com/questions/12914",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/4662/"
] | Depends on the type of game. For a platformer, your movement is basically a large series of hacks, comprised of a lot of ray tests and highly specialized logic. For a top-down action game, you should have a basic physics/collision system that's independent of any particular type of game object, and player controls just set velocities/forces that move the avatar in the indicated direction. Other types of games have other types of controls and use less or more physics engine as appropriate.
In general, though, you have some kind of concept of a physics object. Players, enemies, items, etc. all have a physics component. This component in tandem with the physics engine (which knows about the level layout, among other things) will "correct" motion in response to collision. The player component then responds to movement requests by setting forces, which push the player around, and the physics component deals with things like "the object is moving into a wall, push it back out again and adjust accordingly."
If you provide more information on what exactly you're trying to do, someone may be able to provide a more detailed answer. | you can implement your object control in map calss, but that would create some god object and it'll be too difficult to debug or add features your program in future. there are two approches for this problem :
1- one is to handle all the inputs in the player class, and then ask map to move your player as player calculated, this can also be applied for any other dynamic or static objects in your scene. note that in this apprich map class only check for collision or other similar things , and doesn't care how or why is the object moving.
2- you can also share all map data with other dynamic classes (using physics engine and creating an object for every child of map is a way of doing this) and check all the thing you have to consider while moving objects inside their own class.
the first approch is much more simpler and take less time to implement, because you consider all your object the same and move them in a same manner, but in the second aproch you have much more flexibility with your obejects moving (for example you can make some object to ignore collsion) |
12,914 | I've got a Windows API Window. I've managed to make it windowed, and "fullscreen windowed". But I can't seem to make the window just full-screen- for example, even if I specify `WS_POPUP`, I can still freely move the cursor to my second screen. However, if I launch full-screen games, they don't permit it, and exist only exclusively or minimized. How can I make a window that behaves this way?
Edit: Excuse me. Limiting the mouse movement was just one example of how a full-screen window behaves differently to a non-fullscreen window, I'm not looking for how to limit the mouse movement.
Or is this *set* in the D3DPRESENT\_PARAMETERS, rather than just communicated? | 2011/05/30 | [
"https://gamedev.stackexchange.com/questions/12914",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/4662/"
] | My approach to this is two-fold
* First, I abstract the source of control for every game object. That lets me implement a "user controlled entity", which takes input from gamepad/keyboard/mouse/etc for the player entity, and implement AI routines for NPCs.
* Second, each entity can only express an "intent" ... so the output from the abstracted control object in point 1 only says, "I want to move in that direction". That intent is then implemented by some other routine which checks against map details, does collision, etc. | There is really no right way to do this, because you are talking collisions and character control it really boils down to exactly the problem you are trying to solve, as no matter your approach you are going to have to look at your game mechanics and see if it fits ( order of resolution etc.. ).
My approach to this problem and as it seems you want to keep a certain level of abstraction would be to different layers that process the player data, so if your player controlled object has a velocity and a direction then these would be updated based on the input, these inputs are not checked against anything, the input knows about the player object and simply updates its values, this would be your first layer.
After this first stage is done, your map system then needs to scan through the data and validate it, at this point all controller input has been resolved, the map can then decide to accept or change the forces ( velocity and direction ) associated with player entity based on collisions, wind, whatever can affect the player movement, this is your second layer that affects the data.
The last and final stage would be to simply update any entities position based on their direction and velocity ( as is ), this movement is assumed correct, this could mean the velocity is 0 because he collided with a wall, or he is now being taken on the opposite direction by a river he stepped in.
The key in this type of solution is to have different layers that affect your data, after each successive stage your data is assumed correct and valid for all previous stages, which means each processing stage could be designed to be plugged in and out at run time and even re-ordered. |
12,914 | I've got a Windows API Window. I've managed to make it windowed, and "fullscreen windowed". But I can't seem to make the window just full-screen- for example, even if I specify `WS_POPUP`, I can still freely move the cursor to my second screen. However, if I launch full-screen games, they don't permit it, and exist only exclusively or minimized. How can I make a window that behaves this way?
Edit: Excuse me. Limiting the mouse movement was just one example of how a full-screen window behaves differently to a non-fullscreen window, I'm not looking for how to limit the mouse movement.
Or is this *set* in the D3DPRESENT\_PARAMETERS, rather than just communicated? | 2011/05/30 | [
"https://gamedev.stackexchange.com/questions/12914",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/4662/"
] | My approach to this is two-fold
* First, I abstract the source of control for every game object. That lets me implement a "user controlled entity", which takes input from gamepad/keyboard/mouse/etc for the player entity, and implement AI routines for NPCs.
* Second, each entity can only express an "intent" ... so the output from the abstracted control object in point 1 only says, "I want to move in that direction". That intent is then implemented by some other routine which checks against map details, does collision, etc. | Depends on the type of game. For a platformer, your movement is basically a large series of hacks, comprised of a lot of ray tests and highly specialized logic. For a top-down action game, you should have a basic physics/collision system that's independent of any particular type of game object, and player controls just set velocities/forces that move the avatar in the indicated direction. Other types of games have other types of controls and use less or more physics engine as appropriate.
In general, though, you have some kind of concept of a physics object. Players, enemies, items, etc. all have a physics component. This component in tandem with the physics engine (which knows about the level layout, among other things) will "correct" motion in response to collision. The player component then responds to movement requests by setting forces, which push the player around, and the physics component deals with things like "the object is moving into a wall, push it back out again and adjust accordingly."
If you provide more information on what exactly you're trying to do, someone may be able to provide a more detailed answer. |
12,914 | I've got a Windows API Window. I've managed to make it windowed, and "fullscreen windowed". But I can't seem to make the window just full-screen- for example, even if I specify `WS_POPUP`, I can still freely move the cursor to my second screen. However, if I launch full-screen games, they don't permit it, and exist only exclusively or minimized. How can I make a window that behaves this way?
Edit: Excuse me. Limiting the mouse movement was just one example of how a full-screen window behaves differently to a non-fullscreen window, I'm not looking for how to limit the mouse movement.
Or is this *set* in the D3DPRESENT\_PARAMETERS, rather than just communicated? | 2011/05/30 | [
"https://gamedev.stackexchange.com/questions/12914",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/4662/"
] | Depends on the type of game. For a platformer, your movement is basically a large series of hacks, comprised of a lot of ray tests and highly specialized logic. For a top-down action game, you should have a basic physics/collision system that's independent of any particular type of game object, and player controls just set velocities/forces that move the avatar in the indicated direction. Other types of games have other types of controls and use less or more physics engine as appropriate.
In general, though, you have some kind of concept of a physics object. Players, enemies, items, etc. all have a physics component. This component in tandem with the physics engine (which knows about the level layout, among other things) will "correct" motion in response to collision. The player component then responds to movement requests by setting forces, which push the player around, and the physics component deals with things like "the object is moving into a wall, push it back out again and adjust accordingly."
If you provide more information on what exactly you're trying to do, someone may be able to provide a more detailed answer. | There is really no right way to do this, because you are talking collisions and character control it really boils down to exactly the problem you are trying to solve, as no matter your approach you are going to have to look at your game mechanics and see if it fits ( order of resolution etc.. ).
My approach to this problem and as it seems you want to keep a certain level of abstraction would be to different layers that process the player data, so if your player controlled object has a velocity and a direction then these would be updated based on the input, these inputs are not checked against anything, the input knows about the player object and simply updates its values, this would be your first layer.
After this first stage is done, your map system then needs to scan through the data and validate it, at this point all controller input has been resolved, the map can then decide to accept or change the forces ( velocity and direction ) associated with player entity based on collisions, wind, whatever can affect the player movement, this is your second layer that affects the data.
The last and final stage would be to simply update any entities position based on their direction and velocity ( as is ), this movement is assumed correct, this could mean the velocity is 0 because he collided with a wall, or he is now being taken on the opposite direction by a river he stepped in.
The key in this type of solution is to have different layers that affect your data, after each successive stage your data is assumed correct and valid for all previous stages, which means each processing stage could be designed to be plugged in and out at run time and even re-ordered. |
159,945 | For your info, I am a 17yr high school student currently live in Australia. As the applications for university entry is just around the corner, I am searching for an appropriate degree that suits my interest.
I frequently come across these various "honours" degree and not "straight forward" degree.
For instance, I want to do civil engineering, I see:
"Bachelors of Engineering (Honours)" and not "Bachelor of Civil Engineering" or "Bachelor of computer science"
**SO What exactly is an honours degree?**
I would normally ask these questions to the career guidance and career fest at my local school, but due to the current situation, I hardly get the opportunity to ask any questions about universities and whatnot. So I apologise if this question is too basic or inappropriate for this site. Thanks.
Here is the site I am looking at:
<https://future-students.uq.edu.au/study/programs/bachelor-engineering-honours-2342> | 2020/07/02 | [
"https://workplace.stackexchange.com/questions/159945",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/119336/"
] | Caveat that this is not a definitive answer, and probably could be considered a long comment, but it's the best I could do.
After some research I discovered a description of [Bachelor of Engineering (BE)](https://my.uq.edu.au/programs-courses/program.html?acad_prog=2001) at UQ (from 2015), which has this note:
>
> Please refer to: BACHELOR OF ENGINEERING (HONOURS)
>
>
> The Bachelor of
> Engineering is no longer available.
>
>
>
This is the equivalent [Bachelor of Engineering (Honours) (BE(Hons))](https://my.uq.edu.au/programs-courses/program.html?acad_prog=2342&year=2021) from 2021. Note that both course requirements include 64 units.
So for some reason UQ has dropped the non-Honours degree, but I don't know how this would work. For example, from Monash (of which I am very familiar) [they state](https://www.monash.edu/engineering/current-students/enrolment-and-re-enrolment/course-information/honours-weighted-average):
>
> In Engineering we award our bachelor degrees with honours for
> meritorious performance. An additional honours year is not required.
>
>
>
And I expected that UQ would have been the same given that the the course load for both honours and non-honours was the same. I can't explain how they can just hand out an honours degree.
Please note that on the UQ pages they have a direct contact number for enquiries. I'd highly recommend that given you can't speak to your local guidance people that you call the university directly:
>
> Enquiries for Australian students
>
>
> Manager, Academic Administration, Faculty of Engineering, Architecture
> and Information Technology
>
>
> Email: enquiries@eait.uq.edu.au
>
>
> Phone: 07
> 3365 4777
>
>
>
Finally, in the Australian way of things you get a degree in Engineering with the speciality in parenthesis. EG "Bachelor of Engineering (Civil)" and not "Bachelor of Civil Engineering" | The reason is they found it was hard to get students jobs without the additional honours degree component. It's just an extra year of study with an internship component.
the internship component is the actual reason it's an honours, as they cannot cram everything into 3 years, and a bachelor's full time is not meant to be 4 years.
Yes, they add a thesis and some extra subjects, but it is really about the internship.
Better advice is to do a double degree - it adds an extra year into your time there, but it broadens your knowledge, makes you more versatile and gives you a better idea of what you want to do. |
159,945 | For your info, I am a 17yr high school student currently live in Australia. As the applications for university entry is just around the corner, I am searching for an appropriate degree that suits my interest.
I frequently come across these various "honours" degree and not "straight forward" degree.
For instance, I want to do civil engineering, I see:
"Bachelors of Engineering (Honours)" and not "Bachelor of Civil Engineering" or "Bachelor of computer science"
**SO What exactly is an honours degree?**
I would normally ask these questions to the career guidance and career fest at my local school, but due to the current situation, I hardly get the opportunity to ask any questions about universities and whatnot. So I apologise if this question is too basic or inappropriate for this site. Thanks.
Here is the site I am looking at:
<https://future-students.uq.edu.au/study/programs/bachelor-engineering-honours-2342> | 2020/07/02 | [
"https://workplace.stackexchange.com/questions/159945",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/119336/"
] | A honors degree might involve a combination of the following:
* Attain a certain GPA or average (e.g. an average of 85% or greater)
* More advanced topics in the discipline
* Steeper requirements in terms of prerequisites (e.g. GPA/Average, knowledge)
* Taking more courses
Here is the description of honors degrees provided by my university:
>
> **Honours**. This option usually involves intense specialization in a single field. An honours B.Sc. requires maintenance of a high academic standing and may involve preparation of a graduating thesis.
>
>
>
>
> Honours candidates are required to follow the course of study as set out in the Calendar, to pass all courses completed, and to maintain a minimum overall 68% average in each academic session (higher in some disciplines).
>
>
>
>
> Candidates for honours must meet the credit requirements below both before entering honours and while in honours. Honours specializations that have corresponding majors specializations normally admit qualified students upon promotion to third year (see Course and Specialization Approval).
>
>
> | The reason is they found it was hard to get students jobs without the additional honours degree component. It's just an extra year of study with an internship component.
the internship component is the actual reason it's an honours, as they cannot cram everything into 3 years, and a bachelor's full time is not meant to be 4 years.
Yes, they add a thesis and some extra subjects, but it is really about the internship.
Better advice is to do a double degree - it adds an extra year into your time there, but it broadens your knowledge, makes you more versatile and gives you a better idea of what you want to do. |
5,793 | In addition to the general topical interest indicated in the title, I am also curious about British mathematics of the period viewed through the lens of competition with the rest of Europe (and perhaps even the rest of the world). For example, I have often heard it argued (informally) that the whole Newton-Leibniz-calculus kerfuffle can only be properly viewed in terms of a more general conflict between British and European mathematicians and scientists in that time period. Are there any books which address the existence and historical importance of such a conflict? | 2017/03/15 | [
"https://hsm.stackexchange.com/questions/5793",
"https://hsm.stackexchange.com",
"https://hsm.stackexchange.com/users/5505/"
] | The standard literature on the Mayan calendar is:
J. Eric S. Thompson, Maya Hieroglyphic Writing, Washington : Carnegie Institution, 1950.
Lis Brack-Bernsen, 'Die Basler Mayatafeln', Verhandlungen der Naturforschenden Gesellschaft in Basel, Vol.86 (1977), 1-76.
Floyd G. Lounsbury, 'Maya numeration, computation, and calendrical astronomy', Dictionary of Scientific Biography, Volume XV, Supplement I, 1978.
All of these agree that the year had 18 “months” of 20 days each and one “month” of 5 days, making a year of 365 days. There is no evidence for intercalation. | FWIW: ~ Gama, 1790, P.25 indicates "...pero la corrección no se hacia hasta el fin del Ciclo, en que se intercalaban juntos los 13 días, que gastaban en fiestas, en honor de los dioses seculares, de los quales era uno el mismo Xiuhteuctli TletL...", P.110 "...pues sabiendo ellos bien, que al fin del Ciclo había retrocedido el principio de su año civil 13 dias; y para igualarlo con el solar, los intercalaban; les era fácil saber, en qualquier año, los dias que debían computar en su cuenta, para verificar en ellos el preciso tiempo de los Equinoccios y Solsticios, y del tránsito del Sol por el vértice de la Ciudad..", and a few other references to intercalations (in addition to the intro which gives a primary purpose of the book as including ~"Explícase el sistema de los Calendarios de los Indios, el método que tenian de dividir el tiempo, y la corrección que hacian de él para igualar el año civil, de que usaban, con el año solar trópico..."). |
45,127,835 | I try to initialize my terminal in Android Studio but it gives me error. It says
>
> Can't open local terminal. java.io.IOException: Couldn't create PTY
>
>
>
My computer runs on windows 10. How to fix this? | 2017/07/16 | [
"https://Stackoverflow.com/questions/45127835",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/8309503/"
] | This problem was occured when, you clone one android project from Github Repository..
To solve this Problem, Open Settings, Go To Tools -> Terminal
Then, In the Project Settings, change the path of Start Directory to Your SDK's path like, For My System,
G:\MySoftware\Sdk\platform-tools | you could run studio.exe as administrator, if the solution not work, try another solution: add a Shell path(git-bash or others) to Android Studio Settings(setting--->Tools--->Terminal--->Shell path) |
1,865,515 | Your triplestore contains a lot of nodes, and you have to make accessible this database via a REST interface.
Now, my solution would be that every named (not-anonymous) node is exported as a resource, whose representation is:
1. all the triples having the node as a subject
2. all the triples having the node as an object
3. all the connected anonymous nodes.
I am a little reluctant on point number 2: this would basically give both incoming and outgoing triples.
What is your take on a REST sytle representation of a purely RDF-oriented data store ?
Do you share my point of view or not, and if not, what is your take on it ? | 2009/12/08 | [
"https://Stackoverflow.com/questions/1865515",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/78374/"
] | Depends on what the the data is and what the interface users want to do with it. This question is similar to what the SPARQL DESCRIBE query form should return. (It's determined by the implementation.)
For the use cases I've had with RDF data, I'd go with 1 and 3, producing a blank node closure of the resource. Additionally, you could have a separate interface for case 2, returning the incoming arcs of the resource. | (disclaimer: this may not correspond exactly to the content of your question, but it corresponds to the title)
I think that about the topic of Rest representation of RDF data is a general problem of inverting the order of concepts. For me the normal would be to have a collection of Rest documents with RDF data and use a RDF database for indexing and making global querys.
In this situation you can organize your resources in the way you prefer.
Also (if you pretend to use the URI of the node as the exported resource) your approach will have subtle problems about what is the meaning of your resources: the Rest resources you propose here are "[information resources](http://lists.w3.org/Archives/Public/www-tag/2005Jun/0039.html)" and then they cannot be abstract resources. There will be a conflict between information and meta-information.
I published an article [here](http://blog.notifychanged.com/2010/02/24/semantic-web-rdf-rest/) explaining this view in more detail. |
1,865,515 | Your triplestore contains a lot of nodes, and you have to make accessible this database via a REST interface.
Now, my solution would be that every named (not-anonymous) node is exported as a resource, whose representation is:
1. all the triples having the node as a subject
2. all the triples having the node as an object
3. all the connected anonymous nodes.
I am a little reluctant on point number 2: this would basically give both incoming and outgoing triples.
What is your take on a REST sytle representation of a purely RDF-oriented data store ?
Do you share my point of view or not, and if not, what is your take on it ? | 2009/12/08 | [
"https://Stackoverflow.com/questions/1865515",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/78374/"
] | Depends on what the the data is and what the interface users want to do with it. This question is similar to what the SPARQL DESCRIBE query form should return. (It's determined by the implementation.)
For the use cases I've had with RDF data, I'd go with 1 and 3, producing a blank node closure of the resource. Additionally, you could have a separate interface for case 2, returning the incoming arcs of the resource. | One easy way to make an RDF dataset REST traversible is to use URLs for all traversible elements.
When the URL is accessed, for example via HTTP GET, then result shows the connected nodes (connected as properties and/or inverse properties).
More formally the returned representation could be a [Concise Bounded Description](http://www.w3.org/Submission/CBD/) of the resource. |
1,865,515 | Your triplestore contains a lot of nodes, and you have to make accessible this database via a REST interface.
Now, my solution would be that every named (not-anonymous) node is exported as a resource, whose representation is:
1. all the triples having the node as a subject
2. all the triples having the node as an object
3. all the connected anonymous nodes.
I am a little reluctant on point number 2: this would basically give both incoming and outgoing triples.
What is your take on a REST sytle representation of a purely RDF-oriented data store ?
Do you share my point of view or not, and if not, what is your take on it ? | 2009/12/08 | [
"https://Stackoverflow.com/questions/1865515",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/78374/"
] | One easy way to make an RDF dataset REST traversible is to use URLs for all traversible elements.
When the URL is accessed, for example via HTTP GET, then result shows the connected nodes (connected as properties and/or inverse properties).
More formally the returned representation could be a [Concise Bounded Description](http://www.w3.org/Submission/CBD/) of the resource. | (disclaimer: this may not correspond exactly to the content of your question, but it corresponds to the title)
I think that about the topic of Rest representation of RDF data is a general problem of inverting the order of concepts. For me the normal would be to have a collection of Rest documents with RDF data and use a RDF database for indexing and making global querys.
In this situation you can organize your resources in the way you prefer.
Also (if you pretend to use the URI of the node as the exported resource) your approach will have subtle problems about what is the meaning of your resources: the Rest resources you propose here are "[information resources](http://lists.w3.org/Archives/Public/www-tag/2005Jun/0039.html)" and then they cannot be abstract resources. There will be a conflict between information and meta-information.
I published an article [here](http://blog.notifychanged.com/2010/02/24/semantic-web-rdf-rest/) explaining this view in more detail. |
297,405 | When we say "**I am wearing a Tshirt**" (might last a day), we mean "**I have that Tshirt on my body**"
When we don't say "**I am wearing a Tshirt**" to mean "**I am putting on a Tshirt**" (which might last 1 or 2 minutes).
Does "**I am renting a house**" mean I am negotiating with the landlord now (which might last many hours or many days)?
Does "**I am renting a house**" mean I signed the contract already and am living in the house and paying rent monthly now (which might last many months or years)? | 2021/09/10 | [
"https://ell.stackexchange.com/questions/297405",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/22478/"
] | "I am renting a house" would generally mean you have completed the negotiations and have a contract, whether you are actually moved in is a bit more questionable though in most cases could safely be assumed. This is from a western US perspective, I suppose this could vary by location. | "I am renting a house / a room" -> that negotiations and contracts are all done would be the most common interpretation.
To mean on-going negotiation, you could say "I am trying to rent a house". And "There's a house I'm trying to rent" would be more unambiguous. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | I thought [SpinRite](http://www.grc.com/spinrite.htm) was free, but I guess it isn't any more. Based on that I would use the manufacturers diagnostics tool, see samsungs [here](http://www.samsung.com/global/business/hdd/support/utilities/Support_HUTIL.html). | Wikipedia has a list of [harddisk diagnostic software](http://en.wikipedia.org/wiki/Comparison_of_S.M.A.R.T._tools). These programs are called S.M.A.R.T. tools, they detect if some of the reliability parameters are not okay. I used the free [Active@ Hard Disk Monitor](http://www.ntfs.com/disk-monitor.htm) to run checks on my harddisks. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | The [Ultimate Boot CD](http://www.ultimatebootcd.com/) has some Hard Drive tools on it that should detect any errors on the drive. Just make sure that you back up all your data FIRST. Doing intense testing on a bad drive may lead to worse failure and data lost. | If you do find out that a HDD is failing, you may want to check out [Recovery Is Possible](http://freshmeat.net/projects/recoveryispossible/). |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Wikipedia has a list of [harddisk diagnostic software](http://en.wikipedia.org/wiki/Comparison_of_S.M.A.R.T._tools). These programs are called S.M.A.R.T. tools, they detect if some of the reliability parameters are not okay. I used the free [Active@ Hard Disk Monitor](http://www.ntfs.com/disk-monitor.htm) to run checks on my harddisks. | If you do find out that a HDD is failing, you may want to check out [Recovery Is Possible](http://freshmeat.net/projects/recoveryispossible/). |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Along with all the other suggested options, you may want to look at more basic hardware factors:
* try using a different usb cable
* try a different usb port on the computer
* try using the hard drive in a different external drive case
* if available, try using esata
* try connecting the hard drive directly to your motherboard and see if the problem occurs
And as always, if you think the drive is flaky, make a backup of the data on it asap. It will protect your data from a drive that may be dying and give you the option of doing destructive data tests if need be. | You can grab Hiren's boot cd which contains some free tools for this purpose. [Website is here](http://www.hiren.info/pages/bootcd)
Maybe you'll have to attach your drive to the IDE bus for doing your maintenance work. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Wikipedia has a list of [harddisk diagnostic software](http://en.wikipedia.org/wiki/Comparison_of_S.M.A.R.T._tools). These programs are called S.M.A.R.T. tools, they detect if some of the reliability parameters are not okay. I used the free [Active@ Hard Disk Monitor](http://www.ntfs.com/disk-monitor.htm) to run checks on my harddisks. | Along with all the other suggested options, you may want to look at more basic hardware factors:
* try using a different usb cable
* try a different usb port on the computer
* try using the hard drive in a different external drive case
* if available, try using esata
* try connecting the hard drive directly to your motherboard and see if the problem occurs
And as always, if you think the drive is flaky, make a backup of the data on it asap. It will protect your data from a drive that may be dying and give you the option of doing destructive data tests if need be. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Personnaly, I have used the following in the past:
* HD Tach from Simpli Software (<http://www.simplisoftware.com/Public/index.php?request=HdTach>)
or
* HD Tune from EFD Software (www.hdtune.com)
Cheers,
Thomas | If you do find out that a HDD is failing, you may want to check out [Recovery Is Possible](http://freshmeat.net/projects/recoveryispossible/). |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | I thought [SpinRite](http://www.grc.com/spinrite.htm) was free, but I guess it isn't any more. Based on that I would use the manufacturers diagnostics tool, see samsungs [here](http://www.samsung.com/global/business/hdd/support/utilities/Support_HUTIL.html). | If you do find out that a HDD is failing, you may want to check out [Recovery Is Possible](http://freshmeat.net/projects/recoveryispossible/). |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Personnaly, I have used the following in the past:
* HD Tach from Simpli Software (<http://www.simplisoftware.com/Public/index.php?request=HdTach>)
or
* HD Tune from EFD Software (www.hdtune.com)
Cheers,
Thomas | You can grab Hiren's boot cd which contains some free tools for this purpose. [Website is here](http://www.hiren.info/pages/bootcd)
Maybe you'll have to attach your drive to the IDE bus for doing your maintenance work. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | The [Ultimate Boot CD](http://www.ultimatebootcd.com/) has some Hard Drive tools on it that should detect any errors on the drive. Just make sure that you back up all your data FIRST. Doing intense testing on a bad drive may lead to worse failure and data lost. | Wikipedia has a list of [harddisk diagnostic software](http://en.wikipedia.org/wiki/Comparison_of_S.M.A.R.T._tools). These programs are called S.M.A.R.T. tools, they detect if some of the reliability parameters are not okay. I used the free [Active@ Hard Disk Monitor](http://www.ntfs.com/disk-monitor.htm) to run checks on my harddisks. |
20,392 | Are there any free tools available to perform hard drive diagnostic?
I have an old 3.5 IDE Samsung HDD which I use as an external drive through USB connection. I recently started to have problems with accessing some of the data. | 2009/06/05 | [
"https://serverfault.com/questions/20392",
"https://serverfault.com",
"https://serverfault.com/users/244/"
] | Wikipedia has a list of [harddisk diagnostic software](http://en.wikipedia.org/wiki/Comparison_of_S.M.A.R.T._tools). These programs are called S.M.A.R.T. tools, they detect if some of the reliability parameters are not okay. I used the free [Active@ Hard Disk Monitor](http://www.ntfs.com/disk-monitor.htm) to run checks on my harddisks. | You can grab Hiren's boot cd which contains some free tools for this purpose. [Website is here](http://www.hiren.info/pages/bootcd)
Maybe you'll have to attach your drive to the IDE bus for doing your maintenance work. |
34,927 | A wide gamut screen displays more than the sRGB color space.
So why, with softwares that do not deal with ICC profiles, with a given color that has it's own place / value in the global color space, is rendered differently on a wide gamut screen than on a standard screen.
I mean, the color that corresponds for example to the #E58C4E color, is at the same place on a reduced color space, or an extended color space. So a monitor that can display a large color space should be able to display correctly a color that is in a color space that is inside it, whatever the way it is adressed to it. | 2013/02/20 | [
"https://photo.stackexchange.com/questions/34927",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/6541/"
] | Actually, `#E58C4E`, if you mean the web color, is [defined to be in sRGB](http://www.w3.org/TR/css3-color/#svg-color). However, if you didn't mean that particular convention but rather "red:229, green:140, blue:78", it's a different matter, because the extremes (the "primaries") of each channel are different in different color spaces, so those numbers actually *do* represent something different in each space. (Arguably, that's what a color space *is*.)
The monitor has its own primaries (the native color space of the monitor), and a color managed system translates between a known profile for a given image and the native output. If the color space of the original isn't known, there's no way to guess at the proper translation.
So, another way of asking the question might be: why aren't colors from un-color-managed-applications just *assumed* to be sRGB and translated on the fly? From the monitor's point of view, that's easy: it has no idea about which pixel came from which application, so it wouldn't know what to translate and what to leave alone. The graphics driver level has the same basic problem. Going from the other side, un-aware applications can't do it, because they're unaware. So, that leaves the operating system's display manager to do it, and that's hard.
So here we are. | [afaik] Unless the display driver or application being used applies a color profile or similar to compensate for the wider gamut, the color codes you use in your application will be sent to the display as-is. And in the display a specific color value (say E5) is no longer interpreted in the srgb scale but in the wider gamut scale.
So why does the display not interpret E5 in the srgb scale? This is because the values that the display adapter can send out are (typically) in the range 0-255 so in order for the wider gamut to be achievable the value 255 means maximum intensity in the wider gamut rather than the maximum intensity in the srgb gamut.
Should the display driver or application use a color profile, it can map the srgb values to the appropriate (smaller) wide gamut values and thus look "normal" on the display. Unless a wider scale is used than the typical 0-255 however then you will lose resolution in the process since a subset of the 0-255 scale will be used for srgb, for example 0-160. |
34,927 | A wide gamut screen displays more than the sRGB color space.
So why, with softwares that do not deal with ICC profiles, with a given color that has it's own place / value in the global color space, is rendered differently on a wide gamut screen than on a standard screen.
I mean, the color that corresponds for example to the #E58C4E color, is at the same place on a reduced color space, or an extended color space. So a monitor that can display a large color space should be able to display correctly a color that is in a color space that is inside it, whatever the way it is adressed to it. | 2013/02/20 | [
"https://photo.stackexchange.com/questions/34927",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/6541/"
] | Actually, `#E58C4E`, if you mean the web color, is [defined to be in sRGB](http://www.w3.org/TR/css3-color/#svg-color). However, if you didn't mean that particular convention but rather "red:229, green:140, blue:78", it's a different matter, because the extremes (the "primaries") of each channel are different in different color spaces, so those numbers actually *do* represent something different in each space. (Arguably, that's what a color space *is*.)
The monitor has its own primaries (the native color space of the monitor), and a color managed system translates between a known profile for a given image and the native output. If the color space of the original isn't known, there's no way to guess at the proper translation.
So, another way of asking the question might be: why aren't colors from un-color-managed-applications just *assumed* to be sRGB and translated on the fly? From the monitor's point of view, that's easy: it has no idea about which pixel came from which application, so it wouldn't know what to translate and what to leave alone. The graphics driver level has the same basic problem. Going from the other side, un-aware applications can't do it, because they're unaware. So, that leaves the operating system's display manager to do it, and that's hard.
So here we are. | **There is no global color space.** A color as you specify is just a triplet of Red, Green and Blue. Each component has a value between 0 and 255 (0 and FF in hex) which indicates how much power to give each LED for a given pixel (or phosphor in the days of CRTs).
**The scale is relative to your monitor and its current settings.** This is why monitors need to be calibrated. On a wide-gamut monitor the values are just spread along a wider color-space.
If you expected the color to display to be transformed into the display's color-space then you need to have at least a monitor profile for your monitor and tag your images with an ICC profile so that the operating system knows how to do the transformation. This of course will result in banding artifacts since precision is lost. |
34,927 | A wide gamut screen displays more than the sRGB color space.
So why, with softwares that do not deal with ICC profiles, with a given color that has it's own place / value in the global color space, is rendered differently on a wide gamut screen than on a standard screen.
I mean, the color that corresponds for example to the #E58C4E color, is at the same place on a reduced color space, or an extended color space. So a monitor that can display a large color space should be able to display correctly a color that is in a color space that is inside it, whatever the way it is adressed to it. | 2013/02/20 | [
"https://photo.stackexchange.com/questions/34927",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/6541/"
] | Actually, `#E58C4E`, if you mean the web color, is [defined to be in sRGB](http://www.w3.org/TR/css3-color/#svg-color). However, if you didn't mean that particular convention but rather "red:229, green:140, blue:78", it's a different matter, because the extremes (the "primaries") of each channel are different in different color spaces, so those numbers actually *do* represent something different in each space. (Arguably, that's what a color space *is*.)
The monitor has its own primaries (the native color space of the monitor), and a color managed system translates between a known profile for a given image and the native output. If the color space of the original isn't known, there's no way to guess at the proper translation.
So, another way of asking the question might be: why aren't colors from un-color-managed-applications just *assumed* to be sRGB and translated on the fly? From the monitor's point of view, that's easy: it has no idea about which pixel came from which application, so it wouldn't know what to translate and what to leave alone. The graphics driver level has the same basic problem. Going from the other side, un-aware applications can't do it, because they're unaware. So, that leaves the operating system's display manager to do it, and that's hard.
So here we are. | As Raymond Chen would put it, [developers hate to pay taxes](https://devblogs.microsoft.com/oldnewthing/20050822-11/?p=34483). Colour management is a very big tax on Windows. Developers are expected to ask Windows what colour profile to use and then do all the RGB conversions themselves. (And let's not even get started on the added challenges of multiple monitors!)
Most developers don't know any better so they just draw the image without any conversion, so the raw RGB numbers get sent to the monitor without any correction. It looks good on their screen so they don't care.
Now, the monitor HAS to interpret the RGB colours in the widest possible gamut it supports, or else you can't use those colours at all, defeating the purpose of the wide gamut monitor, so you're left with oversaturated colours. (Most wide gamut monitors do, in fact, include an sRGB mode out of necessity.)
On Mac OS X, everything is colour managed and sRGB profiles are assumed unless the program opts into something else, so it avoids this problem. |
34,927 | A wide gamut screen displays more than the sRGB color space.
So why, with softwares that do not deal with ICC profiles, with a given color that has it's own place / value in the global color space, is rendered differently on a wide gamut screen than on a standard screen.
I mean, the color that corresponds for example to the #E58C4E color, is at the same place on a reduced color space, or an extended color space. So a monitor that can display a large color space should be able to display correctly a color that is in a color space that is inside it, whatever the way it is adressed to it. | 2013/02/20 | [
"https://photo.stackexchange.com/questions/34927",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/6541/"
] | [afaik] Unless the display driver or application being used applies a color profile or similar to compensate for the wider gamut, the color codes you use in your application will be sent to the display as-is. And in the display a specific color value (say E5) is no longer interpreted in the srgb scale but in the wider gamut scale.
So why does the display not interpret E5 in the srgb scale? This is because the values that the display adapter can send out are (typically) in the range 0-255 so in order for the wider gamut to be achievable the value 255 means maximum intensity in the wider gamut rather than the maximum intensity in the srgb gamut.
Should the display driver or application use a color profile, it can map the srgb values to the appropriate (smaller) wide gamut values and thus look "normal" on the display. Unless a wider scale is used than the typical 0-255 however then you will lose resolution in the process since a subset of the 0-255 scale will be used for srgb, for example 0-160. | As Raymond Chen would put it, [developers hate to pay taxes](https://devblogs.microsoft.com/oldnewthing/20050822-11/?p=34483). Colour management is a very big tax on Windows. Developers are expected to ask Windows what colour profile to use and then do all the RGB conversions themselves. (And let's not even get started on the added challenges of multiple monitors!)
Most developers don't know any better so they just draw the image without any conversion, so the raw RGB numbers get sent to the monitor without any correction. It looks good on their screen so they don't care.
Now, the monitor HAS to interpret the RGB colours in the widest possible gamut it supports, or else you can't use those colours at all, defeating the purpose of the wide gamut monitor, so you're left with oversaturated colours. (Most wide gamut monitors do, in fact, include an sRGB mode out of necessity.)
On Mac OS X, everything is colour managed and sRGB profiles are assumed unless the program opts into something else, so it avoids this problem. |
34,927 | A wide gamut screen displays more than the sRGB color space.
So why, with softwares that do not deal with ICC profiles, with a given color that has it's own place / value in the global color space, is rendered differently on a wide gamut screen than on a standard screen.
I mean, the color that corresponds for example to the #E58C4E color, is at the same place on a reduced color space, or an extended color space. So a monitor that can display a large color space should be able to display correctly a color that is in a color space that is inside it, whatever the way it is adressed to it. | 2013/02/20 | [
"https://photo.stackexchange.com/questions/34927",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/6541/"
] | **There is no global color space.** A color as you specify is just a triplet of Red, Green and Blue. Each component has a value between 0 and 255 (0 and FF in hex) which indicates how much power to give each LED for a given pixel (or phosphor in the days of CRTs).
**The scale is relative to your monitor and its current settings.** This is why monitors need to be calibrated. On a wide-gamut monitor the values are just spread along a wider color-space.
If you expected the color to display to be transformed into the display's color-space then you need to have at least a monitor profile for your monitor and tag your images with an ICC profile so that the operating system knows how to do the transformation. This of course will result in banding artifacts since precision is lost. | As Raymond Chen would put it, [developers hate to pay taxes](https://devblogs.microsoft.com/oldnewthing/20050822-11/?p=34483). Colour management is a very big tax on Windows. Developers are expected to ask Windows what colour profile to use and then do all the RGB conversions themselves. (And let's not even get started on the added challenges of multiple monitors!)
Most developers don't know any better so they just draw the image without any conversion, so the raw RGB numbers get sent to the monitor without any correction. It looks good on their screen so they don't care.
Now, the monitor HAS to interpret the RGB colours in the widest possible gamut it supports, or else you can't use those colours at all, defeating the purpose of the wide gamut monitor, so you're left with oversaturated colours. (Most wide gamut monitors do, in fact, include an sRGB mode out of necessity.)
On Mac OS X, everything is colour managed and sRGB profiles are assumed unless the program opts into something else, so it avoids this problem. |
44,139 | I'm making an app that can search Metacritic (screen-scraper app). In the search box, I have an autocomplete function, which essentially does exactly what pressing "Search" would do, which is query Metacritic and populate a listview with the results. The problem is, since the app is a screen-scraping tool, it is not instant on autocomplete like Google.
Since the autocomplete is fired on keyup (every time the value of the text field changes essentially), since it is slow, when the user finishes typing, it updates the listview a bunch of times as it keeps searching, so it keeps changing which is quite annoying as it fetches the next results. Is there a good way to get around this?
I was thinking having it only query every *other* keypress, so that it wouldn't update as much and therefore wouldn't be as blinky.
**Update:**
This fix isn't necessarily perfect, but it helps a tremendous amount. What I did was:
a) Not fire the query until the length of the text field is at least 3
b) Fire *every other* keypress
This removed most of the blinking. | 2013/08/25 | [
"https://ux.stackexchange.com/questions/44139",
"https://ux.stackexchange.com",
"https://ux.stackexchange.com/users/12299/"
] | If auto-complete is slow, you can do several things:
1. best option is to switch it off. Each user knows better than computer, what he wants to write
2. Optimize software which is providing the auto-complete method. Common problem is, that values are read from files, then parsed, sorted, searched and after all that displayed every time a-c is accessed.
3. Enable caching on auto-complete. Usually there are serious slowdowns, when auto-complete list should be read from the hard-drive or downloaded from Internet.
4. **Edited** Enable pre-indexing. Let's have list of the a-c words extracted in the memory or quick access table in db with simple alphabetic or weight index, include background actualization according dynamic changes of the application. Enable most searched words first etc.
Unfortunately, you did not describe exactly what is the application with slow a-c, therefore it is hard to give more detailed answer.
**Edited**: Question is what makes your query slow. Target optimizations to that. | Do you perform the search in a separate thread, so that it doesn't block the UI? If not, do this first.
Then, when results come in, don't show them if the user has already typed further characters. This should prevent the annoying multiple updates. |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | You might describe such a person or their worldview as **transactional**.
In other words, this person makes an effort in a certain context because compensation is offered. Either broadly or in a certain situation, they might interpret their own and others' actions and expectations through the lens of a balance sheet—effort for its own sake is neither extended nor expected.
The term is not especially complimentary but is far less pejorative than the alternatives "greedy," "desperate," ""avaricious," "rapacious," "materialistic," or "mercenary" (!). It includes anyone who expects to receive a paycheck from their employer, for example. One could frame the behavior in a positive way by saying that the person is businesslike or practical in expecting a suitably compensated, mutually fulfilling professional relationship. | I suppose you could say that one is "economically ambitious". To get it into one word you could say "greedy" but this implies a sense of self-centered focus without regard to others.
A term we use to describe effective salespeople is often "coin-operated" |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | The word you are probably seeking is ***mercenary***, which exists both as noun and adjective.
It tends to carry negative connotations and is applied more specifically than in any other context to soldiers of fortune, who fight, not for a national flag, but for financial reward. indeed that is how the term is applied as a noun. ***A mercenary*** is such a soldier.
Its etymology is from Latin via Norman French, though the earliest example the OED has, is from Chaucer:
>
> ▸ c1387–95 Chaucer Canterbury Tales Prol. 514 He [sc. the parson]
> was a shepherde and noght a mercenarye [v.r. mersenarye].
>
>
>
The full range of OED entries for ***mercenary***, without the multiple examples under each sense, is as follows:
>
> A.
> 1. A person who works merely for money or other material reward; a hireling. In later use (probably influenced also by sense A. 2): a
> person whose actions are motivated primarily by personal gain, often
> at the expense of ethics. c1387–95—1998
>
>
> 2.a. A person who receives payment for his or her services. Chiefly and now only: spec. a soldier paid to serve in a foreign army or other
> military organization. 1523—1974
> b. In extended use, with modifying word. 1861—1987
>
>
> B. adj. 1 a. Of a person, organization, etc.: working or acting
> merely for money or other material reward; motivated by self-interest;
> materialistic. 1532—1997 b. Of conduct, a course of action, etc.,
> or its motivation: characterized by self-interest or the pursuit of
> personal gain; prompted by the desire for money or other material
> reward; undertaken only for personal gain. 1532—1990
>
>
> 2. a. Hired, serving for wages. Now: spec. designating a soldier paid to serve in a foreign army or other military organization; (of an
> army) composed of such soldiers. 1569—1974 †b. Salaried,
> stipendiary; profit-making. Obsolete. 1656—1782
>
>
> 3. Of or belonging to a mercenary. a1616—1922
>
>
> | I suppose you could say that one is "economically ambitious". To get it into one word you could say "greedy" but this implies a sense of self-centered focus without regard to others.
A term we use to describe effective salespeople is often "coin-operated" |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | **avaricious** [TFD](https://www.thefreedictionary.com/avaricious)
adj
>
> immoderately desirous of wealth or gain; greedy.
>
>
>
As in:
>
> When you put the most **avaricious** in charge, they use the tools of
> governance to further enrich themselves. New York Times Dec 7, 2018
>
>
> | I suppose you could say that one is "economically ambitious". To get it into one word you could say "greedy" but this implies a sense of self-centered focus without regard to others.
A term we use to describe effective salespeople is often "coin-operated" |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | The word you are probably seeking is ***mercenary***, which exists both as noun and adjective.
It tends to carry negative connotations and is applied more specifically than in any other context to soldiers of fortune, who fight, not for a national flag, but for financial reward. indeed that is how the term is applied as a noun. ***A mercenary*** is such a soldier.
Its etymology is from Latin via Norman French, though the earliest example the OED has, is from Chaucer:
>
> ▸ c1387–95 Chaucer Canterbury Tales Prol. 514 He [sc. the parson]
> was a shepherde and noght a mercenarye [v.r. mersenarye].
>
>
>
The full range of OED entries for ***mercenary***, without the multiple examples under each sense, is as follows:
>
> A.
> 1. A person who works merely for money or other material reward; a hireling. In later use (probably influenced also by sense A. 2): a
> person whose actions are motivated primarily by personal gain, often
> at the expense of ethics. c1387–95—1998
>
>
> 2.a. A person who receives payment for his or her services. Chiefly and now only: spec. a soldier paid to serve in a foreign army or other
> military organization. 1523—1974
> b. In extended use, with modifying word. 1861—1987
>
>
> B. adj. 1 a. Of a person, organization, etc.: working or acting
> merely for money or other material reward; motivated by self-interest;
> materialistic. 1532—1997 b. Of conduct, a course of action, etc.,
> or its motivation: characterized by self-interest or the pursuit of
> personal gain; prompted by the desire for money or other material
> reward; undertaken only for personal gain. 1532—1990
>
>
> 2. a. Hired, serving for wages. Now: spec. designating a soldier paid to serve in a foreign army or other military organization; (of an
> army) composed of such soldiers. 1569—1974 †b. Salaried,
> stipendiary; profit-making. Obsolete. 1656—1782
>
>
> 3. Of or belonging to a mercenary. a1616—1922
>
>
> | You might describe such a person or their worldview as **transactional**.
In other words, this person makes an effort in a certain context because compensation is offered. Either broadly or in a certain situation, they might interpret their own and others' actions and expectations through the lens of a balance sheet—effort for its own sake is neither extended nor expected.
The term is not especially complimentary but is far less pejorative than the alternatives "greedy," "desperate," ""avaricious," "rapacious," "materialistic," or "mercenary" (!). It includes anyone who expects to receive a paycheck from their employer, for example. One could frame the behavior in a positive way by saying that the person is businesslike or practical in expecting a suitably compensated, mutually fulfilling professional relationship. |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | **avaricious** [TFD](https://www.thefreedictionary.com/avaricious)
adj
>
> immoderately desirous of wealth or gain; greedy.
>
>
>
As in:
>
> When you put the most **avaricious** in charge, they use the tools of
> governance to further enrich themselves. New York Times Dec 7, 2018
>
>
> | You might describe such a person or their worldview as **transactional**.
In other words, this person makes an effort in a certain context because compensation is offered. Either broadly or in a certain situation, they might interpret their own and others' actions and expectations through the lens of a balance sheet—effort for its own sake is neither extended nor expected.
The term is not especially complimentary but is far less pejorative than the alternatives "greedy," "desperate," ""avaricious," "rapacious," "materialistic," or "mercenary" (!). It includes anyone who expects to receive a paycheck from their employer, for example. One could frame the behavior in a positive way by saying that the person is businesslike or practical in expecting a suitably compensated, mutually fulfilling professional relationship. |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | You might describe such a person or their worldview as **transactional**.
In other words, this person makes an effort in a certain context because compensation is offered. Either broadly or in a certain situation, they might interpret their own and others' actions and expectations through the lens of a balance sheet—effort for its own sake is neither extended nor expected.
The term is not especially complimentary but is far less pejorative than the alternatives "greedy," "desperate," ""avaricious," "rapacious," "materialistic," or "mercenary" (!). It includes anyone who expects to receive a paycheck from their employer, for example. One could frame the behavior in a positive way by saying that the person is businesslike or practical in expecting a suitably compensated, mutually fulfilling professional relationship. | *Financially motivated* typically refers to cash-strapped sellers and suggests you can get what ever they are selling at a bargain. In this sense, *desperate* is about right. But you seem to be looking for something like a person who's *all business*.
I suggest the idiom *sees dollar signs* - he does his job well when he sees dollar signs.
>
> The White House Sees Only Dollar Signs in the Arctic
>
>
>
[The New York Times](https://www.nytimes.com/2017/10/18/opinion/the-white-house-sees-only-dollar-signs-in-the-arctic.html)
>
> Twista Sees Dollar Signs on 'Stackin Paper': Premierete
>
>
>
<https://www.zerchoo.com/music/twista-sees-dollar-signs-on-stackin-paperpremiere/>
>
> Law firm sees dollar signs in drone technology
>
>
>
<https://finance-commerce.com/2014/08/law-firm-sees-dollar-signs-in-drone-technology/> |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | The word you are probably seeking is ***mercenary***, which exists both as noun and adjective.
It tends to carry negative connotations and is applied more specifically than in any other context to soldiers of fortune, who fight, not for a national flag, but for financial reward. indeed that is how the term is applied as a noun. ***A mercenary*** is such a soldier.
Its etymology is from Latin via Norman French, though the earliest example the OED has, is from Chaucer:
>
> ▸ c1387–95 Chaucer Canterbury Tales Prol. 514 He [sc. the parson]
> was a shepherde and noght a mercenarye [v.r. mersenarye].
>
>
>
The full range of OED entries for ***mercenary***, without the multiple examples under each sense, is as follows:
>
> A.
> 1. A person who works merely for money or other material reward; a hireling. In later use (probably influenced also by sense A. 2): a
> person whose actions are motivated primarily by personal gain, often
> at the expense of ethics. c1387–95—1998
>
>
> 2.a. A person who receives payment for his or her services. Chiefly and now only: spec. a soldier paid to serve in a foreign army or other
> military organization. 1523—1974
> b. In extended use, with modifying word. 1861—1987
>
>
> B. adj. 1 a. Of a person, organization, etc.: working or acting
> merely for money or other material reward; motivated by self-interest;
> materialistic. 1532—1997 b. Of conduct, a course of action, etc.,
> or its motivation: characterized by self-interest or the pursuit of
> personal gain; prompted by the desire for money or other material
> reward; undertaken only for personal gain. 1532—1990
>
>
> 2. a. Hired, serving for wages. Now: spec. designating a soldier paid to serve in a foreign army or other military organization; (of an
> army) composed of such soldiers. 1569—1974 †b. Salaried,
> stipendiary; profit-making. Obsolete. 1656—1782
>
>
> 3. Of or belonging to a mercenary. a1616—1922
>
>
> | *Financially motivated* typically refers to cash-strapped sellers and suggests you can get what ever they are selling at a bargain. In this sense, *desperate* is about right. But you seem to be looking for something like a person who's *all business*.
I suggest the idiom *sees dollar signs* - he does his job well when he sees dollar signs.
>
> The White House Sees Only Dollar Signs in the Arctic
>
>
>
[The New York Times](https://www.nytimes.com/2017/10/18/opinion/the-white-house-sees-only-dollar-signs-in-the-arctic.html)
>
> Twista Sees Dollar Signs on 'Stackin Paper': Premierete
>
>
>
<https://www.zerchoo.com/music/twista-sees-dollar-signs-on-stackin-paperpremiere/>
>
> Law firm sees dollar signs in drone technology
>
>
>
<https://finance-commerce.com/2014/08/law-firm-sees-dollar-signs-in-drone-technology/> |
482,362 | I need an adjective for the term "financially motivated." The word I'm looking for isn't exactly greedy or rapacious. Although it can have connotations like that, the word needs to directly mean something more similar to "he does his job well if there's money (or perhaps just incentive) on the line."
Thanks in anticipation. | 2019/01/22 | [
"https://english.stackexchange.com/questions/482362",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/333107/"
] | **avaricious** [TFD](https://www.thefreedictionary.com/avaricious)
adj
>
> immoderately desirous of wealth or gain; greedy.
>
>
>
As in:
>
> When you put the most **avaricious** in charge, they use the tools of
> governance to further enrich themselves. New York Times Dec 7, 2018
>
>
> | *Financially motivated* typically refers to cash-strapped sellers and suggests you can get what ever they are selling at a bargain. In this sense, *desperate* is about right. But you seem to be looking for something like a person who's *all business*.
I suggest the idiom *sees dollar signs* - he does his job well when he sees dollar signs.
>
> The White House Sees Only Dollar Signs in the Arctic
>
>
>
[The New York Times](https://www.nytimes.com/2017/10/18/opinion/the-white-house-sees-only-dollar-signs-in-the-arctic.html)
>
> Twista Sees Dollar Signs on 'Stackin Paper': Premierete
>
>
>
<https://www.zerchoo.com/music/twista-sees-dollar-signs-on-stackin-paperpremiere/>
>
> Law firm sees dollar signs in drone technology
>
>
>
<https://finance-commerce.com/2014/08/law-firm-sees-dollar-signs-in-drone-technology/> |
364,649 | A typical advice before any production deployments is backup the DB first. This way, if the new update has some issue that can lead to potential data loss or logical data corruption, then you still have a backup to compare and correct old records.
However, this can work well till DB size is in few GBs. Once the DB size is huge, backups take a long time complete. What are some best practices that should be followed in such situations, so as to avoid logical data corruption because of logical issues in a code deployment? | 2018/01/25 | [
"https://softwareengineering.stackexchange.com/questions/364649",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/23888/"
] | As someone who regularly dealt with updating production database for customers for our software upgrades, I tell you that the best way to minimize errors is to make updates as straightforward as possible.
If you can perform a change to all records rather than specific records, it is preferable.
------------------------------------------------------------------------------------------
In other words, if you're given a list of ids of records which need their state changed, you should be asking yourself why the update is being done in the context of the program. It may be that of the 10 records you need to update, the table only *has* 10 elements. Therefore you should be asking yourself if conceptually all you're doing is updating the state of all records.
If you can insert, it is preferable.
------------------------------------
The act of adding a record is self-contained. By this I mean there is only one side effect of adding a record, and that is the existence of a record that didn't exist prior. Therefore unless you're adding a record which shouldn't be there, there should be no issues.
If you can avoid deletion, it is preferable.
--------------------------------------------
If you're performing a deletion, you're removing data which would otherwise be unrecoverable without a backup. If possible, try to organize the data in such a way that you can disable records by changing its state rather than physically deleting the record. The excess of data can be put in a partition or it can be removed entirely in a later moment once you're sure there are no problems.
Have a consistent update policy.
--------------------------------
If you need to update a record, one of several things can happen:
1. Your record doesn't exist.
2. Your record exists but it has already been changed.
3. Your record exists and requires the change.
You need to have a policy to determine the course of action should something not go as planned. For simplicity sake, you should be consistent across the board and apply this policy in *any* situation of this type, not just for specific tables. This makes it easier to be able to recover data later. Generally, my policy is to write the script in such a way as to be able to re-execute it later. Should the script fail, it is nice to know you can make the proper adjustments and re-execute, however you're free to pick your own policy that suits you best.
Backups
-------
This by no means excuses you from performing a backup prior to performing any update in a production environment! Though even with a backup, I consider it a failure to have to use the backup. Losing data cannot be a possibility even in the *worst-case* scenario.
Conclusion
----------
You're not always going to be able to have it your way. The table schema is not likely going to be determined by you, and as such it means the types of updates you can expect to perform will be both complicated and risky. Though if you have any say-so in the matter, it helps to keep these points in mind as they make any updates straightforward and without significant risk.
Good luck! | At that point, you should be using a commercial grade DB system that supports [snapshots](https://docs.microsoft.com/en-us/sql/relational-databases/databases/database-snapshots-sql-server) (Oracles calls it [Flashback](https://docs.oracle.com/cd/B28359_01/appdev.111/b28424/adfns_flashback.htm#g1026131)) - that's exactly the kind of thing they are for.
Keep in mind that you need a backup concept anyway - having more data doesn't mean you drop backups because they become difficult, quite the opposite. You need some kind of continuous backup, e.g. based on replication with automatic failover. |
364,649 | A typical advice before any production deployments is backup the DB first. This way, if the new update has some issue that can lead to potential data loss or logical data corruption, then you still have a backup to compare and correct old records.
However, this can work well till DB size is in few GBs. Once the DB size is huge, backups take a long time complete. What are some best practices that should be followed in such situations, so as to avoid logical data corruption because of logical issues in a code deployment? | 2018/01/25 | [
"https://softwareengineering.stackexchange.com/questions/364649",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/23888/"
] | At that point, you should be using a commercial grade DB system that supports [snapshots](https://docs.microsoft.com/en-us/sql/relational-databases/databases/database-snapshots-sql-server) (Oracles calls it [Flashback](https://docs.oracle.com/cd/B28359_01/appdev.111/b28424/adfns_flashback.htm#g1026131)) - that's exactly the kind of thing they are for.
Keep in mind that you need a backup concept anyway - having more data doesn't mean you drop backups because they become difficult, quite the opposite. You need some kind of continuous backup, e.g. based on replication with automatic failover. | This is a massive area - so expect this question to be closed in fairly short order but, off the top of my head (as a former DBA on yuge databases):
**Mart/Repository**
You can mitigate some risk if you have a separate database for updates and a separate database that everyone uses. Then it is just a case of copying the data from one DB to the other once various checks have taken place. Mart/repository is how it is sometimes described but you might have primary/secondary, master/slave etc.
**Source code**
For everything that can change, have a source code which relates to *how* the data was updated. How many of these you have varies from DB to DB but you might have one for each user, role, data feed, code module etc.
**Create/update date**
Something that can assist greatly when tracking where things have gone wrong is having a creation and update data for every row. Then you can see at a glance which rows have been updated.
**ETL**
If the database update takes part as part of a data factory, you may be able to restore a previous vintage from flat files.
**Backup**
Full backups do of course take lots of space but the usual scenario is for a full backup to happen at regular intervals (says, weekly) and partial ones on a more frequent basis (daily etc).
**Point in time recovery**
Depending on which RDBMS you are using, some support point in time recovery. This allows you to roll back to the time when a good state was known. This does however require a large amount of storage which increases for how far you want to go back.
**Audit**
Having audit tables will tell you who (or what) made an update to a row. This can give you a good starting point for investigation.
**History**
For some critical tables, a copy of the pertinent row is taken at the time of the update so the data can be restored if need be.
**Data validation**
Ensure basic validation checks are carried out on the data before it is stored - over and above basic data type checks.
**Referential integrity**
Referential integrity isn't a silver bullet but it can help ensure the data is well structured. |
364,649 | A typical advice before any production deployments is backup the DB first. This way, if the new update has some issue that can lead to potential data loss or logical data corruption, then you still have a backup to compare and correct old records.
However, this can work well till DB size is in few GBs. Once the DB size is huge, backups take a long time complete. What are some best practices that should be followed in such situations, so as to avoid logical data corruption because of logical issues in a code deployment? | 2018/01/25 | [
"https://softwareengineering.stackexchange.com/questions/364649",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/23888/"
] | At that point, you should be using a commercial grade DB system that supports [snapshots](https://docs.microsoft.com/en-us/sql/relational-databases/databases/database-snapshots-sql-server) (Oracles calls it [Flashback](https://docs.oracle.com/cd/B28359_01/appdev.111/b28424/adfns_flashback.htm#g1026131)) - that's exactly the kind of thing they are for.
Keep in mind that you need a backup concept anyway - having more data doesn't mean you drop backups because they become difficult, quite the opposite. You need some kind of continuous backup, e.g. based on replication with automatic failover. | Many times if we are doing a "one shot" update we take a back up of production and restore it to a test server. Then we create a suit of tests and run the one shot. We verify the data has changed via the tests and become comfortable the that update will succeed and modify the data in a way that we expect it to. This is called a dry or trial run. I recommend doing this.
This gives everyone a good sense that the one shot will succeed. We cannot guarantee 100% because the data will be updated from the date of the trial run, but we boost confidence and success factors. This also gives a real idea of any issues that will occur since we are using a copy of production. Now if for some reason the update fails, we can always go to the back run prior to restore if needed, but we should have found and remediated any issues with the dry run.
If you can't take the entire database (if really large) try and export a smaller sample size and run the update (small dry run) against the actual data. I'd prefer the entire data set if possible to ensure the test is as complete as possible. |
364,649 | A typical advice before any production deployments is backup the DB first. This way, if the new update has some issue that can lead to potential data loss or logical data corruption, then you still have a backup to compare and correct old records.
However, this can work well till DB size is in few GBs. Once the DB size is huge, backups take a long time complete. What are some best practices that should be followed in such situations, so as to avoid logical data corruption because of logical issues in a code deployment? | 2018/01/25 | [
"https://softwareengineering.stackexchange.com/questions/364649",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/23888/"
] | As someone who regularly dealt with updating production database for customers for our software upgrades, I tell you that the best way to minimize errors is to make updates as straightforward as possible.
If you can perform a change to all records rather than specific records, it is preferable.
------------------------------------------------------------------------------------------
In other words, if you're given a list of ids of records which need their state changed, you should be asking yourself why the update is being done in the context of the program. It may be that of the 10 records you need to update, the table only *has* 10 elements. Therefore you should be asking yourself if conceptually all you're doing is updating the state of all records.
If you can insert, it is preferable.
------------------------------------
The act of adding a record is self-contained. By this I mean there is only one side effect of adding a record, and that is the existence of a record that didn't exist prior. Therefore unless you're adding a record which shouldn't be there, there should be no issues.
If you can avoid deletion, it is preferable.
--------------------------------------------
If you're performing a deletion, you're removing data which would otherwise be unrecoverable without a backup. If possible, try to organize the data in such a way that you can disable records by changing its state rather than physically deleting the record. The excess of data can be put in a partition or it can be removed entirely in a later moment once you're sure there are no problems.
Have a consistent update policy.
--------------------------------
If you need to update a record, one of several things can happen:
1. Your record doesn't exist.
2. Your record exists but it has already been changed.
3. Your record exists and requires the change.
You need to have a policy to determine the course of action should something not go as planned. For simplicity sake, you should be consistent across the board and apply this policy in *any* situation of this type, not just for specific tables. This makes it easier to be able to recover data later. Generally, my policy is to write the script in such a way as to be able to re-execute it later. Should the script fail, it is nice to know you can make the proper adjustments and re-execute, however you're free to pick your own policy that suits you best.
Backups
-------
This by no means excuses you from performing a backup prior to performing any update in a production environment! Though even with a backup, I consider it a failure to have to use the backup. Losing data cannot be a possibility even in the *worst-case* scenario.
Conclusion
----------
You're not always going to be able to have it your way. The table schema is not likely going to be determined by you, and as such it means the types of updates you can expect to perform will be both complicated and risky. Though if you have any say-so in the matter, it helps to keep these points in mind as they make any updates straightforward and without significant risk.
Good luck! | This is a massive area - so expect this question to be closed in fairly short order but, off the top of my head (as a former DBA on yuge databases):
**Mart/Repository**
You can mitigate some risk if you have a separate database for updates and a separate database that everyone uses. Then it is just a case of copying the data from one DB to the other once various checks have taken place. Mart/repository is how it is sometimes described but you might have primary/secondary, master/slave etc.
**Source code**
For everything that can change, have a source code which relates to *how* the data was updated. How many of these you have varies from DB to DB but you might have one for each user, role, data feed, code module etc.
**Create/update date**
Something that can assist greatly when tracking where things have gone wrong is having a creation and update data for every row. Then you can see at a glance which rows have been updated.
**ETL**
If the database update takes part as part of a data factory, you may be able to restore a previous vintage from flat files.
**Backup**
Full backups do of course take lots of space but the usual scenario is for a full backup to happen at regular intervals (says, weekly) and partial ones on a more frequent basis (daily etc).
**Point in time recovery**
Depending on which RDBMS you are using, some support point in time recovery. This allows you to roll back to the time when a good state was known. This does however require a large amount of storage which increases for how far you want to go back.
**Audit**
Having audit tables will tell you who (or what) made an update to a row. This can give you a good starting point for investigation.
**History**
For some critical tables, a copy of the pertinent row is taken at the time of the update so the data can be restored if need be.
**Data validation**
Ensure basic validation checks are carried out on the data before it is stored - over and above basic data type checks.
**Referential integrity**
Referential integrity isn't a silver bullet but it can help ensure the data is well structured. |
364,649 | A typical advice before any production deployments is backup the DB first. This way, if the new update has some issue that can lead to potential data loss or logical data corruption, then you still have a backup to compare and correct old records.
However, this can work well till DB size is in few GBs. Once the DB size is huge, backups take a long time complete. What are some best practices that should be followed in such situations, so as to avoid logical data corruption because of logical issues in a code deployment? | 2018/01/25 | [
"https://softwareengineering.stackexchange.com/questions/364649",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/23888/"
] | As someone who regularly dealt with updating production database for customers for our software upgrades, I tell you that the best way to minimize errors is to make updates as straightforward as possible.
If you can perform a change to all records rather than specific records, it is preferable.
------------------------------------------------------------------------------------------
In other words, if you're given a list of ids of records which need their state changed, you should be asking yourself why the update is being done in the context of the program. It may be that of the 10 records you need to update, the table only *has* 10 elements. Therefore you should be asking yourself if conceptually all you're doing is updating the state of all records.
If you can insert, it is preferable.
------------------------------------
The act of adding a record is self-contained. By this I mean there is only one side effect of adding a record, and that is the existence of a record that didn't exist prior. Therefore unless you're adding a record which shouldn't be there, there should be no issues.
If you can avoid deletion, it is preferable.
--------------------------------------------
If you're performing a deletion, you're removing data which would otherwise be unrecoverable without a backup. If possible, try to organize the data in such a way that you can disable records by changing its state rather than physically deleting the record. The excess of data can be put in a partition or it can be removed entirely in a later moment once you're sure there are no problems.
Have a consistent update policy.
--------------------------------
If you need to update a record, one of several things can happen:
1. Your record doesn't exist.
2. Your record exists but it has already been changed.
3. Your record exists and requires the change.
You need to have a policy to determine the course of action should something not go as planned. For simplicity sake, you should be consistent across the board and apply this policy in *any* situation of this type, not just for specific tables. This makes it easier to be able to recover data later. Generally, my policy is to write the script in such a way as to be able to re-execute it later. Should the script fail, it is nice to know you can make the proper adjustments and re-execute, however you're free to pick your own policy that suits you best.
Backups
-------
This by no means excuses you from performing a backup prior to performing any update in a production environment! Though even with a backup, I consider it a failure to have to use the backup. Losing data cannot be a possibility even in the *worst-case* scenario.
Conclusion
----------
You're not always going to be able to have it your way. The table schema is not likely going to be determined by you, and as such it means the types of updates you can expect to perform will be both complicated and risky. Though if you have any say-so in the matter, it helps to keep these points in mind as they make any updates straightforward and without significant risk.
Good luck! | Many times if we are doing a "one shot" update we take a back up of production and restore it to a test server. Then we create a suit of tests and run the one shot. We verify the data has changed via the tests and become comfortable the that update will succeed and modify the data in a way that we expect it to. This is called a dry or trial run. I recommend doing this.
This gives everyone a good sense that the one shot will succeed. We cannot guarantee 100% because the data will be updated from the date of the trial run, but we boost confidence and success factors. This also gives a real idea of any issues that will occur since we are using a copy of production. Now if for some reason the update fails, we can always go to the back run prior to restore if needed, but we should have found and remediated any issues with the dry run.
If you can't take the entire database (if really large) try and export a smaller sample size and run the update (small dry run) against the actual data. I'd prefer the entire data set if possible to ensure the test is as complete as possible. |
197,850 | On which of the Stack Exchange site should I ask a question regarding copyrights?
Example question:
>
> Do I have the permission to include the content of a document in my website?
>
>
> The copyright in question is from the SWEBOK and the licence says:
>
>
>
> >
> > This document may be copied, in whole or in part, in any form or by any means, as is, or with alterations, provided that (1) alterations are clearly marked as alterations and (2) this copyright notice is included unmodified in any copy.
> >
> >
> >
>
>
> SWEBOK book copyright: <http://www.computer.org/portal/web/swebok/html/copyright>
>
>
> | 2013/09/22 | [
"https://meta.stackexchange.com/questions/197850",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/228224/"
] | Questions about technical writing and documentation copyright should be on topic on [Writers Stack Exchange](https://writers.stackexchange.com/).
Before you post your question, you should take some time to read the [site's help articles](https://writers.stackexchange.com/help) and search for similar questions, to make sure your question will be welcomed by the site's community. | You could ask that question at [Law Stack Exchange](https://law.stackexchange.com/) because it deals with the law (clearly) and [copyrights](https://law.stackexchange.com/questions/tagged/copyright) as well. In its [Help Center for on-topic topics](https://law.stackexchange.com/help/on-topic) it lists:
>
> * Statutes or court decisions
> * Legal terms and language, doctrines and theory
> * Legal process and procedure
> * Historical legal applications
> * Dealing with legal professionals
>
>
> |
197,850 | On which of the Stack Exchange site should I ask a question regarding copyrights?
Example question:
>
> Do I have the permission to include the content of a document in my website?
>
>
> The copyright in question is from the SWEBOK and the licence says:
>
>
>
> >
> > This document may be copied, in whole or in part, in any form or by any means, as is, or with alterations, provided that (1) alterations are clearly marked as alterations and (2) this copyright notice is included unmodified in any copy.
> >
> >
> >
>
>
> SWEBOK book copyright: <http://www.computer.org/portal/web/swebok/html/copyright>
>
>
> | 2013/09/22 | [
"https://meta.stackexchange.com/questions/197850",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/228224/"
] | Currently, there are no Stack Exchange sites that cover copyrights or licensing *specifically*.
There was a proposal [here](http://area51.stackexchange.com/proposals/51301/copyrights?referrer=bllUdnErZXVMZd2Nw4wT3Q2) that would have dealt with the topic but it has been deleted. | You could ask that question at [Law Stack Exchange](https://law.stackexchange.com/) because it deals with the law (clearly) and [copyrights](https://law.stackexchange.com/questions/tagged/copyright) as well. In its [Help Center for on-topic topics](https://law.stackexchange.com/help/on-topic) it lists:
>
> * Statutes or court decisions
> * Legal terms and language, doctrines and theory
> * Legal process and procedure
> * Historical legal applications
> * Dealing with legal professionals
>
>
> |
83,661 | I want to have the Pi create its own Access Point for other devices to connect to,
but only when the Pi is out of range from a decent connection to a 'base / home' router?
I want to be able to have the Pi at home, and connect,
but when I'm out and about, and have the Pi on a mobile connection, I want it to create the Access Point, or Wifi Hotspot. | 2018/05/06 | [
"https://raspberrypi.stackexchange.com/questions/83661",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/80508/"
] | This is a very common need. There are lots of solutions out there, and lots of tutorials (that I find to be dated and inaccurate the older they are).
The solution I have been using with success is one of the Auto Hotspot scripts from [Raspberry Connect](https://www.raspberryconnect.com/projects/157-raspberry-pi-auto-wifi-hotspot-switch-internet). Look at the different hotspot articles on that page.
Alternative options that I have not tried but came across in my search include those below; the -turnkey one seems to be well documented.
* [raspberry-wifi-conf](https://github.com/sabhiram/raspberry-wifi-conf)
* [raspberry-pi-turnkey](https://github.com/schollz/raspberry-pi-turnkey)
* [kupiki-hotspot-script](https://github.com/pihomeserver/Kupiki-Hotspot-Script) | I tried the solutions above with some success but nothing as good as I wanted. I found ComitUp as being another method of doing this. It was a lot easier.
<https://davesteele.github.io/comitup/> (Thanks to Dave!)
>
> Comitup is a software package that provides a service to establish Wifi networking on a headless computer (that is, one with no video, keyboard, or mouse).
>
>
> The Comitup Image is a microSD disk image for the Raspberry Pi 3, providing an operating system with the comitup service included. The Comitup Image is an extension of standard Raspbian OS...
>
>
> |
654,860 | I have a new APU and was wondering if putting a small amount of paste on the APU would be too much. if there is too much, will this damage the APU,or will it be fine? | 2013/10/05 | [
"https://superuser.com/questions/654860",
"https://superuser.com",
"https://superuser.com/users/248744/"
] | Use a thin layer - remember, all you need to do is to fill in the microscopically small air-spaces between the two surfaces.
Also, don't forget to clean off all the old thermal compound - you can use alcohol or there are cleaners made for the purpose.
Too much thermal paste will lead to increased temperatures, but unless you add a really huge amount, it wouldn't be enough to cause damage via heat. (Besides, modern processors should be able to shut themselves down quickly (theoretically, fast enough to prevent damage) if they detect sufficiently high temperatures.) | The idea of Thermal Paste is to help distribute the heat from one side to the other side. 2 Metal Contact (APU or CPU to heatsink) are not really distributing heat that nicely. The Thermal Paste helps to create a transferring link between the 2 metal contact.
If you do have a heatsink for the APU, then before applying additional thermal paste, check if the heatsink already have thermal paste. If yes, no need to add more. If none, then add very small amount and spread it out to ensure it is thin and even across the metal contact to the heatsink.
If the APU does not use any heatsink, then why are you applying thermal paste? |
16,729,102 | I've recently been running some benchmarks trying to find the "best" serialization frameworks for C++ and also in Java. The factors that make up "best" for me are
the speed of de/serializing and also the resulting size of the serialized object.
If I look at the results of various frameworks in Java, I see that the resulting byte[] is generally smaller than the object size in memory. This is even the case with the built in Java serialization. If you then look at some of the other offerings (protobuf etc.) the size decreases even more.
I was quite surprised that when I looked at things on the C++ size (boost, protobuf) that the resulting object is generally no smaller (and in some cases bigger) than the original object.
Am I missing something here? Why do I get a fair amount of "compression" for free in Java but not in C++?
n.b for measuring the size of the objects in Java, I'm using Instrumentation <http://docs.oracle.com/javase/6/docs/api/java/lang/instrument/Instrumentation.html> | 2013/05/24 | [
"https://Stackoverflow.com/questions/16729102",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1943093/"
] | Did you compare the absolute size of the data? I would say that Java has more overhead, so if you "compress" the data into a serialized buffer, the amount of overhead decreases a lot more. In C/C++ you have almost the bare minimum required for the physical data size, so there is not much room for compression. And in fact, you have to add additional information to deserialize it, which could even result in a growth. | Object size can be observed to be bigger than the actual data size due to the offset bits between data members.
When an object is serialized, these offset bits are discarded and as a result, serialized object memory is smaller.
Because java is a managed environment, it will need more of such offset data to control memory and ownership, therefore, their compression rate is bigger. |
129,806 | I'm curious about the possibility of sonifying MIDI data, which has a bitrate of 31250. The range of human hearing extends to about 20kHz, which means that the fastest bits of the signal would lie near the top of this range: not ideal for my purposes. I envision stretching the signal, and then passing the result to a small speaker.
I'm not sure if there is a more precise term for this operation than "stretching" (I'm not super familiar with digital signal processing). At any rate, I can provide a concrete example of what I mean:
00101110100001 >> stretch by 2 >> 0000110011111100110000000011
Not unlikely, the resulting sound will seem be rather atonal/noisy, and that's OK by me. I imagine that I could accomplish this task with Arduino, but I'm wondering about whether this could be done with clever circuit design. For instance, something akin to using a 4018 divide-by-N counter IC? I recognize that the 4018 isn't exactly what I'm looking for, since this acts on a strictly periodic clock signal.
Any ideas? | 2014/09/18 | [
"https://electronics.stackexchange.com/questions/129806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/28191/"
] | Since MIDI data appears to be a series of 8-bit data bytes as sent by a standard UART, you could use a microcontroller to receive the data at the standard MIDI rate, and re-transmit it at a lower baud rate. This assumes that there are sufficient gaps in the data that you won't over-run the receive buffer.
This would probably require a microcontroller with two UARTs, as the UARTs I've used don't allow different transmit and receive speeds. | How about something like a 4027 JK flip-flop configured to toggle its output (J & K pulled high) ?
This would effective reduce the maximum possible frequency by half.
The result using your example would be:
00101110100001 >> flip-flop >> 00110000111110
Not the same bit-stream by any stretch of the imagination, but I suspect you're not really looking for accuracy here ... |
129,806 | I'm curious about the possibility of sonifying MIDI data, which has a bitrate of 31250. The range of human hearing extends to about 20kHz, which means that the fastest bits of the signal would lie near the top of this range: not ideal for my purposes. I envision stretching the signal, and then passing the result to a small speaker.
I'm not sure if there is a more precise term for this operation than "stretching" (I'm not super familiar with digital signal processing). At any rate, I can provide a concrete example of what I mean:
00101110100001 >> stretch by 2 >> 0000110011111100110000000011
Not unlikely, the resulting sound will seem be rather atonal/noisy, and that's OK by me. I imagine that I could accomplish this task with Arduino, but I'm wondering about whether this could be done with clever circuit design. For instance, something akin to using a 4018 divide-by-N counter IC? I recognize that the 4018 isn't exactly what I'm looking for, since this acts on a strictly periodic clock signal.
Any ideas? | 2014/09/18 | [
"https://electronics.stackexchange.com/questions/129806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/28191/"
] | So essentially you want to make a high-frequency random bit stream (just some binary data) audible. If it is literally played as bits, without further processing, it will sounds like random 'white' noise.
To slow it down, either :
1. extend its duration, for example 1 second of bits at 31,250b/s will
last, say 3 seconds, or
2. throw some of the data away. Say 2 out of every 3 bits will get it
into an audible range
For a simple, but interesting way, to throw bits away you could use a technique used in some 'bat listening devices'. They use a shift register, initialised with a single '1' and the rest is '0's.
The shift registers output is fed back to the input, and the output would drive the audio. The clock for the shift register is the signal. So a sequence of '1's in the signal will be merged into 1/2 clock, and a sequence of '0's will be merged into 1/2 clock.
So it tends to preserve changes. By careful choice of the length of the shift register, you can divide by a suitable amount.
To extend the duration, so that all bits are eventually played, will require memory. That could be because the MIDI data can be read and re-read, e.g. from a storage device.
Otherwise, the 'delay line' needs enough storage to hold the data. Five minutes at 31250b/s (almost 4kB/s) = 1,144kB, i.e. over 1MB.
It would be straightforward to make something like that with any device with enough RAM, like an R-Pi or BeagleBone black. It could be made using a MCU, like an Arduino, with the addition of some external storage. Don't use Flash memory for the storage unless you only plan on using this rarely, or the external storage is much bigger than a 'track', because Flash memory will eventually wear out. | How about something like a 4027 JK flip-flop configured to toggle its output (J & K pulled high) ?
This would effective reduce the maximum possible frequency by half.
The result using your example would be:
00101110100001 >> flip-flop >> 00110000111110
Not the same bit-stream by any stretch of the imagination, but I suspect you're not really looking for accuracy here ... |
129,806 | I'm curious about the possibility of sonifying MIDI data, which has a bitrate of 31250. The range of human hearing extends to about 20kHz, which means that the fastest bits of the signal would lie near the top of this range: not ideal for my purposes. I envision stretching the signal, and then passing the result to a small speaker.
I'm not sure if there is a more precise term for this operation than "stretching" (I'm not super familiar with digital signal processing). At any rate, I can provide a concrete example of what I mean:
00101110100001 >> stretch by 2 >> 0000110011111100110000000011
Not unlikely, the resulting sound will seem be rather atonal/noisy, and that's OK by me. I imagine that I could accomplish this task with Arduino, but I'm wondering about whether this could be done with clever circuit design. For instance, something akin to using a 4018 divide-by-N counter IC? I recognize that the 4018 isn't exactly what I'm looking for, since this acts on a strictly periodic clock signal.
Any ideas? | 2014/09/18 | [
"https://electronics.stackexchange.com/questions/129806",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/28191/"
] | Since MIDI data appears to be a series of 8-bit data bytes as sent by a standard UART, you could use a microcontroller to receive the data at the standard MIDI rate, and re-transmit it at a lower baud rate. This assumes that there are sufficient gaps in the data that you won't over-run the receive buffer.
This would probably require a microcontroller with two UARTs, as the UARTs I've used don't allow different transmit and receive speeds. | So essentially you want to make a high-frequency random bit stream (just some binary data) audible. If it is literally played as bits, without further processing, it will sounds like random 'white' noise.
To slow it down, either :
1. extend its duration, for example 1 second of bits at 31,250b/s will
last, say 3 seconds, or
2. throw some of the data away. Say 2 out of every 3 bits will get it
into an audible range
For a simple, but interesting way, to throw bits away you could use a technique used in some 'bat listening devices'. They use a shift register, initialised with a single '1' and the rest is '0's.
The shift registers output is fed back to the input, and the output would drive the audio. The clock for the shift register is the signal. So a sequence of '1's in the signal will be merged into 1/2 clock, and a sequence of '0's will be merged into 1/2 clock.
So it tends to preserve changes. By careful choice of the length of the shift register, you can divide by a suitable amount.
To extend the duration, so that all bits are eventually played, will require memory. That could be because the MIDI data can be read and re-read, e.g. from a storage device.
Otherwise, the 'delay line' needs enough storage to hold the data. Five minutes at 31250b/s (almost 4kB/s) = 1,144kB, i.e. over 1MB.
It would be straightforward to make something like that with any device with enough RAM, like an R-Pi or BeagleBone black. It could be made using a MCU, like an Arduino, with the addition of some external storage. Don't use Flash memory for the storage unless you only plan on using this rarely, or the external storage is much bigger than a 'track', because Flash memory will eventually wear out. |
69,497,342 | In CloudRun, my developer can't see the metrics on the CloudRun service dashboard, but they can see the metrics on the metrics dashboard. What policy am I missing?
**The CloudRun Dashboard Missing Metrics Charts:**
[](https://i.stack.imgur.com/tRJsz.png)
**The IAM Policy:**
[](https://i.stack.imgur.com/Z1PSr.png)
**Expected Dashboard:**
[](https://i.stack.imgur.com/a4kYW.png) | 2021/10/08 | [
"https://Stackoverflow.com/questions/69497342",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/8430249/"
] | check if you have a pop-up blocker or any kind of blocker extension | I was able to solve this and get some data on the graphs by granting the user the role of **Monitoring Viewer**.
Since you already have that, @valentin22 answer might still be worth a try.
[](https://i.stack.imgur.com/712ZS.png) |
34,433 | My kitty has some liver issues so is on a bunch of pills, and it is a lot of work to pill her.
Two big pills are Enrofloxacin/Baytril (left) and Clavamox (right). Pictures below with ruler for scale. They are about 1 cm wide as you can see in the picture below.
I have been splitting them in half to make them easier to swallow, but that of course increases the number of things I have to force down her throat. :( Also the Enrofloxacin is bitter and splitting it breaks the coating and exposes the bitter insides which makes it harder for the cat.
Is it ok to pill her on these whole pills or should I split them and do two doses?
I might try the Enrofloxacin first since it is thinner and rounded while the Clavamox is thicker and more rectangular.
[](https://i.stack.imgur.com/1xE2T.jpg) | 2022/02/21 | [
"https://pets.stackexchange.com/questions/34433",
"https://pets.stackexchange.com",
"https://pets.stackexchange.com/users/9150/"
] | You can try any of these:
* Crush the pill, then mix into the wet food (if in a wet food diet)
* Crush the pill, mix with little water, then use a pet-friendly syringe to force-feed the cat. I always use this method, as I'm afraid big pill can choke.
There are many pill-crusher you can find at any online store. | Here's the easy way - get a bit of water (maybe 5 ml) in a small cup. Drop the pill in it for a while (couple hours or more). The pill will start to either
1. dissolve into the water
2. breakdown into powder form and settle in a heap at the bottom of the water.
Stir the water thoroughly and add it to some tasty wet food or gravy. The cat will be happy to eat it all up. No fuss, no struggling with crushing, etc. |
34,433 | My kitty has some liver issues so is on a bunch of pills, and it is a lot of work to pill her.
Two big pills are Enrofloxacin/Baytril (left) and Clavamox (right). Pictures below with ruler for scale. They are about 1 cm wide as you can see in the picture below.
I have been splitting them in half to make them easier to swallow, but that of course increases the number of things I have to force down her throat. :( Also the Enrofloxacin is bitter and splitting it breaks the coating and exposes the bitter insides which makes it harder for the cat.
Is it ok to pill her on these whole pills or should I split them and do two doses?
I might try the Enrofloxacin first since it is thinner and rounded while the Clavamox is thicker and more rectangular.
[](https://i.stack.imgur.com/1xE2T.jpg) | 2022/02/21 | [
"https://pets.stackexchange.com/questions/34433",
"https://pets.stackexchange.com",
"https://pets.stackexchange.com/users/9150/"
] | **Both of those pills are fine to give to a cat whole.**
Both medications are specifically developed and sized for dosing cats (as cats and dogs are different species and therefore have different dosing requirements, as well as different sensitivities to both drugs and other ingredients used). There's no reason a manufacturer of a tablet specifically for dosing cats would make said tablet too large for a cat to swallow.
Additionally, giving the full tablet will be a more comfortable experience for your cat. Splitting the pill adds sharp, jagged edges and corners that can scrape and poke when she swallows the pill, as well as adding the bitter flavor from breaking the coating on the Enrofloxacin tablet. Intact, the pills have smooth, rounded edges and no sharp corners, making it easier for them to go down.
I have dosed average-sized cats with Clavamox many times over the years; they've never had any problems with the size of the tablet. Go ahead and dose with the entire thing. | You can try any of these:
* Crush the pill, then mix into the wet food (if in a wet food diet)
* Crush the pill, mix with little water, then use a pet-friendly syringe to force-feed the cat. I always use this method, as I'm afraid big pill can choke.
There are many pill-crusher you can find at any online store. |
34,433 | My kitty has some liver issues so is on a bunch of pills, and it is a lot of work to pill her.
Two big pills are Enrofloxacin/Baytril (left) and Clavamox (right). Pictures below with ruler for scale. They are about 1 cm wide as you can see in the picture below.
I have been splitting them in half to make them easier to swallow, but that of course increases the number of things I have to force down her throat. :( Also the Enrofloxacin is bitter and splitting it breaks the coating and exposes the bitter insides which makes it harder for the cat.
Is it ok to pill her on these whole pills or should I split them and do two doses?
I might try the Enrofloxacin first since it is thinner and rounded while the Clavamox is thicker and more rectangular.
[](https://i.stack.imgur.com/1xE2T.jpg) | 2022/02/21 | [
"https://pets.stackexchange.com/questions/34433",
"https://pets.stackexchange.com",
"https://pets.stackexchange.com/users/9150/"
] | **Both of those pills are fine to give to a cat whole.**
Both medications are specifically developed and sized for dosing cats (as cats and dogs are different species and therefore have different dosing requirements, as well as different sensitivities to both drugs and other ingredients used). There's no reason a manufacturer of a tablet specifically for dosing cats would make said tablet too large for a cat to swallow.
Additionally, giving the full tablet will be a more comfortable experience for your cat. Splitting the pill adds sharp, jagged edges and corners that can scrape and poke when she swallows the pill, as well as adding the bitter flavor from breaking the coating on the Enrofloxacin tablet. Intact, the pills have smooth, rounded edges and no sharp corners, making it easier for them to go down.
I have dosed average-sized cats with Clavamox many times over the years; they've never had any problems with the size of the tablet. Go ahead and dose with the entire thing. | Here's the easy way - get a bit of water (maybe 5 ml) in a small cup. Drop the pill in it for a while (couple hours or more). The pill will start to either
1. dissolve into the water
2. breakdown into powder form and settle in a heap at the bottom of the water.
Stir the water thoroughly and add it to some tasty wet food or gravy. The cat will be happy to eat it all up. No fuss, no struggling with crushing, etc. |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | This can be done from within the Steam client
=============================================
There is an option within Steam that toggles whether or not Steam starts with your computer. To change it, open up the main Steam window and select 'Preferences' from the 'Steam' menu, followed by selecting the 'Interface' tab, which brings up the following screen;

Simply uncheck 'Run Steam when my computer starts' to prevent Steam from starting with your computer. (Obviously if you wanted to Steam to start with your PC, you'd check this box instead of unchecking it!)
macOS
=====
If, on macOS, this doesn't work for you, check for Steam in "System Preferences" > "Users & Groups" > [Your user] > "Login Items", and remove it via the '-' sign.
[](https://i.stack.imgur.com/MroEC.png) | An alternative solution on Windows (starting from 8) is to open task manager and disable Steam Client Bootstrapper from the Start-up tab.

**Note:** This is an override switch to Steam's setting. You cannot re-enable it from Steam settings after disabling it here. |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | This can be done from within the Steam client
=============================================
There is an option within Steam that toggles whether or not Steam starts with your computer. To change it, open up the main Steam window and select 'Preferences' from the 'Steam' menu, followed by selecting the 'Interface' tab, which brings up the following screen;

Simply uncheck 'Run Steam when my computer starts' to prevent Steam from starting with your computer. (Obviously if you wanted to Steam to start with your PC, you'd check this box instead of unchecking it!)
macOS
=====
If, on macOS, this doesn't work for you, check for Steam in "System Preferences" > "Users & Groups" > [Your user] > "Login Items", and remove it via the '-' sign.
[](https://i.stack.imgur.com/MroEC.png) | In Windows 7 click the Start menu and type `msconfig` in the Search box. In the System Configuration program that opens, click the Startup tab, and then uncheck Steam. This does pretty much same thing as the Windows 8 tip. |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | This can be done from within the Steam client
=============================================
There is an option within Steam that toggles whether or not Steam starts with your computer. To change it, open up the main Steam window and select 'Preferences' from the 'Steam' menu, followed by selecting the 'Interface' tab, which brings up the following screen;

Simply uncheck 'Run Steam when my computer starts' to prevent Steam from starting with your computer. (Obviously if you wanted to Steam to start with your PC, you'd check this box instead of unchecking it!)
macOS
=====
If, on macOS, this doesn't work for you, check for Steam in "System Preferences" > "Users & Groups" > [Your user] > "Login Items", and remove it via the '-' sign.
[](https://i.stack.imgur.com/MroEC.png) | OSX = open steam then next to apple icon in top left corner
steam > preferences > interface
then uncheck - 'run steam when my computer starts' |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | This can be done from within the Steam client
=============================================
There is an option within Steam that toggles whether or not Steam starts with your computer. To change it, open up the main Steam window and select 'Preferences' from the 'Steam' menu, followed by selecting the 'Interface' tab, which brings up the following screen;

Simply uncheck 'Run Steam when my computer starts' to prevent Steam from starting with your computer. (Obviously if you wanted to Steam to start with your PC, you'd check this box instead of unchecking it!)
macOS
=====
If, on macOS, this doesn't work for you, check for Steam in "System Preferences" > "Users & Groups" > [Your user] > "Login Items", and remove it via the '-' sign.
[](https://i.stack.imgur.com/MroEC.png) | For Mac users:
1. Go to the Apple menu and choose “System Preferences”, then go to “Users & Groups”
2. Select the user that is active in OS X, then choose the “Login Items” tab
3. Select “Steam” from this list and then hit the Delete key on the keyboard to remove steam from the automatic launch on login list
4. Close out of System Preferences |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | An alternative solution on Windows (starting from 8) is to open task manager and disable Steam Client Bootstrapper from the Start-up tab.

**Note:** This is an override switch to Steam's setting. You cannot re-enable it from Steam settings after disabling it here. | In Windows 7 click the Start menu and type `msconfig` in the Search box. In the System Configuration program that opens, click the Startup tab, and then uncheck Steam. This does pretty much same thing as the Windows 8 tip. |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | An alternative solution on Windows (starting from 8) is to open task manager and disable Steam Client Bootstrapper from the Start-up tab.

**Note:** This is an override switch to Steam's setting. You cannot re-enable it from Steam settings after disabling it here. | OSX = open steam then next to apple icon in top left corner
steam > preferences > interface
then uncheck - 'run steam when my computer starts' |
102,996 | Every time I boot into OSX or Windows, Steam automatically starts.
Is there a way to stop Steam starting whenever I boot up my computer? | 2013/01/28 | [
"https://gaming.stackexchange.com/questions/102996",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/12936/"
] | An alternative solution on Windows (starting from 8) is to open task manager and disable Steam Client Bootstrapper from the Start-up tab.

**Note:** This is an override switch to Steam's setting. You cannot re-enable it from Steam settings after disabling it here. | For Mac users:
1. Go to the Apple menu and choose “System Preferences”, then go to “Users & Groups”
2. Select the user that is active in OS X, then choose the “Login Items” tab
3. Select “Steam” from this list and then hit the Delete key on the keyboard to remove steam from the automatic launch on login list
4. Close out of System Preferences |
152,969 | I am going through Trevor Hastie's Classification Techniques.
>
> Its says Odds are traditionally used instead of probabilities in horse-racing.
>
>
>
I still don't understand how they relate more naturally to the correct betting strategy? | 2015/05/19 | [
"https://stats.stackexchange.com/questions/152969",
"https://stats.stackexchange.com",
"https://stats.stackexchange.com/users/77344/"
] | It's not that they relate more naturally to the correct betting strategy per se, it's that they're much more easily interpretable while on the horse track.
Consider the following example: the probability of horse A to win is 66% percent. In conventional odds notation, this is represented as a 2:1 bet, which is very easily interpretable as a £2 win for each £1 bet, if the horse does win. For the layman, the 66% probability does not lend itself so easily to interpretation in terms of potential winnings. | In a logistic regression model, odds ratio provide a more coherent solution as compared to probabilities. Odds ratio represent the constant effect of an independent variable on a dependent variable. Here, being constant means that this ratio does not change with a change in the independent (predictor) variable. Odds ratio in this sense provide a much easier way of comprehending the marginal effects. |
3,817,607 | Let's say I have to implement a program for a small clinic company that allows its users(this is, the doctors) to schedule consults, log clients medical histories, etc. So, this will probably be the standard 3-layer application: Presentation, Controller and a Data Layer (that'll connect to a database).
I see 3 possibilities:
1. My first idea was to put the validating code right in the Domain Layer. But I feel that then I might be tempted to do the checking on class A, then the same check on B that uses A, then on C that uses B, etc. It on the other hand is good as it is easy to unit-test the validation logic.
2. There's a second school of thought that'll say that the best place to validate user input is as soon as possible, i.e., probably on the Presentation Layer (or in the Controller). This seems like a good idea, generally. If on the Controller, it will probably be easy to unit-test, too. It also allows one to switch the Views or the Data Layer and still have everything right.
3. Try to put the most validating logic possible on the database itself. This seems like a good idea, as it enforces that no data corrupts the database. The problem I see is that if I want to use different data repositories, I'll have to data validation logic again for the new one. Having that kind of logic at the Domain Layer, for example, would not have this problem.
How do you generally approach this problem? | 2010/09/28 | [
"https://Stackoverflow.com/questions/3817607",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/130758/"
] | As you've noted, there's more than one place to validate data.
There are also several levels of validation:
1. Correct format and type; all required values present (e.g., if date of birth is required, make sure it appears and is of type Date; if it's a String, make sure it conforms to the expected format such as 'yyyy-MM-dd')
2. Level 1 plus "business correctness": completed transaction is valid against your business rules (e.g., "date of birth must be at least eighteen years earlier than today").
There's a school of thought that says you should consider all of them:
1. Validation on the client to ensure the best experience for users. Don't make them wait for a round trip to the server to find out something's wrong. Put a JavaScript validation in place that will tell them the level 1 validity right away.
2. Validate again on the server side, because your service layer might not have a user interface in front of it. Bind and validate all values coming into your service tier.
3. Perform all level 2 validations as part of the transaction in the service layer. Make sure that inputs are correct from a business point of view.
4. If the database is shared by more than one application, put business logic into constraints, stored procedures, and triggers to ensure data integrity.
I don't think these should be "either or" decisions. | Do not mix up "when" and "where".
"When" should be as early as possible, maybe triggered by the presetation layer.
"Where" should be close to the domain logic.
In conjunction this e.g. could mean to have the presentation layer calling a verification service offered by the domain logic. |
3,817,607 | Let's say I have to implement a program for a small clinic company that allows its users(this is, the doctors) to schedule consults, log clients medical histories, etc. So, this will probably be the standard 3-layer application: Presentation, Controller and a Data Layer (that'll connect to a database).
I see 3 possibilities:
1. My first idea was to put the validating code right in the Domain Layer. But I feel that then I might be tempted to do the checking on class A, then the same check on B that uses A, then on C that uses B, etc. It on the other hand is good as it is easy to unit-test the validation logic.
2. There's a second school of thought that'll say that the best place to validate user input is as soon as possible, i.e., probably on the Presentation Layer (or in the Controller). This seems like a good idea, generally. If on the Controller, it will probably be easy to unit-test, too. It also allows one to switch the Views or the Data Layer and still have everything right.
3. Try to put the most validating logic possible on the database itself. This seems like a good idea, as it enforces that no data corrupts the database. The problem I see is that if I want to use different data repositories, I'll have to data validation logic again for the new one. Having that kind of logic at the Domain Layer, for example, would not have this problem.
How do you generally approach this problem? | 2010/09/28 | [
"https://Stackoverflow.com/questions/3817607",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/130758/"
] | As you've noted, there's more than one place to validate data.
There are also several levels of validation:
1. Correct format and type; all required values present (e.g., if date of birth is required, make sure it appears and is of type Date; if it's a String, make sure it conforms to the expected format such as 'yyyy-MM-dd')
2. Level 1 plus "business correctness": completed transaction is valid against your business rules (e.g., "date of birth must be at least eighteen years earlier than today").
There's a school of thought that says you should consider all of them:
1. Validation on the client to ensure the best experience for users. Don't make them wait for a round trip to the server to find out something's wrong. Put a JavaScript validation in place that will tell them the level 1 validity right away.
2. Validate again on the server side, because your service layer might not have a user interface in front of it. Bind and validate all values coming into your service tier.
3. Perform all level 2 validations as part of the transaction in the service layer. Make sure that inputs are correct from a business point of view.
4. If the database is shared by more than one application, put business logic into constraints, stored procedures, and triggers to ensure data integrity.
I don't think these should be "either or" decisions. | You often need to do the valiation at several layers:
* The client layer to ensure that the user experience is as good as in can be. For example avoiding inputing 10's of fields and then having to start over again due to invalid input.
* At the service layer, since you are not 100% sure that the request has been processed by the client (may be some one has sent a message directly to the service layer.
* At the service layer because some things cannot be checked on the client.
The good news is that some technologies allow you to specify the validation rules once and they will generate validation code for both the client and the server. See <http://weblogs.asp.net/scottgu/archive/2010/01/15/asp-net-mvc-2-model-validation.aspx>
This helps for the first 2 cases above. |
3,817,607 | Let's say I have to implement a program for a small clinic company that allows its users(this is, the doctors) to schedule consults, log clients medical histories, etc. So, this will probably be the standard 3-layer application: Presentation, Controller and a Data Layer (that'll connect to a database).
I see 3 possibilities:
1. My first idea was to put the validating code right in the Domain Layer. But I feel that then I might be tempted to do the checking on class A, then the same check on B that uses A, then on C that uses B, etc. It on the other hand is good as it is easy to unit-test the validation logic.
2. There's a second school of thought that'll say that the best place to validate user input is as soon as possible, i.e., probably on the Presentation Layer (or in the Controller). This seems like a good idea, generally. If on the Controller, it will probably be easy to unit-test, too. It also allows one to switch the Views or the Data Layer and still have everything right.
3. Try to put the most validating logic possible on the database itself. This seems like a good idea, as it enforces that no data corrupts the database. The problem I see is that if I want to use different data repositories, I'll have to data validation logic again for the new one. Having that kind of logic at the Domain Layer, for example, would not have this problem.
How do you generally approach this problem? | 2010/09/28 | [
"https://Stackoverflow.com/questions/3817607",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/130758/"
] | Do not mix up "when" and "where".
"When" should be as early as possible, maybe triggered by the presetation layer.
"Where" should be close to the domain logic.
In conjunction this e.g. could mean to have the presentation layer calling a verification service offered by the domain logic. | You often need to do the valiation at several layers:
* The client layer to ensure that the user experience is as good as in can be. For example avoiding inputing 10's of fields and then having to start over again due to invalid input.
* At the service layer, since you are not 100% sure that the request has been processed by the client (may be some one has sent a message directly to the service layer.
* At the service layer because some things cannot be checked on the client.
The good news is that some technologies allow you to specify the validation rules once and they will generate validation code for both the client and the server. See <http://weblogs.asp.net/scottgu/archive/2010/01/15/asp-net-mvc-2-model-validation.aspx>
This helps for the first 2 cases above. |
463,911 | I have a rather large (80k loc) java desktop app that talks to a database. We're now looking at exposing some parts of the database via a web application, using the existing codebase and preferably not having to modify it.
I have good separation between the data access, business logic and presentation layers, but we haven't used enterprise java beans or anything like that (if that's important).
What's the best way forward? Which of the java web frameworks will be best suited to the problem? Learning curve isn't terribly important, since I haven't done any java development on the web... | 2009/01/21 | [
"https://Stackoverflow.com/questions/463911",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246/"
] | To be true, it depends what you already have, and how well is the design of your current desktop application. You might not be able to use any or may be minimal of your existing code without modifying it, if its designed badly, and everything is tightly coupled.
Assuming that you are having a system with a good design, everything is de-coupled well enough. You can look into **Stripes** to make your presentation for the web, and use your existing data access and business code. I wish you all the luck.
Few **other goodies** to look into are, **Groovy on Grail**, **Wicket**.
I **don't recommend** anything like **Seam** and **Spring** they are more of a container and sophisticated large frameworks, which give you almost everything, solution for almost all of your problems. As you mentioned that you already have a complete system, and you just need to make a web interface to publish it for the web, these are not recommended, IMO.
**JSF**, is a **good framework**, but it might **drive you nuts** and has a **big learning curve**, according to few folks. | The two frameworks I would recommend would be [Grails](http://grails.org/) and [Struts 2](http://struts.apache.org/2.1.2/index.html).
Grails comes with a whole bunch of stuff that it configures under the covers including Hibernate and Spring. It makes generating dynamic pages to send to the browser ridiculously easy. What you are probably going to need to do is set up controllers that call Grails services which reference your existing code as you probably don't want Grails managing your database interactions. The disadvantage with Grails is not so much that it is written in Groovy, which is easy to learn for Java programmers, but that the IDE support for Groovy is still maturing. Still if you want quick productivity this is the route to go down.
Struts 2 offers a clean command pattern framework implementation that talks to JSPs (or velocity or FreeMarker templates) on the front end. To use this you would configure actions to call your existing code. You may want to investigate adding Spring to the mix depending on what you need to do.
There are other choices but these are two that I have had some success with. |
5,099,215 | So after reading about Blender (A freely downloadable application) [being sold by some company](http://news.slashdot.org/story/11/02/23/0011250/Trying-To-Lure-Suckers-Company-Resells-Open-Source-Blender?from=rss&utm_source=twitterfeed&utm_medium=twitter&utm_campaign=Slashdot) I learned that this is apparently is legal because of the GNU GPL license allows for this.
I'm wondering what exactly constitutes 'code' in this case? Do you also need to include any build scripts or project files? What about resources (or assets) like static logos and images? | 2011/02/24 | [
"https://Stackoverflow.com/questions/5099215",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/43792/"
] | There's no rule in the GPL against (re)selling free/open soure software, but the Slashdot article you linked to suggests that this company was obscuring the fact that the code was GPL'd. The GPL clearly states that you have to make the license clear, and make it easy to get a copy of the source.
And I'm pretty sure that "code" in this case means anything you need to rebuild the project, possibly with modifications. However, a project or company can trademark a name or logo (Mozilla famously does this with Firefox) so that no one else can call their fork by the same name. | I suppose the whole codebase is covered by GPL, and selling it does not violate its license (unless you don't *require* to pay), but one who bought it can freely redistribute it then. And you cannot develop software that is not GPL, but is using GPL libraries. See also: [related question](https://stackoverflow.com/questions/1749672/do-you-have-to-pay-for-gnu-gpl-software-that-is-for-sale), [GPL FAQ](http://www.gnu.org/licenses/gpl-faq.html) |
36,912,408 | As per my understanding.
A query operation will seek results on the mentioned index until one of the following condition is met
* The result set is exhausted.
* The number of items retrieved reaches the value of the Limit parameter, if specified.
* The amount of data retrieved reaches the maximum result set size limit of 1 MB.
Documented [Here](http://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html#ScanQueryPerformance)
So DynamoDB query will fetch results as per above criteria and then it will apply the FilterExpression so it is quite possible that it might not return any results to you, so it will return empty set and a LastEvaluatedKey
But I also read the following in the [documentation](http://docs.aws.amazon.com/amazondynamodb/latest/APIReference/API_Query.html)
>
> Unlike a Scan operation, a Query operation never returns both an empty result set and a LastEvaluatedKey value.
>
>
>
Can someone please help in explaining what does the above documentation statement actually mean?
Because in practice when I user **queryPage** API with **limit** and **FilterExpression** I am getting opposite of it, i.e. I am getting an empty set as well as LastEvaluatedKey.
Is my above understanding correct? It is possible to get both an empty results and lastEvaluatedKey value? Or I am missing something because of which I am getting empty results? (as per documentation I should not get it. It would be great if I don't get empty results) | 2016/04/28 | [
"https://Stackoverflow.com/questions/36912408",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2033008/"
] | I am also struggling with same problems as you, after reading the document, I find the following description explaining why. According to Amazon document, "In a Query operation, DynamoDB retrieves the items in sorted order, and then processes the items using KeyConditionExpression and any FilterExpression that might be present."
So it explained why you are getting those results. The filter operation is processed after the query results, so no mater you apply FilterExpression or not, the Query operations are no difference, that's why you see the lastEvaluatedKey is presented, and it's no difference with the one without apply FilterExpression.
So the only possible way to achive what you want is using Global Secondary Indexes, the column you want to filter, you can put it as sort key. In this case, you can query withRangeKeyCondition(). | They have updated the documentation, and its inline with what your experience is.
From the documentation link in the question :
Note:
A Query operation can return an empty result set and a LastEvaluatedKey if all the items read for the page of results are filtered out. |
188,376 | There's the following sentence in my Grammar book:
>
> It was a lovely surprise to find that all the washing-up **HAD BEEN DONE** while I was asleep.
>
>
>
The only answer mentioned in the answer key is past perfect (in capitals). Since today, I've thought that we use past perfect to clarify the order of actions but if it's clear by itself we don't need one.
I feel, in this sentence the order of actions is obvious, so **WAS DONE** is also possible. Even more—the past perfect isn't necessary.
Would someone please explain which tense should be used and why? | 2018/12/08 | [
"https://ell.stackexchange.com/questions/188376",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/86659/"
] | "by a month" is simply not idiomatic at all in English here.
**For time**, one does things or things are done **in a month, week, year, day, hour etc**. That is the period of time that will elapse.
**He beat the deadline by a month**.
There, **by a month is used to measure the number of months he won by.**
Some project has to be handed in in six months. The guy hands the project in in five months. He has beat the deadline ***by a month***.
For time that will elapse (go by) we use in: in a month.
To measure some amount of time in relation to a set time, we use by:
He beat me by ten minutes. They beat us (in the sailing race) by a month. | "In a month" is correct. It means that it will take 1 month by the time he'll speak fluent French. "By" is used to indicate the end point of an event. For example, you can say "He'll speak fluent French by February". February is the deadline by which the activity will be completed. |
188,376 | There's the following sentence in my Grammar book:
>
> It was a lovely surprise to find that all the washing-up **HAD BEEN DONE** while I was asleep.
>
>
>
The only answer mentioned in the answer key is past perfect (in capitals). Since today, I've thought that we use past perfect to clarify the order of actions but if it's clear by itself we don't need one.
I feel, in this sentence the order of actions is obvious, so **WAS DONE** is also possible. Even more—the past perfect isn't necessary.
Would someone please explain which tense should be used and why? | 2018/12/08 | [
"https://ell.stackexchange.com/questions/188376",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/86659/"
] | "by a month" is simply not idiomatic at all in English here.
**For time**, one does things or things are done **in a month, week, year, day, hour etc**. That is the period of time that will elapse.
**He beat the deadline by a month**.
There, **by a month is used to measure the number of months he won by.**
Some project has to be handed in in six months. The guy hands the project in in five months. He has beat the deadline ***by a month***.
For time that will elapse (go by) we use in: in a month.
To measure some amount of time in relation to a set time, we use by:
He beat me by ten minutes. They beat us (in the sailing race) by a month. | The primary difference is that one of those two sentences is grammatical, and the other is not.
>
> "He's sure that he'll speak French fluently **in a month**"
>
>
>
is correct. It is equivalent to "He's sure that he'll speak French fluently within a month", or "He's sure that before a month has passed, he'll speak French fluently". In this context, "[with]in [some amount of time]" means "before [that much time] has elapsed".
>
> "He's sure that he'll speak French fluently **by a month**"
>
>
>
is not grammatical, because in this context "by [something]" means "before [something] has *happened*". "In" needs a *span* - a *length* of time, which might begin now, or at some other point already established by context ("once he starts the course, he's sure that he'll speak French fluently in a month" would mean the month started when the course did, at some point in the future, rather than right now).
"**By**" needs an **point** in time: "a month" isn't a point, but "next month" could be (in this context it would be taken to mean "the **start** of next month"). Hence you might say "he's sure that he'll speak French fluently by summer", because "[the start of] summer" is a point. So is "dinner", though expecting someone to gain fluency between now and dinner is probably unreasonable. You can use "by the time [something happens]" in a similar way, so you could say "by the time a month is up", where "the time a month is up" means the point in time at which one complete month has passed. |
304,651 | [](https://i.stack.imgur.com/liZpX.png)

I installed high Sierra a few months ago and nowadays there seems to be a problem with System Information.app, it always shows the disk usage wrong like for example System 40GB but it shows photos to be 0KB even though I have 19GB worth photos and DaisyDisk reports correct disk usage.
I have tried booting into safe mode and rebooting, this persists after restarts
Screenshot:
[](https://i.stack.imgur.com/BwsKp.png)
[](https://i.stack.imgur.com/MbKKj.png)
I have run this scan as root user not admin.
so how do fix this missing photo sizing in system information.app? | 2017/11/05 | [
"https://apple.stackexchange.com/questions/304651",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/-1/"
] | Most likely the space is taken by Time Machine. It keeps making local snapshots when the persistent TM storage is not available. The local snapshots can pile up to as much as 80% of disk space. After the next backup, or when the disk space runs low, the local snapshots are automatically removed by macOS. So this is not really a problem that needs a fix, the system will take care of it.
That said, I'd agree that the System Information app needs a fix on High Sierra. It seems to include the local snapshots into "System" category and therefore it appears as used space, while Finder counts the snapshots as "purgeable", including it into the "available" space. This inconsistency is confusing.
DaisyDisk currently includes the snapshots into the used space (as opposed to Finder). Previously the snapshots could be revealed (./MobileBackups folder) when scanning "As Administrator", but starting from High Sierra and APFS, the local snapshots are saved outside the scannable area, so DaisyDisk simply shows a bigger "hidden space" segment. An update of DaisyDisk is said to be coming that will display purgeable space more consistently with Finder: <https://daisydiskapp.com/manual/4/en/Topics/HiddenSpace.html?subtopic=StillHiddenSpace> | You need to reindex your Spotlight cache. To do this
1. Go to System Preferences → Spotlight → Privacy
2. Click on the plus symbol and add your macOS hard drive you're having the wrong status/info problem
3. After adding your particular hard drive, quit System Preferences then wait for a few seconds.
4. Again open System Preferences → Spotlight → Privacy and this time remove your previously added hard drive.
What this does is it reindexes and creates the .Spotlight-V100 files on your hard drive. There are also various sudo commands to perform this step but for now try this! |
304,651 | [](https://i.stack.imgur.com/liZpX.png)

I installed high Sierra a few months ago and nowadays there seems to be a problem with System Information.app, it always shows the disk usage wrong like for example System 40GB but it shows photos to be 0KB even though I have 19GB worth photos and DaisyDisk reports correct disk usage.
I have tried booting into safe mode and rebooting, this persists after restarts
Screenshot:
[](https://i.stack.imgur.com/BwsKp.png)
[](https://i.stack.imgur.com/MbKKj.png)
I have run this scan as root user not admin.
so how do fix this missing photo sizing in system information.app? | 2017/11/05 | [
"https://apple.stackexchange.com/questions/304651",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/-1/"
] | I had a similar problem just now: Finder, "Get Info", and Disk Utility were all reporting 18GB free on a 250GB SSD that should have had plenty of free space. Under "About this Mac"->Storage, it reported 130GB of "System" usage. None of the space was listed as "purgeable", and macOS started complaining about lack of disk space.
I found a solution in [this Apple discussion thread](https://discussions.apple.com/message/32537764#message32537764). In my case, the problem was fixed by going to Time Machine Preferences and removing an old backup drive (that had long ago died) from the list of backup drives. Immediately, the available disk space on my SSD started being reported correctly as 120GB. | I solved this issue just relaunch finder.
Click on apple logo->Force Quit->Finder and click Relaunch.
So I opened my finder and it shows the correct size. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.