qid int64 1 74.7M | question stringlengths 12 33.8k | date stringlengths 10 10 | metadata list | response_j stringlengths 0 115k | response_k stringlengths 2 98.3k |
|---|---|---|---|---|---|
120,132 | I am trying to practice Smell like a teen spirit on drums and there is something I don’t understand with dotted notes.
Here is the sheet:[](https://i.stack.imgur.com/XvWGI.jpg)
I don’t understand the use of dotted notes in this sheet. If there were no dots I would have played this measure the same way, that is to say crash and kick on the 1 and kick on the d. For me the dotted notes don’t bring any additional information on how I am supposed to play the song. Is there something I don’t understand ?
Thank you. | 2021/12/16 | [
"https://music.stackexchange.com/questions/120132",
"https://music.stackexchange.com",
"https://music.stackexchange.com/users/82856/"
] | This drum part is written as a single voice, only one rhythm which incorporates different drums and cymbals being played at different times, sometimes only one, sometimes two simultaneously. Regardless of the vertical position of the notes or how many notes are played the overall rhythm in the bar has to add up to 4 beats in any combination of quarters, eighths, sixteenths, etc. Because there is a 16th at the end of the 1st beat the cymbal must be a dotted eighth since they are rhythmically linked.
This is one way of writing drum parts. Another way is to use two voices, one for drums and one for cymbals. You do this by using stems up and down. Here is an example of the same exact rhythm written with 2 voices:
[](https://i.stack.imgur.com/K1pbO.jpg)
(Disclaimer: I wrote this on my iPad with my index finger)
As you can see, with two voices the cymbal and hat are just playing quarters (stems up)and the back and forth rhythm between the bass and snare (stems down) is notated separately. One important note especially if writing by hand, everything MUST line up vertically. Some say (myself included) this is a better way to write drum parts because it is more accurate and separates the top and bottom of the kit (i.e. in version one you don’t REALLY play a dotted 16th on the cymbal, it just rings through). It takes a little more time and effort to write 2 voices so many people have adopted the one voice method of writing drum parts like in your example. Ask other drummers you know who are good readers about their preferences. Also check drum method books and see how they are written, one or two voice. | I appreciate that a lot of drum sounds cannot be made to sound longer than the fraction of a second they're hit, but there's more to music - particularly written down - .
Written music follows technical rules, one of which is that there must be the requisite number of notes in each bar.
Those dotted quavers can't make the actual sound any longer, but they need to be written as such, otherwise the semi-quavers that follow could be played too early or too soon. Without the dots, they're too short as notes. |
45,539,068 | I'm experiencing a bug in my production app and my best guess as to what is occurring is two separate users are clicking on the same item on the site and both proceed to create an order. When they get to the order page and submit the form it takes them to PayPal. Both users pay and the orders show up in the database but the inventory of only one item is marked as sold. Basically, multiple orders and payments are being created from only one item.
Anyone have any idea where to start on fixing this issue? Thanks | 2017/08/07 | [
"https://Stackoverflow.com/questions/45539068",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/4552848/"
] | * Can't there be more than 1 order for same item?
* Instead you can check for inventory before redirecting to PayPal, and once user is back to your app, you can check inventory again before placing the order.
* While checking for inventory, also consider the item in other users carts as well. | I figured out the problem. My item model had the has\_many association for orders instead of has\_one and was allowing multiple orders to be created. |
17,509 | I'll admit to being a complete newbie to Bitcoins.
The whole thing has up until now passed me by, until my interest was peaked by [this news story](http://www.theguardian.com/technology/2013/nov/27/hard-drive-bitcoin-landfill-site) about a gentleman in the UK who supposedly threw away a hard drive containing £4 million GBP of Bitcoins, and is now apparently searching through his local landfill.
I'm a little skeptical to say the least, and have a few questions:
* Is this technically feasable? I have seen comments about the place suggesting that one individual wouldn't have had access to the computing power needed to mine this many bitcoins in 2009
* Is there any way to verify this guys story? Surely there is a record somewhere of which bitcoins are registered to who
* Doesn't Bitcoin have a 'Forgot my password' feature or something similar? Surely something like this exists, or there is some administrative body who could be contacted in this scenario.
Thanks | 2013/11/29 | [
"https://bitcoin.stackexchange.com/questions/17509",
"https://bitcoin.stackexchange.com",
"https://bitcoin.stackexchange.com/users/9647/"
] | * Yes - it's feasible.
Bitcoins are released at a constant rate determined by the protocol. At the time, 50 Bitcoins were being generated for every block(this is now 25); and blocks are supposed to be found every 10 minutes. Miners compete amongst each other for this prize.
In 2009, you could mine using your computer's CPU and you were only competing with very few other people. There weren't large mining pools and dedicated hardware, so it was easy for anyone to run the client and mine hundreds or thousands of Bitcoins. It was so easy that you wouldn't think that losing your wallet was a big deal.
* Maybe, but it'd be a lot of work.
All transactions - including mined Bitcoins - are recorded in a publicly accessible global ledger. Many people download the entire transaction history on their own computer, and there are tools for working with it. It might be possible with some detective work to try to track down what address they were likely stored in using the clues given, but it's not easy.
* Nope - if you lose your wallet those Bitcoins are lost forever.
That's terrible for the person who owned them, but it doesn't impair the rest of the Bitcoin economy. It makes the value of the remaining Bitcoins go up to compensate.
There are already exchanges that will hold your wallet for you if you're worried about things like this. They're more secure than holding them on your personal computer, and in the future it's possible some of them might offer insurance or other guarantees in case they lose your money.
It's a design feature that there's no administrative body who can restore your Bitcoins. You would have to trust this administrative organization not to abuse their power. The developers of the original Bitcoin client have no more power than anyone else using Bitcoin. | If the private key for a Bitcoin address is lost, the coins can't be spent and are effectively considered lost.
Some wallet programs have password protection but Bitcoin itself does not have a "forgot my password" feature.
There is no administrative body. Bitcoin is explicitly designed to not rely on any trusted authority to create the coins or keep track of who has how many. That is recorded in the block chain, which is a public ledger that exists in many copies around the world. |
30,999 | I have seen some trees with a thick layer of moss growing on their branches, does this harm the tree? Should this layer of moss be removed or can it safely be left on the tree? | 2017/02/28 | [
"https://gardening.stackexchange.com/questions/30999",
"https://gardening.stackexchange.com",
"https://gardening.stackexchange.com/users/-1/"
] | Moss will not harm the tree! Good news indeed because the forests around here would be in BIG trouble! Moss does not have true roots - their "roots" are just for the purpose anchoring themselves to things like trees, rocks and whatever really. The "roots" don't penetrate the tree or steal any nutrients or water from the tree. Leave the moss, and the lichens =]
Edit:
It is possible that moss can cause bark rot and also be a holding ground for some fungi, bacteria and diseases. This is very dependent on the specie of tree and the specie of moss we are talking about. Tree bark comes in a range of thicknesses and rot resistance. I think it is rather extreme of stormy to instill such fear in moss on your tree though. 99.9% of the time, moss is not a problem for the tree. If you are looking for your answer, look at nature. Many of our oldest trees and old growth forests have an abundance of moss. There are many billions of trees that live to a ripe old age coexisting with moss. Moss thrives in the shade a dense forest or canopy provides. These trees make it to ripe old age, even with a solid coat of moss. Moss has always been around while trees evolved. If moss was such a threat to trees, they would have evolved a strategy to deal with them a long time ago. If moss was a real threat to trees, it would be considered a parasite because it is benefiting from the tree at its expense. Alas, I have never seen anything calling moss a parasite. The relationship between moss and trees dates back millions of years and it is a neutral relationship. Moss has its own purpose and function in the ecosystem. Plus it is beautiful.
So to reiterate my answer, no, moss is almost never a problem for trees. Don't worry about it. | If moss GIRDLES the tree YES it can. If moss is completely surrounding the tree and stays there long enough to hold moisture next to the bark, that will allow bacteria to begin decomposing the bark and compromising the vascular system just below. Somehow, in moist climates, moss on the north side only, seems to be just fine and I think the trees just thicken the bark or get used to using half the vascular system. Or get by until winter kills the moss long enough to allow the bark to dry.
This is more important on the trunk near the soil surface than higher up because there is more moisture available from the soil to keeps the moss alive. |
30,999 | I have seen some trees with a thick layer of moss growing on their branches, does this harm the tree? Should this layer of moss be removed or can it safely be left on the tree? | 2017/02/28 | [
"https://gardening.stackexchange.com/questions/30999",
"https://gardening.stackexchange.com",
"https://gardening.stackexchange.com/users/-1/"
] | Moss will not harm the tree! Good news indeed because the forests around here would be in BIG trouble! Moss does not have true roots - their "roots" are just for the purpose anchoring themselves to things like trees, rocks and whatever really. The "roots" don't penetrate the tree or steal any nutrients or water from the tree. Leave the moss, and the lichens =]
Edit:
It is possible that moss can cause bark rot and also be a holding ground for some fungi, bacteria and diseases. This is very dependent on the specie of tree and the specie of moss we are talking about. Tree bark comes in a range of thicknesses and rot resistance. I think it is rather extreme of stormy to instill such fear in moss on your tree though. 99.9% of the time, moss is not a problem for the tree. If you are looking for your answer, look at nature. Many of our oldest trees and old growth forests have an abundance of moss. There are many billions of trees that live to a ripe old age coexisting with moss. Moss thrives in the shade a dense forest or canopy provides. These trees make it to ripe old age, even with a solid coat of moss. Moss has always been around while trees evolved. If moss was such a threat to trees, they would have evolved a strategy to deal with them a long time ago. If moss was a real threat to trees, it would be considered a parasite because it is benefiting from the tree at its expense. Alas, I have never seen anything calling moss a parasite. The relationship between moss and trees dates back millions of years and it is a neutral relationship. Moss has its own purpose and function in the ecosystem. Plus it is beautiful.
So to reiterate my answer, no, moss is almost never a problem for trees. Don't worry about it. | Moss is of no health concern to trees. Spanish Moss (which isn't a moss) hangs from branches and its weight can be harmful, but actual moss is nothing to worry about.
Proper care is necessary to have healthy trees. Sometimes an abundance of moss or harmless lichens is an indication of poor air circulation. Check the proper pruning techniques for your tree. (This is seen a lot here in WI when people fail to perform necessary maintenance on their maples.). The tree isn't in danger, but without pruning, limbs will die and fall. This is normal for the tree, of course, but it's better to dictate falling limbs when you can.
The only detriment of moss and lichen on bark is it obscures your vision of the tree. This can, in rare cases, prevent an early diagnosis of a problem. However, as the moss' rhizoids provide a boost to the structural integrity of the bark, its more likely to prevent an issue than prevent an issue's diagnosis. Couple that with most issues that it could obscure being terminal (for the most part, if you've got something bubbling through the bark, the tree's a goner) and you're left with it being a question of aesthetics.
NOTE: If you see "moss" dangling from branches and not plastered on the bark, you're most likely looking at Spanish Moss (not a moss) that can snap branches when saturated due to its weight. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | Metaphorically, it may be due to the fact that the brain doesn't "see" a single image, but composes one based on a series of continuous "shots" from the eyes as they move around the scene.
Each of these "shots" are "taken" with variable "apertures", in order to maximize the overall dynamic range of the final "image".
You can think of the mental process as a mix of a panorama and HDR if you prefer. :o) | The main reason for this is that the human eye registers brightness on a logarithmic scale, whereas digital sensors are linear. Take a look [at this site](http://www.petapixel.com/2011/05/05/biology-for-photographers-why-is-the-aperture-scale-logarithmic/) about halfway down for more info. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This is a very good question, and the answer could fill hundreds of pages - and, in fact, the answer already DOES fill hundreds of pages.
The short answer is that the figures you are citing do not agree with apparent reality because the commonly quoted figures are wrong :-). Read on ...
Much is available on the internet on this subject and the quality is, as ever, widely variable. There is also a lot of parroting of "facts" between sites and figures like those in Wikipedia seem common enough BUT there are some very reasoned arguments which seem to suggest that the Wikipedia figure is extremely wrong and underestimates the figure very substantially.
It's important to note that the eye acts as a contrast detector rather than an absolute level detector (such as a digital camera sensor uses) so comparisons need care.
With irising, chemical adaptation and every other trick it can pull it seems that the absolute dynamic range of the whole eye system is well over 20 stops. As each stop is a factor of 2, that's 2^20 or about "well over 1,000,000:1". At the top end, the sun is too bright!!!. At the bottom end the dark adapted eye can detect a single photon. A D3S (better performance than a D4) may have trouble with that. (Note that that is not EVERY photon - when you get down to the few photons per second level a lot of them will hit non-sensor areas and not be detected. But when one DOES strike a sensitive retina area it will produce a signal that can be recorded.)
But, I digress :-). An extremely good (it seems) page that discusses eye dynamic range and more is
* [Notes on the Resolution and Other Details of the Human Eye](http://clarkvision.com/imagedetail/eye-resolution.html).
Paragraph headings are worth noting:
Notes on the Resolution of the Human Eye
Visual Acuity and Resolving Detail on Prints
How many megapixels equivalent does the eye have?
The Sensitivity of the Human Eye (ISO Equivalent)
The Dynamic Range of the Eye
The Focal Length of the Eye
The writer argues that the dynamic range of the eye without changing sensitivity by adaptation or irising is about 1,000,000:1 in low light conditions. That is, as great as the "well over" lower limit mentioned above. Then he justifies this claim as copied below. This sounds fairly convincing at first glance. There may be flaws in the argument, but it seems OK, and this does not mean that it applies in all light levels.
>
> Here is a simple experiment you can do. Go out with a star chart on a clear night with a full moon. Wait a few minutes for your eyes to adjust. Now find the faintest stars you can detect when the you can see the full moon in your field of view. Try and limit the moon and stars to within about 45 degrees of straight up (the zenith).
>
>
> If you have clear skies away from city lights, you will probably be able to see magnitude 3 stars.
>
>
> The full moon has a stellar magnitude of -12.5.
>
>
> If you can see magnitude 2.5 stars, the magnitude range you are seeing is 15.
>
>
> Every 5 magnitudes is a factor of 100, so 15 is 100 \* 100 \* 100 = 1,000,000.
>
>
> Thus, the dynamic range in this relatively low light condition is about 1 million to one, perhaps higher!
>
>
>
But, here's a suggestion from me for an experiment at normal daylight light levels.
* Find a scene that has a good mixture of dark areas and very bright areas - ideally with some dark areas as isolated islands near islands of brightness. An example may be sunlight shining through trees into a heavily shaded area - a few cavelets or deeply shaded areas will help.
* Allow your eyes to adapt to the general lighting level - do not stare at the bright spots near where the sun is shining through and do not focus on any especially dark areas.
* Note how well you can see detail in the darkest of dark areas - at what level of darkness does is fade to black.
* Try the same with bright areas - as you look toward the sun there will be a place where details washes out and you cannot reasonably see more.
* Cast your eyes to and fro across the scene between dark and light to try to stop your adaptation mechanism changing f-stop on you.
* Now, take photos of the scene. Expose "correctly" and then so the darkest areas that you could see can be seen in the photo and then so that the brightest highlights you could distinguish are not washed out.
* If you have the equipment, take an HDR photo with maximum f-stop variation between photos. (My Sony A77 allows 5ev steps.)
My experience is that my eye can always see a wider brightness range than my camera (Minolta 7Hi, A200, 5D, 7D, A700, A77, other)
On maximum HDR image (10 ev range between centers) my eye can see as well as or better than the camera.
The area where this does not APPEAR to be so is in extremely low light when I may need to allow the eye to integrate (which it does for up to about 4 seconds!) whereas I can look at a low light photo and see the image immediately. The fact that I may have needed a 10 second exposure is then irrelevant for viewing.
---
Other variably good stuff:
* [Wikipedia - human eye](http://en.wikipedia.org/wiki/Human_eye)
* [Wikipedia - dynamic range](http://en.wikipedia.org/wiki/Dynamic_range)
* [The Online Photographer - Dynamic Range](http://theonlinephotographer.typepad.com/the_online_photographer/2009/02/dynamic-range.html)
+ more re photos than eyes, but good.
* [Making fine prints in your digital darkroom - Tonal quality and dynamic range in digital cameras](http://www.normankoren.com/digital_tonality.html)
* [HDR FAQ](http://www.hdrsoft.com/resources/dri.html)
* [Discussion. Good.](http://reduser.net/forum/showthread.php?53152-The-Dynamic-Range-of-the-Human-Eye)
* [Panoramas - dynamic range discussion](http://wiki.panotools.org/Dynamic_range)
* [Cameras and Vision](http://www.rags-int-inc.com/PhotoTechStuff/CameraEye/)
+ Good. Claims day 15,000:1 and night 10,000,000:1.
* [Cameras vs the human eye](http://www.cambridgeincolour.com/tutorials/cameras-vs-human-eye.htm)
* [What is the dynamic range of the human eye](http://www.quora.com/What-is-the-dynamic-range-of-the-human-eye)
+ amateur experts opine.
* [Eye and camera differences](http://www.pixiq.com/article/eyes-vs-cameras)
* [Maini's Mind - Our Eyes vs Cameras](https://maini.live/2016/11/26/our-eyes-vs-camera/)
+ an ophthalmologist and a photographer compares. | Metaphorically, it may be due to the fact that the brain doesn't "see" a single image, but composes one based on a series of continuous "shots" from the eyes as they move around the scene.
Each of these "shots" are "taken" with variable "apertures", in order to maximize the overall dynamic range of the final "image".
You can think of the mental process as a mix of a panorama and HDR if you prefer. :o) |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | Metaphorically, it may be due to the fact that the brain doesn't "see" a single image, but composes one based on a series of continuous "shots" from the eyes as they move around the scene.
Each of these "shots" are "taken" with variable "apertures", in order to maximize the overall dynamic range of the final "image".
You can think of the mental process as a mix of a panorama and HDR if you prefer. :o) | This question cannot be standardized because the eye's dynamic range is always shifting to adjust to the intensity of light, not only by the ''human aperture'' but also with the brain's sensitivity to what the eye is looking at. It's like a camera with different processors, using the most sensitive to light when it wants and using the highest sensitivity to dark when it wants. I think the dynamic range of the eye is somewhere around 22 to 24 EV.
I have been intrigued by this question for a while now. Try to take a photo of a milk white exhibition stand with sheets of lightboxes from different angles without having to bracket for exposure and then bracket for white balance separately and then post-processing them later. It is physically impossible.
Just like the eye adjusts to white balance psychologically and that's why the term ''need a fresh eye'' that is because visual perception is also a factor. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | Metaphorically, it may be due to the fact that the brain doesn't "see" a single image, but composes one based on a series of continuous "shots" from the eyes as they move around the scene.
Each of these "shots" are "taken" with variable "apertures", in order to maximize the overall dynamic range of the final "image".
You can think of the mental process as a mix of a panorama and HDR if you prefer. :o) | The top answer here is the best, meanwhile there are several incorrect comments. The eye does not get its massive dynamic range because of eye movements and quick adjustments. Try the experiment where you keep your eyes fixed on a point, and with your eyes fixed note what you can see in your close peripheral vision in areas much brighter or darker. Try fixing on points of varying lightness to see that indeed pretty much everything that falls in the normal light levels is clearly visible to you. Since you are focused and fixed on one spot, eye movements cannot account for the fact you can still easily perceive light and dark objects in your near periphery. Take a picture with the very best cameras and this will not be remotely true.
Of course the sun and other bright sources are too bright when they are close to the center of your view, and going from bright indoor light into pitch dark is also too much. Based on comparisons with the very high dollar videos cameras used for sports, as well as high dollar digital cameras, the 24 stops figure is most likely correct. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This is a very good question, and the answer could fill hundreds of pages - and, in fact, the answer already DOES fill hundreds of pages.
The short answer is that the figures you are citing do not agree with apparent reality because the commonly quoted figures are wrong :-). Read on ...
Much is available on the internet on this subject and the quality is, as ever, widely variable. There is also a lot of parroting of "facts" between sites and figures like those in Wikipedia seem common enough BUT there are some very reasoned arguments which seem to suggest that the Wikipedia figure is extremely wrong and underestimates the figure very substantially.
It's important to note that the eye acts as a contrast detector rather than an absolute level detector (such as a digital camera sensor uses) so comparisons need care.
With irising, chemical adaptation and every other trick it can pull it seems that the absolute dynamic range of the whole eye system is well over 20 stops. As each stop is a factor of 2, that's 2^20 or about "well over 1,000,000:1". At the top end, the sun is too bright!!!. At the bottom end the dark adapted eye can detect a single photon. A D3S (better performance than a D4) may have trouble with that. (Note that that is not EVERY photon - when you get down to the few photons per second level a lot of them will hit non-sensor areas and not be detected. But when one DOES strike a sensitive retina area it will produce a signal that can be recorded.)
But, I digress :-). An extremely good (it seems) page that discusses eye dynamic range and more is
* [Notes on the Resolution and Other Details of the Human Eye](http://clarkvision.com/imagedetail/eye-resolution.html).
Paragraph headings are worth noting:
Notes on the Resolution of the Human Eye
Visual Acuity and Resolving Detail on Prints
How many megapixels equivalent does the eye have?
The Sensitivity of the Human Eye (ISO Equivalent)
The Dynamic Range of the Eye
The Focal Length of the Eye
The writer argues that the dynamic range of the eye without changing sensitivity by adaptation or irising is about 1,000,000:1 in low light conditions. That is, as great as the "well over" lower limit mentioned above. Then he justifies this claim as copied below. This sounds fairly convincing at first glance. There may be flaws in the argument, but it seems OK, and this does not mean that it applies in all light levels.
>
> Here is a simple experiment you can do. Go out with a star chart on a clear night with a full moon. Wait a few minutes for your eyes to adjust. Now find the faintest stars you can detect when the you can see the full moon in your field of view. Try and limit the moon and stars to within about 45 degrees of straight up (the zenith).
>
>
> If you have clear skies away from city lights, you will probably be able to see magnitude 3 stars.
>
>
> The full moon has a stellar magnitude of -12.5.
>
>
> If you can see magnitude 2.5 stars, the magnitude range you are seeing is 15.
>
>
> Every 5 magnitudes is a factor of 100, so 15 is 100 \* 100 \* 100 = 1,000,000.
>
>
> Thus, the dynamic range in this relatively low light condition is about 1 million to one, perhaps higher!
>
>
>
But, here's a suggestion from me for an experiment at normal daylight light levels.
* Find a scene that has a good mixture of dark areas and very bright areas - ideally with some dark areas as isolated islands near islands of brightness. An example may be sunlight shining through trees into a heavily shaded area - a few cavelets or deeply shaded areas will help.
* Allow your eyes to adapt to the general lighting level - do not stare at the bright spots near where the sun is shining through and do not focus on any especially dark areas.
* Note how well you can see detail in the darkest of dark areas - at what level of darkness does is fade to black.
* Try the same with bright areas - as you look toward the sun there will be a place where details washes out and you cannot reasonably see more.
* Cast your eyes to and fro across the scene between dark and light to try to stop your adaptation mechanism changing f-stop on you.
* Now, take photos of the scene. Expose "correctly" and then so the darkest areas that you could see can be seen in the photo and then so that the brightest highlights you could distinguish are not washed out.
* If you have the equipment, take an HDR photo with maximum f-stop variation between photos. (My Sony A77 allows 5ev steps.)
My experience is that my eye can always see a wider brightness range than my camera (Minolta 7Hi, A200, 5D, 7D, A700, A77, other)
On maximum HDR image (10 ev range between centers) my eye can see as well as or better than the camera.
The area where this does not APPEAR to be so is in extremely low light when I may need to allow the eye to integrate (which it does for up to about 4 seconds!) whereas I can look at a low light photo and see the image immediately. The fact that I may have needed a 10 second exposure is then irrelevant for viewing.
---
Other variably good stuff:
* [Wikipedia - human eye](http://en.wikipedia.org/wiki/Human_eye)
* [Wikipedia - dynamic range](http://en.wikipedia.org/wiki/Dynamic_range)
* [The Online Photographer - Dynamic Range](http://theonlinephotographer.typepad.com/the_online_photographer/2009/02/dynamic-range.html)
+ more re photos than eyes, but good.
* [Making fine prints in your digital darkroom - Tonal quality and dynamic range in digital cameras](http://www.normankoren.com/digital_tonality.html)
* [HDR FAQ](http://www.hdrsoft.com/resources/dri.html)
* [Discussion. Good.](http://reduser.net/forum/showthread.php?53152-The-Dynamic-Range-of-the-Human-Eye)
* [Panoramas - dynamic range discussion](http://wiki.panotools.org/Dynamic_range)
* [Cameras and Vision](http://www.rags-int-inc.com/PhotoTechStuff/CameraEye/)
+ Good. Claims day 15,000:1 and night 10,000,000:1.
* [Cameras vs the human eye](http://www.cambridgeincolour.com/tutorials/cameras-vs-human-eye.htm)
* [What is the dynamic range of the human eye](http://www.quora.com/What-is-the-dynamic-range-of-the-human-eye)
+ amateur experts opine.
* [Eye and camera differences](http://www.pixiq.com/article/eyes-vs-cameras)
* [Maini's Mind - Our Eyes vs Cameras](https://maini.live/2016/11/26/our-eyes-vs-camera/)
+ an ophthalmologist and a photographer compares. | The main reason for this is that the human eye registers brightness on a logarithmic scale, whereas digital sensors are linear. Take a look [at this site](http://www.petapixel.com/2011/05/05/biology-for-photographers-why-is-the-aperture-scale-logarithmic/) about halfway down for more info. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This question cannot be standardized because the eye's dynamic range is always shifting to adjust to the intensity of light, not only by the ''human aperture'' but also with the brain's sensitivity to what the eye is looking at. It's like a camera with different processors, using the most sensitive to light when it wants and using the highest sensitivity to dark when it wants. I think the dynamic range of the eye is somewhere around 22 to 24 EV.
I have been intrigued by this question for a while now. Try to take a photo of a milk white exhibition stand with sheets of lightboxes from different angles without having to bracket for exposure and then bracket for white balance separately and then post-processing them later. It is physically impossible.
Just like the eye adjusts to white balance psychologically and that's why the term ''need a fresh eye'' that is because visual perception is also a factor. | The main reason for this is that the human eye registers brightness on a logarithmic scale, whereas digital sensors are linear. Take a look [at this site](http://www.petapixel.com/2011/05/05/biology-for-photographers-why-is-the-aperture-scale-logarithmic/) about halfway down for more info. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | The main reason for this is that the human eye registers brightness on a logarithmic scale, whereas digital sensors are linear. Take a look [at this site](http://www.petapixel.com/2011/05/05/biology-for-photographers-why-is-the-aperture-scale-logarithmic/) about halfway down for more info. | The top answer here is the best, meanwhile there are several incorrect comments. The eye does not get its massive dynamic range because of eye movements and quick adjustments. Try the experiment where you keep your eyes fixed on a point, and with your eyes fixed note what you can see in your close peripheral vision in areas much brighter or darker. Try fixing on points of varying lightness to see that indeed pretty much everything that falls in the normal light levels is clearly visible to you. Since you are focused and fixed on one spot, eye movements cannot account for the fact you can still easily perceive light and dark objects in your near periphery. Take a picture with the very best cameras and this will not be remotely true.
Of course the sun and other bright sources are too bright when they are close to the center of your view, and going from bright indoor light into pitch dark is also too much. Based on comparisons with the very high dollar videos cameras used for sports, as well as high dollar digital cameras, the 24 stops figure is most likely correct. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This is a very good question, and the answer could fill hundreds of pages - and, in fact, the answer already DOES fill hundreds of pages.
The short answer is that the figures you are citing do not agree with apparent reality because the commonly quoted figures are wrong :-). Read on ...
Much is available on the internet on this subject and the quality is, as ever, widely variable. There is also a lot of parroting of "facts" between sites and figures like those in Wikipedia seem common enough BUT there are some very reasoned arguments which seem to suggest that the Wikipedia figure is extremely wrong and underestimates the figure very substantially.
It's important to note that the eye acts as a contrast detector rather than an absolute level detector (such as a digital camera sensor uses) so comparisons need care.
With irising, chemical adaptation and every other trick it can pull it seems that the absolute dynamic range of the whole eye system is well over 20 stops. As each stop is a factor of 2, that's 2^20 or about "well over 1,000,000:1". At the top end, the sun is too bright!!!. At the bottom end the dark adapted eye can detect a single photon. A D3S (better performance than a D4) may have trouble with that. (Note that that is not EVERY photon - when you get down to the few photons per second level a lot of them will hit non-sensor areas and not be detected. But when one DOES strike a sensitive retina area it will produce a signal that can be recorded.)
But, I digress :-). An extremely good (it seems) page that discusses eye dynamic range and more is
* [Notes on the Resolution and Other Details of the Human Eye](http://clarkvision.com/imagedetail/eye-resolution.html).
Paragraph headings are worth noting:
Notes on the Resolution of the Human Eye
Visual Acuity and Resolving Detail on Prints
How many megapixels equivalent does the eye have?
The Sensitivity of the Human Eye (ISO Equivalent)
The Dynamic Range of the Eye
The Focal Length of the Eye
The writer argues that the dynamic range of the eye without changing sensitivity by adaptation or irising is about 1,000,000:1 in low light conditions. That is, as great as the "well over" lower limit mentioned above. Then he justifies this claim as copied below. This sounds fairly convincing at first glance. There may be flaws in the argument, but it seems OK, and this does not mean that it applies in all light levels.
>
> Here is a simple experiment you can do. Go out with a star chart on a clear night with a full moon. Wait a few minutes for your eyes to adjust. Now find the faintest stars you can detect when the you can see the full moon in your field of view. Try and limit the moon and stars to within about 45 degrees of straight up (the zenith).
>
>
> If you have clear skies away from city lights, you will probably be able to see magnitude 3 stars.
>
>
> The full moon has a stellar magnitude of -12.5.
>
>
> If you can see magnitude 2.5 stars, the magnitude range you are seeing is 15.
>
>
> Every 5 magnitudes is a factor of 100, so 15 is 100 \* 100 \* 100 = 1,000,000.
>
>
> Thus, the dynamic range in this relatively low light condition is about 1 million to one, perhaps higher!
>
>
>
But, here's a suggestion from me for an experiment at normal daylight light levels.
* Find a scene that has a good mixture of dark areas and very bright areas - ideally with some dark areas as isolated islands near islands of brightness. An example may be sunlight shining through trees into a heavily shaded area - a few cavelets or deeply shaded areas will help.
* Allow your eyes to adapt to the general lighting level - do not stare at the bright spots near where the sun is shining through and do not focus on any especially dark areas.
* Note how well you can see detail in the darkest of dark areas - at what level of darkness does is fade to black.
* Try the same with bright areas - as you look toward the sun there will be a place where details washes out and you cannot reasonably see more.
* Cast your eyes to and fro across the scene between dark and light to try to stop your adaptation mechanism changing f-stop on you.
* Now, take photos of the scene. Expose "correctly" and then so the darkest areas that you could see can be seen in the photo and then so that the brightest highlights you could distinguish are not washed out.
* If you have the equipment, take an HDR photo with maximum f-stop variation between photos. (My Sony A77 allows 5ev steps.)
My experience is that my eye can always see a wider brightness range than my camera (Minolta 7Hi, A200, 5D, 7D, A700, A77, other)
On maximum HDR image (10 ev range between centers) my eye can see as well as or better than the camera.
The area where this does not APPEAR to be so is in extremely low light when I may need to allow the eye to integrate (which it does for up to about 4 seconds!) whereas I can look at a low light photo and see the image immediately. The fact that I may have needed a 10 second exposure is then irrelevant for viewing.
---
Other variably good stuff:
* [Wikipedia - human eye](http://en.wikipedia.org/wiki/Human_eye)
* [Wikipedia - dynamic range](http://en.wikipedia.org/wiki/Dynamic_range)
* [The Online Photographer - Dynamic Range](http://theonlinephotographer.typepad.com/the_online_photographer/2009/02/dynamic-range.html)
+ more re photos than eyes, but good.
* [Making fine prints in your digital darkroom - Tonal quality and dynamic range in digital cameras](http://www.normankoren.com/digital_tonality.html)
* [HDR FAQ](http://www.hdrsoft.com/resources/dri.html)
* [Discussion. Good.](http://reduser.net/forum/showthread.php?53152-The-Dynamic-Range-of-the-Human-Eye)
* [Panoramas - dynamic range discussion](http://wiki.panotools.org/Dynamic_range)
* [Cameras and Vision](http://www.rags-int-inc.com/PhotoTechStuff/CameraEye/)
+ Good. Claims day 15,000:1 and night 10,000,000:1.
* [Cameras vs the human eye](http://www.cambridgeincolour.com/tutorials/cameras-vs-human-eye.htm)
* [What is the dynamic range of the human eye](http://www.quora.com/What-is-the-dynamic-range-of-the-human-eye)
+ amateur experts opine.
* [Eye and camera differences](http://www.pixiq.com/article/eyes-vs-cameras)
* [Maini's Mind - Our Eyes vs Cameras](https://maini.live/2016/11/26/our-eyes-vs-camera/)
+ an ophthalmologist and a photographer compares. | This question cannot be standardized because the eye's dynamic range is always shifting to adjust to the intensity of light, not only by the ''human aperture'' but also with the brain's sensitivity to what the eye is looking at. It's like a camera with different processors, using the most sensitive to light when it wants and using the highest sensitivity to dark when it wants. I think the dynamic range of the eye is somewhere around 22 to 24 EV.
I have been intrigued by this question for a while now. Try to take a photo of a milk white exhibition stand with sheets of lightboxes from different angles without having to bracket for exposure and then bracket for white balance separately and then post-processing them later. It is physically impossible.
Just like the eye adjusts to white balance psychologically and that's why the term ''need a fresh eye'' that is because visual perception is also a factor. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This is a very good question, and the answer could fill hundreds of pages - and, in fact, the answer already DOES fill hundreds of pages.
The short answer is that the figures you are citing do not agree with apparent reality because the commonly quoted figures are wrong :-). Read on ...
Much is available on the internet on this subject and the quality is, as ever, widely variable. There is also a lot of parroting of "facts" between sites and figures like those in Wikipedia seem common enough BUT there are some very reasoned arguments which seem to suggest that the Wikipedia figure is extremely wrong and underestimates the figure very substantially.
It's important to note that the eye acts as a contrast detector rather than an absolute level detector (such as a digital camera sensor uses) so comparisons need care.
With irising, chemical adaptation and every other trick it can pull it seems that the absolute dynamic range of the whole eye system is well over 20 stops. As each stop is a factor of 2, that's 2^20 or about "well over 1,000,000:1". At the top end, the sun is too bright!!!. At the bottom end the dark adapted eye can detect a single photon. A D3S (better performance than a D4) may have trouble with that. (Note that that is not EVERY photon - when you get down to the few photons per second level a lot of them will hit non-sensor areas and not be detected. But when one DOES strike a sensitive retina area it will produce a signal that can be recorded.)
But, I digress :-). An extremely good (it seems) page that discusses eye dynamic range and more is
* [Notes on the Resolution and Other Details of the Human Eye](http://clarkvision.com/imagedetail/eye-resolution.html).
Paragraph headings are worth noting:
Notes on the Resolution of the Human Eye
Visual Acuity and Resolving Detail on Prints
How many megapixels equivalent does the eye have?
The Sensitivity of the Human Eye (ISO Equivalent)
The Dynamic Range of the Eye
The Focal Length of the Eye
The writer argues that the dynamic range of the eye without changing sensitivity by adaptation or irising is about 1,000,000:1 in low light conditions. That is, as great as the "well over" lower limit mentioned above. Then he justifies this claim as copied below. This sounds fairly convincing at first glance. There may be flaws in the argument, but it seems OK, and this does not mean that it applies in all light levels.
>
> Here is a simple experiment you can do. Go out with a star chart on a clear night with a full moon. Wait a few minutes for your eyes to adjust. Now find the faintest stars you can detect when the you can see the full moon in your field of view. Try and limit the moon and stars to within about 45 degrees of straight up (the zenith).
>
>
> If you have clear skies away from city lights, you will probably be able to see magnitude 3 stars.
>
>
> The full moon has a stellar magnitude of -12.5.
>
>
> If you can see magnitude 2.5 stars, the magnitude range you are seeing is 15.
>
>
> Every 5 magnitudes is a factor of 100, so 15 is 100 \* 100 \* 100 = 1,000,000.
>
>
> Thus, the dynamic range in this relatively low light condition is about 1 million to one, perhaps higher!
>
>
>
But, here's a suggestion from me for an experiment at normal daylight light levels.
* Find a scene that has a good mixture of dark areas and very bright areas - ideally with some dark areas as isolated islands near islands of brightness. An example may be sunlight shining through trees into a heavily shaded area - a few cavelets or deeply shaded areas will help.
* Allow your eyes to adapt to the general lighting level - do not stare at the bright spots near where the sun is shining through and do not focus on any especially dark areas.
* Note how well you can see detail in the darkest of dark areas - at what level of darkness does is fade to black.
* Try the same with bright areas - as you look toward the sun there will be a place where details washes out and you cannot reasonably see more.
* Cast your eyes to and fro across the scene between dark and light to try to stop your adaptation mechanism changing f-stop on you.
* Now, take photos of the scene. Expose "correctly" and then so the darkest areas that you could see can be seen in the photo and then so that the brightest highlights you could distinguish are not washed out.
* If you have the equipment, take an HDR photo with maximum f-stop variation between photos. (My Sony A77 allows 5ev steps.)
My experience is that my eye can always see a wider brightness range than my camera (Minolta 7Hi, A200, 5D, 7D, A700, A77, other)
On maximum HDR image (10 ev range between centers) my eye can see as well as or better than the camera.
The area where this does not APPEAR to be so is in extremely low light when I may need to allow the eye to integrate (which it does for up to about 4 seconds!) whereas I can look at a low light photo and see the image immediately. The fact that I may have needed a 10 second exposure is then irrelevant for viewing.
---
Other variably good stuff:
* [Wikipedia - human eye](http://en.wikipedia.org/wiki/Human_eye)
* [Wikipedia - dynamic range](http://en.wikipedia.org/wiki/Dynamic_range)
* [The Online Photographer - Dynamic Range](http://theonlinephotographer.typepad.com/the_online_photographer/2009/02/dynamic-range.html)
+ more re photos than eyes, but good.
* [Making fine prints in your digital darkroom - Tonal quality and dynamic range in digital cameras](http://www.normankoren.com/digital_tonality.html)
* [HDR FAQ](http://www.hdrsoft.com/resources/dri.html)
* [Discussion. Good.](http://reduser.net/forum/showthread.php?53152-The-Dynamic-Range-of-the-Human-Eye)
* [Panoramas - dynamic range discussion](http://wiki.panotools.org/Dynamic_range)
* [Cameras and Vision](http://www.rags-int-inc.com/PhotoTechStuff/CameraEye/)
+ Good. Claims day 15,000:1 and night 10,000,000:1.
* [Cameras vs the human eye](http://www.cambridgeincolour.com/tutorials/cameras-vs-human-eye.htm)
* [What is the dynamic range of the human eye](http://www.quora.com/What-is-the-dynamic-range-of-the-human-eye)
+ amateur experts opine.
* [Eye and camera differences](http://www.pixiq.com/article/eyes-vs-cameras)
* [Maini's Mind - Our Eyes vs Cameras](https://maini.live/2016/11/26/our-eyes-vs-camera/)
+ an ophthalmologist and a photographer compares. | The top answer here is the best, meanwhile there are several incorrect comments. The eye does not get its massive dynamic range because of eye movements and quick adjustments. Try the experiment where you keep your eyes fixed on a point, and with your eyes fixed note what you can see in your close peripheral vision in areas much brighter or darker. Try fixing on points of varying lightness to see that indeed pretty much everything that falls in the normal light levels is clearly visible to you. Since you are focused and fixed on one spot, eye movements cannot account for the fact you can still easily perceive light and dark objects in your near periphery. Take a picture with the very best cameras and this will not be remotely true.
Of course the sun and other bright sources are too bright when they are close to the center of your view, and going from bright indoor light into pitch dark is also too much. Based on comparisons with the very high dollar videos cameras used for sports, as well as high dollar digital cameras, the 24 stops figure is most likely correct. |
21,579 | According to [DxO tests](http://www.dxomark.com/index.php/Cameras/Camera-Sensor-Ratings/%28type%29/usecase_landscape), cameras have 10 to 12 stops of dynamic range. Is that correct? Noise can completely screw some lowers values (easily resulting in loss of some stops).
Also [Norman Koren says](http://www.normankoren.com/digital_tonality.html) that a digital camera's original dynamic range can be 9 to 11 stops, but prints have "only" 6.5 stops.
In a section on dynamic range, Wikipedia says the human eye has a contrast ratio of around [6.5 stops](http://en.wikipedia.org/wiki/Human_eye#Dynamic_range). If that is the case, why is the human eye clearly much better than cameras to record scenes with high dynamic range? | 2012/03/22 | [
"https://photo.stackexchange.com/questions/21579",
"https://photo.stackexchange.com",
"https://photo.stackexchange.com/users/8794/"
] | This question cannot be standardized because the eye's dynamic range is always shifting to adjust to the intensity of light, not only by the ''human aperture'' but also with the brain's sensitivity to what the eye is looking at. It's like a camera with different processors, using the most sensitive to light when it wants and using the highest sensitivity to dark when it wants. I think the dynamic range of the eye is somewhere around 22 to 24 EV.
I have been intrigued by this question for a while now. Try to take a photo of a milk white exhibition stand with sheets of lightboxes from different angles without having to bracket for exposure and then bracket for white balance separately and then post-processing them later. It is physically impossible.
Just like the eye adjusts to white balance psychologically and that's why the term ''need a fresh eye'' that is because visual perception is also a factor. | The top answer here is the best, meanwhile there are several incorrect comments. The eye does not get its massive dynamic range because of eye movements and quick adjustments. Try the experiment where you keep your eyes fixed on a point, and with your eyes fixed note what you can see in your close peripheral vision in areas much brighter or darker. Try fixing on points of varying lightness to see that indeed pretty much everything that falls in the normal light levels is clearly visible to you. Since you are focused and fixed on one spot, eye movements cannot account for the fact you can still easily perceive light and dark objects in your near periphery. Take a picture with the very best cameras and this will not be remotely true.
Of course the sun and other bright sources are too bright when they are close to the center of your view, and going from bright indoor light into pitch dark is also too much. Based on comparisons with the very high dollar videos cameras used for sports, as well as high dollar digital cameras, the 24 stops figure is most likely correct. |
118,088 | Why is it that chiral biological molecules are enantiomerically pure? The other enantiomer would have the same reactivity, and the only difference is their angle of rotation of plane polarized light. Why, then, is one enantiomer preferred over the other?
Is it that the enantiomer found in our bodies has some advantage over the other enantiomer, or is it random and luck that the structure we see naturally was chosen?
Are all the biochemicals that our body uses enantiomerically pure or are racemic mixtures too? | 2019/07/16 | [
"https://chemistry.stackexchange.com/questions/118088",
"https://chemistry.stackexchange.com",
"https://chemistry.stackexchange.com/users/79319/"
] | >
> Are all the biochemicals that our body uses enantiomerically pure or are racemic mixtures too?
>
>
>
Many molecules exist in both forms in nature. One fun example are the enantiomeric terpenoids R-(–)-carvone and S-(+)-carvone. The R-form smells like spearmint while the S-form smells like caraway. The difference in smell shows that properties other than the optical activity are different for two enantiomers.
>
> Why, then, is one enantiomer preferred over the other?
>
>
>
**Nucleic acids**
The\_Vinz stated correctly in the comments that for replicating structures like RNA, choice of the enatiomer was random; once established, one chiral form prevailed. DNA building blocks have the same chirality as those of RNA because they are made by the same biochemical pathway.
**Amino acids**
One amino acid, glycine, is not chiral. Many amino acids are found in both forms. L-amino acids are used to make proteins, but D-amino acids are made in bacteria and used in the context of cell walls and natural antibiotics.
**Proteins**
Proteins made by ribosomes (from amino acids attached to tRNA by tRNA-synthetases) use L-amino acids exclusively. The tRNA-synthetases are highly specific (including stereo specific), and they don't link tRNA to D-amino acids. It helps that there are very little D-amino acids made in a typical cell. Why one form was chosen over the other is probably luck again. Why all amino acids have the same chirality at the alpha carbon is more intriguing. Some are made from the same precursor, so that will contribute. [Right-handed alpha helices](http://proteopedia.org/wiki/index.php/Alpha_helix) require that the amino acids in them be L-amino acids. If proteins had a mixture of L- and D-amino acids (e.g. all alanines are D-alanines but all aspartates are L-aspartate), alpha helices would be more constrained in the possible sequences, and some of them would be left-handed.
**All other molecules**
Most steps in the synthesis of biomolecules are catalyzed by enzymes. Enzymes, as chiral catalysts that have lots of interactions with reactants, are often highly stereospecific. So the presence or absence of enzymes catalyzing certain reactions largely determines which products are made, and there is no additional cost of making a chiral product from non-chiral precursors (very different from a typical lab synthesis). | Enzymes are very specific for the reaction to enantiomers. For certain products you have only one enantiomer in nature.
It may be the case that there is just one enantiomers which can react with an enzyme because of steric hindrance. But it is also possible that a racemic mixture can occur in fermentation products. So there is no general answer.
The biochemical reaction of two enantiomers can be very different in the human body. For example the contergan / thalidomid can be indicing sleep in the one form and in the other form can cause birth defects. |
38,832 | I’ve seen it done in Harry Potter I just don’t understand when its okay to do it, like can I say something like “John walked away from the house and was now walking along the road” or not? When is it acceptable? | 2018/09/11 | [
"https://writers.stackexchange.com/questions/38832",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/31659/"
] | Think of "now" as a thumbtack that refers to a specific moment. That moment can be in the past.
Some examples:
>
> Jane had been a student. Now she was a programmer.
>
>
>
or
>
> Jane looked up, drearily. Now, *now*, of all times, he was going to interrupt her?
>
>
>
or
>
> Jane ran round the house. Now the car was gone. Where could it be?
>
>
> | If English if not your first language the usage of "now" is difficult to fully comprehend. It has many nuanced translations but is most often used to mean "in this moment". It is also used to emphasise a statement.
"Now Daddy didn't take too kindly his only daughter dating a black boy."
"They put the homeless kids in cages, now that ain't right."
Your sentence is acceptable. |
38,832 | I’ve seen it done in Harry Potter I just don’t understand when its okay to do it, like can I say something like “John walked away from the house and was now walking along the road” or not? When is it acceptable? | 2018/09/11 | [
"https://writers.stackexchange.com/questions/38832",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/31659/"
] | ### It is acceptable when relating a sequence of actions or events.
>
> Jake fed the chickens, then walked the rows of the tomato garden and pulled six new weeds that had sprouted, and now he was throwing a ball for the dog, Reggie. Reggie, panting and waiting for another throw, closed his mouth and turned his attention sharply toward the dirt road that led to the house. Jake turned, too. It was that green truck again. Curtis from the dry grocer, Mama's friend.
>
>
>
The whole thing is past tense, but some of it is more past than others. In the example, I don't want to spend a lot of time describing how Jake fed chickens or pulled weeds, those aren't important at all. I just want to indicate he did his chores and time went by. He didn't just step outside, throw a ball and hear a truck.
I could have said that, "He did a few chores and was playing with the dog, throwing a ball for the dog to fetch." But to me that sounds too vague, for readers I want them to see Jake doing those mundane chores without boring them to tears.
When you are telling a story in the past tense, there is still a "present" in the novel from the viewpoint of the **characters,** not the narrator. On a given page there is stuff they have a past, deeds done and things learned, and a future, deed to do and things to learn. That is the "Now" being referred to.
"Now" is used to return the reader to the present state of the character (from the character's point of view) after you the author have glossed over some time (from minutes to decades) by reciting a short history of that time. | If English if not your first language the usage of "now" is difficult to fully comprehend. It has many nuanced translations but is most often used to mean "in this moment". It is also used to emphasise a statement.
"Now Daddy didn't take too kindly his only daughter dating a black boy."
"They put the homeless kids in cages, now that ain't right."
Your sentence is acceptable. |
38,832 | I’ve seen it done in Harry Potter I just don’t understand when its okay to do it, like can I say something like “John walked away from the house and was now walking along the road” or not? When is it acceptable? | 2018/09/11 | [
"https://writers.stackexchange.com/questions/38832",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/31659/"
] | In a past tense story, "now" means "at the moment in question," not at the "present." | If English if not your first language the usage of "now" is difficult to fully comprehend. It has many nuanced translations but is most often used to mean "in this moment". It is also used to emphasise a statement.
"Now Daddy didn't take too kindly his only daughter dating a black boy."
"They put the homeless kids in cages, now that ain't right."
Your sentence is acceptable. |
38,832 | I’ve seen it done in Harry Potter I just don’t understand when its okay to do it, like can I say something like “John walked away from the house and was now walking along the road” or not? When is it acceptable? | 2018/09/11 | [
"https://writers.stackexchange.com/questions/38832",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/31659/"
] | Think of "now" as a thumbtack that refers to a specific moment. That moment can be in the past.
Some examples:
>
> Jane had been a student. Now she was a programmer.
>
>
>
or
>
> Jane looked up, drearily. Now, *now*, of all times, he was going to interrupt her?
>
>
>
or
>
> Jane ran round the house. Now the car was gone. Where could it be?
>
>
> | ### It is acceptable when relating a sequence of actions or events.
>
> Jake fed the chickens, then walked the rows of the tomato garden and pulled six new weeds that had sprouted, and now he was throwing a ball for the dog, Reggie. Reggie, panting and waiting for another throw, closed his mouth and turned his attention sharply toward the dirt road that led to the house. Jake turned, too. It was that green truck again. Curtis from the dry grocer, Mama's friend.
>
>
>
The whole thing is past tense, but some of it is more past than others. In the example, I don't want to spend a lot of time describing how Jake fed chickens or pulled weeds, those aren't important at all. I just want to indicate he did his chores and time went by. He didn't just step outside, throw a ball and hear a truck.
I could have said that, "He did a few chores and was playing with the dog, throwing a ball for the dog to fetch." But to me that sounds too vague, for readers I want them to see Jake doing those mundane chores without boring them to tears.
When you are telling a story in the past tense, there is still a "present" in the novel from the viewpoint of the **characters,** not the narrator. On a given page there is stuff they have a past, deeds done and things learned, and a future, deed to do and things to learn. That is the "Now" being referred to.
"Now" is used to return the reader to the present state of the character (from the character's point of view) after you the author have glossed over some time (from minutes to decades) by reciting a short history of that time. |
38,832 | I’ve seen it done in Harry Potter I just don’t understand when its okay to do it, like can I say something like “John walked away from the house and was now walking along the road” or not? When is it acceptable? | 2018/09/11 | [
"https://writers.stackexchange.com/questions/38832",
"https://writers.stackexchange.com",
"https://writers.stackexchange.com/users/31659/"
] | In a past tense story, "now" means "at the moment in question," not at the "present." | ### It is acceptable when relating a sequence of actions or events.
>
> Jake fed the chickens, then walked the rows of the tomato garden and pulled six new weeds that had sprouted, and now he was throwing a ball for the dog, Reggie. Reggie, panting and waiting for another throw, closed his mouth and turned his attention sharply toward the dirt road that led to the house. Jake turned, too. It was that green truck again. Curtis from the dry grocer, Mama's friend.
>
>
>
The whole thing is past tense, but some of it is more past than others. In the example, I don't want to spend a lot of time describing how Jake fed chickens or pulled weeds, those aren't important at all. I just want to indicate he did his chores and time went by. He didn't just step outside, throw a ball and hear a truck.
I could have said that, "He did a few chores and was playing with the dog, throwing a ball for the dog to fetch." But to me that sounds too vague, for readers I want them to see Jake doing those mundane chores without boring them to tears.
When you are telling a story in the past tense, there is still a "present" in the novel from the viewpoint of the **characters,** not the narrator. On a given page there is stuff they have a past, deeds done and things learned, and a future, deed to do and things to learn. That is the "Now" being referred to.
"Now" is used to return the reader to the present state of the character (from the character's point of view) after you the author have glossed over some time (from minutes to decades) by reciting a short history of that time. |
192,594 | We are given the following circuit and told to calculate the equivalent voltage of everything to the left of the a and b terminals:
[](https://i.stack.imgur.com/ECjj0.png)
It omits the load resistor and lays the circuit out as such for mesh analysis:
[](https://i.stack.imgur.com/SQkc2.png)
It claims that voltage across terminals a and b, the voltage we are looking for, is the same as the potential in between the 4 ohm and 1 ohm resistor. Shouldn't that 1 ohm resistor cause the potential to change a little? | 2015/09/28 | [
"https://electronics.stackexchange.com/questions/192594",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/38282/"
] | You're trying to find the Thevenin equivalent voltage of the network.
This is the voltage that the network produces when the output is open.
When the output is open, no current flows out or in to the `a` terminal.
Therefore no current flows through the 1 ohm resistor.
Since no current is flowing through the 1 ohm resistor, the voltage across it is 0, by Ohm's law. | To add a little bit to the accepted answer: in real life, the voltage after the 1 ohm resistor will not equal the voltage before it for only a split second during the charging of the circuit, when you first turn on the 32V source, or for the split second just after you ground the 32V source. The reason is because during these short transient times, "a" is charging up through the resistor (since it has a tiny bit of capacitance..as all wires and conductors do in real life), so a tiny current does exist. Once "a" is charged, and you reach the steady-state condition, the current ceases and the voltage on each side of the 1 ohm resistor *is* the same.
In theory, "a" has no capacitance, so you can ignore the above real-life situation. :)
The point, however, is that in this case, the real-life *steady-state* solution is the same as the theoretical one, though the real-life *transient* solution is momentarily different.
Now, the above is just some extra info. For the most direct and correct answer, see "The Photon"'s answer. |
192,594 | We are given the following circuit and told to calculate the equivalent voltage of everything to the left of the a and b terminals:
[](https://i.stack.imgur.com/ECjj0.png)
It omits the load resistor and lays the circuit out as such for mesh analysis:
[](https://i.stack.imgur.com/SQkc2.png)
It claims that voltage across terminals a and b, the voltage we are looking for, is the same as the potential in between the 4 ohm and 1 ohm resistor. Shouldn't that 1 ohm resistor cause the potential to change a little? | 2015/09/28 | [
"https://electronics.stackexchange.com/questions/192594",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/38282/"
] | You're trying to find the Thevenin equivalent voltage of the network.
This is the voltage that the network produces when the output is open.
When the output is open, no current flows out or in to the `a` terminal.
Therefore no current flows through the 1 ohm resistor.
Since no current is flowing through the 1 ohm resistor, the voltage across it is 0, by Ohm's law. | I think its an optical illusion and that's why it looks odd. the 4 ohm and 12 ohm resistors are whats important here. You connect the RL at that junction, so all that matters is that you get the 4 + 12 = 16, and 32/16 = 2A, confirmed by the 2A symbol there.
then its whatever voltage ratio you get. if E = IR then voltage is 2 amps x 12 ohms = 24v at your RL point there
the 1 ohm will cause the current to decrease but i think the voltage will be 24v. in real equipment they will only be able to supply so much current, so in real circuits you would need to watch the 1 ohm there, but in theory with no load defined, its assumed the other parts of this equation are perfect (right?) |
192,594 | We are given the following circuit and told to calculate the equivalent voltage of everything to the left of the a and b terminals:
[](https://i.stack.imgur.com/ECjj0.png)
It omits the load resistor and lays the circuit out as such for mesh analysis:
[](https://i.stack.imgur.com/SQkc2.png)
It claims that voltage across terminals a and b, the voltage we are looking for, is the same as the potential in between the 4 ohm and 1 ohm resistor. Shouldn't that 1 ohm resistor cause the potential to change a little? | 2015/09/28 | [
"https://electronics.stackexchange.com/questions/192594",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/38282/"
] | To add a little bit to the accepted answer: in real life, the voltage after the 1 ohm resistor will not equal the voltage before it for only a split second during the charging of the circuit, when you first turn on the 32V source, or for the split second just after you ground the 32V source. The reason is because during these short transient times, "a" is charging up through the resistor (since it has a tiny bit of capacitance..as all wires and conductors do in real life), so a tiny current does exist. Once "a" is charged, and you reach the steady-state condition, the current ceases and the voltage on each side of the 1 ohm resistor *is* the same.
In theory, "a" has no capacitance, so you can ignore the above real-life situation. :)
The point, however, is that in this case, the real-life *steady-state* solution is the same as the theoretical one, though the real-life *transient* solution is momentarily different.
Now, the above is just some extra info. For the most direct and correct answer, see "The Photon"'s answer. | I think its an optical illusion and that's why it looks odd. the 4 ohm and 12 ohm resistors are whats important here. You connect the RL at that junction, so all that matters is that you get the 4 + 12 = 16, and 32/16 = 2A, confirmed by the 2A symbol there.
then its whatever voltage ratio you get. if E = IR then voltage is 2 amps x 12 ohms = 24v at your RL point there
the 1 ohm will cause the current to decrease but i think the voltage will be 24v. in real equipment they will only be able to supply so much current, so in real circuits you would need to watch the 1 ohm there, but in theory with no load defined, its assumed the other parts of this equation are perfect (right?) |
21,239 | Three weeks ago, I adopted a 9-month-old female cat, Kira. On the ride home and for a little while after, she seemed pretty calm, though withdrawn. She's set up camp in my bedroom and has food/water/litterbox there. When I'm awake in the room, she stays under my bed almost exclusively. Sometimes my roommate can coax her out, but I never seem to be able to. Typically, I'd go under the bed to retrieve her, which I've now learned is the very wrong thing to do. Now she's terrified of me, and sprints away whenever I approach her, desperately searching for a place to hide from me. She still doesn't do this to my roommate. Did I screw this up beyond hope of repair? | 2018/08/31 | [
"https://pets.stackexchange.com/questions/21239",
"https://pets.stackexchange.com",
"https://pets.stackexchange.com/users/12873/"
] | There could be more going on that might deteriorate your relationship with your cat. Looking directly in their eyes for example means a threat in cat language. Moving directly towards her **and** looking her in the eyes is the equivalent of an attack.
Repairing the damage might take some time. Start by offering good smelling, tasty treats in a passive way.
* Move calmly into the room
* Sit down on the floor. Lean sideways and prop yourself up on one ellbow if you can hold the position for some time. This broadcasts relaxation.
* Place a treat at arms length beside you
* Close your eyes to narrow slits, do the lazy cat blink occasionally.
* Completely ignore your cat until she (hopefully) ate the treat and went away.
If this first step was successfull, you can step it up by placing the treat in front of you, then closer, until you hold it in your hand or place it on your lap.
Don't forget to not look directly in her eyes, but close your eyes a little and blink slowly instead. That way you tell your cat "I'm calm and relaxed, don't be afraid". | I had a similar experience with my cat, fortunately it lasted only a day. So not sure if this will help, he is usually under the sofa.
When you are calling their name, tap the couch with fingers or hand rhythmically. “Come
here baby 1...2...1..2” they will become focused on your finger movment and come towards you. My cats take it as a signal to play or to come to me. At this point I’ll excitedly say “good boy”, “good job” and spout nonsense about how much I love him in a coaxing tone . He digs that.
* I softly call his name and speak to him , while waving a feathered toy (distracting him with a toy helps).
* Treats, leave treats out near where the kitty hides (scatter them if you want), just make sure to ignore your cat when he sniffs about. Don’t be in a hurry to hold him.
* Make sure you are the one feeding the cat, cats are often more open to individuals who feed them.
* Don’t be aggressive in words or actions and I dont mean with the intention to do harm, but in terms of raising your voice, getting frustrated or hitting the surfacing to harshly to get their attention. |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | I know this is an old question but PowerPoint now supports SVG and I recently tested a PPT2GIFa converted to ascertain that PowerPoint 2016 renders animated GIF images at a maximum frame rate of 50fps (20ms delay). | **Here is the simple answer...**
Just get a newer version of powerpoint.
**A bit more of a helpful answer**
There is a way of accelerating the GIF by going to this website:
<https://ezgif.com/speed>
Or maybe trying to speed up the FPS in Piskel or Photoshop or something. I don't know *anything* about photoshop so i would just say use Piskel or ezgif |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | I know this is an old question but PowerPoint now supports SVG and I recently tested a PPT2GIFa converted to ascertain that PowerPoint 2016 renders animated GIF images at a maximum frame rate of 50fps (20ms delay). | Scribblemacher answer is in my opinion a good one.
I am just puting an alternative if for some reason you need to stick to the animaged gif.
There is an interesting program: <http://slidedog.com/> for Windows and there is a mac beta.
Basicly what it does is that integrates quite well diferent media in the same presentation.
You can put for example the first slides of a power point, then put a webpage as a next slide and then again a power point, or video, or a pdf into a seamless presentation.
So you could insert your gif into a simple html document as a slide.
If you go for the video option, you can use VirtualDub to convert your gif into video. <http://virtualdub.sourceforge.net/>
Try to use a good codec for avi, like Xvid <https://www.xvid.com/> |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Different applications (mainly browsers, but in this case Powerpoint), interpret frame delays differently. IE, for example, historically [seemed to render gifs slower](https://blogs.msdn.microsoft.com/ieinternals/2010/06/07/trivia-animated-gif-timing/) than other browsers.
As you said in one of your comments, changing the timing in the gif solved the problem. While browsers seem to have tricks for normalizing speed in gifs, Powerpoint's render is probably very basic, considering its basic (or non-existant) support for other image formats like SVG. So this seems to be a case where you need to modify your gif to meet Powerpoint's needs.
The other solution is to convert the gif to a video format and use the video instead. That would be my approach if the animation was very important to the presentation. *(If it wasn't important to the presentation, then I'd actually remove it--personally, I think animations and transitions in slideshows are distracting and pull attention away from the speaker)*
**See also:** [Why is this gifs animation speed different in Firefox vs IE?](https://webmasters.stackexchange.com/questions/26994/why-is-this-gifs-animation-speed-different-in-firefox-vs-ie) | I know this is an old question but PowerPoint now supports SVG and I recently tested a PPT2GIFa converted to ascertain that PowerPoint 2016 renders animated GIF images at a maximum frame rate of 50fps (20ms delay). |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Scribblemacher answer is in my opinion a good one.
I am just puting an alternative if for some reason you need to stick to the animaged gif.
There is an interesting program: <http://slidedog.com/> for Windows and there is a mac beta.
Basicly what it does is that integrates quite well diferent media in the same presentation.
You can put for example the first slides of a power point, then put a webpage as a next slide and then again a power point, or video, or a pdf into a seamless presentation.
So you could insert your gif into a simple html document as a slide.
If you go for the video option, you can use VirtualDub to convert your gif into video. <http://virtualdub.sourceforge.net/>
Try to use a good codec for avi, like Xvid <https://www.xvid.com/> | Switch to Keynote.
Have been going through this nightmare recently when I was tasked to make a Powerpoint for a client. Powerpoint simply can't handle GIFS, unless playback is being done on an extremely powerful gaming rig. Keynote plays them flawlessly even on terrible systems. |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Only because it hasn't been mentioned: Are you inserting the GIF file as an image? From experience, this results in slow, laggy playback with lots of dropped frames.
The solution?
Insert the GIF as a video file (Insert > Video from file) then set its properties to play automatically and loop until stopped.
That's it! ;) | I know this is an old question but PowerPoint now supports SVG and I recently tested a PPT2GIFa converted to ascertain that PowerPoint 2016 renders animated GIF images at a maximum frame rate of 50fps (20ms delay). |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Only because it hasn't been mentioned: Are you inserting the GIF file as an image? From experience, this results in slow, laggy playback with lots of dropped frames.
The solution?
Insert the GIF as a video file (Insert > Video from file) then set its properties to play automatically and loop until stopped.
That's it! ;) | **Here is the simple answer...**
Just get a newer version of powerpoint.
**A bit more of a helpful answer**
There is a way of accelerating the GIF by going to this website:
<https://ezgif.com/speed>
Or maybe trying to speed up the FPS in Piskel or Photoshop or something. I don't know *anything* about photoshop so i would just say use Piskel or ezgif |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | **Here is the simple answer...**
Just get a newer version of powerpoint.
**A bit more of a helpful answer**
There is a way of accelerating the GIF by going to this website:
<https://ezgif.com/speed>
Or maybe trying to speed up the FPS in Piskel or Photoshop or something. I don't know *anything* about photoshop so i would just say use Piskel or ezgif | Switch to Keynote.
Have been going through this nightmare recently when I was tasked to make a Powerpoint for a client. Powerpoint simply can't handle GIFS, unless playback is being done on an extremely powerful gaming rig. Keynote plays them flawlessly even on terrible systems. |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Different applications (mainly browsers, but in this case Powerpoint), interpret frame delays differently. IE, for example, historically [seemed to render gifs slower](https://blogs.msdn.microsoft.com/ieinternals/2010/06/07/trivia-animated-gif-timing/) than other browsers.
As you said in one of your comments, changing the timing in the gif solved the problem. While browsers seem to have tricks for normalizing speed in gifs, Powerpoint's render is probably very basic, considering its basic (or non-existant) support for other image formats like SVG. So this seems to be a case where you need to modify your gif to meet Powerpoint's needs.
The other solution is to convert the gif to a video format and use the video instead. That would be my approach if the animation was very important to the presentation. *(If it wasn't important to the presentation, then I'd actually remove it--personally, I think animations and transitions in slideshows are distracting and pull attention away from the speaker)*
**See also:** [Why is this gifs animation speed different in Firefox vs IE?](https://webmasters.stackexchange.com/questions/26994/why-is-this-gifs-animation-speed-different-in-firefox-vs-ie) | Switch to Keynote.
Have been going through this nightmare recently when I was tasked to make a Powerpoint for a client. Powerpoint simply can't handle GIFS, unless playback is being done on an extremely powerful gaming rig. Keynote plays them flawlessly even on terrible systems. |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Only because it hasn't been mentioned: Are you inserting the GIF file as an image? From experience, this results in slow, laggy playback with lots of dropped frames.
The solution?
Insert the GIF as a video file (Insert > Video from file) then set its properties to play automatically and loop until stopped.
That's it! ;) | Scribblemacher answer is in my opinion a good one.
I am just puting an alternative if for some reason you need to stick to the animaged gif.
There is an interesting program: <http://slidedog.com/> for Windows and there is a mac beta.
Basicly what it does is that integrates quite well diferent media in the same presentation.
You can put for example the first slides of a power point, then put a webpage as a next slide and then again a power point, or video, or a pdf into a seamless presentation.
So you could insert your gif into a simple html document as a slide.
If you go for the video option, you can use VirtualDub to convert your gif into video. <http://virtualdub.sourceforge.net/>
Try to use a good codec for avi, like Xvid <https://www.xvid.com/> |
64,709 | As the title suggests. I have to put a GIF I made on Flash in a PowerPoint presentation. But when I run the presentation, the GIF runs on a dreadfully low frame rate. Is this issue fixed in later versions of PowerPoint? Or is there a fix to it in the 2010 version itself? Or if there's another software I could use to make presentations which run the GIF properly?
I did search for this, but I could only find forum queries which were unanswered. | 2015/12/25 | [
"https://graphicdesign.stackexchange.com/questions/64709",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/48770/"
] | Different applications (mainly browsers, but in this case Powerpoint), interpret frame delays differently. IE, for example, historically [seemed to render gifs slower](https://blogs.msdn.microsoft.com/ieinternals/2010/06/07/trivia-animated-gif-timing/) than other browsers.
As you said in one of your comments, changing the timing in the gif solved the problem. While browsers seem to have tricks for normalizing speed in gifs, Powerpoint's render is probably very basic, considering its basic (or non-existant) support for other image formats like SVG. So this seems to be a case where you need to modify your gif to meet Powerpoint's needs.
The other solution is to convert the gif to a video format and use the video instead. That would be my approach if the animation was very important to the presentation. *(If it wasn't important to the presentation, then I'd actually remove it--personally, I think animations and transitions in slideshows are distracting and pull attention away from the speaker)*
**See also:** [Why is this gifs animation speed different in Firefox vs IE?](https://webmasters.stackexchange.com/questions/26994/why-is-this-gifs-animation-speed-different-in-firefox-vs-ie) | **Here is the simple answer...**
Just get a newer version of powerpoint.
**A bit more of a helpful answer**
There is a way of accelerating the GIF by going to this website:
<https://ezgif.com/speed>
Or maybe trying to speed up the FPS in Piskel or Photoshop or something. I don't know *anything* about photoshop so i would just say use Piskel or ezgif |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | **Resume Considerations**
You may wish to phrase things in such a way to make this clear. If you are sending out resumes, consider how it would look to to see something like:
* Widget Engineer at MegaCorp A
* Widget Engineer at MegaCorp B
It will not be clear at the outset that you have been doing contract work. If you include bullets under each job and include something like "6-month contract to build Awesome Widgets" and have the dates match accordingly, I don't think this is an issue.
In fact, there may be companies which see this as a plus - because if they are looking for a short-term contract employee and see someone with only 1+ year periods at each of their job, it will be difficult for the employer to judge how you would take a contract offer and if you would be happy to accept a shorter contract and have the employment end after this period.
**Interview Considerations**
If you get to an interview, this should be an easy thing to talk about as you are able to give explanations as to why you have had so many jobs. It is not a given people will see this as a negative, however, if you run into older hiring managers or even HR people they may (regardless of this being more typical in software) care and so it is probably a good idea to at least talk about your work experience to answer the unspoken feeling of "does this guy just have a hard time liking any company; he probably won't like us either" a lot of those people likely will have.
Again, keep in mind if you are applying through the traditional route, you and your resume/information will interact with people who are not part of the software-engineering role, and they likely will view things with a somewhat different perspective.
---
Just a note, if you are simply changing jobs because you never like them or because the company doesn't want to renew any of the contracts because of performance, then yes, it probably will be seen as a negative. | Most hiring managers wouldn't consider that to be a black mark. Deliberately changing jobs every year or couple years is viewed very negatively by a lot of hiring managers though. (A survey of recruiters and hiring managers found that 39% of them listed "job-hopping" as the biggest obstacle for job seekers. <http://www.askamanager.org/2012/10/job-hopping-is-killing-your-career.html>)
Because of that, I think it's worth listing the temporary jobs as such on your resume, for example by putting "(6-Month Contract)" after the dates of employment. This can prevent someone who's worked a lot of contract jobs from being viewed as someone who gets bored after a few months and moves on. |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | **Resume Considerations**
You may wish to phrase things in such a way to make this clear. If you are sending out resumes, consider how it would look to to see something like:
* Widget Engineer at MegaCorp A
* Widget Engineer at MegaCorp B
It will not be clear at the outset that you have been doing contract work. If you include bullets under each job and include something like "6-month contract to build Awesome Widgets" and have the dates match accordingly, I don't think this is an issue.
In fact, there may be companies which see this as a plus - because if they are looking for a short-term contract employee and see someone with only 1+ year periods at each of their job, it will be difficult for the employer to judge how you would take a contract offer and if you would be happy to accept a shorter contract and have the employment end after this period.
**Interview Considerations**
If you get to an interview, this should be an easy thing to talk about as you are able to give explanations as to why you have had so many jobs. It is not a given people will see this as a negative, however, if you run into older hiring managers or even HR people they may (regardless of this being more typical in software) care and so it is probably a good idea to at least talk about your work experience to answer the unspoken feeling of "does this guy just have a hard time liking any company; he probably won't like us either" a lot of those people likely will have.
Again, keep in mind if you are applying through the traditional route, you and your resume/information will interact with people who are not part of the software-engineering role, and they likely will view things with a somewhat different perspective.
---
Just a note, if you are simply changing jobs because you never like them or because the company doesn't want to renew any of the contracts because of performance, then yes, it probably will be seen as a negative. | It depends.
Earlier in my career I used to take a skeptical view of people who never stayed in the same gig for more than a few months; my concern was perhaps they couldn't. Later in my career, I started taking a skeptical view of people who never left their jobs; my concern was perhaps they couldn't. Since I've interviewed tons of candidates (and have interviewed for jobs at a couple of companies where many people "never" leave), I can find plenty of supporting evidence for both positions, but realistically, it's actually very hard to predict how people will react.
In software, however, many people are quite used to people doing a string of contract gigs. If you aren't working on a 1099 basis, but through agencies, you may want to be more explicit about the contract nature (FooCorp via BarAgency) so that it doesn't raise any hackles by people who might just assume you're quitting whenever your code starts to have consequences. If you're working on a 1099 basis, I recommend using your company as the employer, with "Engaged at XCorp on Project Y. Delivered Q solution to..." as bullet points instead of showing each project as a different "job".
The biggest risk you might have as a contractor is that people will worry about candidates who have never been working on a project long enough to have to think about maintenance considerations, because they won't have suffered enough from their own previous design decisions.
My current contract gig has actually lasted longer than my last two full time arrangements; I left those jobs because it was clear that either the company was about to implode or that I would stop learning things that would allow me to progress professionally. If anyone asks, I just answer honestly and tactfully. They rarely ask.
Your biggest risk is the pre-interview filtering that may happen. You won't know if you're being tossed out of the pile of resumes because of your career history, but you also won't know if it's because you share the last name of the hiring manager's psychotic ex-girlfriend or because they don't want generalists, or don't want specialists.
The best thing you can do for your career prospects is to build a reputation of acting professionally, working smart, and solving problems. The tech job market is small enough that you're likely to work with someone you've worked with before, and as long as the wake you leave resulted in mostly positive impressions, you'll be able to find future gigs without much of a struggle. |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | **Resume Considerations**
You may wish to phrase things in such a way to make this clear. If you are sending out resumes, consider how it would look to to see something like:
* Widget Engineer at MegaCorp A
* Widget Engineer at MegaCorp B
It will not be clear at the outset that you have been doing contract work. If you include bullets under each job and include something like "6-month contract to build Awesome Widgets" and have the dates match accordingly, I don't think this is an issue.
In fact, there may be companies which see this as a plus - because if they are looking for a short-term contract employee and see someone with only 1+ year periods at each of their job, it will be difficult for the employer to judge how you would take a contract offer and if you would be happy to accept a shorter contract and have the employment end after this period.
**Interview Considerations**
If you get to an interview, this should be an easy thing to talk about as you are able to give explanations as to why you have had so many jobs. It is not a given people will see this as a negative, however, if you run into older hiring managers or even HR people they may (regardless of this being more typical in software) care and so it is probably a good idea to at least talk about your work experience to answer the unspoken feeling of "does this guy just have a hard time liking any company; he probably won't like us either" a lot of those people likely will have.
Again, keep in mind if you are applying through the traditional route, you and your resume/information will interact with people who are not part of the software-engineering role, and they likely will view things with a somewhat different perspective.
---
Just a note, if you are simply changing jobs because you never like them or because the company doesn't want to renew any of the contracts because of performance, then yes, it probably will be seen as a negative. | I wouldn't view you as a job hopper and disqualify you on that basis, but, to be honest, I would have difficulty considering you qualified for a senior position either. This is because you haven't had to live with the results of what you created. Further, many people I have interviewed and worked with as contractors with this type of experience never develop the depth to become senior developers, they are repeating the same level of experience over and over. Now if each contract takes you a little further in depth on something and you are clearly developing significant expertise in an area, then it would not be a deal breaker particularly if you are a specialist of some kind. But the interview would be quite pointed towards how you have grown and learned from your mistakes. |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | It depends.
Earlier in my career I used to take a skeptical view of people who never stayed in the same gig for more than a few months; my concern was perhaps they couldn't. Later in my career, I started taking a skeptical view of people who never left their jobs; my concern was perhaps they couldn't. Since I've interviewed tons of candidates (and have interviewed for jobs at a couple of companies where many people "never" leave), I can find plenty of supporting evidence for both positions, but realistically, it's actually very hard to predict how people will react.
In software, however, many people are quite used to people doing a string of contract gigs. If you aren't working on a 1099 basis, but through agencies, you may want to be more explicit about the contract nature (FooCorp via BarAgency) so that it doesn't raise any hackles by people who might just assume you're quitting whenever your code starts to have consequences. If you're working on a 1099 basis, I recommend using your company as the employer, with "Engaged at XCorp on Project Y. Delivered Q solution to..." as bullet points instead of showing each project as a different "job".
The biggest risk you might have as a contractor is that people will worry about candidates who have never been working on a project long enough to have to think about maintenance considerations, because they won't have suffered enough from their own previous design decisions.
My current contract gig has actually lasted longer than my last two full time arrangements; I left those jobs because it was clear that either the company was about to implode or that I would stop learning things that would allow me to progress professionally. If anyone asks, I just answer honestly and tactfully. They rarely ask.
Your biggest risk is the pre-interview filtering that may happen. You won't know if you're being tossed out of the pile of resumes because of your career history, but you also won't know if it's because you share the last name of the hiring manager's psychotic ex-girlfriend or because they don't want generalists, or don't want specialists.
The best thing you can do for your career prospects is to build a reputation of acting professionally, working smart, and solving problems. The tech job market is small enough that you're likely to work with someone you've worked with before, and as long as the wake you leave resulted in mostly positive impressions, you'll be able to find future gigs without much of a struggle. | Most hiring managers wouldn't consider that to be a black mark. Deliberately changing jobs every year or couple years is viewed very negatively by a lot of hiring managers though. (A survey of recruiters and hiring managers found that 39% of them listed "job-hopping" as the biggest obstacle for job seekers. <http://www.askamanager.org/2012/10/job-hopping-is-killing-your-career.html>)
Because of that, I think it's worth listing the temporary jobs as such on your resume, for example by putting "(6-Month Contract)" after the dates of employment. This can prevent someone who's worked a lot of contract jobs from being viewed as someone who gets bored after a few months and moves on. |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | I wouldn't view you as a job hopper and disqualify you on that basis, but, to be honest, I would have difficulty considering you qualified for a senior position either. This is because you haven't had to live with the results of what you created. Further, many people I have interviewed and worked with as contractors with this type of experience never develop the depth to become senior developers, they are repeating the same level of experience over and over. Now if each contract takes you a little further in depth on something and you are clearly developing significant expertise in an area, then it would not be a deal breaker particularly if you are a specialist of some kind. But the interview would be quite pointed towards how you have grown and learned from your mistakes. | Most hiring managers wouldn't consider that to be a black mark. Deliberately changing jobs every year or couple years is viewed very negatively by a lot of hiring managers though. (A survey of recruiters and hiring managers found that 39% of them listed "job-hopping" as the biggest obstacle for job seekers. <http://www.askamanager.org/2012/10/job-hopping-is-killing-your-career.html>)
Because of that, I think it's worth listing the temporary jobs as such on your resume, for example by putting "(6-Month Contract)" after the dates of employment. This can prevent someone who's worked a lot of contract jobs from being viewed as someone who gets bored after a few months and moves on. |
4,423 | Is there a downside about taking too many short-term contract positions? I'm still early in my career (I'm 25) but I've only been in a perm full-time job once for 1.5 years. My other 2 jobs have been been 6 month contracts. My current job is a 6-month contract-to-hire position which has been ok but I'd like to move elsewhere when the contract finishes. Are all these short-term jobs seen as a black mark?
I've read that it's recommended to stay at a permanent full-time job for a year but don't contractor jobs have different guidelines? Isn't it common for contractors to jump around jobs like this? | 2012/10/08 | [
"https://workplace.stackexchange.com/questions/4423",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/3467/"
] | I wouldn't view you as a job hopper and disqualify you on that basis, but, to be honest, I would have difficulty considering you qualified for a senior position either. This is because you haven't had to live with the results of what you created. Further, many people I have interviewed and worked with as contractors with this type of experience never develop the depth to become senior developers, they are repeating the same level of experience over and over. Now if each contract takes you a little further in depth on something and you are clearly developing significant expertise in an area, then it would not be a deal breaker particularly if you are a specialist of some kind. But the interview would be quite pointed towards how you have grown and learned from your mistakes. | It depends.
Earlier in my career I used to take a skeptical view of people who never stayed in the same gig for more than a few months; my concern was perhaps they couldn't. Later in my career, I started taking a skeptical view of people who never left their jobs; my concern was perhaps they couldn't. Since I've interviewed tons of candidates (and have interviewed for jobs at a couple of companies where many people "never" leave), I can find plenty of supporting evidence for both positions, but realistically, it's actually very hard to predict how people will react.
In software, however, many people are quite used to people doing a string of contract gigs. If you aren't working on a 1099 basis, but through agencies, you may want to be more explicit about the contract nature (FooCorp via BarAgency) so that it doesn't raise any hackles by people who might just assume you're quitting whenever your code starts to have consequences. If you're working on a 1099 basis, I recommend using your company as the employer, with "Engaged at XCorp on Project Y. Delivered Q solution to..." as bullet points instead of showing each project as a different "job".
The biggest risk you might have as a contractor is that people will worry about candidates who have never been working on a project long enough to have to think about maintenance considerations, because they won't have suffered enough from their own previous design decisions.
My current contract gig has actually lasted longer than my last two full time arrangements; I left those jobs because it was clear that either the company was about to implode or that I would stop learning things that would allow me to progress professionally. If anyone asks, I just answer honestly and tactfully. They rarely ask.
Your biggest risk is the pre-interview filtering that may happen. You won't know if you're being tossed out of the pile of resumes because of your career history, but you also won't know if it's because you share the last name of the hiring manager's psychotic ex-girlfriend or because they don't want generalists, or don't want specialists.
The best thing you can do for your career prospects is to build a reputation of acting professionally, working smart, and solving problems. The tech job market is small enough that you're likely to work with someone you've worked with before, and as long as the wake you leave resulted in mostly positive impressions, you'll be able to find future gigs without much of a struggle. |
34,833,010 | My font has an empty "interior". As show below:
Font with white background:
[](https://i.stack.imgur.com/q83HW.png)
Font with blue background:
[](https://i.stack.imgur.com/DGm9J.png)
How can we make the "interior" of the text always be white, without using a background? I'll be using the Unity game engine.
And of course, I can't edit the .ttf file. As @itchee pointed out: A TTF font is always in one color only. It can only specify opaque and transparent parts (technically, only the opaque parts are specified and everything else is transparent). The application that displays that font can chose one single color for the opaque part. But it cannot make "some parts black and some parts white". My question is how to achieve what I want **using the Unity game engine**. Any C# scripts are welcome.
As @fafase said, a filling algorithm, for each character, looking for surrounded zone and filling them would be ideal. | 2016/01/16 | [
"https://Stackoverflow.com/questions/34833010",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/-1/"
] | You could use bitmap fonts.
They aren't supported natively by unity, so you can use an asset from the asset store called [CJFinc: Bitmap Font Tools](https://www.assetstore.unity3d.com/en/#!/content/27216). If you aren't willing to pay that much I found a nice tutorial [here](http://www.benoitfreslon.com/unity-generate-and-import-a-bitmap-font-with-a-free-tool).
But so far, my most flexible idea, and probably what I would do is to create a second font that you can place behind your first font. You would keep your original font at a black color. The second font would be simple shapes that fit the shape of your fill, and that you set to whatever color you want. Here's a quick sketch of what I mean:
[](https://i.stack.imgur.com/oe9dE.png)
You can use [fontforge](https://fontforge.github.io/en-US/) to create your custom fonts. It might take you quite a bit of time, but It will then work consistently, simply, and without many changes to your scripts. | While it is possible to edit TTF fonts, you cannot make the "background" white.
TTF fonts can only specify opaque and transparent parts (technically, only the opaque parts are specified and everything else is transparent). The application that displays that font can chose one single color for the opaque part. But it cannot make "some parts black and some parts white". |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | Plants give off CO2 at night, when they start to respirate using the 'fuel' (glucose) synthesised by photosynthesis in the day. Check out the [Calvin-Benson cycle](https://en.wikipedia.org/wiki/Light-independent_reactions#Coupling_to_other_metabolic_pathways) | You might also look into *photorespiration*, an error in photosynthesis which causes a plant to produce CO2 instead of O2. This only occurs when CO2 supply to the plants is low, which might happen when the stomata are closed to prevent water loss. Research suggests that photorespiration might have to do with nitrogen assimilation as well, meaning a plant in a nitrogen-poor environment might be more likely to photorespire. |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | Plants give off CO2 at night, when they start to respirate using the 'fuel' (glucose) synthesised by photosynthesis in the day. Check out the [Calvin-Benson cycle](https://en.wikipedia.org/wiki/Light-independent_reactions#Coupling_to_other_metabolic_pathways) | As has been pointed out, plants can give off CO2 at night. If you pick a [plant that only blooms at night](http://themysteriousworld.com/10-most-beautiful-night-blooming-flowers/), you'd have a plant that is using up much more energy at night, thus producing more CO2, and the scent of the flower could add to the insect repellent function. My mother used to grow these orchids in our home (back in my home-country) that would only bloom once every few weeks, and only at midnight and the bloom would be wilted by morning. They were so beautiful, when my mom knew they would be blooming, we'd stay up to see them. |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | All living things respire, and when they respire they break down 'fuel' to produce energy, which gives off CO2. A plant makes its source of fuel through collecting water, CO2 and light, the by-products of which are oxygen (output/waste) and glucose (the fuel). At night the plants can't collect light and instead they increase respiration, creating the energy needed for the night and the rest of the next day (stored in neat little packages called ATP).
Because of this, it seems you have no problem. Since you use the plants at night, they will naturally be giving off CO2, you only need to make a plant that respires heavily at night.
Of course, unless you want to hand-wave, you'll need a reason for this. Perhaps the simplest solution would be to have a fast-growing plant, maybe herbaceous. The rapid growth demands more energy, which means more CO2 intake during the day and therefore more CO2 output during the night. | You might also look into *photorespiration*, an error in photosynthesis which causes a plant to produce CO2 instead of O2. This only occurs when CO2 supply to the plants is low, which might happen when the stomata are closed to prevent water loss. Research suggests that photorespiration might have to do with nitrogen assimilation as well, meaning a plant in a nitrogen-poor environment might be more likely to photorespire. |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | All living things respire, and when they respire they break down 'fuel' to produce energy, which gives off CO2. A plant makes its source of fuel through collecting water, CO2 and light, the by-products of which are oxygen (output/waste) and glucose (the fuel). At night the plants can't collect light and instead they increase respiration, creating the energy needed for the night and the rest of the next day (stored in neat little packages called ATP).
Because of this, it seems you have no problem. Since you use the plants at night, they will naturally be giving off CO2, you only need to make a plant that respires heavily at night.
Of course, unless you want to hand-wave, you'll need a reason for this. Perhaps the simplest solution would be to have a fast-growing plant, maybe herbaceous. The rapid growth demands more energy, which means more CO2 intake during the day and therefore more CO2 output during the night. | Follow the side effect. When things evolve, it's because some random side effect proved increasingly beneficial. If this plant of yours found these deadly mosquitoes to be a good source of food, it could have evolved to attract and "fumigate" the mosquitoes, using their dead carcasses as an alternative source of energy, making the plant less dependent on photosynthesis, causing the plant to give off increased levels of CO2, which further served to attract more mosquitoes... |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | Plants give off CO2 at night, when they start to respirate using the 'fuel' (glucose) synthesised by photosynthesis in the day. Check out the [Calvin-Benson cycle](https://en.wikipedia.org/wiki/Light-independent_reactions#Coupling_to_other_metabolic_pathways) | All living things respire, and when they respire they break down 'fuel' to produce energy, which gives off CO2. A plant makes its source of fuel through collecting water, CO2 and light, the by-products of which are oxygen (output/waste) and glucose (the fuel). At night the plants can't collect light and instead they increase respiration, creating the energy needed for the night and the rest of the next day (stored in neat little packages called ATP).
Because of this, it seems you have no problem. Since you use the plants at night, they will naturally be giving off CO2, you only need to make a plant that respires heavily at night.
Of course, unless you want to hand-wave, you'll need a reason for this. Perhaps the simplest solution would be to have a fast-growing plant, maybe herbaceous. The rapid growth demands more energy, which means more CO2 intake during the day and therefore more CO2 output during the night. |
39,639 | In my alternate history (1800s), I have troop hiking through tropical sub-Saharan Africa. For plot reasons, I have man-killing mosquitos, but I also have a plant that exudes CO2 and other attractants. These flowering plants are placed in the room at night to attract the mosquitos away from people.
Evolution-wise, however, 'consuming' CO2 and giving off O2 is so deeply ingrained in a plant's DNA, it is hardly believable to be the other way around.
I'd easily hand-wave it away as some special scent, but the CO2 is critical to the plot. **Can a plant evolve to give off CO2?** Or rather, how difficult would that be to explain. | 2016/04/07 | [
"https://worldbuilding.stackexchange.com/questions/39639",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/3276/"
] | As has been pointed out, plants can give off CO2 at night. If you pick a [plant that only blooms at night](http://themysteriousworld.com/10-most-beautiful-night-blooming-flowers/), you'd have a plant that is using up much more energy at night, thus producing more CO2, and the scent of the flower could add to the insect repellent function. My mother used to grow these orchids in our home (back in my home-country) that would only bloom once every few weeks, and only at midnight and the bloom would be wilted by morning. They were so beautiful, when my mom knew they would be blooming, we'd stay up to see them. | You might also look into *photorespiration*, an error in photosynthesis which causes a plant to produce CO2 instead of O2. This only occurs when CO2 supply to the plants is low, which might happen when the stomata are closed to prevent water loss. Research suggests that photorespiration might have to do with nitrogen assimilation as well, meaning a plant in a nitrogen-poor environment might be more likely to photorespire. |
274,404 | How can I know in a C#-Application, in which direction the screen of the mobile device is orientated? (i.e. horizontal or vertical). | 2008/11/08 | [
"https://Stackoverflow.com/questions/274404",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/26070/"
] | In Microsoft.WindowsMobile.Status there is a class which keeps track of all kinds of properties of your device.
Besides the one you need, DisplayRotation, it also contains properties about phone coverage, Nr of missed calls, next appointment and many more. See [msdn](http://msdn.microsoft.com/en-us/library/microsoft.windowsmobile.status.systemproperty.aspx) for more info.
You can also add an event-handler to be notified of changes of these properties. | Add a reference to Microsoft.WindowsCE.Forms to your project. Then you can reference the
Microsoft.WindowsCE.Forms.**SystemSettings.ScreenOrientation** property, which will give you what you need.
Incidentally, you can set this property, so it can be used to set your screen orientation also. |
114,795 | I have a warlock PC, who has left the nature of their pact fuzzy, and I'm going to fill in the gaps to help drive the story along as necessary.
How powerful does it make sense for the fiend to be? Could it be any devil, or does it need be one near the top of the tree (so it can potentially power a 20th level Warlock)? I can see the potential humour in the patron being an imp with delusions of grandeur, but this could cause some headaches with the Warlock being vastly more powerful than their patron.
I'm also considering having a devil antagonist (Rakshasa) to the party buying out the Warlock's contract, so I want to gauge if this would necessarily be a step up/down.
I understand the rules are vague on this, so I'm mainly looking for what would narratively make sense. | 2018/02/03 | [
"https://rpg.stackexchange.com/questions/114795",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/42170/"
] | It needs to be one near the top
-------------------------------
From the PHB p.109:
>
> Fiends powerful enough to
> forge a pact include demon lords such as Demogorgon,
> Orcus, Fraz’Urb-luu, and Baphomet; archdevils such
> as Asmodeus, Dispater, Mephistopheles, and Belial;
> pit fiends and balors that are especially mighty; and
> ultroloths and other lords of the yugoloths.
>
>
>
Your world, your rules
----------------------
If you don't like that, change it. If you think it would be fun for the pact to be held by an imp or rakshasa then do it. Just think through the implications - a warlock without a patron isn't a warlock - I can see a lot of stress on a warlock with a weak patron trying to keep it alive.
Notwithstanding, your warlock is not important enough to deal with the boss
---------------------------------------------------------------------------
I have a contract with Microsoft - when Microsoft deals with me I don't talk to the CEO. I can certainly see devils running hell as a multi-planar corporation. Sure your contract is with Asmodeus but *your* case manager is Roger, imp 3rd class who has Thursdays and Saturdays off and will get back to you within 10 business days after you lodge the correct paperwork. You can, of course, call customer support to hear "We are currently dealing with other warlocks. Your pact is important to us so please hold the line. Current wait time is approximately 342 years. Thank you."
Demons of course do things far more chaotically. Yes, your pact is with Fraz’Urb-luu whose filing system consists of dropping things wherever he happens to be and then moving on. The guys *you* deal with are never consistent depending on who happened to take an interest at the time, assuming, of course, that *someone* bothered. You never know if any given demon who *says* they are working for Fraz’Urb-luu actually is or isn't or if Fraz’Urb-luu wants them to do whatever they want you to do or not. You wanted a quiet life? Don't go making deals with the most chaotic and evil beings in the multiverse.
Lords of the yugoloths are more likely to give the personal touch. For them, corrupting mortals to evil is neither a bureaucratic process nor something to do when they feel like it. Each corrupted mortal is a work of art to be admired for all eternity and they constantly strive for perfection in that art. Fashions change of course, this millennia the fashion is to corrupt as quickly as possible, a soul corrupted in a day is better than one that takes a week. Next millennia, the slow burn is admired, slowly stripping away everything and everyone the mortal cares about until the sink, at the end of their life, into the depths of evil despair - more satisfying than a cold beer on a hot day.
I can also see a secondary market in warlock pacts (run by devils, of course) where warlock pacts, and options and derivatives of them, are traded back and forth. Perhaps angels have to come to buy out contracts of warlocks who died in a state of grace? | Extraordinarily powerful
------------------------
From Pact of the Fiend in the PHB pg. 109:
>
> Fiends powerful enough to forge a pact include demon lords such as Demogorgon, Orcus, Fraz’Urb-luu, and Baphomet; archdevils such as Asmodeus, Dispater, Mephistopheles, and Belial; pit fiends and balors that are especially mighty; and ultroloths and other lords of the yugoloths.
>
>
> |
459,536 | I've changed my mind a number of times about which of these variants works best:
1. >
> the **second most northerly** coffee shop in Seattle
>
>
>
2. >
> the **second-most-northerly** coffee shop in Seattle
>
>
>
3. >
> the **second-most northerly** coffee shop in Seattle
>
>
>
4. >
> the **second most-northerly** coffee shop in Seattle
>
>
>
I've tried to do a little research, but I'm not even sure what to call something like this, so I haven't had much luck.
Any guidance on what is considered most clear/correct would be much appreciated! | 2018/08/09 | [
"https://english.stackexchange.com/questions/459536",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/4675/"
] | A hyphen should be used to force the parsing "(second most) northerly" rather than "second (most northerly)". Thus, it should be "second-most northerly". | Surely the primary variable here is the 'northerliness' of the coffee shops' locations. Consider the following:
>
> '***The Coffee Cup***' is a coffee shop in Seattle.
>
>
> '***The Barista Bar***' is another coffee shop in Seattle.
> It is further North than '*The Coffee Cup*'.
>
>
> '***Luscious Lattes***' is another coffee shop in Seattle.
> It is further North than '*The Barista Bar*'.
>
>
>
The simplest way to make clear their relative northerliness is to say:
>
> *Luscious Lattes* is the most northerly coffee shop in Seattle.
>
>
> *The Barista Bar* is the second most-northerly coffee shop in Seattle.
>
>
> *The Coffee Cup* is the third most-northerly coffee shop in Seattle.
>
>
> *etc*.
>
>
> |
459,536 | I've changed my mind a number of times about which of these variants works best:
1. >
> the **second most northerly** coffee shop in Seattle
>
>
>
2. >
> the **second-most-northerly** coffee shop in Seattle
>
>
>
3. >
> the **second-most northerly** coffee shop in Seattle
>
>
>
4. >
> the **second most-northerly** coffee shop in Seattle
>
>
>
I've tried to do a little research, but I'm not even sure what to call something like this, so I haven't had much luck.
Any guidance on what is considered most clear/correct would be much appreciated! | 2018/08/09 | [
"https://english.stackexchange.com/questions/459536",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/4675/"
] | I'd be inclined to write this as "the second most northerly coffee shop in Seattle", with no hyphens. But hyphenation is far from a definite area of English punctuation. I don't think any of the options that you list is unclear or ambiguous in practice.
Google Ngram Viewer data
========================
While I'm not entirely sure of the accuracy of the [Google Ngram Viewer](https://books.google.com/ngrams/graph?content=the+second+most+common%2Cthe+second-most+common%2Cthe+second-most-common%2C+the+second+most-common&year_start=1800&year_end=2000&corpus=15&smoothing=3&share=&direct_url=t1%3B%2Cthe%20second%20most%20common%3B%2Cc0%3B.t1%3B%2Cthe%20second%20-%20most%20common%3B%2Cc0%3B.t1%3B%2Cthe%20second%20most%20-%20common%3B%2Cc0) with hyphenated phrases, it seems to suggest that "the second most common" is much more frequent than "the second-most common":
[](https://i.stack.imgur.com/wba4F.png)
(The double-hyphenated spelling "the second-most-common" was so infrequent that it did not show up at all on the Ngram Viewer chart.)
CMOS rules that seem applicable
===============================
The Chicago Manual of Style (17th ed) gives somewhat complicated rules for hyphenating phrases like this.
It recommends using a hyphen between a ordinal number and a superlative
-----------------------------------------------------------------------
In the original version of this answer, I missed the fact that there is actually a special section in the CMOS hyphenation table for "number, ordinal, + superlative". This says to use a hyphen after words like "second" when they come before an attributive superlative adjective. The examples are
>
> a *second-best* decision
>
> *third-largest* town
>
> *fourth-to-last* contestant
>
> he arrived *fourth to last*
>
>
>
Unfortunately, none of these examples show how to hyphenate a phrase that uses *most* rather than *-est* to form a superlative adjective. The example of *fourth-to-last* shows hyphenation between all words in the phrase, not just after the word *fourth.*
The CMOS's recommendation here seems a bit unusual to me, actually. A number of dictionaries mention this use of *second* and give examples without hyphenation:
* OED *second, adj. and n.2* 2b : "With following superlative: Having only one superior in the specified attribute."
>
> 1979 *Nature* 15 Feb. 561/2 *Secernosaurus* is the second most primitive hadrosaur known.
>
>
>
(the OED entry does give one example quotation that uses hyphenation, "1977 *Word* **28** 104 The second-youngest of the fluent speakers.")
* [MW](https://www.merriam-webster.com/dictionary/second) *3second* adverb 2 : "before all others with one exception · the nation's *second* largest city · They are my *second* favorite band."
* [AHD](https://www.ahdictionary.com/word/search.html?q=second)
*second2* adv. 2 : "But for one other; save one: the *second highest peak*."
So I don't think it's universally regarded as unacceptable to use a space rather than a hyphen after *second* in this context. The preference for a hyphen between *second* and an immediately following attributive superlative adjective seems to be something that is particular to the *Chicago Manual of Style* (I'm not sure what other style guides say about this topic).
But CMOS also recommends not using a hyphen with *most,* or with an adverb that is modified by another adverb
-------------------------------------------------------------------------------------------------------------
CMOS 17 says
>
> certain compounds, including those with *more, most, less, least,* and *very,* can usually be left open unless ambiguity threatens.
>
>
>
For example, it recommends writing "the *most efficient* method" without a hyphen.
What it means by "unless ambiguity threatens" is indicated with the following pair of examples:
>
> the *most skilled workers* (most in number) but
>
> the *most-skilled* workers (most in skill)
>
>
>
So my interpretation is that, according to CMOS 17, it might be acceptable to write something like "the most-northerly coffee shops" to avoid the alleged ambiguity of "the most northerly coffee shops" (e.g. a sentence like "This street has the most northerly coffee shops in Seattle" is theoretically ambiguous, although not really in practice because "northerly" isn't an adjective that typically can apply to a bunch of different shops).
I can't find support in CMOS for using a hyphen before but not after *most.* It says
>
> When the adverb rather than the compound as a whole is modified by another adverb, the entire expression is open.
>
>
>
The relevant pair of examples is "a *much-needed* addition" vs. "a *very much needed* addition". (It's a bit unfortunate that this uses the adverb *very*). Based on this, it seems to me that if *second* is viewed as an adverb modifying *most,* then you should write the whole thing with spaces according to CMOS.
Now, it's not clear to me that this is the actual function of *second* in this phrase. The [AHD](https://www.ahdictionary.com/word/search.html?q=second) does have an entry for *second* as an adverb, but the OED does not, and lists this use of *second* under the entry for the adjective.
The fact that we can say things like "the second highest mountain", where the equivalent of the word *most* is expressed by a suffix *-est* on the adjective, suggests to me that *second* in this context is not used as a modifier of the adverb *most* (since that adverb is not even present in "the second highest mountain"). On the other hand, I guess some people can say things like "the most and second most important reason", which might indicate that "second most" is treated as a phrase (I'm not sure).
Two hyphens: recommended by "Daily Writing Tips"
================================================
Although the use of two hyphens seems to be uncommon, as I mentioned in the Google Ngram Viewer section, it has been recommended by at least one person. The Daily Writing Tips article "[5 Examples of Insufficient Hyphenation](https://www.dailywritingtips.com/5-examples-of-insufficient-hyphenation/)", by Mark Nichol, criticizes the punctuation of “The adviser some call the world’s second-most powerful man prefers to work behind the scenes,” saying
>
> *Second-most* is a nonsensical modification of “powerful man.” *Powerful* is part of the ranking, so it should be part of the phrasal adjective: “The adviser some call the world’s second-most-powerful man prefers to work behind the scenes.”
>
>
> | Surely the primary variable here is the 'northerliness' of the coffee shops' locations. Consider the following:
>
> '***The Coffee Cup***' is a coffee shop in Seattle.
>
>
> '***The Barista Bar***' is another coffee shop in Seattle.
> It is further North than '*The Coffee Cup*'.
>
>
> '***Luscious Lattes***' is another coffee shop in Seattle.
> It is further North than '*The Barista Bar*'.
>
>
>
The simplest way to make clear their relative northerliness is to say:
>
> *Luscious Lattes* is the most northerly coffee shop in Seattle.
>
>
> *The Barista Bar* is the second most-northerly coffee shop in Seattle.
>
>
> *The Coffee Cup* is the third most-northerly coffee shop in Seattle.
>
>
> *etc*.
>
>
> |
23,673,076 | I have a strange problem. My ios game integrated with Facebook. I invited to my game a friend, who added to testers in Facebook app, but my friend did not receive invitation. But when he logged in Facebook on his mobile phone, he received invitation immediately. My friend opened browser and tried to find invitation in Notifications but there is no any one.
Can somebody explain me why it happened? | 2014/05/15 | [
"https://Stackoverflow.com/questions/23673076",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/987632/"
] | You have to go to developer.facebook account of yours and make some changes.
Steps:-
1.Select your app from apps.
2.Go to setting option
3.Press Add Platform and add App On Facebook
When Same problem i have occur its solve by this | If your game only supports mobile then your friends will not get notification on their browsers |
39,547 | There are a variety of other questions that address elements of this security issue, but none that seem (to a non-expert) to address the substantial core of the issues raised in this article:
* [Tinder's privacy breach lasted much longer than the company claimed](http://qz.com/107739/tinders-privacy-breach-lasted-much-longer-than-the-company-claimed/)
From a mobile perspective, what are the correct steps the company should have taken to prevent, other than using a CA-issued SSL certificate? | 2013/07/25 | [
"https://security.stackexchange.com/questions/39547",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/19951/"
] | **General SSL-related answer:** To do SSL (HTTPS) securely, the really important point is that the *client system* (smartphone, PC... it does not matter) can make sure that the *server public key* it is using for the SSL handshake is really the genuine one from the intended server. CA-issued certificates are about that, really. Without a validation of the server public key of some kind, [MitM attacks](http://en.wikipedia.org/wiki/Man-in-the-middle_attack) are possible. Otherwise, they are not (unless the client and/or server implementation goofs up, but that's another matter).
With certificates, the MitM attack can work only if the attacker can get a fake certificate, with the genuine server name but the attacker's public key in it, from a CA that the client trusts. Incidents with fake certificates being issued are rare (we hear about one such incident per year). Much more often, a gullible user decides to disregard the scary warning from his browser, and clicks through it. If you pay heed to your browser warnings, you should be fine.
Unfortunately, a lot of SSL connections are not between a *Web browser* and a server, but between an *application* and a server, and it is up to the application to not do the stupid click-through-warning (metaphorically). There has been a lot of reports of application who do not check the server certificate properly, or at all. This is app-specific and, as a user, this is hard to check.
---
**About Tinder:** none of the above applies to the issues described in the article you link to. Apparently, Tinder runs a dating service, with customer running a specific app on their phone; the app talks to a central server. The central server then computes "possible encounters", i.e. warns customers about the nearby presence of other customers with whom romantic compatibility is heuristically estimated to be above average.
It so happens that the dialog between the app and the server involves the server telling the app a lot about other customers -- actually a lot more than should be needed for the described service, including leaking Facebook IDs and other similar information. The app user interface won't show it, but a custom application could obtain such data from the server and record it.
SSL would have done nothing for or against this privacy issue. Maybe they *are* using SSL; it does not matter. The problem is that the server leaks too much private information to whoever asks for it. A MitM attack is relevant only when the attacker tries to intercept a data transmission containing data that an authorized client may legitimately obtain, but attackers should not; here, the client should not be able to obtain the information in the first place, and *that* is the problem the article talks about. | The article you mention has nothing at all to do with HTTPS. If an HTTPS connection is used, regardless of desktop or mobile, then the connection is protected between the client system and the server that holds the private key for the certificate being used.
In this case, the API appears to have been operating over a non-encrypted link. Simply encrypting the link and/or not including unnecessary personal information in the API requests would have been sufficient. |
18,901,451 | I'm upgrading a website from Sitecore 6.5 to Sitecore 7. The upgrade has been smooth and I was able to get everything up and running except for WFFM.
After installing WFFM 2.3 latest, I keep getting the error
>
> Could not load file or assembly 'HtmlAgilityPack, Version=1.4.0.0,
> Culture=neutral, PublicKeyToken=bd319b19eaf3b43a' or one of its
> dependencies. The located assembly's manifest definition does not
> match the assembly reference. (Exception from HRESULT: 0x80131040)
>
>
>
Sitecore 7 upgrades it's 3rd party dll for HtmlAgilityPack to 1.4.3, but WFFM seems to still reference 1.4.0
Has anyone dealt with this? Am I missing something?
Thanks for the help. | 2013/09/19 | [
"https://Stackoverflow.com/questions/18901451",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2796525/"
] | This is a known bug for Sitecore. Check out [this link](http://sdn.sitecore.net/Products/Web%20Forms%20for%20Marketers/Web%20Forms%20for%20Marketers%202,-d-,3/Release%20Notes/Known%20issues.aspx) for a fix (382886). | I've run into HtmlAgilityPack version mismatches during previous Sitecore upgrades when WFFM was installed. I've just been able to update the reference by hand to the new file without suffering any consequences.
It seems that a separate team works on WFFM and the work is lagged several months behind the Sitecore core. |
332,788 | >
> I told you never to go nowhere without me!
>
>
>
According to Grammarly, the above sentence contains a double negative and should be modified to something like, "I told you never to go anywhere without me!"
Doesn't the original sentence mean the person addressed as "you" in the sentence must stay here? | 2023/02/14 | [
"https://ell.stackexchange.com/questions/332788",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/168464/"
] | The sentence contains a double-negative, (or perhaps actually a triple negative) and this means it is unclear.
It would be taken to mean "I told you never to go *anywhere* without me." This does mean "stay here". This is because the "logical" removal of the double negative (don't go nowhere -> do go somewhere) doesn't work here.
If you try to apply "logic" you reach something like "Always go somewhere without me." Which is clearly nonsense.
So you would assume a dialect in which the double negative is used and the meaning is "I told you to stay here until I come back."
Learners should **avoid** this kind of double-negative, because it is unclear and considered to be a mistake in careful English. | In lots of non-standard English dialects the double negative is fine.
Not in standard English, though: in standard English, Grammarly is correct.
Double negatives are possible in standard English, but they often add up to an affirmative. They're sometimes useful in conversation, often to contradict negative assertions by one's conversation partner.
Grumpy teacher to student: "You're going nowhere, Jimmy."
Student: "I'm *not* going *nowhere*; I'm just going nowhere you'd like me to go."
They can also be seen as two single negatives, each in a separate clause:
"She did*n't* say you could*n't* go; she only said you couldn't bring your laptop with you."
I'm not sure I understand the last sentence in your question. |
245,584 | "Each time he made the turn, he saw only the white, unused, buckling sidewalk, with perhaps, on one night, something vanishing swiftly across a lawn"... p.11 Fahreheit 451
Does it mean that the sidewalk is bent (upward), as when a tree grows next to a sidewalk and distorts it to make room for its roots? But then again, it seems paradoxical that it's buckling but also unused... Any thoughts? | 2020/04/26 | [
"https://ell.stackexchange.com/questions/245584",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/112428/"
] | “Going on vacation” refers to the change of state from “not on vacation to “on vacation”, similar to a light bulb “going out” when we flip its switch.
This usage may have originated from the idea that people do often go *away* on their vacations, but it is not limited to that meaning today. | "On vacation" simply means "not at work" or "not in school", etc. - you're taking a break from your usual duties. If someone is on vacation now, it simply means they're taking time off. Some people travel for their vacations, but not everyone does - in fact, there's a colloquial term "staycation" for those who *stay* home rather than go somewhere else.
I agree with Kate Bunting that "going on vacation" may imply traveling, but in the end it really is a question for Susan ;) |
245,584 | "Each time he made the turn, he saw only the white, unused, buckling sidewalk, with perhaps, on one night, something vanishing swiftly across a lawn"... p.11 Fahreheit 451
Does it mean that the sidewalk is bent (upward), as when a tree grows next to a sidewalk and distorts it to make room for its roots? But then again, it seems paradoxical that it's buckling but also unused... Any thoughts? | 2020/04/26 | [
"https://ell.stackexchange.com/questions/245584",
"https://ell.stackexchange.com",
"https://ell.stackexchange.com/users/112428/"
] | "On vacation" is an *adjective phrase* describing Susan's current state. "Go on vacation" is a *verb phrase*. They're different parts of speech, which makes it difficult to compare them in any meaningful way.
Perhaps some similar example pairs will make it clearer:
"*fall asleep*" and "*asleep*"
"*get on a train*" and "*on a train*"
"*sit in a chair*" and "*in a chair*" | "On vacation" simply means "not at work" or "not in school", etc. - you're taking a break from your usual duties. If someone is on vacation now, it simply means they're taking time off. Some people travel for their vacations, but not everyone does - in fact, there's a colloquial term "staycation" for those who *stay* home rather than go somewhere else.
I agree with Kate Bunting that "going on vacation" may imply traveling, but in the end it really is a question for Susan ;) |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | It sounds like for this app, you could use [Microsoft Dynamic Data](http://www.asp.net/dynamicdata/) or [Castle Active Record](http://www.castleproject.org/activerecord/index.html), and have the application working a few minutes after you finished the database. These tools connect to a database and generate forms for inputing data. Take a look at them.
Access is probably your best choice for database. MS Sql 2005/2008 Express would also work well, but that would require an install. | SqlExpress would be your best bet, simply because you already have all the support that you need in System.Data.SqlClient. Other then that, there are some decent help on the MSDN. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | It sounds like for this app, you could use [Microsoft Dynamic Data](http://www.asp.net/dynamicdata/) or [Castle Active Record](http://www.castleproject.org/activerecord/index.html), and have the application working a few minutes after you finished the database. These tools connect to a database and generate forms for inputing data. Take a look at them.
Access is probably your best choice for database. MS Sql 2005/2008 Express would also work well, but that would require an install. | If you're programming in C#, Visual Studio comes with an added install for SQL Server Express. If you're looking to get something up quick and dirty, it would pretty easy to leverage that database in building your app. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | It sounds like for this app, you could use [Microsoft Dynamic Data](http://www.asp.net/dynamicdata/) or [Castle Active Record](http://www.castleproject.org/activerecord/index.html), and have the application working a few minutes after you finished the database. These tools connect to a database and generate forms for inputing data. Take a look at them.
Access is probably your best choice for database. MS Sql 2005/2008 Express would also work well, but that would require an install. | How about using SQLlite instead of access for db? I have never used it but have heard that its good for some light weight and quick db tasks. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | If you're programming in C#, Visual Studio comes with an added install for SQL Server Express. If you're looking to get something up quick and dirty, it would pretty easy to leverage that database in building your app. | How about using SQLlite instead of access for db? I have never used it but have heard that its good for some light weight and quick db tasks. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | If concurrency is *not* an issue, then I'd say go with SQLite. An ADO.NET provider can be found here: <http://sourceforge.net/projects/sqlite-dotnet2/files>. As for the concurrency issue, when performing a bunch of inserts, that operation needs to be enclosed within a transaction. However, that transaction places an exclusive lock on the database until the transaction is either commited or rolled back. At least that's been my experience, but I've only been using it for a week or two myself.
Hope that helps! | If you already have MS Access installed, then yes the mdb is probably your quickest way to get started.
Also you will want to just started with a quick ADO.NET tutorial. There are hundreds of these (well almost hundreds, I haven't counted). |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | If you're programming in C#, Visual Studio comes with an added install for SQL Server Express. If you're looking to get something up quick and dirty, it would pretty easy to leverage that database in building your app. | SqlExpress would be your best bet, simply because you already have all the support that you need in System.Data.SqlClient. Other then that, there are some decent help on the MSDN. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | It sounds like for this app, you could use [Microsoft Dynamic Data](http://www.asp.net/dynamicdata/) or [Castle Active Record](http://www.castleproject.org/activerecord/index.html), and have the application working a few minutes after you finished the database. These tools connect to a database and generate forms for inputing data. Take a look at them.
Access is probably your best choice for database. MS Sql 2005/2008 Express would also work well, but that would require an install. | I would suggest using [SubSonic](http://subsonicproject.com/) to generate your data access code and scaffold your screens. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | If concurrency is *not* an issue, then I'd say go with SQLite. An ADO.NET provider can be found here: <http://sourceforge.net/projects/sqlite-dotnet2/files>. As for the concurrency issue, when performing a bunch of inserts, that operation needs to be enclosed within a transaction. However, that transaction places an exclusive lock on the database until the transaction is either commited or rolled back. At least that's been my experience, but I've only been using it for a week or two myself.
Hope that helps! | SqlExpress would be your best bet, simply because you already have all the support that you need in System.Data.SqlClient. Other then that, there are some decent help on the MSDN. |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | It sounds like for this app, you could use [Microsoft Dynamic Data](http://www.asp.net/dynamicdata/) or [Castle Active Record](http://www.castleproject.org/activerecord/index.html), and have the application working a few minutes after you finished the database. These tools connect to a database and generate forms for inputing data. Take a look at them.
Access is probably your best choice for database. MS Sql 2005/2008 Express would also work well, but that would require an install. | If concurrency is *not* an issue, then I'd say go with SQLite. An ADO.NET provider can be found here: <http://sourceforge.net/projects/sqlite-dotnet2/files>. As for the concurrency issue, when performing a bunch of inserts, that operation needs to be enclosed within a transaction. However, that transaction places an exclusive lock on the database until the transaction is either commited or rolled back. At least that's been my experience, but I've only been using it for a week or two myself.
Hope that helps! |
76,206 | Okay, I've seen but haven't programmed in C# before. You can assume I'm competent in C++, and advanced in C (for what good that'll do me). I understand inheritance, polymorphism, etc so OO concepts aren't going to be a huge problem.
Let's say I've been given a task to prototype a quick and dirty program that won't be much different than what I could do in access in a short time.
* It'll have a DB with 5-6 tables (mostly small, a few with have several thousand rows but only 4 or so columns, etc)
* I'll need to have forms generated dynamically from one of the DBs
* The results of the forms will be stored in another table
* The DB isn't multiuser
Basically your run of the mill access app... except without access. I'm sure I can muddle my way through and create horrendously bad code, but I'm equally sure lots of people here can give me a push in the right direction (tutorials, wizards, info, differences and killers moving from C/C++ to C#, etc).
Is there a simple DB I can plug in to get started aside from mdb, or is that the best choice for this particular nail? I'm aiming for a quick and dependency-less install.
Thanks!
-Adam | 2008/09/16 | [
"https://Stackoverflow.com/questions/76206",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2915/"
] | If concurrency is *not* an issue, then I'd say go with SQLite. An ADO.NET provider can be found here: <http://sourceforge.net/projects/sqlite-dotnet2/files>. As for the concurrency issue, when performing a bunch of inserts, that operation needs to be enclosed within a transaction. However, that transaction places an exclusive lock on the database until the transaction is either commited or rolled back. At least that's been my experience, but I've only been using it for a week or two myself.
Hope that helps! | How about using SQLlite instead of access for db? I have never used it but have heard that its good for some light weight and quick db tasks. |
34,647 | Im currently researching how to handle lead generation forms.
I came across a website which has 3 slightly different forms for downloading a white paper, downloading an ebook, and downloading the trial.The trial download form is the shortes.
Trial Download fields:
* name
* company
* phone
* email
White paper form fields are the same but have 2 more fields:
* job positon
* country
The ebook form has trials form fields + state field.
I can not wrap my head around this.
Why would they have those differences in forms? Does this strategy have a name?
Also, i want to add on this particular site, if you fill out a form once, the info you already entered is auto filed on the next form you would fill out.
So, if you get a trial first, you would only need to fill out 2 more fields for the white paper.
But who is to say that most people would first get the trial, and later a whitepaper? Does this even make sense? | 2013/02/13 | [
"https://ux.stackexchange.com/questions/34647",
"https://ux.stackexchange.com",
"https://ux.stackexchange.com/users/25350/"
] | Assuming there is a good reason for this (which we will never know):
If they have a **different expected audience** for each product, it may be that their business logic has dictated that they need to know the additional information for whitepapers - for marketing or sales reasons. This actually makes sense if they want to be able to say "349 companies have used our whitepapers to inform them of *X*".
It may also be that each item is handled by a **different department**, and so the information requirements have simply been given to the developers to include.
From a UX perspective, it is best to ask for as little information as possible, but **UX needs have to always be balanced against other business needs**. | This strategy looks like a reverse-common-sense implementation. The most guarded of my personal information is my phone number, followed by email, and then anything else from the list.
If I had to do this, I'd tag the most valuable asset with direct contact information of the person downloading it, not the other way around. |
34,647 | Im currently researching how to handle lead generation forms.
I came across a website which has 3 slightly different forms for downloading a white paper, downloading an ebook, and downloading the trial.The trial download form is the shortes.
Trial Download fields:
* name
* company
* phone
* email
White paper form fields are the same but have 2 more fields:
* job positon
* country
The ebook form has trials form fields + state field.
I can not wrap my head around this.
Why would they have those differences in forms? Does this strategy have a name?
Also, i want to add on this particular site, if you fill out a form once, the info you already entered is auto filed on the next form you would fill out.
So, if you get a trial first, you would only need to fill out 2 more fields for the white paper.
But who is to say that most people would first get the trial, and later a whitepaper? Does this even make sense? | 2013/02/13 | [
"https://ux.stackexchange.com/questions/34647",
"https://ux.stackexchange.com",
"https://ux.stackexchange.com/users/25350/"
] | For anyone interested I found out what this is all about. Turns out this is called **progressive profiling** in marketing and it's a way to ask a prospect lots of questions without the need to stick them all on a single form.
sources:
<http://www.hubspot.com/blog/bid/33993/HubSpot-Forms-Now-Feature-Progressive-Profiling-and-a-New-Interface>
<http://www.4thoughtmarketing.com/Eloqua-Progressive-Profiling-Forms/>
Edit:
Does going through all this just to have 1-2 less fields even make sense?
It seems so:
<http://www.marketingexperiments.com/blog/internet-marketing-strategy/lead-generation-testing-form-field-length-reduces-cost-per-lead-by-10-66.html>
This study suggests 1 less field = 1% more conversions. | This strategy looks like a reverse-common-sense implementation. The most guarded of my personal information is my phone number, followed by email, and then anything else from the list.
If I had to do this, I'd tag the most valuable asset with direct contact information of the person downloading it, not the other way around. |
9,291,071 | I am making a website which involves lot of computing at the back end...I'm using Java Script at the front end and Java for the back end. Could there be any security problems because of Java Script or can I use any better front end instead of Java Script to make website better? | 2012/02/15 | [
"https://Stackoverflow.com/questions/9291071",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1210720/"
] | The biggest potential security problem with Javascript has nothing to do with Javascript (and also applies for "dumb" websites): If you think that the end user will only use your "official" client code, and as a result blindly trust the Javascript code with security-relevant information, then you have a security problem.
* You need to validate all data on the server. The user can make his browser send whatever data he wants, whenever he wants. He does not even have to use a browser (could be a completely hacked-together tool).
* You must not put "secret" data into the client code. Even if it is not directly visible in the browser, the resourceful user can see it. | >
> Could there be any security problems because of JavaScript
>
>
>
Not intrinsically. You could introduce security problems by writing insecure code, but that is try of any language.
>
> can I use any better front end instead of JavaScript to make website better?
>
>
>
Other options for client-side programming require browser plugins (such as Flash) or specific browsers (such as IE for VBScript).
You might not need any client side programming, and any JS you do write should be [progressive](http://en.wikipedia.org/wiki/Progressive_enhancement) and [unobtrusive](http://en.wikipedia.org/wiki/Unobtrusive_JavaScript). |
2,265,586 | I was working on a problem for my job that involved transitioning from a binary vector of length 7 to another binary vector of length 7. The only rule in the transition is that no "1" element in the "from vector" can become a "0" in the "to vector". If I create all possible states and check for valid transitions (like shown here:[http://imgur.com/gcJELC1](https://imgur.com/gcJELC1)) I end up with this (when scaled): [http://imgur.com/BqhqIOj](https://imgur.com/BqhqIOj).
Once I listed all possible binary states I but a 1 if it was an allowed transition, and a 0 if it was not allowed (i colored the 0's red). Here are some examples with smaller vectors:
(0,0,0)->(0,0,1) is allowed
(0,0,0)->(0,0,0) is allowed
(1,0,0)->(1,0,1) is allowed
(1,0,0)->(0,0,1) is not allowed. The first 1 went to a 0, breaking the only rule, so the whole transition is listed in the matrix as a 0 because it is an invalid transition. | 2017/05/04 | [
"https://math.stackexchange.com/questions/2265586",
"https://math.stackexchange.com",
"https://math.stackexchange.com/users/353610/"
] | I think the emergence of the 2D fractal is a property of the ordering of the coordinates, which themselves have a fractal nature in binary notation. See the edges of this figure:
[](https://i.stack.imgur.com/vIYLU.png)
Source code for the figure, a visualisation of $2017$ in the context of [OEIS A134169](https://oeis.org/A134169), is on my [blog post](https://mathr.co.uk/blog/2017-01-08_rollover_2017.html). | It's a cellular automata modelizing a form of Sierpinski triangle.
See rule 60 in (<http://mathworld.wolfram.com/SierpinskiSieve.html>).
See this paper considering it as a biological model (<http://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.0020424>) |
63,351 | I would like to be pointed to information / resources for creating algorithms like the one illustrated on this blog, which is a subdivision of a polygon (in my case a voronoi cell) into several boxes of varying size:
<http://procworld.blogspot.nl/2011/07/city-lots.html>
In the comments a paper by among others the author of the blog can be found, however the only formula listed is about candidate location suitability:
<http://www.groenewegen.de/delft/thesis-final/ProceduralCityLayoutGeneration-Preprint.pdf>
Any language will do, but if examples can be given Javascript is preferred (as it is the language i am currently working with)
A similar question is this one: [What is an efficient packing algorithm for packing rectangles into a polygon?](https://gamedev.stackexchange.com/questions/27055/what-is-an-efficient-packing-algorithm-for-packing-rectangles-into-a-polygon) | 2013/10/10 | [
"https://gamedev.stackexchange.com/questions/63351",
"https://gamedev.stackexchange.com",
"https://gamedev.stackexchange.com/users/35970/"
] | Sorry for the long delay since my comment. I'd been trying to reverse-engineer the ProcWorld example and couldn't work out why the algorithm was making some of the choices it did.
In any case though, here's my best idea so far, using the center block from the ProcWorld example as a guide:

Start by finding an oriented bounding rectangle of the polygon. (For this example it happens to be aligned with the grid, which made these diagrams a bit easier) This rectangle defines the axes for the partition inside (you may want to transform all of your vertices into this new coordinate space to make the math simpler)

Next, split the polygon along one of these axes, from one vertex all the way to the far side. (I'm not sure how this vertex is chosen. Most of the time, it looks like it's the one that gives the longest cut, but not always. It might be chosen to roughly equalize the areas on each side of the cut, or to ensure that the max distance on one side does not exceed a maximum building depth constraint)

For each sub-polygon formed, select another vertex from the original polygon boundary and cut across to the previous cut. (This vertex may be chosen as the one furthest from the previous cut - which I assume here - or again in an area-balancing fashion)
 
Continue until the remaining polygon is a quadrilateral, or the next cut would exceed a maximum-skinniness constraint (eg. the cut shown in green).

Split the remaining polygons into quadrilaterals by cutting parallel to the previous successful cut.
If cutting at a vertex would result in an excessively skinny building, discard it.
 
Finally, we have a set of quadrilaterals. Several will be bounded by cuts on three sides - cap them off at their obtuse angle to make them rectangular. For quads that are bounded by cuts on only two edges, find the largest rectangle that fits inside.

If ever the next proposed cut would go outside the polygon (shouldn't happen with Voronoi cells since they're always convex), generate a new oriented bounding box for that sub-polygon and start cutting its remaining boundary vertices as though you were starting from scratch. (That's the best way I can think of to explain how the alignment axis changes in places like the one pictured - I think it's because the algorithm wanted to connect the bottom-right vertex to the bold blue cut, and couldn't because of the concavity in-between)
I hope that may be of some use, even if you don't use this algorithm exactly. Thinking about the problem in terms of cutting rather than filling certainly seems to make it easier to get a clean, tightly-packed result. | I have solved my problem in a different way.
As i was looking for my problem, it turned out to be a fairly complex one, both measured in difficulty to implement as algorithm, as algorithm complexity.
If anyone is having a similar problem, these problems are classified as 'packing problems' in general, with specific problems like the 'pallet loading problem'.
The problem i was interested in, is illustrated at the bottom of this page:
<https://www.ime.usp.br/~egbirgin/packing/>
and a paper about this problem, with algorithm descriptions of how to solve the packing problem for convex polygons and curved shapes:
<http://www.ime.usp.br/~egbirgin/publications/bmnr.pdf>
Some more information on these kinds of problems:
<http://lagrange.ime.usp.br/~lobato/utdc/>
<http://mathworld.wolfram.com/SquarePacking.html> |
35,638 | I am flying to Berlin Tegel Airport (TXL). I need to get to Leipzig. Is it possible to get a train from the Airport? Or do I have to get a connection train? I read something about Rail& Fly. But I am not really sure about the train from Berlin I will be carrying luggage, which will be relatively heavy (~30KG) | 2014/08/24 | [
"https://travel.stackexchange.com/questions/35638",
"https://travel.stackexchange.com",
"https://travel.stackexchange.com/users/7345/"
] | In Berlin, the bigger problem is that there is no long-distance train station in any of the airports. In spite of [the official website's rhetoric](http://www.berlin-airport.de/en/travellers-txl/to-and-from/buses-and-trains/local-public-transport/index.php), Tegel is actually relatively poorly connected to the local and national transport networks, you need to take a bus to get anywhere, there isn't even a good S-Bahn or U-Bahn connection. Since it's much closer to the center of the city than many other airports, it's not that bad but you do need to get at least one bus connection.
Once you get to one of Berlin's major train stations (e.g. the new “Hauptbahnhof”), you can easily catch a direct train to Leipzig, possibly with a Rail&Fly ticket. [Rail&Fly](https://www.bahn.de/p/view/service/flug/rail_und_fly.shtml) fares can only be booked through your airline but if there is a train, you can always take it with a regular train ticket as well, you don't necessarily need Rail&Fly or any other special fare.
Incidentally, Tegel was slated to be closed some time ago but Berlin's new airport (built around the old Schönefeld airport) also had many problems with it's connection to the rail network and isn't even operating yet, several years after the planned opening so Berlin is a very poor choice if you are looking for an easy train connection. | You can reach the "Hauptbahnhof" from Tegel by riding the TXL-Bus (public transportation) which runs every 5 minutes. The fare is about 2.7€ and it will take about 15-20 minutes.
Take a direct train to Leipzig from there. This is the Link to the german railway authority called [Deutsche Bahn](http://www.deutschebahn.com/en/start.html) |
2,882,172 | I have some javascript code that includes an ANTLR-generated lexer and parser, and some associated syntax tree evaluation functionality. This code runs in the browser in my web app to support users who author code snippets which process scientific data.
Now I'd like to do some additional background processing on the server using the same generated parser. I would prefer not to have to re-implement this stuff in C# and have multiple bits of code that did the exact same thing. Performance isn't as critical to me as eliminating duplication, since this is a background process. So, how can I call into my javascript code from C#? And how can I format my script so that it plays nicely with my .NET web app? | 2010/05/21 | [
"https://Stackoverflow.com/questions/2882172",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/404/"
] | Anyone who wants to do this who isn't running Windows should definitely check out [node.js](http://nodejs.org/). Unfortunately for your purposes, I don't think the Windows port is working yet.
For server-side JS usage in general, [CommonJS](http://commonjs.org/) is trying to standardize how libraries are created and used. You can also find links there to a lot of server-side Javascript implementations.
Update: node.js [will now work under CygWin](http://github.com/ry/node/commit/b3b81d67ff1bc7e795cff7baa00dd716eca95eba). I'm not sure how much work it would be to integrate it into your environment though. | You could [have a look](http://msdn.microsoft.com/en-us/library/ms974588.aspx) at [JScript.NET](http://msdn.microsoft.com/en-us/library/x85xxsf4(VS.71).aspx) although there will be (at least some) differences with traditional (client-side) JavaScript (ECMA). |
2,882,172 | I have some javascript code that includes an ANTLR-generated lexer and parser, and some associated syntax tree evaluation functionality. This code runs in the browser in my web app to support users who author code snippets which process scientific data.
Now I'd like to do some additional background processing on the server using the same generated parser. I would prefer not to have to re-implement this stuff in C# and have multiple bits of code that did the exact same thing. Performance isn't as critical to me as eliminating duplication, since this is a background process. So, how can I call into my javascript code from C#? And how can I format my script so that it plays nicely with my .NET web app? | 2010/05/21 | [
"https://Stackoverflow.com/questions/2882172",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/404/"
] | Anyone who wants to do this who isn't running Windows should definitely check out [node.js](http://nodejs.org/). Unfortunately for your purposes, I don't think the Windows port is working yet.
For server-side JS usage in general, [CommonJS](http://commonjs.org/) is trying to standardize how libraries are created and used. You can also find links there to a lot of server-side Javascript implementations.
Update: node.js [will now work under CygWin](http://github.com/ry/node/commit/b3b81d67ff1bc7e795cff7baa00dd716eca95eba). I'm not sure how much work it would be to integrate it into your environment though. | I have two idea's:
* using [Rhino Shell](http://www.mozilla.org/rhino/shell.html)
* generating html page with your javascript code and starting this page using [WatiN](http://watin.sourceforge.net) or [Selenium](http://seleniumhq.org/) |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | Yes, this is an error at Apple's side. I have submitted my app with this error and Apple approved the app. | Thats because there still isn't a option to upload screenshots for the iPad Pro. No doubt error on Apple's side. |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | Thats because there still isn't a option to upload screenshots for the iPad Pro. No doubt error on Apple's side. | It's available now in xCode 7.1
Now you can launch your application in iPad Pro iOS simulator.
[xCode 7.1 simulators](http://i.stack.imgur.com/QgpCS.png) |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | >
> **Update:**
>
>
>
**iTunes Connect** now supports uploading **iPad Pro Screenshots**
[](https://i.stack.imgur.com/0TfJt.png)
(german screenshot)
*As mentioned by @MikeKeskinov this may be only visible if the app is **not** already **ready for sale**. (thanks for commenting)* | Thats because there still isn't a option to upload screenshots for the iPad Pro. No doubt error on Apple's side. |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | Yes, this is an error at Apple's side. I have submitted my app with this error and Apple approved the app. | It's available now in xCode 7.1
Now you can launch your application in iPad Pro iOS simulator.
[xCode 7.1 simulators](http://i.stack.imgur.com/QgpCS.png) |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | Yes, this is an error at Apple's side. I have submitted my app with this error and Apple approved the app. | >
> **Update:**
>
>
>
**iTunes Connect** now supports uploading **iPad Pro Screenshots**
[](https://i.stack.imgur.com/0TfJt.png)
(german screenshot)
*As mentioned by @MikeKeskinov this may be only visible if the app is **not** already **ready for sale**. (thanks for commenting)* |
33,131,758 | After submitting the app to app store for review, I got below alert in iTunes Connect. I don't see the option to upload iPad Pro screenshots in iTunes Connect.[](https://i.stack.imgur.com/P4fB2.png)
Please let me know what can be done here. | 2015/10/14 | [
"https://Stackoverflow.com/questions/33131758",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265949/"
] | >
> **Update:**
>
>
>
**iTunes Connect** now supports uploading **iPad Pro Screenshots**
[](https://i.stack.imgur.com/0TfJt.png)
(german screenshot)
*As mentioned by @MikeKeskinov this may be only visible if the app is **not** already **ready for sale**. (thanks for commenting)* | It's available now in xCode 7.1
Now you can launch your application in iPad Pro iOS simulator.
[xCode 7.1 simulators](http://i.stack.imgur.com/QgpCS.png) |
74,078,772 | I'm trying to make a simple navbar, where some of the links link to sections in the same component(page), using react scroll and then have other links link to a different component(route page). Hope the makes sense?
Here is a simple version: <https://codesandbox.io/s/late-firefly-x109rk>
In this example I have four links. The idea is to have Home, About and Contact scroll to the corresponding section and have new page open a new page using the Router.
There are currently two problems with this example:
1: if Home, About or Contact is active when going to the new page they stay active, while the new page also gets active.
2: When trying to go back from NewPage to Home, About or Contact I get "target Element not found" | 2022/10/15 | [
"https://Stackoverflow.com/questions/74078772",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/18523007/"
] | Delete it and start over. Why wonder if you've found everything the cybercreep left behind? And don't enable access except from an allow-list of known IPV4 and IPV6 addresses.
Many database ops people choose never to expose their databases to the public net, and now you know why. | Just turning all the 'trust' to 'md5' and rebooting the OS might be good enough, but there is no way to tell that based on this info. If there is nothing important on the server, just burn it down and rebuild it.
In addition to not using trust, additional simple steps would be to make sure your password is actually good, not allowing non-local connections for the 'postgres' user (or any other superuser), or not allowing non-local connections at all if not necessary or whitelisting connections from specific IP addresses if you do need to allow them. |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | To quote a recently running TV commercial, the answer is "infinity plus 1".
SAS does not have any limitation on the size of a file during read operations, although it's possible you might run into a limit on the number of observations written to a new SAS data set. On a 64-bit system, that would be 2\*\*63-1, about 9.2 quintillion observations. You are more likely to run out of disk space. | "The maximum size of a SAS data set in a Direct Access Bound Library is limited by the maximum size of the library, which is about 2986 GB on 3390 volumes."
<http://support.sas.com/kb/8/213.html> |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | "The maximum size of a SAS data set in a Direct Access Bound Library is limited by the maximum size of the library, which is about 2986 GB on 3390 volumes."
<http://support.sas.com/kb/8/213.html> | Depends on HOW you read the file. If loading a hash table for instance, you are limited by memory.. If you are reading via a set or infile statement, with no or limited size output dataset, you are limited by the 'uptime' of your process - eg if / when you are forcibly logged off, or become disconnected from whatever you are reading.
Also, some datasets (eg with a large number of columns) may be impossible to read, as cannot load all the variables into the PDV due to memory limits. Again, this is an OS limitation and not SAS. |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | "The maximum size of a SAS data set in a Direct Access Bound Library is limited by the maximum size of the library, which is about 2986 GB on 3390 volumes."
<http://support.sas.com/kb/8/213.html> | Reading from a text file, you are limited by the OS typically. Windows allows 1 million characters wide, for example, although SAS can theoretically handle a lrecl of over that.
In general, if you're asking about observations, you will be disk space limited before you are limited by anything else. I've seen a file created that were denominated in PB (petabytes, or 1 thousand terabytes) when on a compressed volume. |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | To quote a recently running TV commercial, the answer is "infinity plus 1".
SAS does not have any limitation on the size of a file during read operations, although it's possible you might run into a limit on the number of observations written to a new SAS data set. On a 64-bit system, that would be 2\*\*63-1, about 9.2 quintillion observations. You are more likely to run out of disk space. | Depends on HOW you read the file. If loading a hash table for instance, you are limited by memory.. If you are reading via a set or infile statement, with no or limited size output dataset, you are limited by the 'uptime' of your process - eg if / when you are forcibly logged off, or become disconnected from whatever you are reading.
Also, some datasets (eg with a large number of columns) may be impossible to read, as cannot load all the variables into the PDV due to memory limits. Again, this is an OS limitation and not SAS. |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | To quote a recently running TV commercial, the answer is "infinity plus 1".
SAS does not have any limitation on the size of a file during read operations, although it's possible you might run into a limit on the number of observations written to a new SAS data set. On a 64-bit system, that would be 2\*\*63-1, about 9.2 quintillion observations. You are more likely to run out of disk space. | Reading from a text file, you are limited by the OS typically. Windows allows 1 million characters wide, for example, although SAS can theoretically handle a lrecl of over that.
In general, if you're asking about observations, you will be disk space limited before you are limited by anything else. I've seen a file created that were denominated in PB (petabytes, or 1 thousand terabytes) when on a compressed volume. |
17,114,306 | What is the max file size that SAS software can successfully read.
(without considering hardware limit)
Thanks in advance. | 2013/06/14 | [
"https://Stackoverflow.com/questions/17114306",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2487084/"
] | Depends on HOW you read the file. If loading a hash table for instance, you are limited by memory.. If you are reading via a set or infile statement, with no or limited size output dataset, you are limited by the 'uptime' of your process - eg if / when you are forcibly logged off, or become disconnected from whatever you are reading.
Also, some datasets (eg with a large number of columns) may be impossible to read, as cannot load all the variables into the PDV due to memory limits. Again, this is an OS limitation and not SAS. | Reading from a text file, you are limited by the OS typically. Windows allows 1 million characters wide, for example, although SAS can theoretically handle a lrecl of over that.
In general, if you're asking about observations, you will be disk space limited before you are limited by anything else. I've seen a file created that were denominated in PB (petabytes, or 1 thousand terabytes) when on a compressed volume. |
2,029,736 | Firefox has the Sandbox and evalInSandbox(). Chrome has sandboxed execution in their content scripts (they call it isolated execution). I'm looking for the same thing in an IE browser extension.
I can load a javascript file, then call evalScript(), but the code executes in the same environment as javascript that exists on the page. I need a way to run my library (which includes and is based on jQuery) in an sandboxed/isolated environment, but still allow it to modify the DOM as if it were running on the page.
Jint looks promising, but cannot currently evaluate jQuery. (They can parse it.)
How can I do this? | 2010/01/08 | [
"https://Stackoverflow.com/questions/2029736",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246624/"
] | You might find it worthwhile looking at the Microsoft Web Sandbox at the Live Labs:
<http://websandbox.livelabs.com/>
Although it's more fully fledged and related to mashups and the like it might point you in the right direction. | Can you elaborate on your goals a bit?
I don't think you've clearly defined what you're trying to accomplish. If your code has the ability to modify the page's DOM, then it effectively *is* executing in the context of the page. It can create new script blocks to perform any unsafe or unreliable action that it could accomplish were it operating in the same execution environment.
IE doesn't offer a feature to do what you're asking, and I'm not convinced that the Firefox and Chrome features work the way that you expect that they do. |
2,029,736 | Firefox has the Sandbox and evalInSandbox(). Chrome has sandboxed execution in their content scripts (they call it isolated execution). I'm looking for the same thing in an IE browser extension.
I can load a javascript file, then call evalScript(), but the code executes in the same environment as javascript that exists on the page. I need a way to run my library (which includes and is based on jQuery) in an sandboxed/isolated environment, but still allow it to modify the DOM as if it were running on the page.
Jint looks promising, but cannot currently evaluate jQuery. (They can parse it.)
How can I do this? | 2010/01/08 | [
"https://Stackoverflow.com/questions/2029736",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246624/"
] | Looks like you are in a pickle with this one.
* Some Microsoft people say that IE doesn't offer this functionality
* Some who say they have been able to pull this off hold it close and protect it as their secret sauce
* "Professional" IE extension shops, when contacted about building what you are talking about turn down the job
I really wish I had better news but it looks like it might take a small miracle to accomplish what you are looking for... or maybe a lot of money. : )
Your best bet is probably going to be finding one of the few who claim they have been able to do it and pay them a lot to share the secret or reconsider why you really want what you want and see if you can't accomplish it another way. | You might find it worthwhile looking at the Microsoft Web Sandbox at the Live Labs:
<http://websandbox.livelabs.com/>
Although it's more fully fledged and related to mashups and the like it might point you in the right direction. |
2,029,736 | Firefox has the Sandbox and evalInSandbox(). Chrome has sandboxed execution in their content scripts (they call it isolated execution). I'm looking for the same thing in an IE browser extension.
I can load a javascript file, then call evalScript(), but the code executes in the same environment as javascript that exists on the page. I need a way to run my library (which includes and is based on jQuery) in an sandboxed/isolated environment, but still allow it to modify the DOM as if it were running on the page.
Jint looks promising, but cannot currently evaluate jQuery. (They can parse it.)
How can I do this? | 2010/01/08 | [
"https://Stackoverflow.com/questions/2029736",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246624/"
] | Looks like you are in a pickle with this one.
* Some Microsoft people say that IE doesn't offer this functionality
* Some who say they have been able to pull this off hold it close and protect it as their secret sauce
* "Professional" IE extension shops, when contacted about building what you are talking about turn down the job
I really wish I had better news but it looks like it might take a small miracle to accomplish what you are looking for... or maybe a lot of money. : )
Your best bet is probably going to be finding one of the few who claim they have been able to do it and pay them a lot to share the secret or reconsider why you really want what you want and see if you can't accomplish it another way. | Can you elaborate on your goals a bit?
I don't think you've clearly defined what you're trying to accomplish. If your code has the ability to modify the page's DOM, then it effectively *is* executing in the context of the page. It can create new script blocks to perform any unsafe or unreliable action that it could accomplish were it operating in the same execution environment.
IE doesn't offer a feature to do what you're asking, and I'm not convinced that the Firefox and Chrome features work the way that you expect that they do. |
2,029,736 | Firefox has the Sandbox and evalInSandbox(). Chrome has sandboxed execution in their content scripts (they call it isolated execution). I'm looking for the same thing in an IE browser extension.
I can load a javascript file, then call evalScript(), but the code executes in the same environment as javascript that exists on the page. I need a way to run my library (which includes and is based on jQuery) in an sandboxed/isolated environment, but still allow it to modify the DOM as if it were running on the page.
Jint looks promising, but cannot currently evaluate jQuery. (They can parse it.)
How can I do this? | 2010/01/08 | [
"https://Stackoverflow.com/questions/2029736",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246624/"
] | Thing you are looking for is ActiveScript engine (<http://en.wikipedia.org/wiki/Active_Scripting>)
I can`t find any usefull links now, MSDN contains only interfaces definition (<http://msdn.microsoft.com/en-us/library/ccd0zt2w(v=vs.85).aspx>)
Try to Google "Active Scripting" (not "Action scripting"!)
You need to implement interface IActiveScriptHost in your extension, create an object of «JScript», call SetSite and pass your host object, then you can load your js code into this engine and run it.
This technique is a quite difficult due to lack of documentation about it.
If you are still interested in this, I can send you some examples in C++/ATL. | Can you elaborate on your goals a bit?
I don't think you've clearly defined what you're trying to accomplish. If your code has the ability to modify the page's DOM, then it effectively *is* executing in the context of the page. It can create new script blocks to perform any unsafe or unreliable action that it could accomplish were it operating in the same execution environment.
IE doesn't offer a feature to do what you're asking, and I'm not convinced that the Firefox and Chrome features work the way that you expect that they do. |
2,029,736 | Firefox has the Sandbox and evalInSandbox(). Chrome has sandboxed execution in their content scripts (they call it isolated execution). I'm looking for the same thing in an IE browser extension.
I can load a javascript file, then call evalScript(), but the code executes in the same environment as javascript that exists on the page. I need a way to run my library (which includes and is based on jQuery) in an sandboxed/isolated environment, but still allow it to modify the DOM as if it were running on the page.
Jint looks promising, but cannot currently evaluate jQuery. (They can parse it.)
How can I do this? | 2010/01/08 | [
"https://Stackoverflow.com/questions/2029736",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/246624/"
] | Looks like you are in a pickle with this one.
* Some Microsoft people say that IE doesn't offer this functionality
* Some who say they have been able to pull this off hold it close and protect it as their secret sauce
* "Professional" IE extension shops, when contacted about building what you are talking about turn down the job
I really wish I had better news but it looks like it might take a small miracle to accomplish what you are looking for... or maybe a lot of money. : )
Your best bet is probably going to be finding one of the few who claim they have been able to do it and pay them a lot to share the secret or reconsider why you really want what you want and see if you can't accomplish it another way. | Thing you are looking for is ActiveScript engine (<http://en.wikipedia.org/wiki/Active_Scripting>)
I can`t find any usefull links now, MSDN contains only interfaces definition (<http://msdn.microsoft.com/en-us/library/ccd0zt2w(v=vs.85).aspx>)
Try to Google "Active Scripting" (not "Action scripting"!)
You need to implement interface IActiveScriptHost in your extension, create an object of «JScript», call SetSite and pass your host object, then you can load your js code into this engine and run it.
This technique is a quite difficult due to lack of documentation about it.
If you are still interested in this, I can send you some examples in C++/ATL. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | Follow the model of the [Hutterites](https://en.wikipedia.org/wiki/Hutterite).
They are in many ways like your Grey: a community (a communistic community!) of religious devotees who marry among themselves, and who are possessed with a sense of purpose.
Hutterites have lots of children. Children are baptized into the church. Not all of them are. Some might not want to. Some don't make the cut. They do not have some penal colony for children of Hutterites who do not become part of the church. These kids grow up to be people outside the church: Catholics, or agnostics, or whatever they want to be.
The Grey can do the same as the Hutterites. They have lots of kids. Pay attention to them. Those who are willing and who make the cut can be Grey. The other kids are just regular people. I am sure the population of Greenland is not only the 1000 Grey. There are probably lots of other people doing the things people do. The Grey kids who grow up to be not Grey just mix in with the regular population.
From <http://www.hutterites.org>
>
> When young people feel ready to make a commitment to baptism, they
> meet with the senior minister and make a formal request. The minister
> brings this request to the whole brotherhood and if there are no
> objections, they are accepted for a probation period. Every Sunday
> afternoon for 6 to 7 weeks, the baptismal candidates visit each of the
> witness brothers who provides spiritual and religious teachings
> ranging in length from ten minutes to over an hour. Hutterites accept
> all 12 points of the Apostle’s Creed as Truth. All members publicly
> declare their belief in The Apostle’s Creed upon baptism.
>
>
> Can a non-Hutterite join a Hutterite colony A question that is often
> asked is whether or not an outsider can join a Hutterite colony. It
> has happened that outsiders or non-Hutterites join a Hutterite colony,
> but it is quite rare. Few Hutterite colony are open to outsider
> joining, but there are a handful who would consider it, depending on
> the candidate and the willingness of the candidate to adopt to the
> Hutterian norms. Over the year’s many people have attempted to become
> full members of the Hutterite community, but haven’t been successful
> for a variety of reasons.
>
>
> | I would divide the country into three groups:
1. The Grey.
2. The military/police.
3. Everyone else.
Children of the Grey become officers in the military or the police. Children of current members of the military or the police may become officers as well, although most would stay in the ranks. Officers are expected to exemplify the ascetic nature of the Grey.
Everyone else can join the military or the police but few slots are available.
When a member of the Grey dies, one of the officers is chosen as a replacement. This might be a child of the Grey, of an officer, or of someone from the ranks.
This keeps the Grey on top. It gives them strong roots in the military and police. It allows for merit-based promotion. It does not allow for quick promotion, such that someone would have to live the ascetic lifestyle for years to reach the Grey. This would be hard for someone to fake for that length of time, particularly as their schooling would also have been monitored by the Grey. So a non-believer would likely be dismissed early.
People in the everyone else category can not themselves join the Grey. They can only join the military or police ranks. Their children can potentially join the Grey. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | Follow the model of the [Hutterites](https://en.wikipedia.org/wiki/Hutterite).
They are in many ways like your Grey: a community (a communistic community!) of religious devotees who marry among themselves, and who are possessed with a sense of purpose.
Hutterites have lots of children. Children are baptized into the church. Not all of them are. Some might not want to. Some don't make the cut. They do not have some penal colony for children of Hutterites who do not become part of the church. These kids grow up to be people outside the church: Catholics, or agnostics, or whatever they want to be.
The Grey can do the same as the Hutterites. They have lots of kids. Pay attention to them. Those who are willing and who make the cut can be Grey. The other kids are just regular people. I am sure the population of Greenland is not only the 1000 Grey. There are probably lots of other people doing the things people do. The Grey kids who grow up to be not Grey just mix in with the regular population.
From <http://www.hutterites.org>
>
> When young people feel ready to make a commitment to baptism, they
> meet with the senior minister and make a formal request. The minister
> brings this request to the whole brotherhood and if there are no
> objections, they are accepted for a probation period. Every Sunday
> afternoon for 6 to 7 weeks, the baptismal candidates visit each of the
> witness brothers who provides spiritual and religious teachings
> ranging in length from ten minutes to over an hour. Hutterites accept
> all 12 points of the Apostle’s Creed as Truth. All members publicly
> declare their belief in The Apostle’s Creed upon baptism.
>
>
> Can a non-Hutterite join a Hutterite colony A question that is often
> asked is whether or not an outsider can join a Hutterite colony. It
> has happened that outsiders or non-Hutterites join a Hutterite colony,
> but it is quite rare. Few Hutterite colony are open to outsider
> joining, but there are a handful who would consider it, depending on
> the candidate and the willingness of the candidate to adopt to the
> Hutterian norms. Over the year’s many people have attempted to become
> full members of the Hutterite community, but haven’t been successful
> for a variety of reasons.
>
>
> | StephenG was quite comprehensive; and I agree. So let me get more prescriptive:
Your strategy is wrong. Your Grey are not being very intelligent if they give themselves all the perks of power as an elite ruling class, that is a recipe (and always has been) for the underclass to revolt out of resentment and slaughter them, assassinate them or their children, particularly when large numbers of the underclass have nothing to lose. Such as at times of great poverty, failed crops or other economic collapse.
**Everyone** can be overthrown by sufficient force; there is simply no insurance against it. The best fighter on the planet cannot dodge a thousand bullets (or knives or spears). Your Greys must rule by terror (e.g. Hussein, Ghadafi), or rule by love; meaning there are *much better ways* to make money and have power than being a member of the government.
If you want to filter out opportunists, you must find ways to eliminate monetary reward, so the only rewards are the satisfaction of knowing you have helped others. Who would do that? Altruists. What if a Grey child must take an irreversible vow of average income in order to join the government? They get the average income of their citizens, for life, period. Any more than that due to any reason whatsoever must go to charity, they have no access to their pre-existing property or wealth for life, such pre-existing property or wealth will not increase in value (or all increases will be donated to charities); any cheating is a death sentence; there is zero upward mobility, in money or property, if you choose to enter government. The Greys ensure you **will** live the average life, period.
To counter megalomania (power hungry, not money or property hungry), use WL Gore corporation's inverted management. Here is a slightly modified version of that:
First, every new Grey must start at the bottom, on a one year probation.
In that time they must apprentice to four task teams in government (a task team being 1 to 20 individuals working on a specific thing); the first is chosen by the candidate, each subsequent one is chosen by the team they just finished with (after 3 months); although the team may consider the candidate's stated preferences for type of work.
At the end of the year, all members of the teams must vote on whether to accept the apprentice or not; all four teams must declare them suitable, within each team 2/3 of members (rounded up) must declare them suitable.
If they do, the lifelong vow of average income begins, with ceremony: This is like joining the military, certain transgressions or betrayals can literally cost your life.
On to countering megalomania: The inverted management idea is that every level of management is chosen by their underlings, and their position can be revoked at any time by their underlings. The first level managers similarly choose amongst themselves for their boss, etc, all the way up.
So, say we manage 1000 Greys by groups of 10: 10 greys choose a group manager; resulting in 100 group managers. These organize in groups of 10, and choose 10 section managers (over 100 Greys each). These 10 section managers elect a President. If at any time they want to overrule a Presidential decision, they can; or they can demote the President back to their own ranks and put somebody else in charge.
Only the level beneath a manager can demote a manager.
To adjust this slightly to address the costs of management; Low level groups are actually self-organized in groups of about 12 to do the work of 10. When they elect a boss (A), it is one of them; If they pull their boss back into their ranks; they put someone else there. Now if their boss (A) is then elected to the next management layer up: They elect a new boss, B, and hire a new employee (or recruit one from another team with more than they need). If boss (A) is ever pulled back from an upper level, he falls all the way back to his original, no-boss-at-all level. He can rejoin his original team, or some other team. If no team will have him, and presuming he has not committed any crime, then he is without work. But he still gets paid.
There is a small risk of free-riders in this system; presumably they would be rooted out by the 1-year probation period by one of the four teams that vetted each candidate. But there is little risk of megalomaniacs getting very far, and little risk of corruption due to the limitations on income: The other Greys in the government that **are** working will ensure that is strictly enforced. You have altruists doing this work because they want to help people, and they don't get any special rewards for doing it.
The vast majority of actual humans are not altruistic to the extent of giving up all future possibility of wealth, and making themselves permanently subordinate to the collective. The Citizens don't want this job; it would be like being jealous of garbage collectors being paid an average wage. A small percentage of citizens might feel that way, but not enough to overthrow the system.
If it is religion that motivates the Grey, fine: The restrictions on what it costs them to help people ensures their religious motivation is not a pretense to cover for self enrichment or satisfying their hunger for power. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | **Short version of solution : Mid Ranking Management.**
Long version : lots of things I don't like.
>
> In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
>
>
>
Or something more in keeping with the national language and culture, perhaps. "The Grey" seems a rather uninspiring name for a group leading anyone. It conveys no meaning or association with any historically or socially important ideal. It may be the sort of name an author might pick, but it's not going to help get you to power and at some point you have to start not in power and rise to power.
"The Grey" is a dark name and would convey, if anything, a sense that they might be dangerous. It's not an accident that groups tend to go for names like "New Dawn", "Shining Path and so on. Positive spin.
Rethink the name.
>
> Numbering at 1000 or so
>
>
>
So barely enough to populate even the most senior positions in an administration.
>
> , the grey is a group of highly intelligent
>
>
>
In what way ?
Intelligence can take many forms and if they're just randomly intelligent and individual members can funnel their energies into medicine, art, music, literature, games, programming, engineering and so on, that won't really give you many of the 1000 who are actually skilled at the skills of politics and economics that they'll actually desperately need as a small group to survive.
>
> , genetically enhanced (with slight improvement in brain power and longevity)
>
>
>
Slight ? So if I concentrate hard I can out think them ? So the top 1% of the million or more population can match them or better them ?
They'd better be well beyond "slight".
Longevity isn't much use if you're a minority group that can find itself shot against a wall in a revolution.
So you want long lived individuals or a long lived clan ?
>
> religious devotees.
>
>
>
Dogma rears it head.
Dogma is a weakness. Any form of religion is, by definition, a factor that will be considered reactionary by people outside that religion.
All a religion does is make it focus for resentment. "Why are we being led by this minority ?" is going to be followed, eventually by "Hang the Bar Stewards !". History is pretty clear on this point.
If you have a religion and you want to hold power, it must be a religion the ordinary populace are comfortable with.
Which leads us to ...
>
> Details about their religion is unimportant,
>
>
>
Yes, they are. Details of a practically minute nature are responsible for some of the bloodiest and most ruthlessly pursued wars in history. The devil is in the detail.
>
> but it is not at odds with science/progress,
>
>
>
It must be. Religion requires belief without evidence. There will always be a point where this clashes with science, which require no belief unless there's evidence.
There will always be a point of contention, usually many, between these two fundamentally different outlooks.
>
> and its overall effect is that, among true believers,
>
>
>
Question : how do you tell a true believer from someone who just claims they are ?
Answer : tell them you'll spare the ones who relinquish their faith and burn the believers are the stake.
A system, I might add, which has been used extensively by humans and, hard as it is to believe, the true believers often will go to the stake and you're left with the cynics to feed.
Another way of viewing this is Who Polices The Police ? Who says you're a true believer and how do you prove it ?
>
> this religion inspires a sense of duty,
>
>
>
One man's duty is another man's holocaust. Dogma is a problem and, again, who decides what's duty and what's unworthy personal obsession ?
>
> and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
>
>
>
Surely they'd start seeing each other that way as well ?
Normal greed, envy, lust and all those other human emotions will lead you to look at not just your servant, but your master, as a waste of space you could be using better. :-)
>
> Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool),
>
>
>
Bet that's popular with the vast majority they depend on to not rush teh palace gates !
>
> directly controlling the army
>
>
>
1000 people don't control an army. They control, at best, 1000 people.
The army will be the de facto rulers and, as history proves, if you depend on the army to keep you in power, you're just a moving target.
Something other than a terrible name and a condescending manner is going to be required to stop all these armed and highly trained people from just turning on you.
>
> and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
>
>
>
A mob also holds a veto right. Usually in both hands.
A minority religious caste creaming money off the top and who cannot be democratically removed is, in no way, going to successfully pass itself or it's society off as a "mostly democratic" government.
If it looks like a nobility, acts like a nobility and swims like a nobility in it's private swimming pool, it's a nobility.
>
> When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers.
>
>
>
So not produced with much calculating care at all.
>
> This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
>
>
>
Actually just the ambition of the ones being groomed for future leadership is sufficient to create a lot more instability than anything else.
The ones who don't have ambition and under-perform are, by definition, no threat at all to the ones who have it and do perform.
They are a non-problem who'll be happy to get by on a basic "better than average" job with enough trappings of power to not need real power. Perfect mid ranking managers could hardly be described better.
>
> If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest.
>
>
>
That'd be the same sense of separation that will find them waiting for execution after the revolution, I take it. :-)
>
> It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
>
>
>
As I've pointed out, it's just as difficult to tell what's really going on inside the mind of all these condescending, ambitious people.
It's fundamentally impossible to avoid this problem and still have any free will in a group.
>
> Proposed solution:
>
>
>
There is no perfect solution.
You need a system of exams, formal and legally binding (on both sides) to test for ability, inclination and deception.
How do you test for deception : you use good interrogation methods (and that's the talking and listening type, not the idiotic buckets of water nonsense). You use detectives to check things. You use, in fact, a good Intelligence service. The exams just test technical competence and knowledge. You'd probably also test for ability to lead - check out how the military do that, because if there's one sphere of human endeavor that requires a better than average success at finding good leaders, it's the military !
>
> I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
>
>
>
I'll say this again : mid ranking managers. Practically the same thing without the weird connotation of "culling" - an unfortunate choice of word.
>
> Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey.
>
>
>
Gotta have mid ranking managers with little or no real power. Planet's full of them and while they're a nuisance and get in the way really powerful people (like engineers) they're not going to be able to form and execute any plan, either singly or in a group, that could possibly destabilize anything.
No revolution was ever started by a mid ranking manager. Well, probably. | StephenG was quite comprehensive; and I agree. So let me get more prescriptive:
Your strategy is wrong. Your Grey are not being very intelligent if they give themselves all the perks of power as an elite ruling class, that is a recipe (and always has been) for the underclass to revolt out of resentment and slaughter them, assassinate them or their children, particularly when large numbers of the underclass have nothing to lose. Such as at times of great poverty, failed crops or other economic collapse.
**Everyone** can be overthrown by sufficient force; there is simply no insurance against it. The best fighter on the planet cannot dodge a thousand bullets (or knives or spears). Your Greys must rule by terror (e.g. Hussein, Ghadafi), or rule by love; meaning there are *much better ways* to make money and have power than being a member of the government.
If you want to filter out opportunists, you must find ways to eliminate monetary reward, so the only rewards are the satisfaction of knowing you have helped others. Who would do that? Altruists. What if a Grey child must take an irreversible vow of average income in order to join the government? They get the average income of their citizens, for life, period. Any more than that due to any reason whatsoever must go to charity, they have no access to their pre-existing property or wealth for life, such pre-existing property or wealth will not increase in value (or all increases will be donated to charities); any cheating is a death sentence; there is zero upward mobility, in money or property, if you choose to enter government. The Greys ensure you **will** live the average life, period.
To counter megalomania (power hungry, not money or property hungry), use WL Gore corporation's inverted management. Here is a slightly modified version of that:
First, every new Grey must start at the bottom, on a one year probation.
In that time they must apprentice to four task teams in government (a task team being 1 to 20 individuals working on a specific thing); the first is chosen by the candidate, each subsequent one is chosen by the team they just finished with (after 3 months); although the team may consider the candidate's stated preferences for type of work.
At the end of the year, all members of the teams must vote on whether to accept the apprentice or not; all four teams must declare them suitable, within each team 2/3 of members (rounded up) must declare them suitable.
If they do, the lifelong vow of average income begins, with ceremony: This is like joining the military, certain transgressions or betrayals can literally cost your life.
On to countering megalomania: The inverted management idea is that every level of management is chosen by their underlings, and their position can be revoked at any time by their underlings. The first level managers similarly choose amongst themselves for their boss, etc, all the way up.
So, say we manage 1000 Greys by groups of 10: 10 greys choose a group manager; resulting in 100 group managers. These organize in groups of 10, and choose 10 section managers (over 100 Greys each). These 10 section managers elect a President. If at any time they want to overrule a Presidential decision, they can; or they can demote the President back to their own ranks and put somebody else in charge.
Only the level beneath a manager can demote a manager.
To adjust this slightly to address the costs of management; Low level groups are actually self-organized in groups of about 12 to do the work of 10. When they elect a boss (A), it is one of them; If they pull their boss back into their ranks; they put someone else there. Now if their boss (A) is then elected to the next management layer up: They elect a new boss, B, and hire a new employee (or recruit one from another team with more than they need). If boss (A) is ever pulled back from an upper level, he falls all the way back to his original, no-boss-at-all level. He can rejoin his original team, or some other team. If no team will have him, and presuming he has not committed any crime, then he is without work. But he still gets paid.
There is a small risk of free-riders in this system; presumably they would be rooted out by the 1-year probation period by one of the four teams that vetted each candidate. But there is little risk of megalomaniacs getting very far, and little risk of corruption due to the limitations on income: The other Greys in the government that **are** working will ensure that is strictly enforced. You have altruists doing this work because they want to help people, and they don't get any special rewards for doing it.
The vast majority of actual humans are not altruistic to the extent of giving up all future possibility of wealth, and making themselves permanently subordinate to the collective. The Citizens don't want this job; it would be like being jealous of garbage collectors being paid an average wage. A small percentage of citizens might feel that way, but not enough to overthrow the system.
If it is religion that motivates the Grey, fine: The restrictions on what it costs them to help people ensures their religious motivation is not a pretense to cover for self enrichment or satisfying their hunger for power. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | StephenG was quite comprehensive; and I agree. So let me get more prescriptive:
Your strategy is wrong. Your Grey are not being very intelligent if they give themselves all the perks of power as an elite ruling class, that is a recipe (and always has been) for the underclass to revolt out of resentment and slaughter them, assassinate them or their children, particularly when large numbers of the underclass have nothing to lose. Such as at times of great poverty, failed crops or other economic collapse.
**Everyone** can be overthrown by sufficient force; there is simply no insurance against it. The best fighter on the planet cannot dodge a thousand bullets (or knives or spears). Your Greys must rule by terror (e.g. Hussein, Ghadafi), or rule by love; meaning there are *much better ways* to make money and have power than being a member of the government.
If you want to filter out opportunists, you must find ways to eliminate monetary reward, so the only rewards are the satisfaction of knowing you have helped others. Who would do that? Altruists. What if a Grey child must take an irreversible vow of average income in order to join the government? They get the average income of their citizens, for life, period. Any more than that due to any reason whatsoever must go to charity, they have no access to their pre-existing property or wealth for life, such pre-existing property or wealth will not increase in value (or all increases will be donated to charities); any cheating is a death sentence; there is zero upward mobility, in money or property, if you choose to enter government. The Greys ensure you **will** live the average life, period.
To counter megalomania (power hungry, not money or property hungry), use WL Gore corporation's inverted management. Here is a slightly modified version of that:
First, every new Grey must start at the bottom, on a one year probation.
In that time they must apprentice to four task teams in government (a task team being 1 to 20 individuals working on a specific thing); the first is chosen by the candidate, each subsequent one is chosen by the team they just finished with (after 3 months); although the team may consider the candidate's stated preferences for type of work.
At the end of the year, all members of the teams must vote on whether to accept the apprentice or not; all four teams must declare them suitable, within each team 2/3 of members (rounded up) must declare them suitable.
If they do, the lifelong vow of average income begins, with ceremony: This is like joining the military, certain transgressions or betrayals can literally cost your life.
On to countering megalomania: The inverted management idea is that every level of management is chosen by their underlings, and their position can be revoked at any time by their underlings. The first level managers similarly choose amongst themselves for their boss, etc, all the way up.
So, say we manage 1000 Greys by groups of 10: 10 greys choose a group manager; resulting in 100 group managers. These organize in groups of 10, and choose 10 section managers (over 100 Greys each). These 10 section managers elect a President. If at any time they want to overrule a Presidential decision, they can; or they can demote the President back to their own ranks and put somebody else in charge.
Only the level beneath a manager can demote a manager.
To adjust this slightly to address the costs of management; Low level groups are actually self-organized in groups of about 12 to do the work of 10. When they elect a boss (A), it is one of them; If they pull their boss back into their ranks; they put someone else there. Now if their boss (A) is then elected to the next management layer up: They elect a new boss, B, and hire a new employee (or recruit one from another team with more than they need). If boss (A) is ever pulled back from an upper level, he falls all the way back to his original, no-boss-at-all level. He can rejoin his original team, or some other team. If no team will have him, and presuming he has not committed any crime, then he is without work. But he still gets paid.
There is a small risk of free-riders in this system; presumably they would be rooted out by the 1-year probation period by one of the four teams that vetted each candidate. But there is little risk of megalomaniacs getting very far, and little risk of corruption due to the limitations on income: The other Greys in the government that **are** working will ensure that is strictly enforced. You have altruists doing this work because they want to help people, and they don't get any special rewards for doing it.
The vast majority of actual humans are not altruistic to the extent of giving up all future possibility of wealth, and making themselves permanently subordinate to the collective. The Citizens don't want this job; it would be like being jealous of garbage collectors being paid an average wage. A small percentage of citizens might feel that way, but not enough to overthrow the system.
If it is religion that motivates the Grey, fine: The restrictions on what it costs them to help people ensures their religious motivation is not a pretense to cover for self enrichment or satisfying their hunger for power. | So, what you want is the English house of Lords. Those people are not genetically engineered and they still like to live the posh life and get privilege because their grand-grandx24grandfather 600 years ago had a dinner with a king.
Of course, from time to time there is one person who likes to live a common life. Fortunately they usually have more than one child so the emptiness is not permanent.
Oh, and also, they usually go to the army and become important figures in the command chain.
And the sense of duty is derived from a very low place. Cover your ass or communists will take your castles and palaces and money. They learned the lesson when the French part of the family was given a nice view from the guillotine.
Also your problem is only a problem if you **NEED** them to have exactly 657 members in the body. If you let the number be changeable to suit the able number your problem is non-existent. Source: The House of Lords. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | Follow the model of the [Hutterites](https://en.wikipedia.org/wiki/Hutterite).
They are in many ways like your Grey: a community (a communistic community!) of religious devotees who marry among themselves, and who are possessed with a sense of purpose.
Hutterites have lots of children. Children are baptized into the church. Not all of them are. Some might not want to. Some don't make the cut. They do not have some penal colony for children of Hutterites who do not become part of the church. These kids grow up to be people outside the church: Catholics, or agnostics, or whatever they want to be.
The Grey can do the same as the Hutterites. They have lots of kids. Pay attention to them. Those who are willing and who make the cut can be Grey. The other kids are just regular people. I am sure the population of Greenland is not only the 1000 Grey. There are probably lots of other people doing the things people do. The Grey kids who grow up to be not Grey just mix in with the regular population.
From <http://www.hutterites.org>
>
> When young people feel ready to make a commitment to baptism, they
> meet with the senior minister and make a formal request. The minister
> brings this request to the whole brotherhood and if there are no
> objections, they are accepted for a probation period. Every Sunday
> afternoon for 6 to 7 weeks, the baptismal candidates visit each of the
> witness brothers who provides spiritual and religious teachings
> ranging in length from ten minutes to over an hour. Hutterites accept
> all 12 points of the Apostle’s Creed as Truth. All members publicly
> declare their belief in The Apostle’s Creed upon baptism.
>
>
> Can a non-Hutterite join a Hutterite colony A question that is often
> asked is whether or not an outsider can join a Hutterite colony. It
> has happened that outsiders or non-Hutterites join a Hutterite colony,
> but it is quite rare. Few Hutterite colony are open to outsider
> joining, but there are a handful who would consider it, depending on
> the candidate and the willingness of the candidate to adopt to the
> Hutterian norms. Over the year’s many people have attempted to become
> full members of the Hutterite community, but haven’t been successful
> for a variety of reasons.
>
>
> | So, what you want is the English house of Lords. Those people are not genetically engineered and they still like to live the posh life and get privilege because their grand-grandx24grandfather 600 years ago had a dinner with a king.
Of course, from time to time there is one person who likes to live a common life. Fortunately they usually have more than one child so the emptiness is not permanent.
Oh, and also, they usually go to the army and become important figures in the command chain.
And the sense of duty is derived from a very low place. Cover your ass or communists will take your castles and palaces and money. They learned the lesson when the French part of the family was given a nice view from the guillotine.
Also your problem is only a problem if you **NEED** them to have exactly 657 members in the body. If you let the number be changeable to suit the able number your problem is non-existent. Source: The House of Lords. |
82,676 | **Context:**
In my world, society has progressed to 2050-ish, a new nation rose on Greenland led by an aristocratic community called the "Grey".
Numbering at 1000 or so, the grey is a group of highly intelligent, genetically enhanced (with slight improvement in brain power and longevity) religious devotees. Details about their religion is unimportant, but it is not at odds with science/progress, and its overall effect is that, among true believers, this religion inspires a sense of duty, and gives the Grey a condescending attitude towards nearly all outsiders, seeing them as hedonistic, wasteful primitives.
Also, the Grey exercise their power by collecting a small percentage of national tax (which they share among themselves from a common pool), directly controlling the army and hold veto rights over the more "normal" political activity performed by a mostly democratic government.
**Problem:**
When the Grey replaces dead members, they prefer using their own children (which they produce with calculated care) however, it is possible that the children they raise are not as religious as their parents or are unwilling to act as rulers. This means that either many children have to be produced so that some may suffice for future leadership, and it also generates the problem with the under-performing children, who may become a factor of instability.
If it is necessary to replenish the Grey's members by assimilating outsiders, the group may compromise its ideological/genetic purity and the sense of separation from the rest. It is also difficult to test the new members for their loyalty, more specifically, it's difficult to tell whether the new members are truly assimilated with the Grey's ideology or simply pretending to be so as to secure power.
**Proposed solution:**
I am thinking that culling the Grey's own children for sufficient replacement will work, and the under-performing children can be put in a sort of "pacified confinement" which is equivalent to a highly comfortable prison where the inmates are stripped of all power and cut-off from communication with the outside world.
Nevertheless, I am not confident that this will completely dispel the inmates' capacity as factors of instability, and perhaps a disproportionate number of Grey's children will voluntarily choose this lifestyle, forming a significant burden on the Grey. | 2017/06/03 | [
"https://worldbuilding.stackexchange.com/questions/82676",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/26674/"
] | A religious order of aristocrats would educate their children in a strict way that encourages memorization and conformity. People need two things to be happy: security and love. Deprived the children of parental contact and have them treated warmly by their care-givers only when they follow "the rules". Have the care-givers isolate a child who fails to meet expectations. Allow only the religious doctrine which they are being thought to provide comfort and hope.
Children of inattentive or distant parents grow to seek approval of authority figures or rebel. Those that rebel would do so by the time puberty sets in and could be culled. This would not be a problem if they are never allowed to bond with their parents and vice versa. Since the child has been taken at a young age if not at birth, the parents would simply chose to believe that one of the successful children was theirs and refuse to admit that their genes could have produced one of the bad eggs. Once grown and given social standing and influence, the chances of one of them questioning the status quo would be negligible.
The real problem of socially engineering a caste of people with such rigid prerequisites is that the greys would all lack empathy and social skills. Beyond the completion of their necessary duties, they would be hedonistic robots with an unbalanced view of self. Spoiler (you would just be copying the Vatican and the religious education of the clergy in the middle ages). | So, what you want is the English house of Lords. Those people are not genetically engineered and they still like to live the posh life and get privilege because their grand-grandx24grandfather 600 years ago had a dinner with a king.
Of course, from time to time there is one person who likes to live a common life. Fortunately they usually have more than one child so the emptiness is not permanent.
Oh, and also, they usually go to the army and become important figures in the command chain.
And the sense of duty is derived from a very low place. Cover your ass or communists will take your castles and palaces and money. They learned the lesson when the French part of the family was given a nice view from the guillotine.
Also your problem is only a problem if you **NEED** them to have exactly 657 members in the body. If you let the number be changeable to suit the able number your problem is non-existent. Source: The House of Lords. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.