qid int64 1 74.7M | question stringlengths 12 33.8k | date stringlengths 10 10 | metadata list | response_j stringlengths 0 115k | response_k stringlengths 2 98.3k |
|---|---|---|---|---|---|
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | A few options come to mind:
>
> "he thought this... but after being criticized, he \_\_ that the critics were correct"
>
>
>
* realized
* agreed
* concurred
* admitted
* was persuaded | Regarding the final sentence of the question (as quoted below) related words among others include *[co-opt](http://en.wiktionary.org/wiki/co-opt#English)*,
*[assimilate](http://en.wiktionary.org/wiki/assimilate)*, *[hoodwink](http://en.wiktionary.org/wiki/hoodwink)*, *[cave](http://en.wiktionary.org/wiki/cave#Verb)*, and perhaps *[recant](http://en.wiktionary.org/wiki/recant)*.
>
> It's actually a negative effect, his way of thinking permanently changed due to criticism (and I'm trying to say his original opinion was better).
>
>
>
However, to fit those words into your sample sentence (“he thought this ... but after being criticized, he \_\_ that the critics were correct”) requires adjustments, perhaps as illustrated in the following samples.
>
> • He thought his views were sound but the critics soon co-opted him.
>
> • His views were sound, but he caved under withering criticism.
>
> • Under light but unrelenting peer pressure, he recanted and was assimilated.
>
> • Despite his precise and marvelous mind, the critics soon hoodwinked him.
>
>
> |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | Not a word, but a phrase: "saw the light"
edit (after comment by OP)
***Submit (to)***, then, or "***knuckle under***" | Regarding the final sentence of the question (as quoted below) related words among others include *[co-opt](http://en.wiktionary.org/wiki/co-opt#English)*,
*[assimilate](http://en.wiktionary.org/wiki/assimilate)*, *[hoodwink](http://en.wiktionary.org/wiki/hoodwink)*, *[cave](http://en.wiktionary.org/wiki/cave#Verb)*, and perhaps *[recant](http://en.wiktionary.org/wiki/recant)*.
>
> It's actually a negative effect, his way of thinking permanently changed due to criticism (and I'm trying to say his original opinion was better).
>
>
>
However, to fit those words into your sample sentence (“he thought this ... but after being criticized, he \_\_ that the critics were correct”) requires adjustments, perhaps as illustrated in the following samples.
>
> • He thought his views were sound but the critics soon co-opted him.
>
> • His views were sound, but he caved under withering criticism.
>
> • Under light but unrelenting peer pressure, he recanted and was assimilated.
>
> • Despite his precise and marvelous mind, the critics soon hoodwinked him.
>
>
> |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | Regarding the final sentence of the question (as quoted below) related words among others include *[co-opt](http://en.wiktionary.org/wiki/co-opt#English)*,
*[assimilate](http://en.wiktionary.org/wiki/assimilate)*, *[hoodwink](http://en.wiktionary.org/wiki/hoodwink)*, *[cave](http://en.wiktionary.org/wiki/cave#Verb)*, and perhaps *[recant](http://en.wiktionary.org/wiki/recant)*.
>
> It's actually a negative effect, his way of thinking permanently changed due to criticism (and I'm trying to say his original opinion was better).
>
>
>
However, to fit those words into your sample sentence (“he thought this ... but after being criticized, he \_\_ that the critics were correct”) requires adjustments, perhaps as illustrated in the following samples.
>
> • He thought his views were sound but the critics soon co-opted him.
>
> • His views were sound, but he caved under withering criticism.
>
> • Under light but unrelenting peer pressure, he recanted and was assimilated.
>
> • Despite his precise and marvelous mind, the critics soon hoodwinked him.
>
>
> | * *capitulated*
* *acquiesced*
* *agreed begrudgingly*
* *resigned himself to the fact*
* *caved in under pressure* |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | A few options come to mind:
>
> "he thought this... but after being criticized, he \_\_ that the critics were correct"
>
>
>
* realized
* agreed
* concurred
* admitted
* was persuaded | He thought this . . . but after being criticized, he "**acquiesced**" *and agreed* that the critics were correct.
"acquiesce to": To yield or succumb to one's requests or demands. |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | Regarding the final sentence of the question (as quoted below) related words among others include *[co-opt](http://en.wiktionary.org/wiki/co-opt#English)*,
*[assimilate](http://en.wiktionary.org/wiki/assimilate)*, *[hoodwink](http://en.wiktionary.org/wiki/hoodwink)*, *[cave](http://en.wiktionary.org/wiki/cave#Verb)*, and perhaps *[recant](http://en.wiktionary.org/wiki/recant)*.
>
> It's actually a negative effect, his way of thinking permanently changed due to criticism (and I'm trying to say his original opinion was better).
>
>
>
However, to fit those words into your sample sentence (“he thought this ... but after being criticized, he \_\_ that the critics were correct”) requires adjustments, perhaps as illustrated in the following samples.
>
> • He thought his views were sound but the critics soon co-opted him.
>
> • His views were sound, but he caved under withering criticism.
>
> • Under light but unrelenting peer pressure, he recanted and was assimilated.
>
> • Despite his precise and marvelous mind, the critics soon hoodwinked him.
>
>
> | He thought this . . . but after being criticized, he "**acquiesced**" *and agreed* that the critics were correct.
"acquiesce to": To yield or succumb to one's requests or demands. |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | Not a word, but a phrase: "saw the light"
edit (after comment by OP)
***Submit (to)***, then, or "***knuckle under***" | I think you are looking for something like:
gave in,
backed down,
capitulated,
caved in,
conceded,
resigned,
surrendered,
threw in the towel,
washed his hands of,
yielded |
90,284 | I don't feel like this belongs on SU, so I put it here.
I know that "OS X" is pronounced "oh-ess ten," but how should the common construction "OS X 10.9" be pronounced?
The primary possibility I can think of is:
* The X becomes silent: "oh-ess ten-point-nine"
However, this feels awkward when reading from paper and there are suddenly silent words in the middle of sentences. Because of this, I also see as a possibility:
* This construction is unpronounceable and should be read "oh-ess ten mountain lion"
Of course, my personal favorite (the least accurate) is:
* The 10 becomes slient: "oh-ess eks-point-nine" | 2012/11/06 | [
"https://english.stackexchange.com/questions/90284",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/17688/"
] | Not a word, but a phrase: "saw the light"
edit (after comment by OP)
***Submit (to)***, then, or "***knuckle under***" | He thought this . . . but after being criticized, he "**acquiesced**" *and agreed* that the critics were correct.
"acquiesce to": To yield or succumb to one's requests or demands. |
60,107 | Is there any way to install C&C on Windows 7 without the XP VM?
* I tried using the XP virtual mode but I guess my laptop is not compatible with it.
* I tried changing the compatibility in the properties as stated to Win 95 and I tried all the others too, but I still get the error:
>
> The version is not compatible with the version of windows. Check the systems info to see whether you need a 32 bit or 64 bit version of the program and then contact the software provider."
>
>
>
Is there any way I can get it to install and work? | 2012/03/31 | [
"https://gaming.stackexchange.com/questions/60107",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/22852/"
] | Install the patch by nyerguds - <http://nyerguds.arsaneus-design.com/cnc95upd/cc95p106/>
This patch lets you play C&C on newer OS's and with higher resolutions. Also lots of bug fixes in it.....
EDIT: There is a patched installer available for download here with details... <http://digiex.net/downloads/download-center-2-0/games/2929-command-conquer-gold-full-hard-disk-install.html> | The error message you're getting shows you're trying to run a 16-bit executable (for DOS or Windows 3.1) on a 64-bit version of Windows - which isn't possible due to the way 64-bit processors work when running in 64-bit mode.
To the best of my knowledge, there was no Windows 3.1 version of C&C, so unless the Windows 95 version of C&C used a 16-bit installer (it *could* have - some games did, although I don't think this is one of them), that tells me you're trying to play the DOS version.
If that is the case, I recommend you download and install [DOSBox](http://www.dosbox.com/). DOSBox emulates a complete DOS PC, which makes it ideal for running these old games. If you're not used to working with DOS, it might take you a little while to get used to the program. If you need help, the [DOSBox Wiki](http://www.dosbox.com/wiki/Main_Page) contains information on how to use DOSBox and basic DOS commands.
AFAIK, C&C requires the CD to be present, so you may prefer to create ISO images of your CDs and mount the appropriate one in DOSBox prior to starting the game. You can also use an external image mounting tool and simply point DOSBox at that drive. |
60,107 | Is there any way to install C&C on Windows 7 without the XP VM?
* I tried using the XP virtual mode but I guess my laptop is not compatible with it.
* I tried changing the compatibility in the properties as stated to Win 95 and I tried all the others too, but I still get the error:
>
> The version is not compatible with the version of windows. Check the systems info to see whether you need a 32 bit or 64 bit version of the program and then contact the software provider."
>
>
>
Is there any way I can get it to install and work? | 2012/03/31 | [
"https://gaming.stackexchange.com/questions/60107",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/22852/"
] | Install the patch by nyerguds - <http://nyerguds.arsaneus-design.com/cnc95upd/cc95p106/>
This patch lets you play C&C on newer OS's and with higher resolutions. Also lots of bug fixes in it.....
EDIT: There is a patched installer available for download here with details... <http://digiex.net/downloads/download-center-2-0/games/2929-command-conquer-gold-full-hard-disk-install.html> | Just download the No-CD version of the game from the official site: it works in Windows 7 64-bit edition. However, you have to download the movie clips separately. |
159,197 | SP13 doesn't recognize any list item changes when they are made by a workflow. (2010 workflows did)
Therefore does not trigger any list workflows on "item change". Items need manual change to trigger workflows. not good...
Is there any easy way to bypass this?
(i ve found a complicated way by using call http & get dictionary, but that does not help in my case)
thanks | 2015/10/09 | [
"https://sharepoint.stackexchange.com/questions/159197",
"https://sharepoint.stackexchange.com",
"https://sharepoint.stackexchange.com/users/39104/"
] | Workflows are not triggered by system accounts. If your Task item is created under the System Account (or if the account used is flagged "to act as system") then workflow won't fire.
This can happen based on the account used to trigger the workflow directly or if your workflow uses elevated permissions. I think in 2010 workflow this is called "Impersonation step" or similar. | One solution might be to build the first workflow as a 2010 workflow, instead of a 2013 one.
Another might be to enable the Site Feature "Workflows can use app permissions", and if you are a Site Owner (have Full Control on the site), you should be able to use the "App Step" icon to create an Impersonation Step to do your update to the other list, to kick off its workflow. We did not try this one because the first option was very easy and worked for us, since we did not have that many steps to re-create when we realized this.
Additional info: msdn.microsoft.com/en-us/library/jj822159.aspx |
159,197 | SP13 doesn't recognize any list item changes when they are made by a workflow. (2010 workflows did)
Therefore does not trigger any list workflows on "item change". Items need manual change to trigger workflows. not good...
Is there any easy way to bypass this?
(i ve found a complicated way by using call http & get dictionary, but that does not help in my case)
thanks | 2015/10/09 | [
"https://sharepoint.stackexchange.com/questions/159197",
"https://sharepoint.stackexchange.com",
"https://sharepoint.stackexchange.com/users/39104/"
] | Workflows are not triggered by system accounts. If your Task item is created under the System Account (or if the account used is flagged "to act as system") then workflow won't fire.
This can happen based on the account used to trigger the workflow directly or if your workflow uses elevated permissions. I think in 2010 workflow this is called "Impersonation step" or similar. | I had a very similar problem today. My SP13 WF was generating a task. I was using a SP10 WF to update it after a user Edit action on the task, but the SP13 WF wasn’t pickup the change made by the SP10 WF. Though manually updating the item always resulted in the SP13 WF catching that the item was updated. I tried vapcguy approach of enabling the Site feature but that didn’t solve my problem.
Turns out there is a race condition that if the SP10 WF updates the task item too quickly the original WF wont pick up the change. The approach I used to bypass this race condition was to inject a wait state to the SP10 WF by using the Pause for a Duration action. |
118,920 | As a Ubuntu newbie (11.10) I have a wired desktop, and a laptop on Ubuntu, and another windows 7 laptop I use for college...
How do I get them to "talk" to one another, for sharing files... sharing printing... and the like...
my Win laptop does not see the Ubuntu machines, even though they connect to the same FIOS Gateway...
Help??? | 2012/04/04 | [
"https://askubuntu.com/questions/118920",
"https://askubuntu.com",
"https://askubuntu.com/users/53468/"
] | Install samba
* sudo apt-get install samba
* sudo apt-get install samba samba-common
* sudo apt-get install python-glade2
* sudo apt-get install system-config-samba
do a google search on [ubuntu to windows share + samba](http://www.google.com/webhp?hl=en#hl=en&output=search&q=ubuntu%20to%20windows%20share%20%2b%20samba&oq=ubuntu%20to%20windows%20share%20%2b%20samba&fp=976397350e8ab85) | Im not sure but, I don't think you can do it on campus since they have their own private broadcast.
You can try pogoplug (cloud, shared printer. there's also a desktop software for it, which allows you to view the file from "mycomputer"), i-twin USB drive (allows you to share file from one computer to another, it use the memory from your harddrive, not usb). |
61,908 | Not too long ago, I watched a match on stream from #ggChronicle and saw Alistar Headbutting the blue buff golem over a wall. Based on that observation, I want to know if Alistar can Headbutt other objectives over walls, such as Dragon? | 2012/04/10 | [
"https://gaming.stackexchange.com/questions/61908",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/22860/"
] | Alistar cannot headbutt Dragon nor Baron over the wall, for one simple reason. Baron and Dragon are now both immune to Debuffs and Crowd Control effects, so no displacement spell will ever move them.
Alistar might be able to headbutt the double Golems over the thin wall behind them, but I don't think the walls near Wraiths and Wolves are thin enough to do the same with them. | Nowadays, no body can do anything versus Baron or dragon that would move they position, Debuffs, knock up and Crowd Control effects (Besides kite the dragon a little away from his spot). A couple of years ago that was possible.
Check those links:
As you can see, in both of those videos was possible, but they are very old, and RIOT changed it a while ago.
As for Alistar. He can throw even champions over walls, not only Golem and others monsters. But the push is short, basically you can push over walls where you can flash too. Not big walls/forest. |
52,841 | Vikings wearing helmets with horns has been stereotyped by cartoon shows and animated depictions of vikings. Vikings are also portrayed as beasts and savages who looted and raided innocent shrines and institutions. But archeological evidences prove that vikings wore plain helmets and traded a fair bit as well. So where did this pre-conceived notion come from? | 2019/05/27 | [
"https://history.stackexchange.com/questions/52841",
"https://history.stackexchange.com",
"https://history.stackexchange.com/users/37037/"
] | It started with the vikings themselves...
>
> Helmets with horns?
>
>
> Depictions of an Iron Age date exist featuring people with horned helmets/heads, such as upon the Golden Horns. Similar images are also known from the Viking period itself.
>
>
> In the Oseberg burial from Norway, which dates to the early Viking period, a tapestry was found on which horned helmets are also depicted. Does this prove that all Vikings wore the famous helmets with horns? The answer is probably not. However, there is some evidence to suggest that certain warriors wore such headgear. The horned figures on the Golden Horns are berserkers. These were wild warriors, who threw themselves into battle in a trance-like fury. We are also familiar with them from the Icelandic sagas, in which they are amongst the most feared of all Vikings.
>
>
> It is also possible that such headgear was worn for display or for cultic purposes. In a battle situation, horns on a helmet would get in the way. Such helmets would also have caused problems on board the warships, where space was already at a premium. In addition, none of the contemporary sources mention Vikings wearing horned headgear.
>
>
> [Viking helmets](https://en.natmus.dk/historical-knowledge/denmark/prehistoric-period-until-1050-ad/the-viking-age/weapons/helmets/)
>
>
>
The idea that *all* vikings wore these horned helmets is a more recent invention...
>
> The popular image of the strapping Viking in a horned helmet dates back to the 1800s, when Scandinavian artists like Sweden’s Gustav Malmström included the headgear in their portrayals of the raiders. When Wagner staged his “Der Ring des Nibelungen” opera cycle in the 1870s, costume designer Carl Emil Doepler created horned helmets for the Viking characters, and an enduring stereotype was born.
>
>
> [Did Vikings really wear horned helmets?](https://www.history.com/news/did-vikings-really-wear-horned-helmets)
>
>
> | You can blame the early ages of archaeology when differences between the different northern cultures and time periods were not well understood.
Horned helmets did exist in the bronze age, the Vikso helmets from Denmark an obvious example though these weren't known when the early depictions of Vikings with horned helmets were made.
Mycenaean helmets judging from art from the period also had horns, possibly actual animal horns if the helmets made of boar tusks are any indicator.
Such helmets likely derive from early societies where horned animals were venerated. A series of deer skulls with antlers, modified to be used as hats were discovered at Star Carr in Yorkshire dating back to the stone age. Some had holes drilled into them in front of the antlers, presumably as eye holes.
Horned helmets of heroes or gods appear in Anglo-Saxon art such as on some of the panels on the Sutton Hoo helmet. They also appear in later Norse art.
Horned helmets appear in the Iron Age where Samnite and Celtic warriors wore them, probably an archaism from the earlier Bronze Age. They also appear in Celtic art and Roman depictions of Celtic warriors as well as Etruscan art.
There are also Bronze Age helmets from Greece with bronze wings of the Chalcidian type.
So helmets with horns or similar decoration are well attested.
My guess is that the classic Viking horned helmet is based on finds like those from Bronze Age Crete - the Horned God and Ingot God. Both have headresses/helmets with bulls horns. The examples from Viking art tend to have balls on the ends of the horns whereas these Cretan statues have sharp horns which is what we see in the classic Viking depiction.
I suspect the horned helmets depicted on the Golden Horns of Gallehus are actually horned entities similar to the Celtic Cernunnos. One is holding a circle which is similar to the torc Cernunnos holds and both are next to a collection of animals which Cernunnos is also associated with. There are two warriors nearby holding shields showing that the circle held by the horned figure is not a shield but an empty circle. |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | This is perhaps a borderline example. I heard this story from BCnrd a few years ago, so I hope it hasn't been too mangled in my head since then.
One day, Ken Ribet got a phone call in his office:
* "Is this Professor Ribet?"
* "Yeah."
* "Could you tell me what one tenth of one percent means?"
* "One part in a thousand."
* "Are you willing to give that answer under oath?"
It turns out that someone around Berkeley had rented some property, and the rental contract specified that the landlord could not increase the rent by more than one tenth of one percent each year. When the landlord tried to raise it by more, the tenant sued.
At the trial, the tenant's lawyer called in the expert witness. They swore him in, asked him about his job and his qualifications, then came the key question:
* "Professor Ribet, what, in your expert view, is one tenth of one percent?"
* "It's one part in a thousand."
Before he could give a rigorous proof, the landlord's lawyer said, "Objection! Anyone reading the contract can tell that it *obviously* meant one part in 10." The judge agreed, and threw out the case. | This is not exactly an answer to the question but interesting none-the-less.
In 2005 the British model Kate Moss was filmed putting some kind of white powder, allegedly cocaine, up her nose. The police wanted to press charges for possessing illegal drugs. However the pictures could not demonstrate exactly what the substance was, in particular whether it was a Class A or Class B drug, each of which would require different charges to be brought. The British legal system does not allow a person to be tried on a disjunction of two charges so the case was dropped.
Legal reasoning is often different to mathematical reasoning, but I found it amusing to see the courts employing intuitionist logic in this case, whereas mathematicians would usually be satisfied with a classical proof.
Details: <http://news.bbc.co.uk/2/hi/5082546.stm> |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | This is perhaps a borderline example. I heard this story from BCnrd a few years ago, so I hope it hasn't been too mangled in my head since then.
One day, Ken Ribet got a phone call in his office:
* "Is this Professor Ribet?"
* "Yeah."
* "Could you tell me what one tenth of one percent means?"
* "One part in a thousand."
* "Are you willing to give that answer under oath?"
It turns out that someone around Berkeley had rented some property, and the rental contract specified that the landlord could not increase the rent by more than one tenth of one percent each year. When the landlord tried to raise it by more, the tenant sued.
At the trial, the tenant's lawyer called in the expert witness. They swore him in, asked him about his job and his qualifications, then came the key question:
* "Professor Ribet, what, in your expert view, is one tenth of one percent?"
* "It's one part in a thousand."
Before he could give a rigorous proof, the landlord's lawyer said, "Objection! Anyone reading the contract can tell that it *obviously* meant one part in 10." The judge agreed, and threw out the case. | A new book titled *Math on Trial: How Numbers Get Used and Abused in the Courtroom* was published on March 12, 2013 that exclusively seems to address this question. It may not involve "rigorous proofs" but the book contains cases where math was (ab)used.
Here's the blurb from [Amazon](http://rads.stackoverflow.com/amzn/click/0465032923):
>
> In Math on Trial, mathematicians Leila Schneps and Coralie Colmez describe ten trials spanning from the nineteenth century to today, in which mathematical arguments were used—and disastrously misused—as evidence. They tell the stories of Sally Clark, who was accused of murdering her children by a doctor with a faulty sense of calculation; of nineteenth-century tycoon Hetty Green, whose dispute over her aunt’s will became a signal case in the forensic use of mathematics; and of the case of Amanda Knox, in which a judge’s misunderstanding of probability led him to discount critical evidence—which might have kept her in jail. Offering a fresh angle on cases from the nineteenth-century Dreyfus affair to the murder trial of Dutch nurse Lucia de Berk, Schneps and Colmez show how the improper application of mathematical concepts can mean the difference between walking free and life in prison.
>
>
> A colorful narrative of mathematical
> abuse, Math on Trial blends courtroom
> drama, history, and math to show that
> legal expertise isn’t always enough to
> prove a person innocent.
>
>
> |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | This is not exactly an answer to the question but interesting none-the-less.
In 2005 the British model Kate Moss was filmed putting some kind of white powder, allegedly cocaine, up her nose. The police wanted to press charges for possessing illegal drugs. However the pictures could not demonstrate exactly what the substance was, in particular whether it was a Class A or Class B drug, each of which would require different charges to be brought. The British legal system does not allow a person to be tried on a disjunction of two charges so the case was dropped.
Legal reasoning is often different to mathematical reasoning, but I found it amusing to see the courts employing intuitionist logic in this case, whereas mathematicians would usually be satisfied with a classical proof.
Details: <http://news.bbc.co.uk/2/hi/5082546.stm> | *Note: Firstly, I apologize for asking the question without spending substantial amount of time doing research. After posting OP, I found the following resources to be a good starting point. As it turns out there are mostly game-theoretic approaches through [argumentation theory](http://en.wikipedia.org/wiki/Argumentation_theory).*
* [Strategic Argumentation: A Game
Theoretical Investigation](http://espace.uq.edu.au/eserv/UQ%3A13620/icail.pdf)
* [Heuristics in Argumentation:A
Game-Theoretical Investigation](http://www.cs.uu.nl/groups/IS/archive/henry/sartorprob08.pdf)
* [Game Theory and the Law](http://books.google.com/books?id=ncEJHu35yvQC&printsec=frontcover#v=onepage&q&f=false)
* [Argumentation and Game Theory](http://www.mit.edu/~irahwan/docs/chapter2009a.pdf)
* [Fictitious legal cases in *The Case
of the Speluncean Explorers:Nine New
Opinions*](http://www.earlham.edu/~peters/writing/cse.htm) |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | This is perhaps a borderline example. I heard this story from BCnrd a few years ago, so I hope it hasn't been too mangled in my head since then.
One day, Ken Ribet got a phone call in his office:
* "Is this Professor Ribet?"
* "Yeah."
* "Could you tell me what one tenth of one percent means?"
* "One part in a thousand."
* "Are you willing to give that answer under oath?"
It turns out that someone around Berkeley had rented some property, and the rental contract specified that the landlord could not increase the rent by more than one tenth of one percent each year. When the landlord tried to raise it by more, the tenant sued.
At the trial, the tenant's lawyer called in the expert witness. They swore him in, asked him about his job and his qualifications, then came the key question:
* "Professor Ribet, what, in your expert view, is one tenth of one percent?"
* "It's one part in a thousand."
Before he could give a rigorous proof, the landlord's lawyer said, "Objection! Anyone reading the contract can tell that it *obviously* meant one part in 10." The judge agreed, and threw out the case. | In France some roadside apparatus will take a picture of you at point A and another apparatus will take a picture at point B on the same highway. They will then apply the mean value theorem to determine if you deserve a ticket for speeding. |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | For a case which very nearly came to court disputing a claim made by a company to have built a computer chip mathematically proven to meet its specification, take a look at this section of an article by Donald MacKenzie: <http://books.google.co.uk/books?id=rx3oUTzjh8sC&pg=PA134>. It would have been wonderful to have had barristers contest the nature of mathematical proof. Mackenzie writes about the case at greater length in 'Mechanizing Proof: Computing, Risk, and Trust' MIT Press 2001. | *Note: Firstly, I apologize for asking the question without spending substantial amount of time doing research. After posting OP, I found the following resources to be a good starting point. As it turns out there are mostly game-theoretic approaches through [argumentation theory](http://en.wikipedia.org/wiki/Argumentation_theory).*
* [Strategic Argumentation: A Game
Theoretical Investigation](http://espace.uq.edu.au/eserv/UQ%3A13620/icail.pdf)
* [Heuristics in Argumentation:A
Game-Theoretical Investigation](http://www.cs.uu.nl/groups/IS/archive/henry/sartorprob08.pdf)
* [Game Theory and the Law](http://books.google.com/books?id=ncEJHu35yvQC&printsec=frontcover#v=onepage&q&f=false)
* [Argumentation and Game Theory](http://www.mit.edu/~irahwan/docs/chapter2009a.pdf)
* [Fictitious legal cases in *The Case
of the Speluncean Explorers:Nine New
Opinions*](http://www.earlham.edu/~peters/writing/cse.htm) |
101,893 | *Motivation*: Loius Pojman mentions in *What Can We Know?* (2001) of a certain Carneades (ca. 214-129 B.C>) who must have been a "remarkable dialectician"because " in 155BC he was sent on a diplomatic mission to Rome and in his spare time he gave two lectures. On first day he eulogized justice, making a profound impression on his audience. To their amazement on the second day he gave a diatribe against justice, arguing that there were equally good reasons for not adopting it".
Research yielded the [Carneades argumentation framework](https://www.google.com/search?q=carneades+arugmenttation+framekwrok.&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla%3aen-US%3aofficial&client=firefox-a#hl=en&safe=off&client=firefox-a&hs=m13&rls=org.mozilla%3aen-US%3Aofficial&sclient=psy-ab&q=the+carneades+argumentation+framework+filetype+pdf&oq=carneades+arugmentation+framework&gs_l=serp.1.0.0i13i30.4465.8977.2.11395.4.4.0.0.0.0.202.613.0j3j1.4.0...0.0.ABva4QdLgMU&pbx=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=15f4f2ff325d4b5a&biw=1280&bih=885).
However I am interested in the *more recent* ones. Here's one:
>
> **The Mathematical Proof that got a Physicist out of a Traffic Ticket**
>
>
> Dmitri Krioukov, a UC San Diego
> physicist, was recently given a ticket
> for running a stop sign. He went to
> court to argue the ticket, armed with
> a scientific paper that mathematically
> demonstrated that he really had
> stopped. He won.
>
>
> Krioukov has since posted the entire
> paper, rather immodestly called "The
> Proof of Innocence", on the arXiv
> server. It's probably debatable how
> much his ironclad mathematical
> reasoning really helped determine his
> innocence - it's just as likely the
> judge threw out the ticket when it was
> demonstrated another car had
> obstructed the ticketing police
> officer's view. Still, let's take a
> look at...
>
>
>
[Full ezine article](http://io9.com/5902182/the-mathematical-proof-that-got-a-physicist-out-of-a-traffic-ticket)
[Here is the arXiv paper](http://arxiv.org/pdf/1204.0162v2.pdf).
Are there any similar cases?
**EDIT**: [Peter Suber has a description of the case on the famous Paradox of Court here](http://www.earlham.edu/~peters/writing/psa/sec20.htm#A). | 2012/07/10 | [
"https://mathoverflow.net/questions/101893",
"https://mathoverflow.net",
"https://mathoverflow.net/users/20215/"
] | In France some roadside apparatus will take a picture of you at point A and another apparatus will take a picture at point B on the same highway. They will then apply the mean value theorem to determine if you deserve a ticket for speeding. | This is not exactly an answer to the question but interesting none-the-less.
In 2005 the British model Kate Moss was filmed putting some kind of white powder, allegedly cocaine, up her nose. The police wanted to press charges for possessing illegal drugs. However the pictures could not demonstrate exactly what the substance was, in particular whether it was a Class A or Class B drug, each of which would require different charges to be brought. The British legal system does not allow a person to be tried on a disjunction of two charges so the case was dropped.
Legal reasoning is often different to mathematical reasoning, but I found it amusing to see the courts employing intuitionist logic in this case, whereas mathematicians would usually be satisfied with a classical proof.
Details: <http://news.bbc.co.uk/2/hi/5082546.stm> |
390,299 | I've built this PCB which includes two I2C port expanders, an Arduino nano, inputs for buttons and encoders, and headers to plug in an NRF24L01 board.
[](https://i.imgur.com/jYpFGhE.jpg)
Currently, I just have code running to read two encoders plugged into J12 and J11 on this schematic, and send how far each of them have turned over the serial port.
[](https://i.imgur.com/chz3j7Z.png)
This all works perfectly, the problem comes when I plug in the NRF24L01 into the headers on the board. As soon as I do, the arduino disconnects from the PC; I get the little noise from Windows, both the COM port and the device disappear, and it resets the arduino. However, I can still see the TX LED light up, so the code is still running and the data is transmitting over serial, the PC just doesn't see it anymore. As soon as I remove the NRF board, the nano reconnects and works as usual.
I've tried plugging in an external power source and get the exact same behavior, so I don't believe it's an issue of the NRF drawing too much power. I've tried several different NRF boards as well, so it isn't an issue with the specific board I have plugged in.
What could my problem be? | 2018/08/10 | [
"https://electronics.stackexchange.com/questions/390299",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/195610/"
] | Nevermind, I'm an idiot. My problem was unrelated to any power issues. I noticed that 70mA seemed much too high for a chip that's supposed to draw 12mA at the absolute most, and after taking a closer look at the PCB, I realized I swapped the two rows of four pins horizontally, so every pin was connected wrong. After wiring it correctly with jumper cables, the problem is solved and the NRF works fine. | In your design the 3.3V power to NRF24L01 is coming from Arduino Nano USB interface chip, which is FT232RL. The internal FT232 regulator can only supply 50 mA, "Up to 50mA can be drawn from this pin to power external logic if required". the FT232 specs says.
The 50 mA regulator is a pretty weak source. While the spec says the NRF24L01 draws only "11 mA" in transmit", I suspect it draws much more than that in bursts. More, the NRF24L01 requires the power to be bypassed by "high performance RF capacitor" of 4.7 uF. Your board has a junky aluminum electrolytic cap, which likely has no bypass capability at burst frequencies the NRF24L01 operates at.
So essentially you inject huge noise into USB PHY power rail, and the USB signals (or their frequency) are likely got severely distorted, which causes the host to detect massive errors and shut the port down.
To fix the problem you need to place a 10 uF low-ESR ceramic capacitor right between pins of NRF24L01 2x4 header, and better de-couple it from Arduino pin14 (3.3V) with a good ferrite bead or good inductor, to filter the noise from entering the FT232 IC. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | To begin with, the Constitution implicitly assumes in dozens of instances that the president and vice president are separate people:
>
> The executive power shall be vested in a President of the United States of America. He shall hold his office during the term of four years, and, **together with the Vice President**, chosen for the same term, be elected, as follows:
>
>
>
Indeed, the primary purpose of the vice presidency to the ensure there is an acting president in case the president is incapacitated. Having the same person in both posts would defeat the purpose. But perhaps more importantly, if the president is impeached, the vice president would automatically become president according to the 25th Amendment.
>
> In case of the removal of the President from office or of his death or resignation, the Vice President shall become President.
>
>
>
So if the same person could serve as president AND vice president, it would be effectively impossible to impeach the president.
Additionally, the historical meaning of a 'vice' president is someone who acts "[in place of](http://www.etymonline.com/index.php?term=vice-)" a president. The very term implies there's a separate person serving as president. So I think the same person serving as president and vice president would be unconstitutional.
But as to your question about **what would happen if the electors selected the same person as president and vice president**, I think it depends.
### Scenario 1: The President Elect actually tries to claim both offices
I think this would lead to a serious constitutional crisis. The Supreme Court generally shuns resolving "[political questions](http://en.wikipedia.org/wiki/Political_question)" where intervening would involve the Court too intimately in political matters. More to the point, they can only settle cases and controversies based on actual harm. Someone would have to prove that the Super President arrangement caused them material, judicable harm before the Supreme Court.
But keep in mind, a member of the Supreme Court has to swear in the vice president. So if all the justices declined to swear in the president as vice president, that'd be the end of it.
The more practical answer is that the Congress is required to certify the Electors' results. There is precedent for Congress refusing to accept and discarding certain votes, such as when Horace Greeley died [after the 1872 election](http://en.wikipedia.org/wiki/United_States_presidential_election,_1872#Results_and_disputed_votes). The Senate would then have to choose between the two vice presidential candidates with the most remaining electoral votes.
### Scenario 2: The President Elect declines to be a Super President
The president would be sworn in come January, and it would be treated like any other vacancy. The president would then be able to nominate a vice president (again, [under the 25th Amendment](http://en.wikipedia.org/wiki/Twenty-fifth_Amendment_to_the_United_States_Constitution)), who would be have to confirmed by both houses of Congress. The only question would then be if the Congress decided to play along and certify his election as Vice President, or to throw it to the Senate to possibly select their own Vice President. | The Electoral College could do it but it could be also be done by the House and/or Senate.
The Electoral College votes 270 Washington for President, 268 Adams for President. Washington wins the Presidency.
The Electoral College votes 268 Washington for Vice President, 268 Adman for Vice President, 2 Jefferson for Vice President. The Senate has to choose between Washington and Adams. I don't see why the Senate could not choose Washington.
Alternatively, the Electoral College could fail to elect a President and the House of Representatives elects the Vice-President or the Electoral College could fail to elect anyone and the House and Senate vote for the same person. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | To begin with, the Constitution implicitly assumes in dozens of instances that the president and vice president are separate people:
>
> The executive power shall be vested in a President of the United States of America. He shall hold his office during the term of four years, and, **together with the Vice President**, chosen for the same term, be elected, as follows:
>
>
>
Indeed, the primary purpose of the vice presidency to the ensure there is an acting president in case the president is incapacitated. Having the same person in both posts would defeat the purpose. But perhaps more importantly, if the president is impeached, the vice president would automatically become president according to the 25th Amendment.
>
> In case of the removal of the President from office or of his death or resignation, the Vice President shall become President.
>
>
>
So if the same person could serve as president AND vice president, it would be effectively impossible to impeach the president.
Additionally, the historical meaning of a 'vice' president is someone who acts "[in place of](http://www.etymonline.com/index.php?term=vice-)" a president. The very term implies there's a separate person serving as president. So I think the same person serving as president and vice president would be unconstitutional.
But as to your question about **what would happen if the electors selected the same person as president and vice president**, I think it depends.
### Scenario 1: The President Elect actually tries to claim both offices
I think this would lead to a serious constitutional crisis. The Supreme Court generally shuns resolving "[political questions](http://en.wikipedia.org/wiki/Political_question)" where intervening would involve the Court too intimately in political matters. More to the point, they can only settle cases and controversies based on actual harm. Someone would have to prove that the Super President arrangement caused them material, judicable harm before the Supreme Court.
But keep in mind, a member of the Supreme Court has to swear in the vice president. So if all the justices declined to swear in the president as vice president, that'd be the end of it.
The more practical answer is that the Congress is required to certify the Electors' results. There is precedent for Congress refusing to accept and discarding certain votes, such as when Horace Greeley died [after the 1872 election](http://en.wikipedia.org/wiki/United_States_presidential_election,_1872#Results_and_disputed_votes). The Senate would then have to choose between the two vice presidential candidates with the most remaining electoral votes.
### Scenario 2: The President Elect declines to be a Super President
The president would be sworn in come January, and it would be treated like any other vacancy. The president would then be able to nominate a vice president (again, [under the 25th Amendment](http://en.wikipedia.org/wiki/Twenty-fifth_Amendment_to_the_United_States_Constitution)), who would be have to confirmed by both houses of Congress. The only question would then be if the Congress decided to play along and certify his election as Vice President, or to throw it to the Senate to possibly select their own Vice President. | One thing being missed here. The House of Representatives and the Senate must certify the electoral votes. I find it highly unlikely they would certify that someone was elected to both offices in the same election. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | To begin with, the Constitution implicitly assumes in dozens of instances that the president and vice president are separate people:
>
> The executive power shall be vested in a President of the United States of America. He shall hold his office during the term of four years, and, **together with the Vice President**, chosen for the same term, be elected, as follows:
>
>
>
Indeed, the primary purpose of the vice presidency to the ensure there is an acting president in case the president is incapacitated. Having the same person in both posts would defeat the purpose. But perhaps more importantly, if the president is impeached, the vice president would automatically become president according to the 25th Amendment.
>
> In case of the removal of the President from office or of his death or resignation, the Vice President shall become President.
>
>
>
So if the same person could serve as president AND vice president, it would be effectively impossible to impeach the president.
Additionally, the historical meaning of a 'vice' president is someone who acts "[in place of](http://www.etymonline.com/index.php?term=vice-)" a president. The very term implies there's a separate person serving as president. So I think the same person serving as president and vice president would be unconstitutional.
But as to your question about **what would happen if the electors selected the same person as president and vice president**, I think it depends.
### Scenario 1: The President Elect actually tries to claim both offices
I think this would lead to a serious constitutional crisis. The Supreme Court generally shuns resolving "[political questions](http://en.wikipedia.org/wiki/Political_question)" where intervening would involve the Court too intimately in political matters. More to the point, they can only settle cases and controversies based on actual harm. Someone would have to prove that the Super President arrangement caused them material, judicable harm before the Supreme Court.
But keep in mind, a member of the Supreme Court has to swear in the vice president. So if all the justices declined to swear in the president as vice president, that'd be the end of it.
The more practical answer is that the Congress is required to certify the Electors' results. There is precedent for Congress refusing to accept and discarding certain votes, such as when Horace Greeley died [after the 1872 election](http://en.wikipedia.org/wiki/United_States_presidential_election,_1872#Results_and_disputed_votes). The Senate would then have to choose between the two vice presidential candidates with the most remaining electoral votes.
### Scenario 2: The President Elect declines to be a Super President
The president would be sworn in come January, and it would be treated like any other vacancy. The president would then be able to nominate a vice president (again, [under the 25th Amendment](http://en.wikipedia.org/wiki/Twenty-fifth_Amendment_to_the_United_States_Constitution)), who would be have to confirmed by both houses of Congress. The only question would then be if the Congress decided to play along and certify his election as Vice President, or to throw it to the Senate to possibly select their own Vice President. | It's impossible for the electors to elect same person both president and vice-president. According to the Constitution, the electors have to choose a president and vice-president from different states. The same person can only be from one state (they have to choose one residence as a "primary" residence if they have multiple.)
It is possible that some electors will vote for someone for president, and others for vice president. However, it requires a majority of electors (270+) to win. Otherwise, Congress chooses. Hence, the electors can never elect someone both. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | The Electoral College could do it but it could be also be done by the House and/or Senate.
The Electoral College votes 270 Washington for President, 268 Adams for President. Washington wins the Presidency.
The Electoral College votes 268 Washington for Vice President, 268 Adman for Vice President, 2 Jefferson for Vice President. The Senate has to choose between Washington and Adams. I don't see why the Senate could not choose Washington.
Alternatively, the Electoral College could fail to elect a President and the House of Representatives elects the Vice-President or the Electoral College could fail to elect anyone and the House and Senate vote for the same person. | One thing being missed here. The House of Representatives and the Senate must certify the electoral votes. I find it highly unlikely they would certify that someone was elected to both offices in the same election. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | The Electoral College could do it but it could be also be done by the House and/or Senate.
The Electoral College votes 270 Washington for President, 268 Adams for President. Washington wins the Presidency.
The Electoral College votes 268 Washington for Vice President, 268 Adman for Vice President, 2 Jefferson for Vice President. The Senate has to choose between Washington and Adams. I don't see why the Senate could not choose Washington.
Alternatively, the Electoral College could fail to elect a President and the House of Representatives elects the Vice-President or the Electoral College could fail to elect anyone and the House and Senate vote for the same person. | It's impossible for the electors to elect same person both president and vice-president. According to the Constitution, the electors have to choose a president and vice-president from different states. The same person can only be from one state (they have to choose one residence as a "primary" residence if they have multiple.)
It is possible that some electors will vote for someone for president, and others for vice president. However, it requires a majority of electors (270+) to win. Otherwise, Congress chooses. Hence, the electors can never elect someone both. |
4,729 | As I was reading [this answer](https://politics.stackexchange.com/a/4727/2130), I realized that there's nothing in the Constitution to prevent a majority of Electors from voting for the same person for both President and Vice President.
The [Twelfth Amendment](http://en.wikisource.org/wiki/Additional_amendments_to_the_United_States_Constitution#Amendment_XII) says:
>
> The Electors shall meet in their respective states, and vote by ballot for President and Vice-President, one of whom, at least, shall not be an inhabitant of the same state with themselves; **they shall name in their ballots the person voted for as President, and in distinct ballots the person voted for as Vice-President,** and they shall make distinct lists of all persons voted for as President, and of all persons voted for as Vice-President and of the number of votes for each...
>
>
> The person having the greatest Number of votes for President, shall be the President, if such number be a majority of the whole number of Electors appointed...
>
>
> The person having the greatest number of votes as Vice-President, shall be the Vice-President, if such number be a majority of the whole number of Electors appointed...
>
>
>
The only restriction here is that one of the two people each elector votes for must not be from their own state. There is nothing to say that people from other states can't vote for the same person as both President and VP, and it would only take one such vote to enable it to be mathematically possible.
---
Given all this, **what would actually happen if that were the way the vote played out?**
There have been Presidents without a VP before, but as far as I'm aware, there's no provision for the President to *not* also be the VP. Assuming the candidate doesn't immediately resign as VP, would we have the President as tie-breaker vote in the Senate? Would the President be able to appoint anyone he chose as VP? Would there be grounds for the Supreme Court to get involved? | 2014/08/19 | [
"https://politics.stackexchange.com/questions/4729",
"https://politics.stackexchange.com",
"https://politics.stackexchange.com/users/2130/"
] | It's impossible for the electors to elect same person both president and vice-president. According to the Constitution, the electors have to choose a president and vice-president from different states. The same person can only be from one state (they have to choose one residence as a "primary" residence if they have multiple.)
It is possible that some electors will vote for someone for president, and others for vice president. However, it requires a majority of electors (270+) to win. Otherwise, Congress chooses. Hence, the electors can never elect someone both. | One thing being missed here. The House of Representatives and the Senate must certify the electoral votes. I find it highly unlikely they would certify that someone was elected to both offices in the same election. |
210,461 | A limbless creature obviously can't have conventional wings and could suffer many problems in the blastoff.
Of cousrse I know the existance of the flying snakes, which are actually gliders. But I was thinking problem with the flying snake for this model is that they use flexible ribs to expand the surface area and slow down, using flexible ribs, this looks mutually excluyent with have a powered flying with resistant and strong bones (like a bird keel) for muscle attachment to permit the flapping to impulse the flight.
And other problem that this creature could is comparing with all the real flying all animals is that all them have had short bodys, I remember this is because long and flexible bodys or just tails increase the drag resistance of this animals causing that the required energy to fly increase much more.
[](https://i.stack.imgur.com/XyYzL.png)
So how a limbless creature could evolve powered flight, remembering the blastoff and stay in air?
Other things that could import for are the used bones and muscle changes. | 2021/08/19 | [
"https://worldbuilding.stackexchange.com/questions/210461",
"https://worldbuilding.stackexchange.com",
"https://worldbuilding.stackexchange.com/users/80688/"
] | Presenting the "brolly" -- it flies a little like a swimming jellyfish, but it's actually as smart as an octopus...
First, you have ribs like those in an umbrella, light and somewhat flexible. A skin membrane similar to a bat's wing skin covers the "umbrella" surface, and a central "knot" containing the animal's nervous system, digestive tract, reproductive organs, and musculature to be able to either relax the umbrella to a gentle "inverted" cone (to descend in a controllable fashion -- an umbrella is unstable concave down) or contract to produce a "puff" of higher pressure.
Rapidly repeated contractions will propel the creature up (with horizontal motion via tilting) or allow a hover of sorts; relaxation will let it descend near-vertically but at a pretty slow rate, if only there's enough surface area relative to the mass.
Eyes on top and bottom, of course.
This might be something that could evolve from an aquatic octopoid -- they already have most of the shape, and some stiffening cartilage or bone in the tentacles, followed by a few million years of optimizing for stiffness and lightness, might let them transition from "flying" this way to swim, to actually leaving the water and being able to fly for longer and longer periods in the air. | This problem can be approached from the opposite direction. Rather than "How can a limbless creature fly?" one can ask "What conditions favor flight for limbless creatures?" I'm thinking an extremely dense atmosphere. Nowhere near as dense as water (since that would favor the development of fish-like flight, or outright swimming), but dense enough that limbless land animals could gain a significant advantage by slithering/swimming through the atmosphere. |
111,321 | I'm planning to consult a potential future advisor for a postdoc position. However, they do not put up any advertisement for postdocs, thus there is a large possibility that they cannot afford a postdoc. However, even if they cannot afford any postdocs, I still prefer them to read my email, as our work are close in some degree and I do not wanna lose the opportunity for them to point me to someone else. For this reason, is it a good idea to not mentioning postdoc position upfront in the subject and only mention it much later in the body of the email? | 2018/06/16 | [
"https://academia.stackexchange.com/questions/111321",
"https://academia.stackexchange.com",
"https://academia.stackexchange.com/users/16084/"
] | Professors are often very busy. Sometimes they receive more than 100 emails per day, not including automatic messages. An email from a stranger without essential information in the subject field / first line is not likely to be answered. | Consider a title such as: Postdoc on [subject]
Body:
[Greetings]
I am PhD student / postdoc working with/at/on institute/advisor/subject (whichever you judge would interest the advisor). I am interested in your work on [subject] and maybe mention a particularly interesting article. Would you happen to have open postdoc positions or other funding options I could pursue?
I attached my CV and here is a link to a related publication of mine that I am proud of, or my homepage etc.
Maybe mention a concrete research idea here, if you have one.
[End matter]
---
I would not expect the potential advisor to work as a recruitment service, but if they answer positively but do not have positions open, you might ask if they know someone working on related subjects who could hire you.
Also, if you are willing to co-operate even without, or before, getting the position, you might want to mention that and maybe even add it to the subject. |
80,777 | I've got a bathroom sink that has flow that is faster than I would like when the faucet valves are full open. I know I could go get an low-flow aerator to slow the flow but right now I have just adjusted the wall stop valves to throttle the flow to what I want. This seems to work fine. Is there any problem with this as a long term solution? | 2015/12/29 | [
"https://diy.stackexchange.com/questions/80777",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/9443/"
] | Hmm, those were harder to find than expected. It's extremely unlikely that you'll find an exact match (unless you can go back to the manufacturer of your cabinets), but here's a few possibilities.
Hettich: <https://diy.hettich.com/en/products/connecting-fittings/carcase-and-cabinets-connectors.html>
Blum: <http://www.wwhardware.com/blum-surface-mount-rta-fittings-b040-2>
Lee Valley: <http://www.leevalley.com/en/hardware/page.aspx?p=40080&cat=3,41306,41319&ap=1> (Although these don't look like much, they're extremely solid and set with a satisfying thunk. They'll also come apart without wrecking anything.) | Given that multiples of these have already broken, I would recommend NOT replacing them with the same type of hardware. It is obviously too weak for the intended use.
I would recommend these as a possible alternative:
[Keku Clips](http://rads.stackoverflow.com/amzn/click/B0027Z7T64)
[](https://i.stack.imgur.com/bxFUc.jpg) |
183,017 | When i deploy a rails application in production mode, it appends a date-time string as a query param to the end of all the static asset urls. This is to prevent browsers using old-out of date cahed copies of the assets after I redeploy the application.
Is there a way to make rails use the old time stamps for the assets that have not changed (and **only** the ones that have not changed) since the last deployment. I want to do this to prevent users having to redownload those assets that have not changed. | 2008/10/08 | [
"https://Stackoverflow.com/questions/183017",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/7473/"
] | I think you can use ENV['RAILS\_ASSET\_ID'] to alter the cache-busting asset ID. Unfortunately, this is for all assets.
But if it's not set, it uses the asset's source modification time. If that file hasn't been modified since the last time you used it, it shouldn't be a problem.
If the asset ID is changing when they haven't been changed, it might be because of your deployment process altering the modification time and maybe you could look at tweaking that.
Lastly, you can always [override rails\_asset\_id with your own custom method](http://geekblog.vodpod.com/?p=88).
Hope this helps. | @Aupajo I like this, but I think I can take it a bit further. The problem here is that capistrano does create new copies of all the files on every deploy, so all the cache-busting strings are changed. However an MD5 of the file would only change when the file content changes.
Of course, generating an MD5 is expensive and slow, but you can cache the MD5 of a file in memcache, (keyed on the change time, as if the time has changed the MD5 **may** have changed, but if the timestamp has not changed the MD5 **will not** have changed |
39,311,031 | I am an Android developer and I want to make an app which shows users on a map and performs tasks based on their location.
The whole model of the app has to run in the server. I need an API which:
* Receives user location
* Performs calculations based on the location of the users
* Sends response to specific users with the results
The problem is that I have 0 experience in doing server side programming.
Can you please suggest me a way of making the server?
I checked the *Google Colud Platform* and [this](https://www.youtube.com/watch?v=v5u_Owtbfew&feature=player_embedded) video. The video addresses the connection between the app and the server, but what I really need is coding the model and deploying it on the cloud.
What is the way for me to build the API for such an app, as a developer with no server side programming experience?
Can you suggest me a tutorial which goes trough the process of building a cloud backend for a mobile app? | 2016/09/03 | [
"https://Stackoverflow.com/questions/39311031",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/5126604/"
] | This is a very complex question. I don't recommend using a "ready-made" solution like FireBase because it's even harder to transform it into a "proper" API later on when you need it. If you know Android then you know Java, you'll have no problem learning working with a framework like Spring Framework which I recommend. Java on the backend needs a Java Servlet container, like Tomcat. First you should set up a development environment on your machine for this. I recommend searching for Spring Framework tutorials for this i.e. Spring Framework REST tutorials.
Secondly you'll need a database like MySql or MongoDB to store data. Spring comes with ready made connections to most of the most common databases, so its pretty easy to work with them.
When you're ready to deploy your service I recommend using a PAAS like Heroku.com where you can run your service for free first. In this way you get to control everything yourself and you also learn a useful skill. | Since you are coming from the JAVA background you just need some basic building blocks and you shall be good to go.
* Use PASS: They will take any devops from your side which will be a big relief considering you are not familiar with anything on server side. I recommend using Heroku or App Engine. Here's guide to App Engine with JAVA on top. <https://cloud.google.com/appengine/docs/java/>
* Database: Not sure how you are storing and managing your user data as of now, but if you need database, there are various of those available now. The reason Firebase is top suggestion, cause it leverages realtime and gives you control on your side without spending a lot of time on your side.
* APIs: You will be taking user's geo-location and sending it to server. On server you will need to process that into a real world location and any other logic. You will need to use Google's GeoLocation and reverse geocoding APIs for that. Find apis here : <https://console.cloud.google.com/apis/>
Also if it's any help, App Engine and Heroku both offer free limits and should be pretty sufficient for your use case. |
39,311,031 | I am an Android developer and I want to make an app which shows users on a map and performs tasks based on their location.
The whole model of the app has to run in the server. I need an API which:
* Receives user location
* Performs calculations based on the location of the users
* Sends response to specific users with the results
The problem is that I have 0 experience in doing server side programming.
Can you please suggest me a way of making the server?
I checked the *Google Colud Platform* and [this](https://www.youtube.com/watch?v=v5u_Owtbfew&feature=player_embedded) video. The video addresses the connection between the app and the server, but what I really need is coding the model and deploying it on the cloud.
What is the way for me to build the API for such an app, as a developer with no server side programming experience?
Can you suggest me a tutorial which goes trough the process of building a cloud backend for a mobile app? | 2016/09/03 | [
"https://Stackoverflow.com/questions/39311031",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/5126604/"
] | This is a very complex question. I don't recommend using a "ready-made" solution like FireBase because it's even harder to transform it into a "proper" API later on when you need it. If you know Android then you know Java, you'll have no problem learning working with a framework like Spring Framework which I recommend. Java on the backend needs a Java Servlet container, like Tomcat. First you should set up a development environment on your machine for this. I recommend searching for Spring Framework tutorials for this i.e. Spring Framework REST tutorials.
Secondly you'll need a database like MySql or MongoDB to store data. Spring comes with ready made connections to most of the most common databases, so its pretty easy to work with them.
When you're ready to deploy your service I recommend using a PAAS like Heroku.com where you can run your service for free first. In this way you get to control everything yourself and you also learn a useful skill. | Since you have limited experience on server side development I'd suggest using a backend-as-a-service platform. This approach has the benefits of allowing you to focus your time coding the client where your competitive differentiators are, and lets you leverage years of experience in server-side development without having to ramp up on the technology. It also means you can iterate quickly and test features ideas with less risk.
There are several options out there, but in your case, you'll need something that provides enough flexibility to implement custom logic and a custom data model. Some platforms allow you to create a full backend application with REST api endpoints, and automatically take care of deployment, security and management. Some examples of backend service providers can be found here <https://en.wikipedia.org/wiki/Mobile_backend_as_a_service>.
Some of the platforms actually allow you to build a full custom app without actually writing any code, and some of them provide ready-made components (like chat, leaderboards etc) but are less flexible.
Here's a list of Backend-as-a-service providers:
* <https://quickblox.com/>
* <https://www.appcelerator.com/>
* <http://www.configure.it/>
* <https://cloudboost.io/>
* <http://api.shephertz.com/>
* <https://backendless.com/>
* <http://pipegears.com> (No coding required)
You didn't provide specific details about the tasks you want to implement on the backend, but it's likely that you could satisfy your requirements with a straightforward REST api that accepts your location data, runs business logic, reads or writes from your data model and returns whatever data your client needs. (Sorry I can't get more specific without details on your requirements). If you're dealing in sensitive information like location data, I'd strongly recommend securing the api with TLS/SSL (i.e. https), or make sure the platform you use provides this. |
39,311,031 | I am an Android developer and I want to make an app which shows users on a map and performs tasks based on their location.
The whole model of the app has to run in the server. I need an API which:
* Receives user location
* Performs calculations based on the location of the users
* Sends response to specific users with the results
The problem is that I have 0 experience in doing server side programming.
Can you please suggest me a way of making the server?
I checked the *Google Colud Platform* and [this](https://www.youtube.com/watch?v=v5u_Owtbfew&feature=player_embedded) video. The video addresses the connection between the app and the server, but what I really need is coding the model and deploying it on the cloud.
What is the way for me to build the API for such an app, as a developer with no server side programming experience?
Can you suggest me a tutorial which goes trough the process of building a cloud backend for a mobile app? | 2016/09/03 | [
"https://Stackoverflow.com/questions/39311031",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/5126604/"
] | Since you are coming from the JAVA background you just need some basic building blocks and you shall be good to go.
* Use PASS: They will take any devops from your side which will be a big relief considering you are not familiar with anything on server side. I recommend using Heroku or App Engine. Here's guide to App Engine with JAVA on top. <https://cloud.google.com/appengine/docs/java/>
* Database: Not sure how you are storing and managing your user data as of now, but if you need database, there are various of those available now. The reason Firebase is top suggestion, cause it leverages realtime and gives you control on your side without spending a lot of time on your side.
* APIs: You will be taking user's geo-location and sending it to server. On server you will need to process that into a real world location and any other logic. You will need to use Google's GeoLocation and reverse geocoding APIs for that. Find apis here : <https://console.cloud.google.com/apis/>
Also if it's any help, App Engine and Heroku both offer free limits and should be pretty sufficient for your use case. | Since you have limited experience on server side development I'd suggest using a backend-as-a-service platform. This approach has the benefits of allowing you to focus your time coding the client where your competitive differentiators are, and lets you leverage years of experience in server-side development without having to ramp up on the technology. It also means you can iterate quickly and test features ideas with less risk.
There are several options out there, but in your case, you'll need something that provides enough flexibility to implement custom logic and a custom data model. Some platforms allow you to create a full backend application with REST api endpoints, and automatically take care of deployment, security and management. Some examples of backend service providers can be found here <https://en.wikipedia.org/wiki/Mobile_backend_as_a_service>.
Some of the platforms actually allow you to build a full custom app without actually writing any code, and some of them provide ready-made components (like chat, leaderboards etc) but are less flexible.
Here's a list of Backend-as-a-service providers:
* <https://quickblox.com/>
* <https://www.appcelerator.com/>
* <http://www.configure.it/>
* <https://cloudboost.io/>
* <http://api.shephertz.com/>
* <https://backendless.com/>
* <http://pipegears.com> (No coding required)
You didn't provide specific details about the tasks you want to implement on the backend, but it's likely that you could satisfy your requirements with a straightforward REST api that accepts your location data, runs business logic, reads or writes from your data model and returns whatever data your client needs. (Sorry I can't get more specific without details on your requirements). If you're dealing in sensitive information like location data, I'd strongly recommend securing the api with TLS/SSL (i.e. https), or make sure the platform you use provides this. |
332,704 | I have a couple of questions about the practicality of developing a WPF application that loads data into memory from a text file and then manipulates the resulting object(s) instead of transactions with a database such as SQLite. Multiple views would have bindings on the resulting object(s) in memory. Persisting the updates to the text file on disk might only happen at app shutdown.
The data is a ham radio logbook stored as .adi format. ADI is a textual format that stores tabular data in strings, including column names and their values. Rows of data are terminated with an tag. There are numerous apps in existence that parse these .adi files line by line into a database table where it is read and edited by the app. The entire database can then be exported back to an .adi file whereby each table record is serialized into a line of text in the .adi file. Most of them are utilizing some form of database like SQL.
The largest logbook that I'm aware of is about 120,000 records, but a typical user might reach 60,000 to 70,000 in their lifetime. Each record of data could have as many as 30-40 columns including a mixture of string, integer and datetime values.
I'm interested in developing an app that can handle say 100,000 such records without the use of a database solution such as SQLite or MS SQL. An .adi text file of this magnitude should be under 50mb in size on disk. When the app first loads, it would read the text file into an object for binding to a grid. Edits would be applied to the object(s) until app shutdown when they are persisted to the file on disk.
1. Is an object based on a text file approximately 50mb in size likely to be problematic if the app is running on a semi-modern Windows machine with 4gb-8gb of system memory?
2. The ADI standard includes many enumerations. If I were converting this file standard to a relational database I would have numerous lookup tables representing the enumerations. But given the goal of developing without the need for an external database solution, I am thinking of storing the enumerations as class objects in a library. Am I correct that a hard-coded dictionary or enumeration does not occupy any memory space until it is called?
Thanks for your help. | 2016/10/04 | [
"https://softwareengineering.stackexchange.com/questions/332704",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/248513/"
] | >
> Is an object based on a text file approximately 50mb in size likely to be problematic if the app is running on a semi-modern Windows machine with 4gb-8gb of system memory?
>
>
>
Not the object itself. I'm more concerned with the way you plan on implementing the UI. The grid design you propose will require many, many times the amount of memory that holding the actual ADIF will require, and it won't exactly be user-friendly.
>
> Am I correct that a hard-coded dictionary or enumeration does not occupy any memory space until it is called?
>
>
>
You can certainly make such objects lazy-loading. However, I've had a look at the enumerations, and the amount of memory required for them is almost certainly inconsequential anyway.
>
> ...without the use of a database solution such as SQLite or MS SQL
>
>
>
There are many compelling reasons why those other guys built their apps on top of a database system; you get loads of capabilities for free. Having all that log data in a database suddenly makes it useful for all sorts of purposes, not the least of which is filtering, searching, sorting and querying the resulting data. | If you are talking .NET a single object or collection is limited to 2 GB and you don't get all of that as it needs to be contiguous. If it is only 50 MB on disk then you should be OK.
As Robert said - if your design is to display it ALL in one grid that is going to be more memory than the data. You may want to rethink that design.
You can certainly use collection and enumerations to store your data in memory.
You can search collections with LINQ.
I wrote small application where data was read from XML, manipulated in memory, and then written back to XML. But it would have been a lot easier to write it as a database application. Data in XML was a requirement. |
332,704 | I have a couple of questions about the practicality of developing a WPF application that loads data into memory from a text file and then manipulates the resulting object(s) instead of transactions with a database such as SQLite. Multiple views would have bindings on the resulting object(s) in memory. Persisting the updates to the text file on disk might only happen at app shutdown.
The data is a ham radio logbook stored as .adi format. ADI is a textual format that stores tabular data in strings, including column names and their values. Rows of data are terminated with an tag. There are numerous apps in existence that parse these .adi files line by line into a database table where it is read and edited by the app. The entire database can then be exported back to an .adi file whereby each table record is serialized into a line of text in the .adi file. Most of them are utilizing some form of database like SQL.
The largest logbook that I'm aware of is about 120,000 records, but a typical user might reach 60,000 to 70,000 in their lifetime. Each record of data could have as many as 30-40 columns including a mixture of string, integer and datetime values.
I'm interested in developing an app that can handle say 100,000 such records without the use of a database solution such as SQLite or MS SQL. An .adi text file of this magnitude should be under 50mb in size on disk. When the app first loads, it would read the text file into an object for binding to a grid. Edits would be applied to the object(s) until app shutdown when they are persisted to the file on disk.
1. Is an object based on a text file approximately 50mb in size likely to be problematic if the app is running on a semi-modern Windows machine with 4gb-8gb of system memory?
2. The ADI standard includes many enumerations. If I were converting this file standard to a relational database I would have numerous lookup tables representing the enumerations. But given the goal of developing without the need for an external database solution, I am thinking of storing the enumerations as class objects in a library. Am I correct that a hard-coded dictionary or enumeration does not occupy any memory space until it is called?
Thanks for your help. | 2016/10/04 | [
"https://softwareengineering.stackexchange.com/questions/332704",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/248513/"
] | I agree with the other answers here but I want to add a little detail as to why building this all in-memory and persisting to a file can be problematic.
If you were just building an application to view these files, then building a DB for it would probably be overkill. But you want to make edits and then save everything at the end. This approach is problematic:
* If the application crashes (computer shuts down, user forces it down, etc.), all changes since the last time it was opened will be permanently lost.
* If the application crashes during the save or has an bug in the save routine, you will corrupt the file and potentially lose the entire thing.
I've had the displeasure of working with applications that save changes on exit and I hate that about them. I consider such an approach a bad practice. To fix it with a file design, you need to, at the very least, back up the file before you attempt to write it. Then you'll need to be able to save the file either on change or let the user control when the save is done. Saving the entire file could be a little slow but the files are not very big so it might be OK. That's your minimal implementation.
If you want to make this better, though, you'd need to be able to update sections of the file without writing the entire thing every time. Once you get to this point of sophistication, using a database is much easier than trying to manipulate a file in this way. | If you are talking .NET a single object or collection is limited to 2 GB and you don't get all of that as it needs to be contiguous. If it is only 50 MB on disk then you should be OK.
As Robert said - if your design is to display it ALL in one grid that is going to be more memory than the data. You may want to rethink that design.
You can certainly use collection and enumerations to store your data in memory.
You can search collections with LINQ.
I wrote small application where data was read from XML, manipulated in memory, and then written back to XML. But it would have been a lot easier to write it as a database application. Data in XML was a requirement. |
374,967 | I recently started studying set theory and am having quite a bit of difficulty accepting Cantor's diagonal proof for the uncountability of the reals. I also saw a topological proof via nested sets for uncountability which still does not satisfy me completely, given that just like the diagonal it relies on a never ending process. In fact, the nested sets proof sounds very much like the diagonalization proof to me.
Do all proofs of the uncountability of the reals involve diagonalization? Are there any other proofs I can look at to understand? I couldn't find any on searching stack exchange. Thanks. | 2013/04/28 | [
"https://math.stackexchange.com/questions/374967",
"https://math.stackexchange.com",
"https://math.stackexchange.com/users/74780/"
] | The real numbers are a complete densely ordered set without endpoints. That is, there is no minimum, no maximum, between every two points there is a third, and every set which has an upper bound has a *least* upper bound.
>
> **Theorem:** Every countable dense order without endpoints is order-isomorphic to the rational numbers.
>
>
>
Since the rational numbers are not order complete, the real numbers are not order-isomorphic to the rationals. Therefore the real numbers cannot be countable. | The real numbers are a perfect set, and all perfect sets are uncountable. In particular, this gives a proof of the uncountability of real numbers that does not reference decimal expansions. |
87,183 | Berkeley has a god that forces regularity of sensations/perception on all other minds. So a super-mind is the source of the laws of physics. I assume it's the same with any theistic idealism.
But I assume not all idealist philosophers are/were theists. How do non-theist idealists account for laws of physics?
By theist, I'm referring to someone that believes in the existence of a personal god (a god that can be thought of as a person). | 2021/12/01 | [
"https://philosophy.stackexchange.com/questions/87183",
"https://philosophy.stackexchange.com",
"https://philosophy.stackexchange.com/users/18928/"
] | Going by your definition of theism, there are many idealist philosophers who did not need a God with a persona:
1. Monists like Shankara from Advaita Vedanta use Brahman principle as the foundation (though physical entities are just aberrations of this principle)
2. Pantheism of Spinoza uses Substance as the unifying force (physical attributes are modes of Substance)
3. As commented by Conifold, Confucius's inscrutable Dao or Plato's Forms can also serve the purpose
(In all such theories, physical entities and processes are accidental/unintentional/illusory modifications of the unifying principle)
From a strictly atheistic standpoint:
1. Mahayana Buddhism (Yogacara-Vijnanavada) and Sautrantika schools explain inter-subjective reality (which is what natural phenomenon is as per them) as a causal interaction between various mental streams of different subjects [1](https://en.wikipedia.org/wiki/Buddhist_philosophy), [2](https://en.wikipedia.org/wiki/Yogachara)
2. Similarly, Hume (in between idealism and realism) also explains physical laws as based in human instinct rather than objective fact. [3](https://en.wikipedia.org/wiki/David_Hume#Induction_and_causation) | I am not sure I understand this question... is it not true that non-theist idealists, or anyone else for that matter, may account for the laws of physics in any manner they choose? After all, they are not responsible for their content.
Physicists on the other hand do account for laws of physics as this is their responsibility, and the presence or absence of a personal god has no bearing on what the laws of physics can or may contain. |
87,183 | Berkeley has a god that forces regularity of sensations/perception on all other minds. So a super-mind is the source of the laws of physics. I assume it's the same with any theistic idealism.
But I assume not all idealist philosophers are/were theists. How do non-theist idealists account for laws of physics?
By theist, I'm referring to someone that believes in the existence of a personal god (a god that can be thought of as a person). | 2021/12/01 | [
"https://philosophy.stackexchange.com/questions/87183",
"https://philosophy.stackexchange.com",
"https://philosophy.stackexchange.com/users/18928/"
] | Going by your definition of theism, there are many idealist philosophers who did not need a God with a persona:
1. Monists like Shankara from Advaita Vedanta use Brahman principle as the foundation (though physical entities are just aberrations of this principle)
2. Pantheism of Spinoza uses Substance as the unifying force (physical attributes are modes of Substance)
3. As commented by Conifold, Confucius's inscrutable Dao or Plato's Forms can also serve the purpose
(In all such theories, physical entities and processes are accidental/unintentional/illusory modifications of the unifying principle)
From a strictly atheistic standpoint:
1. Mahayana Buddhism (Yogacara-Vijnanavada) and Sautrantika schools explain inter-subjective reality (which is what natural phenomenon is as per them) as a causal interaction between various mental streams of different subjects [1](https://en.wikipedia.org/wiki/Buddhist_philosophy), [2](https://en.wikipedia.org/wiki/Yogachara)
2. Similarly, Hume (in between idealism and realism) also explains physical laws as based in human instinct rather than objective fact. [3](https://en.wikipedia.org/wiki/David_Hume#Induction_and_causation) | Mathematical Platonists like Tegmark see the mathematical structures as the underlying reality.
Consider just as an example to illustrate this kind of thinking, Lisi's proposal for the regularities of physics as resulting from our universe emerging at a location [in the E8 hyperstructure of possible laws](https://en.m.wikipedia.org/wiki/An_Exceptionally_Simple_Theory_of_Everything).
On a deeper level, for physicists & scientists the regularities are the result of the unity of the cosmos, that everything in some sense is made of the same 'stuff' manifesting in different ways, & shares a larger causal continuity. |
87,183 | Berkeley has a god that forces regularity of sensations/perception on all other minds. So a super-mind is the source of the laws of physics. I assume it's the same with any theistic idealism.
But I assume not all idealist philosophers are/were theists. How do non-theist idealists account for laws of physics?
By theist, I'm referring to someone that believes in the existence of a personal god (a god that can be thought of as a person). | 2021/12/01 | [
"https://philosophy.stackexchange.com/questions/87183",
"https://philosophy.stackexchange.com",
"https://philosophy.stackexchange.com/users/18928/"
] | Going by your definition of theism, there are many idealist philosophers who did not need a God with a persona:
1. Monists like Shankara from Advaita Vedanta use Brahman principle as the foundation (though physical entities are just aberrations of this principle)
2. Pantheism of Spinoza uses Substance as the unifying force (physical attributes are modes of Substance)
3. As commented by Conifold, Confucius's inscrutable Dao or Plato's Forms can also serve the purpose
(In all such theories, physical entities and processes are accidental/unintentional/illusory modifications of the unifying principle)
From a strictly atheistic standpoint:
1. Mahayana Buddhism (Yogacara-Vijnanavada) and Sautrantika schools explain inter-subjective reality (which is what natural phenomenon is as per them) as a causal interaction between various mental streams of different subjects [1](https://en.wikipedia.org/wiki/Buddhist_philosophy), [2](https://en.wikipedia.org/wiki/Yogachara)
2. Similarly, Hume (in between idealism and realism) also explains physical laws as based in human instinct rather than objective fact. [3](https://en.wikipedia.org/wiki/David_Hume#Induction_and_causation) | Plato considered ideas existing in the immaterial realm of forms. Similarly, many Platonists in the realm of mathematics and physics, [consider ideas residing in, for instance, possibility space](https://www.youtube.com/watch?v=QBSwfaKfTMk).
Schopenhauer, building on Plato, held a similar view that such ideas and laws exist eternally outside space and time and only objectify (or instantiate) themselves in the world through the craving of the Will (the thing-in-itself).
Laws appear to materialise themselves in the world during certain circumstances but are themselves eternal blueprints that exist outside space and time. |
87,183 | Berkeley has a god that forces regularity of sensations/perception on all other minds. So a super-mind is the source of the laws of physics. I assume it's the same with any theistic idealism.
But I assume not all idealist philosophers are/were theists. How do non-theist idealists account for laws of physics?
By theist, I'm referring to someone that believes in the existence of a personal god (a god that can be thought of as a person). | 2021/12/01 | [
"https://philosophy.stackexchange.com/questions/87183",
"https://philosophy.stackexchange.com",
"https://philosophy.stackexchange.com/users/18928/"
] | Going by your definition of theism, there are many idealist philosophers who did not need a God with a persona:
1. Monists like Shankara from Advaita Vedanta use Brahman principle as the foundation (though physical entities are just aberrations of this principle)
2. Pantheism of Spinoza uses Substance as the unifying force (physical attributes are modes of Substance)
3. As commented by Conifold, Confucius's inscrutable Dao or Plato's Forms can also serve the purpose
(In all such theories, physical entities and processes are accidental/unintentional/illusory modifications of the unifying principle)
From a strictly atheistic standpoint:
1. Mahayana Buddhism (Yogacara-Vijnanavada) and Sautrantika schools explain inter-subjective reality (which is what natural phenomenon is as per them) as a causal interaction between various mental streams of different subjects [1](https://en.wikipedia.org/wiki/Buddhist_philosophy), [2](https://en.wikipedia.org/wiki/Yogachara)
2. Similarly, Hume (in between idealism and realism) also explains physical laws as based in human instinct rather than objective fact. [3](https://en.wikipedia.org/wiki/David_Hume#Induction_and_causation) | My three cents.
Whatever one takes the *Ultimate* to be, by the mere fact of being the *Ultimate* is necessarily the foundation of all the universe and all reality. That is it.
One does not necessarily have to explain how a certain fact (eg law, being, thing, ..) emerges from the *Ultimate*, as in principle it cannot but be traced back to the *Ultimate* itself (by definition of *Ultimate*).
That is why any philosophy that adheres to some *Ultimate* does not have to reduce everything conceivable to that *Ultimate* explicitly in order to be discussed. Being reducible "*in principle*" is enough for most purposes of philosophical discussion.
That *Ultimate* may be:
1. A personal deity or deities (eg Christian God, Muslim God, ..)
2. An principle or principles (eg the Dao, the Brahman, the Geist, the Forms, the Laws, Mind, Matter, ..)
3. The sum of everything existing (eg the universe itself)
4. so on..
All three candidates for being the *Ultimate* above, can easily accommodate physical laws (a detailed account of each is outside the scope of this answer). |
230,471 | For reasons I'd rather not get into, I want to permanently disable my shield, so unlimited people can attack me one after another. Is this possible to do? | 2015/08/02 | [
"https://gaming.stackexchange.com/questions/230471",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/120208/"
] | Not really possible, but if you wanted to just continuously disable it, launch an attack every time you get raided and they get 50% or your town hall. You could just go with a few barbs/archers and look for a farmer. There isn't a way to disable your shield. | **2016 UPDATE**
While you cant technically disable your shield, you can simply dismiss it every time you have one by clicking on the "-" button on the upper-right part of the active shield icon. Just take note that removing all your shields (including village guard) will affect your total remaining online time. |
11,658,480 | I have made a website with Visual Studio 2010. I have a simple Fortran .dll that I want to call from my project. I have added this DLL in my project but when I run it I get the error:
>
> Unable to load DLL 'Dll1.dll': The specified module could not be found.
>
>
>
which probably means that the DLL can not be found. Do I have to add the .dll somewhere specifically or add the location of the DLL to the code? | 2012/07/25 | [
"https://Stackoverflow.com/questions/11658480",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1365504/"
] | I think using a head request might get you started along the right path.
check this thread:
[HTTP HEAD Request in Javascript/Ajax?](https://stackoverflow.com/questions/333634/http-head-request-in-javascript-ajax)
if you're using jquery, it's a simple type change in the ajax call:
$.ajax({url: "some url", type: "HEAD".....});
personally, I would check the file size & date modified against the previous response, and fetch the new data if it has been updated. I'm not sure if you can fetch only parts of a file via ajax, but I'm sure this can be accomplished via php pretty easily, possibly this thread may help:
[How to read only 5 last line of the text file in PHP?](https://stackoverflow.com/questions/2961618/how-to-read-only-5-last-line-of-the-txt-file) | It depends how your program is made and how does it print your data, but you can use timestamps to reduce the amount of data. If you have some kind of IDs, you should probably use them insteam of timestamps. |
51,860 | When trying to install the Xbox 360 version of *The Darkness II*, to the hard drive, the system does nothing and the game never gets installed.
The install appears to start, but the screen stays at the Installing screen at 0% progress indefinitely, all while showing the spinning wait symbol. I left it for 30 minutes and no change was made, and tried again multiple times. Attempting to then exit out back the Xbox Home fails, and cancelling the install also fails. The only option other than turning the system off that I could find, is to hit the Guide button and start the game itself, or starting another game. The game does play fine so far, and the disc doesn't appear to be scratched in any way.
I did update the game to the latest release when starting it directly from the disc, but that didn't change anything. Is there any fix for this hard drive intall problem, or is this a limited problem with an unknown resolution? | 2012/02/21 | [
"https://gaming.stackexchange.com/questions/51860",
"https://gaming.stackexchange.com",
"https://gaming.stackexchange.com/users/4239/"
] | Usually that means a problem with Xbox Live service. Again, usually, installing while signed out works.
Judging by the link Dean posted, that is the case with Darkness II. | Just found the answer to our problem
<http://www.gamefaqs.com/boards/621308-/61907326> |
352,754 | Actually, in the mathematics sine and cosine functions are defined based on right angled triangles. But how will the representation of a wave or signal say based on these trigonometric functions (we can't draw any right angled triangles in the media, i.e., the air) then how can we say that? | 2017/08/20 | [
"https://physics.stackexchange.com/questions/352754",
"https://physics.stackexchange.com",
"https://physics.stackexchange.com/users/166425/"
] | Mathematics has progressed from geometry to calculus and differential equations. It is established that differential equations whose solutions describe waves have [sinusoidal functions](https://en.wikipedia.org/wiki/Sine_wave) in those solutions. This should not be surprising, as waves are periodic in time or space, and sines and cosines are periodic functions. As a consequence differential equations that have sinusoidal solutions are called wave equations. See also [this link](https://en.wikipedia.org/wiki/Sinusoidal_plane-wave_solutions_of_the_electromagnetic_wave_equation) . | Sines, as functions of time, are not just features of geometric interest. The behavior of some time-dependent electrical components (capacitors, inductors) are simple (linear, in a sense) if and only if one excites those components with a known signal frequency, such as the function sine(omega \* t + phi).
When 'phi' = pi/2, that's cosine(omega \* t).
Much signaling is done with modulated narrow-band signals, because the
narrow band allows for efficient rejection of noise. Noise, unlike a signal,
is present at ALL frequencies simultaneously, thus is mainly outside
the sensitive frequency range of a narrow-band receiver. Radio, television, hard disk
magnetization patterns, Ethernet, Wi-Fi... all the best modern communication
channels rely on frequency-selection to some extent. Understanding
those methods, and the inductors and capacitors that implement them, requires a selected-frequency test signal for each
calculation. A pure frequency being a sine(omega \*t + phi).
This implies a lot of discussion of sinewaves, generators of sinewaves,
and graphs against frequency (meaningful only after decomposition of a complex function into... a multiplicity of sinewaves). |
5,274,449 | I am trying to create a Colorpicker control for my application. I prefer pure actionscript. Does someone know how to create a picker like this: <http://parasolarchives.com/tools/colorpicker/>
Most interestingly I am interested how to draw the gradient because the gradient has a saturation, | 2011/03/11 | [
"https://Stackoverflow.com/questions/5274449",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/609997/"
] | AJAX runs clientside, and the code can be modified by any client that executes it. You should not have a direct connection to your database using AJAX, always put a server-side script in between to protect your database.
Even if it were technically possible, you **should not be doing this**. | AJAX can communicate back to the web server. The web server can then communicate to the db. No direct connection between AJAX and the db. Exposing your db is probably a security risk anyway. |
5,274,449 | I am trying to create a Colorpicker control for my application. I prefer pure actionscript. Does someone know how to create a picker like this: <http://parasolarchives.com/tools/colorpicker/>
Most interestingly I am interested how to draw the gradient because the gradient has a saturation, | 2011/03/11 | [
"https://Stackoverflow.com/questions/5274449",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/609997/"
] | AJAX runs clientside, and the code can be modified by any client that executes it. You should not have a direct connection to your database using AJAX, always put a server-side script in between to protect your database.
Even if it were technically possible, you **should not be doing this**. | No, AJAX can't communicate with MySQL directly. You will always need a driver for MySQL. [CouchDB](http://couchdb.apache.org/) can be accessed by HTTP, so with CouchDB it is possible! (CouchDB+Jquery=[jqCouch](http://plugins.jquery.com/project/jqcouch)) |
5,274,449 | I am trying to create a Colorpicker control for my application. I prefer pure actionscript. Does someone know how to create a picker like this: <http://parasolarchives.com/tools/colorpicker/>
Most interestingly I am interested how to draw the gradient because the gradient has a saturation, | 2011/03/11 | [
"https://Stackoverflow.com/questions/5274449",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/609997/"
] | AJAX runs clientside, and the code can be modified by any client that executes it. You should not have a direct connection to your database using AJAX, always put a server-side script in between to protect your database.
Even if it were technically possible, you **should not be doing this**. | As others have already said, you can't connect to MySQL from JavaScript directly.
If you want a database that can do that then you can try [CouchDB](http://couchdb.apache.org/). In fact, this is how [Futon](http://couchdb.apache.org/screenshots.html) works. |
3,094,941 | I have one [Zen Cart](http://en.wikipedia.org/wiki/Zen_Cart) website in the English language. I want to make the same website in French and Italian languages. How can I do this?
I mean, if I install language packs for Zen Cart, will this convert all the site and data inside the database in that language or do I have to type data again in particular language?
What is the URL to download the above language packs? | 2010/06/22 | [
"https://Stackoverflow.com/questions/3094941",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/172319/"
] | For downloading the Zen Cart language packs, open <http://www.zen-cart.com/index.php?main_page=index&cPath=40_46>. | For download language pack, visit here: first link.
There is all language pack that have this cms, after this read this topic in main forum of cms: second link
I think that you know how upload your language pack in your web server...
<http://nopaste.voric.com/paste.php?f=227xgw> |
3,094,941 | I have one [Zen Cart](http://en.wikipedia.org/wiki/Zen_Cart) website in the English language. I want to make the same website in French and Italian languages. How can I do this?
I mean, if I install language packs for Zen Cart, will this convert all the site and data inside the database in that language or do I have to type data again in particular language?
What is the URL to download the above language packs? | 2010/06/22 | [
"https://Stackoverflow.com/questions/3094941",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/172319/"
] | Most general Zen Cart language text is held in files for versions up to and including Zen Cart 1.5. So installing the language packs for those languages will deal with a lot of your needs for French and Italian. But not all of it.
Category and Product information is held in the database and installing a language pack will not automatically translate that to a different language, you will need to arrange that yourself.
Zen Cart also allows you to insert blocks of HTML (through the define\_pages editor) such as on the home page, or explanatory text on the shopping cart page. If you have used text in these HTML blocks, then you will need to translate that too.
A major exception on many sites would be pages created using Zen Cart's built in CMS, EZ-Pages. If you've done this, you will need to install the Multilingual EZ-Pages add-in module and then translate the pages created this way.
One final, and rare exception, would be if you have created any HTML banners. Most people don't so if you're not sure what this means, it probably doesn't affect you. | For download language pack, visit here: first link.
There is all language pack that have this cms, after this read this topic in main forum of cms: second link
I think that you know how upload your language pack in your web server...
<http://nopaste.voric.com/paste.php?f=227xgw> |
3,094,941 | I have one [Zen Cart](http://en.wikipedia.org/wiki/Zen_Cart) website in the English language. I want to make the same website in French and Italian languages. How can I do this?
I mean, if I install language packs for Zen Cart, will this convert all the site and data inside the database in that language or do I have to type data again in particular language?
What is the URL to download the above language packs? | 2010/06/22 | [
"https://Stackoverflow.com/questions/3094941",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/172319/"
] | For downloading the Zen Cart language packs, open <http://www.zen-cart.com/index.php?main_page=index&cPath=40_46>. | Most general Zen Cart language text is held in files for versions up to and including Zen Cart 1.5. So installing the language packs for those languages will deal with a lot of your needs for French and Italian. But not all of it.
Category and Product information is held in the database and installing a language pack will not automatically translate that to a different language, you will need to arrange that yourself.
Zen Cart also allows you to insert blocks of HTML (through the define\_pages editor) such as on the home page, or explanatory text on the shopping cart page. If you have used text in these HTML blocks, then you will need to translate that too.
A major exception on many sites would be pages created using Zen Cart's built in CMS, EZ-Pages. If you've done this, you will need to install the Multilingual EZ-Pages add-in module and then translate the pages created this way.
One final, and rare exception, would be if you have created any HTML banners. Most people don't so if you're not sure what this means, it probably doesn't affect you. |
25,125 | In pictures of Iron Man 3 we see Tony has a room with what I am guessing are old armors. What are they all?

Here are two other angles:

 | 2012/10/24 | [
"https://scifi.stackexchange.com/questions/25125",
"https://scifi.stackexchange.com",
"https://scifi.stackexchange.com/users/6195/"
] | They are all the past models of Iron Man armor, or at least placeholders for them. At the end of *Iron Man 2*, it's implied that Rhodey kept the War Machine armor, which was the Mark II, and yet there's an armor that looks like the Mark II in that lineup. From left to right, they are:
1. Mark I - This armor was the first Iron Man suit. Stark and Yin-sen built this in a cave while imprisoned by the Ten Rings in *Iron Man*. The suit was used to escape captivity and was destroyed in the process. The Ten Rings recovered the wreckage and it ended up in Obidiah Stane's possession. From there, Tony must have recovered it after the end of *Iron Man*.
2. Mark II - This was the first suit Tony built upon returning from captivity in *Iron Man*. It's a basic Iron Man suit, with minimal weaponry and decoration. In *Iron Man 2*, Rhodey takes the armor and has it outfitted with Hammer weaponry.
3. Mark III - The final suit of armor built during the film *Iron Man*. This model has the familiar red and gold design, with additional icing protection built in. This model also has additional weaponry that the Mark II lacked (i.e stunners, unibeam, wrist rocket). This armor was damaged during the climactic battle at the end of *Iron Man* against Stane's "Iron Monger" armor.
4. Mark IV - Not much is known about this armor. This was the armor that was worn by Tony at the start of *Iron Man 2*, so it was designed and built between the films. He's shown wearing normal clothing underneath this armor, a new feature as the previous armors have required a jumpsuit to be worn underneath. It isn't shown with any significant new weaponry over the Mark III, and since it is replaced in the film by the Mark VI, we'll never know what made this armor special.
5. Mark V - This was the suitcase armor that Tony used in *Iron Man 2* during the battle on the race track. Like the Mark IV, this was designed and built between films. This armor seems to be a more minimal design, prioritizing size and weight over firepower. During his battle with Vanko, the only weapons used were the hand repulsors, so it's possible that's the extent of this suit's weaponry.
6. Mark VI - Tony built this suit during *Iron Man 2* to make use of his new, palladium-free RT unit. It is similar in design to the Mark IV, but has a triangular hole in the chest plate, instead of the traditional round hole. It has numerous new weapons which Stark made use of during the climactic battle in *Iron Man 2*, such as the one-use wrist lasers. This is also the armor used for the first half of *The Avengers*.
7. Mark VII - After the Mark VI sustained heavy damage during the events of *The Avengers*, Stark built this armor. Its design brings back the circular chest plate hole. On this armor's back are supplementary thrusters, which grant the armor additional stability in flight. This armor also adds a new remote-deploy capability, where the armor is capable of being launched and then it seeks out Tony and attaches to him. Since there are 7 suits of armor in the images, it seems this is the latest armor in *Iron Man 3*, until (based on the trailers) Tony designs the new gold on red armor.
8. Unknown new *Iron Man 3* armor - This is the armor we see in one of the shots above in the central podium. This is a notable design change from the previous armors, as it has a gold with red highlights styling, instead of the traditional red with gold highlights. From the first *Iron Man 3* trailer, and some concept art released during Comic-Con, it seems to be a variation of the *Extremis* armor. In the comics,
>
> the *Extremis* story arc featured Tony building a new suit of armor that could be mentally remote controlled by Tony. The armor moves via additional repulsors on each piece of armor, which allows Tony to fly the pieces of armor onto himself. The *Extremis* storyline has this made possible by Tony becoming a cyborg, but the trailer doesn't make it clear if this aspect will be part of *Iron Man 3*.
>
>
> | **They are the Iron Man armors in technological progress from the Mark I to the Mark VII.**
* Iron Man Armor Mark I - far left
* through Iron Man Armor Mark VII - far right
Mark VIII is the newest golden armor in the center panel.
>
> The Iron Man Armor is arguably one of the most powerful forms of technology developed by Tony Stark.
>
>
> * Alone, the Mark I was sufficient enough to allow him to go through armed terrorists in order to escape.
> * The Mark II was sufficiently powerful enough to go toe to toe with a more advanced and updated version of the armor (the Mark IV).
> * The Mark III was able to battle the Iron Monger suit, however, with some difficulty (due to the fact that it was powered by the first generation mini-arc reactor and the Iron Monger was powered by the second arc reactor he made upon his return).
> * The Mark V, while lacking weaponry and flight capabilities, was still able to take on Vanko while just utlizing just the Whiplash harness.
> * The Marks VI and VII proved to be the most powerful generation suits used by Stark. They were powerful enough to battle hordes of Hammer drones, Vanko with his own armor, survive and battle two Asgardian gods (Thor and Loki, respectively) and hordes of alien foot soldiers and large transports (the Levithan).
>
>
> |
150,309 | I managed to read the first chapter of this comic about a year ago, then I lost the tab. It starts with a girl with a tail surveying a landscape which she seems to think is (or will be) hers. It's not clear whether she was actually royal, but she acts like she is, and the uncle she was being escorted to seemed like an important guy.
She argues constantly with her surly zombie escort. She's bad at directions and leads him on some detours. Eventually, she manages to piss of some giant tree elemental sort of thing (I think she panicked and shot a fireball at it or something). Guide kept trying to deescalate because apparently the things are normally peaceful and if she would just stop provoking it they wouldn't have to kill it. She doesn't, and they do, and I think they find a bunch of old stuff that it had eaten. They continue on their trip and the chapter ends.
Then we cut to chapter two, which from the first page seemed to be about some dudes who were smuggling something, but I'm not at all sure about that.
The art was very colorful and looked pretty professional, but still a cartoony rather than realistic style, and not very anime-ish. Not really sure how to describe it beyond that. | 2017/01/16 | [
"https://scifi.stackexchange.com/questions/150309",
"https://scifi.stackexchange.com",
"https://scifi.stackexchange.com/users/73618/"
] | Could be [Unsounded](http://www.casualvillain.com/Unsounded/).
The girl is indeed very annoying to her guide.
It's been a while since I read it but I do seem to remember something about a tree elemental as well.
Smugglers do appear as well though I'm not sure if it's in the second chapter already.
The girl also has a tail. The guide is a zombie. | This is probably not the right one, but a Google search for princesses and trees and fireballs turned up *[Kagerous](http://www.kagerou.org/01/index.html)* which starts with a female protagonist in a tree with an insect guardian, and who throws around fire. It is very colorful. That said, her guide is more nonverbal, and chapter 2 does not open with smugglers. And [the protagonist may be an androgynous male](http://tvtropes.org/pmwiki/pmwiki.php/Webcomic/Kagerou) with multiple personalities.
[](https://i.stack.imgur.com/tPnvS.jpg) |
1,955 | Is there an easy to install Apache graphing tool for Linux? Something which requires minimal setup and configuration? I'm having problems with slow loading times on a LAMP application and would like to check how Apache is performing.
**Requirements**
* Must work on Centos 5.
* Minimal configuration needed.
* Should graph usage to help pinpoint problems like max connections exceeded, etc.
**Nice to haves**
* Create own logs rather than working from just Apache logs | 2014/03/04 | [
"https://softwarerecs.stackexchange.com/questions/1955",
"https://softwarerecs.stackexchange.com",
"https://softwarerecs.stackexchange.com/users/1204/"
] | **[Monitorix](http://www.monitorix.org/ "Monitorix Homepage")** might come in handy here. It does not exactly *focus* on Apache, but of course includes Apache-specific data:
[](https://i.stack.imgur.com/YeJ7g.png)
Monitorix: Apache Statistics (source: [Monitorix.Org](https://i.stack.imgur.com/YeJ7gm.png); click image for larger variant)
As the graph shows, you can easily see a "journal" of Apache resource usage: idle/busy workers, requests, CPU-usage. Together with plenty of other statistics (see [here](http://www.monitorix.org/screenshots.html "Monitorix Screenshots") for more screenshots; there can be other reasons for your "slow loading times", e.g. another process hogging all resources), this gives a pretty good image on your server's performance. Statistics are saved using [RRDTool](http://en.wikipedia.org/wiki/RRDtool "Wikipedia: RRDtool"), and the web front-end allows you to select different intervals to view the stored data.
*Monitorix* is [available for download](http://www.monitorix.org/downloads.html "Monitorix Downloads") for multiple systems, and easily installed on most Linux distros directly via their package management ([APT](http://en.wikipedia.org/wiki/Advanced_Packaging_Tool "Wikipedia: Advanced Packaging Tool") on Debian and its derivates, [RPM](http://en.wikipedia.org/wiki/RPM_Package_Manager "Wikipedia: RPM Package Manager") for RedHat/Mandriva/SuSE etc). For more details, please also see my answer [here](https://softwarerecs.stackexchange.com/a/469/185 "SR: Webbased Network analyzer").
I'm using *Monitorix* on multiple servers, and it saved my day more than once. The dev is quite busy implementing new features, and support is great also on its IRC channel in case you need help quickly.
Disclosure: I'm the maintainer for the *Monitorix* Debian/Ubuntu packages and the corresponding [repository](http://apt.izzysoft.de/ubuntu/dists/generic/ "Debian/Ubuntu repository at IzzySoft"). | Are you looking at how Apache is performing using system resources, or how apache is performing serving web pages? I am unfamiliar with Apache, but if errors like connections exceeded are logged in the Apache logs, perhaps something like Logstash sending to an ElasticSearch db, then viewed with the Kibana front end?
I've used this for Windows IIS logs, and I could graph things like http response codes, response times, bytes sent/received etc, all over a customiseable time.
There is some overhead setting up the grok filter within the logstash agent, but after you get that done, it just kind of works.
edit adding links -
logstash - <http://www.logstash.net>
elasticsearch and kibana - <http://www.elasticsearch.org/overview/kibana/> |
1,854 | I'm making a mobile (in the Calder sense) by laser-cutting from a vector file (Corel Draw). I have a set of outlines that will be the individual pieces.
Is there a way in Corel (or some other program) to calculate the center of mass so I place the holes to suspend the pieces? | 2011/05/01 | [
"https://graphicdesign.stackexchange.com/questions/1854",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/1213/"
] | This answer may not count if you require a strictly programmatic solution, but I found it interesting so I thought I'd share. :)
The math way would be to break the shape into a collection of triangles and do a weighted average of their individual centers of gravity. I don't think any of the popular drawing programs will automate that for you. It can be done manually but it may not be worth your time.
This site has a breakdown of a physical approach that may be a faster yield on your time: <http://www.scn.org/~bh162/center_of_mass.pdf>
The basic premise is that objects with uniform thickness can be determined using 2-3 plumb lines from random points at or near the edges of your shape.
1. Print & cut out a copy of your shape.
2. Select 3 points somewhat close to different edges on your object.
3. Use a pin or nail to poke holes in your 3 points. The holes should end up slightly larger than the diameter of the implement so that the cutout can swing freely.
4. Hang your object using the pin through one of the holes.
5. While your object is hanging, use a string and weight to make a plumb line and trace the line.
6. Repeat the hang & plumb line for the other holes.
7. The point where all your lines intersect is the center of gravity for your shape.
8. If your lines are precise, you should be able to determine an accurate center with 2 lines and the 3rd line is just for additional verification. | I export to a very large PNG file of 3000 pixels. Then I drag and drop into Algodoo (which is now free). Put the axle where you want and play the simulation and it will show you where it's off balance. You can then adjust your drawing until it's balanced where you want. |
1,854 | I'm making a mobile (in the Calder sense) by laser-cutting from a vector file (Corel Draw). I have a set of outlines that will be the individual pieces.
Is there a way in Corel (or some other program) to calculate the center of mass so I place the holes to suspend the pieces? | 2011/05/01 | [
"https://graphicdesign.stackexchange.com/questions/1854",
"https://graphicdesign.stackexchange.com",
"https://graphicdesign.stackexchange.com/users/1213/"
] | This answer may not count if you require a strictly programmatic solution, but I found it interesting so I thought I'd share. :)
The math way would be to break the shape into a collection of triangles and do a weighted average of their individual centers of gravity. I don't think any of the popular drawing programs will automate that for you. It can be done manually but it may not be worth your time.
This site has a breakdown of a physical approach that may be a faster yield on your time: <http://www.scn.org/~bh162/center_of_mass.pdf>
The basic premise is that objects with uniform thickness can be determined using 2-3 plumb lines from random points at or near the edges of your shape.
1. Print & cut out a copy of your shape.
2. Select 3 points somewhat close to different edges on your object.
3. Use a pin or nail to poke holes in your 3 points. The holes should end up slightly larger than the diameter of the implement so that the cutout can swing freely.
4. Hang your object using the pin through one of the holes.
5. While your object is hanging, use a string and weight to make a plumb line and trace the line.
6. Repeat the hang & plumb line for the other holes.
7. The point where all your lines intersect is the center of gravity for your shape.
8. If your lines are precise, you should be able to determine an accurate center with 2 lines and the 3rd line is just for additional verification. | I was stuck with the same dilemma and using Adobe Illustrator wrote a script that places a small hole at the center of gravity of an irregular polygon. It worked great! After this iteration, I also added a few snippets to create the hole just above the center of gravity to ensure that the part balanced well when suspended from just that one point.
Downloadable script: [CentroidFinder.js](https://gist.github.com/scottswaaley/afec2645953a910c02616dfe7a51543f)
Blog Post: [Finding the Center of Gravity (Centroid) of a Polygon in Adobe Illustrator](http://gritlab.org/finding-the-center-of-gravity-centroid-of-a-polygon-in-adobe-illustrator/) |
36,098,149 | I am trying to create a game but before that i am creating **all the graphics that will be used in it** but here is the problem.
Problem
-------
The current games **graphics size is 50MB** and might increase as all the graphic work is not done yet and it is not possible to load 50 MB every time when the user runs the game. So, please tell me this if it is possible to **downloaded all the graphics on the first run and use it again when the user runs the game again**?
If yes then please help me by giving a piece of code snippet or any article related to it. Platform : ASP.NET | 2016/03/19 | [
"https://Stackoverflow.com/questions/36098149",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/5143387/"
] | Take advantage of all the different caching mechanisms:
1. Http Caching. You may read more about it [here](http://www.mobify.com/blog/beginners-guide-to-http-cache-headers/). Especially the section "2. Expires"
2. HTML5 has the concept of local storage and you can possibly take advantage of that as well. You can read more [here](https://developer.chrome.com/apps/offline_storage).
3. Try and compressing the graphics. You can read more [here](http://webappsuccess.com/web-app-performance.html). Especially the section "Minimize payload size".
Good luck and hope that helps. | You can download all the graphics on the first run & **cache it**. This will improve the performance tremendously.
Best practice is to use a CDN/Cache Server.
Refer to the following article for catching: [Link](https://hollyit.net/blog/caching-heavy-queries-good-or-bad) & [Link](https://stackoverflow.com/questions/16362893/caching-data-on-a-heavy-load-web-server) |
64,248 | It is possible to use an inverter to transfer energy from home made produced energy to the main power network. It is not clear to me how this can be achieved. What type of load appear to be the power supply network? What are the strategy to have the power flowing from the inverter to the the load and not vice-versa? | 2013/04/03 | [
"https://electronics.stackexchange.com/questions/64248",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/11079/"
] | [Grid-tied inverters](http://en.wikipedia.org/wiki/Grid-tie_inverter) do exactly what you're describing. Adding power to the grid is far from trivial, and isn't something that a few paragraphs of explanation can fully cover. People have spent decades on this sort of research.
The best way I've heard to describe the grid is 'infinite sink' and 'infinite source'.
An inverter carefully and accurately monitors the phase of the grid voltage and keeps itself in agreement, else the grid will try and push/pull far more current than the inverter can produce. (Mechanical generators have to be similarly synchronized else the grid will correct the generators, sometimes catastrophically)
It also must detect grid interruptions (islanding) and disconnect itself such that linesmen working on an isolated section of grid don't get electrocuted. | Yes, an inverter can be used to transfer energy to the main power network. I'm not about to answer how an inverter works but if you meant how can power transfer be achieved, here is a quote from Mother Nature network -
Most standard utility meters can measure the flow of energy in either direction. The meter spins forward when electricity is flowing from the utility to the customer and spins backward when power is flowing from the customer to the utility. Check with your power company to be sure you have the correct inverter to safely operate your alternative energy system.
This should explain that the strategy is automatic providing you have a meter capable of reverse flow |
142,595 | The YouTube app told me that there was a problem with the network, so I waited for about 2 days assuming that it would have to do something with the network, but my internet has been functional. I'm presenting the screenshot that I took now.
>
> 
>
>
> There was a problem with the network [401]
>
>
>
I'm writing a question from my mobile and there's no network problem as I'm able to write it.
I did a Google search but all the issues were 4 years back, everybody is talking about Gingerbread in those posts, while I'm running Android Lollipop.
What's the problem and how to fix it? | 2016/04/17 | [
"https://android.stackexchange.com/questions/142595",
"https://android.stackexchange.com",
"https://android.stackexchange.com/users/151369/"
] | If your internet is otherwise working fine, it looks like it could be an issue with the YouTube app itself. Try clearing cache and data for the YouTube app. This will effectively reset the app back to default settings and you will have to log back in.
Go to Settings -> Application Manager -> YouTube -> Storage -> Clear Data and Clear Cache. | If you uninstall updates from the Settings menu or the YouTube app, that should fix it. |
142,595 | The YouTube app told me that there was a problem with the network, so I waited for about 2 days assuming that it would have to do something with the network, but my internet has been functional. I'm presenting the screenshot that I took now.
>
> 
>
>
> There was a problem with the network [401]
>
>
>
I'm writing a question from my mobile and there's no network problem as I'm able to write it.
I did a Google search but all the issues were 4 years back, everybody is talking about Gingerbread in those posts, while I'm running Android Lollipop.
What's the problem and how to fix it? | 2016/04/17 | [
"https://android.stackexchange.com/questions/142595",
"https://android.stackexchange.com",
"https://android.stackexchange.com/users/151369/"
] | Go to the app store, search YouTube, then delete it. It will leave you on the last update. Don't update it as it will start showing the error 401 again. | If you uninstall updates from the Settings menu or the YouTube app, that should fix it. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | These are some of the things I have done in the past.
1. Have a humanoid enemy turn coat and join the players.
2. Allow the player to be a "short term" ghost.
3. Have the player's essence enter an object but let the team know that soon the essence will be forced out. | I'm all for providing a **temporary**(till end of session) NPC.
1. Ask the player to think of a short paragraph of how has the character's life and death affected the world. Have them write it down on the character sheet with a non-erasable pen.
This is the de-facto death of the character. Morn it, move the character sheet to the Graveyard if you have one else hand it back to the player.
2. Hand the player one of 4-5 drop-in mini antagonists\* you have prepared. The card should hold sections 'combat stats', 'major skills', 'intent'. Deliberately looking not like a character sheet. With immediate general course of action stated in the last section.
3. After the session do make arrangements with the player to support them with character creation. Have the boring stuff (stats, equipment) finalized before the next session. First half an hour of it everyone works together to weave the past and motivation of the character with that of the existing group (so that their goals are aligned enough to work together but divergent enough to create drama).
Examples (which can happen pretty much anywhere, so are a kind of random encounters; just increase the CR because of the human control):
* Leader of a bandit strike team.
* A rogue hired to assassinate one of the PCs.
* Leader of a pickpocket team.
* Town guard sergeant reacting to a distress bell.
* An invisible spy seeking to misdirect them.
* + having several unused villains, helpers, quests, places, items - is my style of DMing. Anytime the party does something I couldn't have reasonably expected, I whip out a random one of those and let the party have fun for half an hour while I am trying to think how to fit THIS NOW into the overarching plot. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | Here are some suggestions that we have tried for our 4e game. It does not come up much as we do not have a high mortality campaign thus far.
* Have backup characters ready. In 4e we have had backup characters in the character builder ready to be printed if we needed them during a session. This can provide a quick pick up right back to where you left off by just creating some story around the new character(s) and getting them into the game fast. This has an advantage because you may already have some story for the character and why he exists in your world. Keep them leveled with the party and keep the equipped similarly to the character they are replacing.
* Res the character quickly, maybe even cheaply. This has so far been how we've dealt with character deaths in our game. The one encounter where we had multiple PC deaths was at the very end of an extended encounter. We carried our bodies out and were able to find a local healer who was willing to res our friends for a very small amount of gold compared to what it would have cost out of gratitude for saving the town.
* Have them fill NPC roles for the rest of the session. This is not something we've tried, but is something I've seen suggested several places and is a good idea for a group in general to keep engagement when characters are off screen. Let your players have the role of an NPC or two for the rest of the session, or just for this interaction, it keeps everyone engaged and it helps take some of the burden off of you as the DM.
I'm sure there are plenty of better solutions, but these are some I've tried or seen floated pretty regularly. | As far as Character wise:
* Have PCs make backups
* Have sheets made for a few of your NPC characters. Allow the character to take control of an NPC that is a tagalong type (not, perhaps, the archenemy of the story) or support character until session end, then have them make another.
* Resurrection, as mentioned above
* Simply have them make a very basic character, without all the fluff, jump in, and then either flesh the character out later / as story goes along, or make a proper one after session.
As far as in-game hooking goes:
* Again, resurrection - although in my normal setting, PCs don't get resurrection often, and not at lower levels much at all.
* If you're dungeon crawling, finding a new or remade PC as a prisoner is one of my favorites.
* An old friend of another PC, or a guild mate, fellow soldier, etc. joining you for X reason (bringing a message, last survivor of their squad, etc)
* Run into them at the pub (or other relevant place of interaction) |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | As far as Character wise:
* Have PCs make backups
* Have sheets made for a few of your NPC characters. Allow the character to take control of an NPC that is a tagalong type (not, perhaps, the archenemy of the story) or support character until session end, then have them make another.
* Resurrection, as mentioned above
* Simply have them make a very basic character, without all the fluff, jump in, and then either flesh the character out later / as story goes along, or make a proper one after session.
As far as in-game hooking goes:
* Again, resurrection - although in my normal setting, PCs don't get resurrection often, and not at lower levels much at all.
* If you're dungeon crawling, finding a new or remade PC as a prisoner is one of my favorites.
* An old friend of another PC, or a guild mate, fellow soldier, etc. joining you for X reason (bringing a message, last survivor of their squad, etc)
* Run into them at the pub (or other relevant place of interaction) | There is a lot of technics described above, but are you really sure you want to put a player back in fast? I'd like to mention this once more, because it looks very important to me: easy resurrection might damage role-playing.
In my opinion, quick resurrection (by magic, backup characters, deus ex machina) is appropriate only if the PC have died because of a GM mistake (like, choosing too hard monsters and closing the only door) or a very unfortunate coincidence (maybe caused by an unforseen game system pecularity).
Otherwise, being a player I would find it very fair to have episodic NPC roles for quite a long time, before a character gets resurrected (i.e. after the party works hard to earn a huge sum of money to pay for the spell) or remastered. Gandalf's resurrection in LOTR is a masterpiece. And if it was a game, Gandalf's player could play a not so importrant and detailed, but a somewhat lasting character (Eomer comes to my mind). And, well, you can kill the latter or leave him behind (injured or in love) to resurrect the main character of the player. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | Here are some suggestions that we have tried for our 4e game. It does not come up much as we do not have a high mortality campaign thus far.
* Have backup characters ready. In 4e we have had backup characters in the character builder ready to be printed if we needed them during a session. This can provide a quick pick up right back to where you left off by just creating some story around the new character(s) and getting them into the game fast. This has an advantage because you may already have some story for the character and why he exists in your world. Keep them leveled with the party and keep the equipped similarly to the character they are replacing.
* Res the character quickly, maybe even cheaply. This has so far been how we've dealt with character deaths in our game. The one encounter where we had multiple PC deaths was at the very end of an extended encounter. We carried our bodies out and were able to find a local healer who was willing to res our friends for a very small amount of gold compared to what it would have cost out of gratitude for saving the town.
* Have them fill NPC roles for the rest of the session. This is not something we've tried, but is something I've seen suggested several places and is a good idea for a group in general to keep engagement when characters are off screen. Let your players have the role of an NPC or two for the rest of the session, or just for this interaction, it keeps everyone engaged and it helps take some of the burden off of you as the DM.
I'm sure there are plenty of better solutions, but these are some I've tried or seen floated pretty regularly. | I tend to bring in new PCs who had some sort of relationship with the recently deceased PC. The nature of the relationship will vary, but it has to be strong enough that the new PC has a reason to show up hot on the heels of the death of the old PC. A few highly situation-dependent variations on this theme:
* Sibling who wants to prove worth to the family by joining
* Close friend who wanted to come along earlier but couldn't
* Former enemy who heard PC #1 was in the area and wanted to make amends
* Messenger with a warning and decides to stay ("Hey, the Guildmaster says Cromius should avoid the... oh, I see. Never mind.").
One good thing about establishing a relationship between the new and old PCs is that it provides a feeling of story continuity and doesn't feel quite as forced as if the new PC came in from completely out of the blue. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | Here are some suggestions that we have tried for our 4e game. It does not come up much as we do not have a high mortality campaign thus far.
* Have backup characters ready. In 4e we have had backup characters in the character builder ready to be printed if we needed them during a session. This can provide a quick pick up right back to where you left off by just creating some story around the new character(s) and getting them into the game fast. This has an advantage because you may already have some story for the character and why he exists in your world. Keep them leveled with the party and keep the equipped similarly to the character they are replacing.
* Res the character quickly, maybe even cheaply. This has so far been how we've dealt with character deaths in our game. The one encounter where we had multiple PC deaths was at the very end of an extended encounter. We carried our bodies out and were able to find a local healer who was willing to res our friends for a very small amount of gold compared to what it would have cost out of gratitude for saving the town.
* Have them fill NPC roles for the rest of the session. This is not something we've tried, but is something I've seen suggested several places and is a good idea for a group in general to keep engagement when characters are off screen. Let your players have the role of an NPC or two for the rest of the session, or just for this interaction, it keeps everyone engaged and it helps take some of the burden off of you as the DM.
I'm sure there are plenty of better solutions, but these are some I've tried or seen floated pretty regularly. | This varies somewhat by the tone of your campaign and the system/setting.
In some systems/settings, **resurrections** are relatively easy. In that case, the obvious answer is to perform the resurrection with some sort of minor penalty and then go on. This, though, cheapens character death. In a mostly hack 'n slash, lighthearted game, this is perfectly fine and essentially replicates the feeling of most videogames (you died, now you loose a little bit of progress and get to try again). In a more serious game with a deep story, knowing death is cheap can affect the story (sacraficing your life means a lot less if that life comes back at the price of a little gold and XP or whatever) and knowing it is not just available but cheap will definitely affect the way the characters play. It will make them more likely to take large risks. (again whether that is good or bad depends on what you are going for).
An alternative, again available in some but not all settings is that they **come back as a ghost** or other undead form. This was interesting in some oWoD games I played, but probably does not work well in say DnD. You have to be in a setting that supports that and it would have to be such that a ghost could integrate without messing up the story, or this becomes a very bad option.
The next option of course is to have **backups standing by**. This could mean having a new character already generated, or having the player make one quickly while everyone else continues on and then integrating them as soon as ready. This works well in games where character creation is fairly quick, but less well when it takes a while, and it is complicated if the players are expected to make up an elaborate backstory for their character. One way to simplify this, though one that is rarely satisfying, is to let them bring back a variant of their old character. Tweak the name, maybe reassign a skill or two, and tweak the backstory a bit, but basically let them reuse the character that died.
Finally, use **deus ex machina** to prevent the death. Personally, I hate this option. It seems to cheapen death even more than resurrection, and basically changes what just happened. But as long as it is not overused, it can be a way of saving a character that someone put a lot of time and effort into developping, with a long backstory even in settings where resurrection is just not an option. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | The option that worked better for me (especially during a spell when I made my adventures quite hard and therefore I "killed" characters quite often...) was having each player use and develop a number of different characters at the same time.
If one dies and can't be resurrected, you have the replacement ready...
PS the players understood quite quickly the situation (i.e. if you're not careful, your character will die easily) and in general enjoyed the challenge of really risking losing the character almost at every turn and of developing a fairly broad number of different characters. | As far as Character wise:
* Have PCs make backups
* Have sheets made for a few of your NPC characters. Allow the character to take control of an NPC that is a tagalong type (not, perhaps, the archenemy of the story) or support character until session end, then have them make another.
* Resurrection, as mentioned above
* Simply have them make a very basic character, without all the fluff, jump in, and then either flesh the character out later / as story goes along, or make a proper one after session.
As far as in-game hooking goes:
* Again, resurrection - although in my normal setting, PCs don't get resurrection often, and not at lower levels much at all.
* If you're dungeon crawling, finding a new or remade PC as a prisoner is one of my favorites.
* An old friend of another PC, or a guild mate, fellow soldier, etc. joining you for X reason (bringing a message, last survivor of their squad, etc)
* Run into them at the pub (or other relevant place of interaction) |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | The option that worked better for me (especially during a spell when I made my adventures quite hard and therefore I "killed" characters quite often...) was having each player use and develop a number of different characters at the same time.
If one dies and can't be resurrected, you have the replacement ready...
PS the players understood quite quickly the situation (i.e. if you're not careful, your character will die easily) and in general enjoyed the challenge of really risking losing the character almost at every turn and of developing a fairly broad number of different characters. | I tend to bring in new PCs who had some sort of relationship with the recently deceased PC. The nature of the relationship will vary, but it has to be strong enough that the new PC has a reason to show up hot on the heels of the death of the old PC. A few highly situation-dependent variations on this theme:
* Sibling who wants to prove worth to the family by joining
* Close friend who wanted to come along earlier but couldn't
* Former enemy who heard PC #1 was in the area and wanted to make amends
* Messenger with a warning and decides to stay ("Hey, the Guildmaster says Cromius should avoid the... oh, I see. Never mind.").
One good thing about establishing a relationship between the new and old PCs is that it provides a feeling of story continuity and doesn't feel quite as forced as if the new PC came in from completely out of the blue. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | I tend to bring in new PCs who had some sort of relationship with the recently deceased PC. The nature of the relationship will vary, but it has to be strong enough that the new PC has a reason to show up hot on the heels of the death of the old PC. A few highly situation-dependent variations on this theme:
* Sibling who wants to prove worth to the family by joining
* Close friend who wanted to come along earlier but couldn't
* Former enemy who heard PC #1 was in the area and wanted to make amends
* Messenger with a warning and decides to stay ("Hey, the Guildmaster says Cromius should avoid the... oh, I see. Never mind.").
One good thing about establishing a relationship between the new and old PCs is that it provides a feeling of story continuity and doesn't feel quite as forced as if the new PC came in from completely out of the blue. | There is a lot of technics described above, but are you really sure you want to put a player back in fast? I'd like to mention this once more, because it looks very important to me: easy resurrection might damage role-playing.
In my opinion, quick resurrection (by magic, backup characters, deus ex machina) is appropriate only if the PC have died because of a GM mistake (like, choosing too hard monsters and closing the only door) or a very unfortunate coincidence (maybe caused by an unforseen game system pecularity).
Otherwise, being a player I would find it very fair to have episodic NPC roles for quite a long time, before a character gets resurrected (i.e. after the party works hard to earn a huge sum of money to pay for the spell) or remastered. Gandalf's resurrection in LOTR is a masterpiece. And if it was a game, Gandalf's player could play a not so importrant and detailed, but a somewhat lasting character (Eomer comes to my mind). And, well, you can kill the latter or leave him behind (injured or in love) to resurrect the main character of the player. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | This varies somewhat by the tone of your campaign and the system/setting.
In some systems/settings, **resurrections** are relatively easy. In that case, the obvious answer is to perform the resurrection with some sort of minor penalty and then go on. This, though, cheapens character death. In a mostly hack 'n slash, lighthearted game, this is perfectly fine and essentially replicates the feeling of most videogames (you died, now you loose a little bit of progress and get to try again). In a more serious game with a deep story, knowing death is cheap can affect the story (sacraficing your life means a lot less if that life comes back at the price of a little gold and XP or whatever) and knowing it is not just available but cheap will definitely affect the way the characters play. It will make them more likely to take large risks. (again whether that is good or bad depends on what you are going for).
An alternative, again available in some but not all settings is that they **come back as a ghost** or other undead form. This was interesting in some oWoD games I played, but probably does not work well in say DnD. You have to be in a setting that supports that and it would have to be such that a ghost could integrate without messing up the story, or this becomes a very bad option.
The next option of course is to have **backups standing by**. This could mean having a new character already generated, or having the player make one quickly while everyone else continues on and then integrating them as soon as ready. This works well in games where character creation is fairly quick, but less well when it takes a while, and it is complicated if the players are expected to make up an elaborate backstory for their character. One way to simplify this, though one that is rarely satisfying, is to let them bring back a variant of their old character. Tweak the name, maybe reassign a skill or two, and tweak the backstory a bit, but basically let them reuse the character that died.
Finally, use **deus ex machina** to prevent the death. Personally, I hate this option. It seems to cheapen death even more than resurrection, and basically changes what just happened. But as long as it is not overused, it can be a way of saving a character that someone put a lot of time and effort into developping, with a long backstory even in settings where resurrection is just not an option. | There is a lot of technics described above, but are you really sure you want to put a player back in fast? I'd like to mention this once more, because it looks very important to me: easy resurrection might damage role-playing.
In my opinion, quick resurrection (by magic, backup characters, deus ex machina) is appropriate only if the PC have died because of a GM mistake (like, choosing too hard monsters and closing the only door) or a very unfortunate coincidence (maybe caused by an unforseen game system pecularity).
Otherwise, being a player I would find it very fair to have episodic NPC roles for quite a long time, before a character gets resurrected (i.e. after the party works hard to earn a huge sum of money to pay for the spell) or remastered. Gandalf's resurrection in LOTR is a masterpiece. And if it was a game, Gandalf's player could play a not so importrant and detailed, but a somewhat lasting character (Eomer comes to my mind). And, well, you can kill the latter or leave him behind (injured or in love) to resurrect the main character of the player. |
13,743 | As a GM, what kind of techniques could I use to help a player re-join the gaming session after their character dies? I want to avoid any major interruptions to the flow of the adventure. I would also like to avoid having to end the session immediately or making them sit on the side-lines for the rest of the session. | 2012/04/24 | [
"https://rpg.stackexchange.com/questions/13743",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/3436/"
] | Here are some suggestions that we have tried for our 4e game. It does not come up much as we do not have a high mortality campaign thus far.
* Have backup characters ready. In 4e we have had backup characters in the character builder ready to be printed if we needed them during a session. This can provide a quick pick up right back to where you left off by just creating some story around the new character(s) and getting them into the game fast. This has an advantage because you may already have some story for the character and why he exists in your world. Keep them leveled with the party and keep the equipped similarly to the character they are replacing.
* Res the character quickly, maybe even cheaply. This has so far been how we've dealt with character deaths in our game. The one encounter where we had multiple PC deaths was at the very end of an extended encounter. We carried our bodies out and were able to find a local healer who was willing to res our friends for a very small amount of gold compared to what it would have cost out of gratitude for saving the town.
* Have them fill NPC roles for the rest of the session. This is not something we've tried, but is something I've seen suggested several places and is a good idea for a group in general to keep engagement when characters are off screen. Let your players have the role of an NPC or two for the rest of the session, or just for this interaction, it keeps everyone engaged and it helps take some of the burden off of you as the DM.
I'm sure there are plenty of better solutions, but these are some I've tried or seen floated pretty regularly. | I'm all for providing a **temporary**(till end of session) NPC.
1. Ask the player to think of a short paragraph of how has the character's life and death affected the world. Have them write it down on the character sheet with a non-erasable pen.
This is the de-facto death of the character. Morn it, move the character sheet to the Graveyard if you have one else hand it back to the player.
2. Hand the player one of 4-5 drop-in mini antagonists\* you have prepared. The card should hold sections 'combat stats', 'major skills', 'intent'. Deliberately looking not like a character sheet. With immediate general course of action stated in the last section.
3. After the session do make arrangements with the player to support them with character creation. Have the boring stuff (stats, equipment) finalized before the next session. First half an hour of it everyone works together to weave the past and motivation of the character with that of the existing group (so that their goals are aligned enough to work together but divergent enough to create drama).
Examples (which can happen pretty much anywhere, so are a kind of random encounters; just increase the CR because of the human control):
* Leader of a bandit strike team.
* A rogue hired to assassinate one of the PCs.
* Leader of a pickpocket team.
* Town guard sergeant reacting to a distress bell.
* An invisible spy seeking to misdirect them.
* + having several unused villains, helpers, quests, places, items - is my style of DMing. Anytime the party does something I couldn't have reasonably expected, I whip out a random one of those and let the party have fun for half an hour while I am trying to think how to fit THIS NOW into the overarching plot. |
19,553 | I have now written my first e-book in LaTeX but, before I publish it, I wish to hire someone with superior LaTeX skills to fix the design and formatting of the book. At the moment I have tons of issues with the e-book and it doesn't look too great. Where could I find people for hourly hire with good book design skills and good LaTeX skills who could help me? It's easy to find people with good book design skills, but all of them work in Illustrator or CorelDRAW and not LaTeX. | 2011/05/31 | [
"https://tex.stackexchange.com/questions/19553",
"https://tex.stackexchange.com",
"https://tex.stackexchange.com/users/5177/"
] | TUG have pages for both ['consultants for hire'](http://www.tug.org/consultants.html) and ['jobs advertised](http://www.tug.org/jobboard.html). You could consider using these to find the appropriate person. | If it's hard to find a LaTeX capable designer: you could hire a good book designer to develop a great look, a skilled LaTeX user can help to implement that.
Anyway, you can ask LaTeX users in web forums, Usenet groups and Q&A site listed here: [Good resources on-line for information about TeX, LaTeX and friends](https://tex.stackexchange.com/questions/162/what-are-other-good-resources-on-line-for-information-about-tex-latex-and-friend) |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | If you need to implement such functionality yourself, I suggest you to check up [Duff's Device](http://en.wikipedia.org/wiki/Duff%27s_device) if it has to be done efficiently. | Reading your comments, it sounds like you might want to use parallelism. There are instructions to do this, but they only operate on registers, not memory.
This is because of the way the computer architecture is (I'm assuming x86).
You can only be accessing one memory location at a time because the computer only has one address bus. If you tried to access more than one location at a time, you would be overloading the bus and nothing would work properly.
If you can put the data you need in registers, then you can use a lot of cool processor instructions, such as MMX or SSE, to perform parallel calculations. But as for copying memory in parallel, it's not possible.
As others have said, use memcpy. It's reliable, debugged, and fast. |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | Use memmove() if the origin and source overlap. Usually memcpy() and memmove() have been highly optimized already for your compiler's clib. If you do write a replacement, at least benchmark it against the clib versions to make sure you're not slowing down your code.
>
> i have a block consisting of 5000 rows or more and in a function that is called all the time 10000 times
>
>
>
Also, consider changing your data structure. Perhaps instead of a 2D array, you can have a 1D array of Pointers to secondary Arrays (the columns). Then instead of copying the entire rows, you need only copy or move the Pointers. You could Pool the column Arrays in a Free-List so that you're not spending lots of time allocating and freeing them as well. | [memcpy](http://www.cppreference.com/wiki/c/string/memcpy)? |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | Reading your comments, it sounds like you might want to use parallelism. There are instructions to do this, but they only operate on registers, not memory.
This is because of the way the computer architecture is (I'm assuming x86).
You can only be accessing one memory location at a time because the computer only has one address bus. If you tried to access more than one location at a time, you would be overloading the bus and nothing would work properly.
If you can put the data you need in registers, then you can use a lot of cool processor instructions, such as MMX or SSE, to perform parallel calculations. But as for copying memory in parallel, it's not possible.
As others have said, use memcpy. It's reliable, debugged, and fast. | Use memmove() if the origin and source overlap. Usually memcpy() and memmove() have been highly optimized already for your compiler's clib. If you do write a replacement, at least benchmark it against the clib versions to make sure you're not slowing down your code.
>
> i have a block consisting of 5000 rows or more and in a function that is called all the time 10000 times
>
>
>
Also, consider changing your data structure. Perhaps instead of a 2D array, you can have a 1D array of Pointers to secondary Arrays (the columns). Then instead of copying the entire rows, you need only copy or move the Pointers. You could Pool the column Arrays in a Free-List so that you're not spending lots of time allocating and freeing them as well. |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | Reading your comments, it sounds like you might want to use parallelism. There are instructions to do this, but they only operate on registers, not memory.
This is because of the way the computer architecture is (I'm assuming x86).
You can only be accessing one memory location at a time because the computer only has one address bus. If you tried to access more than one location at a time, you would be overloading the bus and nothing would work properly.
If you can put the data you need in registers, then you can use a lot of cool processor instructions, such as MMX or SSE, to perform parallel calculations. But as for copying memory in parallel, it's not possible.
As others have said, use memcpy. It's reliable, debugged, and fast. | [memcpy](http://www.cppreference.com/wiki/c/string/memcpy)? |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | If you need to implement such functionality yourself, I suggest you to check up [Duff's Device](http://en.wikipedia.org/wiki/Duff%27s_device) if it has to be done efficiently. | REP MOVSD in assembly perhaps? Hard to say without more information on exactly what you're trying to copy... Or, you can reprogram the DMA controller to do it too, but it'll actually end up being slower than just using the processor. :-) |
944,302 | I need a suggestion on on how do I copy a block of memory efficiently, in single attempt if possible, in C++ or assembly language.
I have a pointer to memory location and offset. Think of a memory as a 2D array that I need to copy consisting of rows and columns. | 2009/06/03 | [
"https://Stackoverflow.com/questions/944302",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/100751/"
] | Use memmove() if the origin and source overlap. Usually memcpy() and memmove() have been highly optimized already for your compiler's clib. If you do write a replacement, at least benchmark it against the clib versions to make sure you're not slowing down your code.
>
> i have a block consisting of 5000 rows or more and in a function that is called all the time 10000 times
>
>
>
Also, consider changing your data structure. Perhaps instead of a 2D array, you can have a 1D array of Pointers to secondary Arrays (the columns). Then instead of copying the entire rows, you need only copy or move the Pointers. You could Pool the column Arrays in a Free-List so that you're not spending lots of time allocating and freeing them as well. | REP MOVSD in assembly perhaps? Hard to say without more information on exactly what you're trying to copy... Or, you can reprogram the DMA controller to do it too, but it'll actually end up being slower than just using the processor. :-) |
130,321 | I found a database structure for a CMS system and I want to copy it with the EF that this database is done in, is it OK to copy the database structure of open source software which is under GNU v2 lisence?
I don't want the rest of the software only the database structure. | 2012/01/16 | [
"https://softwareengineering.stackexchange.com/questions/130321",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/45291/"
] | The easy answer is "Yes, if your software will be distributed under GPL v2". Likewise, if you do not intend to distribute your software (*e.g.*, you're coding for a company's internal use only), the GPL allows you use it however you want.
The question gets difficult if you intend to distribute the software and not under GPL v2, because then you have to determine whether the database design is sufficiently creative that it would be copyrightable and thus protected under the GPL. The only safe answer, unless you've got good intellectual property lawyers, is "No". | Went through the GNU V2 License copy and found nothing that may restrict your usage of the the db structure. But your end result should have the GNU V2 License as well. That is you must not restrict the source of your code.
Again this answer of mine comes without any warranty whatsoever.
GO [here](http://www.gnu.org/licenses/gpl-2.0.html) for more details. |
130,321 | I found a database structure for a CMS system and I want to copy it with the EF that this database is done in, is it OK to copy the database structure of open source software which is under GNU v2 lisence?
I don't want the rest of the software only the database structure. | 2012/01/16 | [
"https://softwareengineering.stackexchange.com/questions/130321",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/45291/"
] | The easy answer is "Yes, if your software will be distributed under GPL v2". Likewise, if you do not intend to distribute your software (*e.g.*, you're coding for a company's internal use only), the GPL allows you use it however you want.
The question gets difficult if you intend to distribute the software and not under GPL v2, because then you have to determine whether the database design is sufficiently creative that it would be copyrightable and thus protected under the GPL. The only safe answer, unless you've got good intellectual property lawyers, is "No". | The way I read the [GPL v2](http://www.gnu.org/licenses/gpl-2.0.html) Section 1:
>
> You may copy and distribute verbatim copies of the Program's source
> code as you receive it, in any medium, provided that you
> conspicuously and appropriately publish on each copy an appropriate
> copyright notice and disclaimer of warranty; keep intact all the
> notices that refer to this License and to the absence of any
> warranty; and give any other recipients of the Program a copy of this
> License along with the Program.
>
>
>
and Section 2:
>
> You may modify your copy or copies of the Program or any portion of
> it, thus forming a work based on the Program, and copy and distribute
> such modifications or work under the terms of Section 1 above,
> provided that you also meet all of these conditions:
>
>
>
You can do this because what you are doing is basically taking a copy of the entire software distributed under GPL v2 and removing all portions except the database structure which is part of software. The only thing is that anything you will distribute that has this as part of your software will be distributed under GPL v2 license. |
130,321 | I found a database structure for a CMS system and I want to copy it with the EF that this database is done in, is it OK to copy the database structure of open source software which is under GNU v2 lisence?
I don't want the rest of the software only the database structure. | 2012/01/16 | [
"https://softwareengineering.stackexchange.com/questions/130321",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/45291/"
] | The intent of this license is to share developement and its benefits so you can copy what you wish from the GPL software but the "output" needs to also be released as GPL. The license clearly states that you are granted the right to make modifications under the terms of GPL
*You must cause any work that you distribute or publish, that in whole or in part contains or is derived from the Program or any part thereof, to be licensed as a **whole** at no charge to all third parties under the terms of this License*
This term doesn't apply to "separated" parts of the program. But the database design isn't separated (like, for example some external library). The database scheme is protected under copyright (as a whole) i don't see any reason why it should not be. I bet if we spoken about copying some part of paid software no one would have such doubts. | Went through the GNU V2 License copy and found nothing that may restrict your usage of the the db structure. But your end result should have the GNU V2 License as well. That is you must not restrict the source of your code.
Again this answer of mine comes without any warranty whatsoever.
GO [here](http://www.gnu.org/licenses/gpl-2.0.html) for more details. |
130,321 | I found a database structure for a CMS system and I want to copy it with the EF that this database is done in, is it OK to copy the database structure of open source software which is under GNU v2 lisence?
I don't want the rest of the software only the database structure. | 2012/01/16 | [
"https://softwareengineering.stackexchange.com/questions/130321",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/45291/"
] | The intent of this license is to share developement and its benefits so you can copy what you wish from the GPL software but the "output" needs to also be released as GPL. The license clearly states that you are granted the right to make modifications under the terms of GPL
*You must cause any work that you distribute or publish, that in whole or in part contains or is derived from the Program or any part thereof, to be licensed as a **whole** at no charge to all third parties under the terms of this License*
This term doesn't apply to "separated" parts of the program. But the database design isn't separated (like, for example some external library). The database scheme is protected under copyright (as a whole) i don't see any reason why it should not be. I bet if we spoken about copying some part of paid software no one would have such doubts. | The way I read the [GPL v2](http://www.gnu.org/licenses/gpl-2.0.html) Section 1:
>
> You may copy and distribute verbatim copies of the Program's source
> code as you receive it, in any medium, provided that you
> conspicuously and appropriately publish on each copy an appropriate
> copyright notice and disclaimer of warranty; keep intact all the
> notices that refer to this License and to the absence of any
> warranty; and give any other recipients of the Program a copy of this
> License along with the Program.
>
>
>
and Section 2:
>
> You may modify your copy or copies of the Program or any portion of
> it, thus forming a work based on the Program, and copy and distribute
> such modifications or work under the terms of Section 1 above,
> provided that you also meet all of these conditions:
>
>
>
You can do this because what you are doing is basically taking a copy of the entire software distributed under GPL v2 and removing all portions except the database structure which is part of software. The only thing is that anything you will distribute that has this as part of your software will be distributed under GPL v2 license. |
594,978 | If an observer moves in an accelerated frame in flat spacetime, the vacuum looks like a thermal distribution of particles to that observer. This is the Unruh effect.
* Is it a special relativistic (SR) effect or a general relativistic (GR) effect? On one hand, it's flat spacetime which seems that SR is enough. But then it talks about the accelerated observer, which seems to suggest that it is a nontrivial GR effect. Please help. | 2020/11/19 | [
"https://physics.stackexchange.com/questions/594978",
"https://physics.stackexchange.com",
"https://physics.stackexchange.com/users/164488/"
] | It is a SR phenomenon **since it would be true even if GR were false**: no equivalence principle, no Einstein equations, etc. (and thus it is independent of it). The said effect is a consequence of properties of the unique Poincare invariant vacuum state which satisfies a peculiar property (KMS condition) with respect to the boost symmetry. | This is not an *either - or* situation. One should make a distinction between the *physics* of the effect, in this case the accelerating detector coupled to a quantum field registering the thermal spectrum of excitations of that field, and the *conceptual framework* used to explain the effect. Both GR and SR are such frameworks and each could be used to describe this effect.
Since the effect does not depend on the spacetime curvature, one does not *need* to use GR for the description, it is perfectly possible to describe the behaviour of accelerating detector using only notions of QFT in Minkowski spacetime (i.e. the SR theory). At the same time using the set of tools from general relativity (more specifically, from QFT in curved spacetime) also allows one to describe the very same physical situation. Moreover, there are multiple GR techniques that could be used to achieve quantitative description of the Unruh effect (Bogoliubov transformation method, WKB/tunneling method, anomaly cancellation method, etc.) each one offering new insights into the situation. The existence of such dual descriptions (SR and GR) is an important test of the formalisms employed and of our intuition built upon these sets of tools. |
4,428,684 | In my iPhone app, I'm using MPMoviePlayerController to play a movie. I'm hiding all the controls that are by default visible on the movie player. But I placed a "Replay" button over the player control.
At the end of movie, the player is being removed. But I want to stop at the last frame, so that when I click "Replay" button, it will start from beginning. I wrote the functionality for replay and its working good. When ever the movie is playing and on click of "Replay", its restarting the video from starting.
The problem I'm facing is that, at the end of movie its becoming white screen and pressing "Replay" button is not restarting the movie. How to handle this situation? | 2010/12/13 | [
"https://Stackoverflow.com/questions/4428684",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/265341/"
] | Its not necessary to create new instance. I found out the solution. Don't just release the instance of player. When replay button is clicked (either in middle of video or after completion), just pause it, move the location to beginning and play it. That's it.. no retain, no release nothing.... Its upto the programmer/developer when to release the player and remove it from the view. | I never used MPMoviePlayerController but pherhaps it's instance is released when the movie has ended. In this case incrementing the retaincounter and manually releasing it would solve the problem. |
97,471 | Is there a way to mention/tag a user in a question or answer? | 2011/07/05 | [
"https://meta.stackexchange.com/questions/97471",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/162425/"
] | If you just want to mention a user in your question, you should link to [their profile](https://meta.stackexchange.com/users/162425/xyz).
If you're trying to get a specific user to answer your question, that's not a feature of Stack Overflow. Ask the question of the whole community. If the person is an expert in the technology you're asking about (and an active SO visitor), they probably follow the tag and will find your question. It's entirely up to them whether they answer or not.
---
[@midopa](https://meta.stackexchange.com/users/339134/midopa) commented:
>
> I'd like to use mentions to give credit to others' suggestions to my answers. Having to get/make a link to their profile and add it as an explicit link in my response makes it tedious.
>
>
>
If you *only* want to mention/credit someone in a post *without* notification, then yes, it would be nice to just @mention them. The problem is that display names on Stack Overflow are not unique, as they are on many social media platforms. I'd have to go to your profile to get your unique user ID anyway, so at that point I might as well just make a link. | If you want the attention of specific member you can look for him in the chat - you can see the active users list here: <https://chat.stackoverflow.com/users>
Another option is check their profile page, many have website and even email address mentioned there so just use it to contact him/her. |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | The service you meantion is called **"Network Location Awareness" or NLA**. It determines what kind of connectivity you have and makes connection specific information available to other applications or services. The Advanced Firewall in Windows Server 2008 uses the NLA information to apply specific firewall settings.
It's a Windows Service, so you could disable the service. | I do not believe that there is group policy that will let you assign a network profile (it is determined by the Network Location Awareness, more information here: <http://msdn.microsoft.com/en-us/library/ms739931(VS.85).aspx>)
You can, however, apply group policy to the servers to define the behaviour of the Advanced firewall (disabling it, allowing traffic from your administrative workstations, etc). Instructions on doing so availale here: <http://technet.microsoft.com/en-us/library/cc732400.aspx> |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | The service you meantion is called **"Network Location Awareness" or NLA**. It determines what kind of connectivity you have and makes connection specific information available to other applications or services. The Advanced Firewall in Windows Server 2008 uses the NLA information to apply specific firewall settings.
It's a Windows Service, so you could disable the service. | If you want to just disable the service
you can create a custom group policy that disable the NLA service
Since i am a new user i cant provide you with a link so just search in google these words
"disable service from group policy"
The first result is what you are looking for |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | Just ran into this exact problem. Unidentified networks are by default set to type of "Public". This is awkward when you want Windows Firewall to be active on Public networks but not Private ones -- and your internal network is always "Unidentified".
What's an "Unidentified" network to Windows Server 2008?
>
> The Network Lists service (netprofm) works with the Network Location Awareness service (nlasvc) to identify networks and find the associated saved settings for the network, if any. The NLA service will use a Default Gateway or SSID to identify a network, so **if the NIC has neither a Default Gateway or associated SSID, then NLA will determine that the network is Unidentified.**
>
>
>
You can however change the default -- so that so-called "Unidentified" networks will be defaulted to something other than Public:
1. Open Administrative Tools -> Local Security Policy.
2. Highlight the "Network List Manager Policies" item, then double click the "Unidentified Networks" on the right panel.
3. Set the "Location Type" to "Private" or "Public".

Worked for me! | The service you meantion is called **"Network Location Awareness" or NLA**. It determines what kind of connectivity you have and makes connection specific information available to other applications or services. The Advanced Firewall in Windows Server 2008 uses the NLA information to apply specific firewall settings.
It's a Windows Service, so you could disable the service. |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | The service you meantion is called **"Network Location Awareness" or NLA**. It determines what kind of connectivity you have and makes connection specific information available to other applications or services. The Advanced Firewall in Windows Server 2008 uses the NLA information to apply specific firewall settings.
It's a Windows Service, so you could disable the service. | Had the same precise issue; a couple of Windows 2012 servers that would occasionally sulk and decide that their only NIC was a "Public" interface, rather than a "Domain" interface.
Through the power of the interwebs, I came across [this helpful post](http://www.pdconsec.net/windows-2012-unidentified-network-at-boot.aspx), which, summarized, simply says to restart the "Network Location Awareness" service and see if that fixes the issue. If it does, then to prevent the problem from recurring, simply change the start-up type from "Automatic" to "Automatic (Delayed Start)". |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | I do not believe that there is group policy that will let you assign a network profile (it is determined by the Network Location Awareness, more information here: <http://msdn.microsoft.com/en-us/library/ms739931(VS.85).aspx>)
You can, however, apply group policy to the servers to define the behaviour of the Advanced firewall (disabling it, allowing traffic from your administrative workstations, etc). Instructions on doing so availale here: <http://technet.microsoft.com/en-us/library/cc732400.aspx> | If you want to just disable the service
you can create a custom group policy that disable the NLA service
Since i am a new user i cant provide you with a link so just search in google these words
"disable service from group policy"
The first result is what you are looking for |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | Just ran into this exact problem. Unidentified networks are by default set to type of "Public". This is awkward when you want Windows Firewall to be active on Public networks but not Private ones -- and your internal network is always "Unidentified".
What's an "Unidentified" network to Windows Server 2008?
>
> The Network Lists service (netprofm) works with the Network Location Awareness service (nlasvc) to identify networks and find the associated saved settings for the network, if any. The NLA service will use a Default Gateway or SSID to identify a network, so **if the NIC has neither a Default Gateway or associated SSID, then NLA will determine that the network is Unidentified.**
>
>
>
You can however change the default -- so that so-called "Unidentified" networks will be defaulted to something other than Public:
1. Open Administrative Tools -> Local Security Policy.
2. Highlight the "Network List Manager Policies" item, then double click the "Unidentified Networks" on the right panel.
3. Set the "Location Type" to "Private" or "Public".

Worked for me! | I do not believe that there is group policy that will let you assign a network profile (it is determined by the Network Location Awareness, more information here: <http://msdn.microsoft.com/en-us/library/ms739931(VS.85).aspx>)
You can, however, apply group policy to the servers to define the behaviour of the Advanced firewall (disabling it, allowing traffic from your administrative workstations, etc). Instructions on doing so availale here: <http://technet.microsoft.com/en-us/library/cc732400.aspx> |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | Just ran into this exact problem. Unidentified networks are by default set to type of "Public". This is awkward when you want Windows Firewall to be active on Public networks but not Private ones -- and your internal network is always "Unidentified".
What's an "Unidentified" network to Windows Server 2008?
>
> The Network Lists service (netprofm) works with the Network Location Awareness service (nlasvc) to identify networks and find the associated saved settings for the network, if any. The NLA service will use a Default Gateway or SSID to identify a network, so **if the NIC has neither a Default Gateway or associated SSID, then NLA will determine that the network is Unidentified.**
>
>
>
You can however change the default -- so that so-called "Unidentified" networks will be defaulted to something other than Public:
1. Open Administrative Tools -> Local Security Policy.
2. Highlight the "Network List Manager Policies" item, then double click the "Unidentified Networks" on the right panel.
3. Set the "Location Type" to "Private" or "Public".

Worked for me! | If you want to just disable the service
you can create a custom group policy that disable the NLA service
Since i am a new user i cant provide you with a link so just search in google these words
"disable service from group policy"
The first result is what you are looking for |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | Had the same precise issue; a couple of Windows 2012 servers that would occasionally sulk and decide that their only NIC was a "Public" interface, rather than a "Domain" interface.
Through the power of the interwebs, I came across [this helpful post](http://www.pdconsec.net/windows-2012-unidentified-network-at-boot.aspx), which, summarized, simply says to restart the "Network Location Awareness" service and see if that fixes the issue. If it does, then to prevent the problem from recurring, simply change the start-up type from "Automatic" to "Automatic (Delayed Start)". | If you want to just disable the service
you can create a custom group policy that disable the NLA service
Since i am a new user i cant provide you with a link so just search in google these words
"disable service from group policy"
The first result is what you are looking for |
10,014 | I'm not sure what exactly this feature is called. But in Windows Server 2008, it has the Vista Public/Private/Domain locations. This makes sense for laptops, and none at all for servers.
My problem is that sometimes some network adapters decide they are now on a public network. This completely activates the firewall, even for the "domain" networks. So net effect is that I reboot some machines, and then they never come back on the network until we KVM in and tell it that the network is private.
What's the name of this feature? Is there a GP setting I can use to turn it off and make all networks be "domain"?
Edit: Thanks, that's that NLA is. I tried disabling the service on a non-domain machine, and it just flips everything public. On a domain machine, the Network List Service refuses to stop -- I'll try group policy. | 2009/05/19 | [
"https://serverfault.com/questions/10014",
"https://serverfault.com",
"https://serverfault.com/users/2258/"
] | Just ran into this exact problem. Unidentified networks are by default set to type of "Public". This is awkward when you want Windows Firewall to be active on Public networks but not Private ones -- and your internal network is always "Unidentified".
What's an "Unidentified" network to Windows Server 2008?
>
> The Network Lists service (netprofm) works with the Network Location Awareness service (nlasvc) to identify networks and find the associated saved settings for the network, if any. The NLA service will use a Default Gateway or SSID to identify a network, so **if the NIC has neither a Default Gateway or associated SSID, then NLA will determine that the network is Unidentified.**
>
>
>
You can however change the default -- so that so-called "Unidentified" networks will be defaulted to something other than Public:
1. Open Administrative Tools -> Local Security Policy.
2. Highlight the "Network List Manager Policies" item, then double click the "Unidentified Networks" on the right panel.
3. Set the "Location Type" to "Private" or "Public".

Worked for me! | Had the same precise issue; a couple of Windows 2012 servers that would occasionally sulk and decide that their only NIC was a "Public" interface, rather than a "Domain" interface.
Through the power of the interwebs, I came across [this helpful post](http://www.pdconsec.net/windows-2012-unidentified-network-at-boot.aspx), which, summarized, simply says to restart the "Network Location Awareness" service and see if that fixes the issue. If it does, then to prevent the problem from recurring, simply change the start-up type from "Automatic" to "Automatic (Delayed Start)". |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | >
> Is there such a thing?
>
>
>
Absolutely. Feeding malicious input to a parser is one of the most common ways of creating an exploit (and, for a JPEG, "decompression" is "parsing").
>
> Is this description based on some real exploit?
>
>
>
It might be based on the [Microsoft Windows GDI+ buffer overflow vulnerability](http://www.kb.cert.org/vuls/id/297462):
>
> There is a buffer overflow vulnerability in the way the JPEG parsing
> component of GDI+ (Gdiplus.dll) handles **malformed JPEG images**. By
> introducing a **specially crafted JPEG file** to the vulnerable component,
> a remote attacker could trigger a buffer overflow condition.
>
>
> ...
>
>
> A remote, unauthenticated attacker could potentially execute arbitrary
> code on a vulnerable system by introducing a specially crafted JPEG
> file. This **malicious JPEG image may be introduced to the system via a**
> malicious web page, **HTML email, or an email attachment.**
>
>
>
.
>
> This was published in December 2006.
>
>
>
The GDI+ JPEG parsing vulnerability was published in September 2004.
>
> Is it sensible to say "the operating system" was decompressing the image to render it?
>
>
>
Sure; in this case, it was a system library that required an OS vendor patch to correct it. Often such libraries are used by multiple software packages, making them part of the operating system rather than application-specific.
In actuality, "the email application invoked a system library to parse a JPEG," but "the operating system" is close enough for a novel. | Yes this is possible:
>
> A new variant of the nefarious Zeus banking trojan – dubbed ZeusVM – is concealed in JPG image files, according to the collaborative findings of Jerome Segura, senior security researcher with Malwarebytes, and French security researcher Xylitol.
>
>
> The act is known as steganography – concealing messages or images in other messages or images.
>
>
> In the case of ZeusVM, the malware's code is hidden in unassuming JPG images, a Monday blog post by Segura revealed. These photos serve as misdirection for ZeusVM to retrieve its configuration file.
>
>
> “The JPG contains the malware configuration file, which is essentially a list of scripts and financial institutions - but doesn't need to be opened by the victim themselves,” Segura told SCMagazine.com in a Tuesday email correspondence. “In fact, the JPG itself has very little visibility to the user and is largely a cloaking technique to ensure it is undetected from a security software standpoint.”
>
>
>
[Source](http://www.scmagazine.com/new-variant-of-zeus-banking-trojan-concealed-in-jpg-images/article/334477/). |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | Agreeing with others to say yes this is totally possible, but also to add an interesting anecdote:
Joshua Drake (@jduck), discovered a bug based on a very similar concept (images being interpreted by the OS) which ended up being named "Stagefright", and affected a [ridiculous number of Android devices](http://www.androidcentral.com/stagefright).
He also discovered a similar image based bug in [libpng](http://libpng.org/pub/png/libpng.html) that would cause certain devices to crash. He tweeted an example of the exploit basically saying "Hey, check out this cool malicious PNG I made, it'll probably crash your device", without realising that twitter had added automatic rendering of inline images. Needless to say a lot of his followers started having their machines crash the instant the browser tried to load the image thumbnail in their feed. | Yes this is possible:
>
> A new variant of the nefarious Zeus banking trojan – dubbed ZeusVM – is concealed in JPG image files, according to the collaborative findings of Jerome Segura, senior security researcher with Malwarebytes, and French security researcher Xylitol.
>
>
> The act is known as steganography – concealing messages or images in other messages or images.
>
>
> In the case of ZeusVM, the malware's code is hidden in unassuming JPG images, a Monday blog post by Segura revealed. These photos serve as misdirection for ZeusVM to retrieve its configuration file.
>
>
> “The JPG contains the malware configuration file, which is essentially a list of scripts and financial institutions - but doesn't need to be opened by the victim themselves,” Segura told SCMagazine.com in a Tuesday email correspondence. “In fact, the JPG itself has very little visibility to the user and is largely a cloaking technique to ensure it is undetected from a security software standpoint.”
>
>
>
[Source](http://www.scmagazine.com/new-variant-of-zeus-banking-trojan-concealed-in-jpg-images/article/334477/). |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | Unrealistic? There was recent critical bug in font definition parsing: <https://technet.microsoft.com/en-us/library/security/ms15-078.aspx> and libjpeg changenotes are full of security advisories. Parsing files[1] is hard: overflows, underflows, out of bounds access. Recently there were many fuzzing tools developed for semi-automatic detection of input that can cause crash.
[1] or network packets, XML or even SQL queries. | Yes this is possible:
>
> A new variant of the nefarious Zeus banking trojan – dubbed ZeusVM – is concealed in JPG image files, according to the collaborative findings of Jerome Segura, senior security researcher with Malwarebytes, and French security researcher Xylitol.
>
>
> The act is known as steganography – concealing messages or images in other messages or images.
>
>
> In the case of ZeusVM, the malware's code is hidden in unassuming JPG images, a Monday blog post by Segura revealed. These photos serve as misdirection for ZeusVM to retrieve its configuration file.
>
>
> “The JPG contains the malware configuration file, which is essentially a list of scripts and financial institutions - but doesn't need to be opened by the victim themselves,” Segura told SCMagazine.com in a Tuesday email correspondence. “In fact, the JPG itself has very little visibility to the user and is largely a cloaking technique to ensure it is undetected from a security software standpoint.”
>
>
>
[Source](http://www.scmagazine.com/new-variant-of-zeus-banking-trojan-concealed-in-jpg-images/article/334477/). |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | As others have pointed out, such attacks usually exploit buffer overflows.
Regarding the nuts-and-bolts of how, it's called a stack-smashing attack. It involves corrupting the [call stack](https://en.wikipedia.org/wiki/Call_stack), and overwriting an address to legitimate code to be executed with an address to attacker-supplied code, which gets executed instead.
You can find details at [insecure.org/stf/smashstack.html](http://insecure.org/stf/smashstack.html). | Yes this is possible:
>
> A new variant of the nefarious Zeus banking trojan – dubbed ZeusVM – is concealed in JPG image files, according to the collaborative findings of Jerome Segura, senior security researcher with Malwarebytes, and French security researcher Xylitol.
>
>
> The act is known as steganography – concealing messages or images in other messages or images.
>
>
> In the case of ZeusVM, the malware's code is hidden in unassuming JPG images, a Monday blog post by Segura revealed. These photos serve as misdirection for ZeusVM to retrieve its configuration file.
>
>
> “The JPG contains the malware configuration file, which is essentially a list of scripts and financial institutions - but doesn't need to be opened by the victim themselves,” Segura told SCMagazine.com in a Tuesday email correspondence. “In fact, the JPG itself has very little visibility to the user and is largely a cloaking technique to ensure it is undetected from a security software standpoint.”
>
>
>
[Source](http://www.scmagazine.com/new-variant-of-zeus-banking-trojan-concealed-in-jpg-images/article/334477/). |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | >
> Is there such a thing?
>
>
>
Absolutely. Feeding malicious input to a parser is one of the most common ways of creating an exploit (and, for a JPEG, "decompression" is "parsing").
>
> Is this description based on some real exploit?
>
>
>
It might be based on the [Microsoft Windows GDI+ buffer overflow vulnerability](http://www.kb.cert.org/vuls/id/297462):
>
> There is a buffer overflow vulnerability in the way the JPEG parsing
> component of GDI+ (Gdiplus.dll) handles **malformed JPEG images**. By
> introducing a **specially crafted JPEG file** to the vulnerable component,
> a remote attacker could trigger a buffer overflow condition.
>
>
> ...
>
>
> A remote, unauthenticated attacker could potentially execute arbitrary
> code on a vulnerable system by introducing a specially crafted JPEG
> file. This **malicious JPEG image may be introduced to the system via a**
> malicious web page, **HTML email, or an email attachment.**
>
>
>
.
>
> This was published in December 2006.
>
>
>
The GDI+ JPEG parsing vulnerability was published in September 2004.
>
> Is it sensible to say "the operating system" was decompressing the image to render it?
>
>
>
Sure; in this case, it was a system library that required an OS vendor patch to correct it. Often such libraries are used by multiple software packages, making them part of the operating system rather than application-specific.
In actuality, "the email application invoked a system library to parse a JPEG," but "the operating system" is close enough for a novel. | Agreeing with others to say yes this is totally possible, but also to add an interesting anecdote:
Joshua Drake (@jduck), discovered a bug based on a very similar concept (images being interpreted by the OS) which ended up being named "Stagefright", and affected a [ridiculous number of Android devices](http://www.androidcentral.com/stagefright).
He also discovered a similar image based bug in [libpng](http://libpng.org/pub/png/libpng.html) that would cause certain devices to crash. He tweeted an example of the exploit basically saying "Hey, check out this cool malicious PNG I made, it'll probably crash your device", without realising that twitter had added automatic rendering of inline images. Needless to say a lot of his followers started having their machines crash the instant the browser tried to load the image thumbnail in their feed. |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | >
> Is there such a thing?
>
>
>
Absolutely. Feeding malicious input to a parser is one of the most common ways of creating an exploit (and, for a JPEG, "decompression" is "parsing").
>
> Is this description based on some real exploit?
>
>
>
It might be based on the [Microsoft Windows GDI+ buffer overflow vulnerability](http://www.kb.cert.org/vuls/id/297462):
>
> There is a buffer overflow vulnerability in the way the JPEG parsing
> component of GDI+ (Gdiplus.dll) handles **malformed JPEG images**. By
> introducing a **specially crafted JPEG file** to the vulnerable component,
> a remote attacker could trigger a buffer overflow condition.
>
>
> ...
>
>
> A remote, unauthenticated attacker could potentially execute arbitrary
> code on a vulnerable system by introducing a specially crafted JPEG
> file. This **malicious JPEG image may be introduced to the system via a**
> malicious web page, **HTML email, or an email attachment.**
>
>
>
.
>
> This was published in December 2006.
>
>
>
The GDI+ JPEG parsing vulnerability was published in September 2004.
>
> Is it sensible to say "the operating system" was decompressing the image to render it?
>
>
>
Sure; in this case, it was a system library that required an OS vendor patch to correct it. Often such libraries are used by multiple software packages, making them part of the operating system rather than application-specific.
In actuality, "the email application invoked a system library to parse a JPEG," but "the operating system" is close enough for a novel. | Unrealistic? There was recent critical bug in font definition parsing: <https://technet.microsoft.com/en-us/library/security/ms15-078.aspx> and libjpeg changenotes are full of security advisories. Parsing files[1] is hard: overflows, underflows, out of bounds access. Recently there were many fuzzing tools developed for semi-automatic detection of input that can cause crash.
[1] or network packets, XML or even SQL queries. |
97,856 | The novel [Daemon](https://en.wikipedia.org/wiki/Daemon_(novel_series)) is frequently praised for being realistic in its portrayal rather than just mashing buzzwords.
However, this struck me as unrealistic:
>
> Gragg's e-mail contained a poisoned JPEG of the brokerage logo. JPEGs were compressed image files. When the user viewed the e-mail, the operating system ran a decompression algorithm to render the graphic on-screen; it was this decompression algorithm that executed Gragg's malicious script and let him slip inside the user's system—granting him full access. There was a patch available for the decompression flaw, but older, rich folks typically had no clue about security patches.
>
>
>
Is there such a thing? Is this description based on some real exploit?
This was published in December 2006.
Is it sensible to say "the operating system" was decompressing the image to render it?
---
Note this has nothing to do with security of PHP image uploading scripts. I'm asking about the *decoding process of displaying a JPEG*, not scripts taking input from remote users, nor files misnamed as `.jpeg`. The duplicate flagging I'm responding to looks poor even for a buzzword match; really nothing alike other than mentioning image files. | 2015/08/26 | [
"https://security.stackexchange.com/questions/97856",
"https://security.stackexchange.com",
"https://security.stackexchange.com/users/80219/"
] | >
> Is there such a thing?
>
>
>
Absolutely. Feeding malicious input to a parser is one of the most common ways of creating an exploit (and, for a JPEG, "decompression" is "parsing").
>
> Is this description based on some real exploit?
>
>
>
It might be based on the [Microsoft Windows GDI+ buffer overflow vulnerability](http://www.kb.cert.org/vuls/id/297462):
>
> There is a buffer overflow vulnerability in the way the JPEG parsing
> component of GDI+ (Gdiplus.dll) handles **malformed JPEG images**. By
> introducing a **specially crafted JPEG file** to the vulnerable component,
> a remote attacker could trigger a buffer overflow condition.
>
>
> ...
>
>
> A remote, unauthenticated attacker could potentially execute arbitrary
> code on a vulnerable system by introducing a specially crafted JPEG
> file. This **malicious JPEG image may be introduced to the system via a**
> malicious web page, **HTML email, or an email attachment.**
>
>
>
.
>
> This was published in December 2006.
>
>
>
The GDI+ JPEG parsing vulnerability was published in September 2004.
>
> Is it sensible to say "the operating system" was decompressing the image to render it?
>
>
>
Sure; in this case, it was a system library that required an OS vendor patch to correct it. Often such libraries are used by multiple software packages, making them part of the operating system rather than application-specific.
In actuality, "the email application invoked a system library to parse a JPEG," but "the operating system" is close enough for a novel. | As others have pointed out, such attacks usually exploit buffer overflows.
Regarding the nuts-and-bolts of how, it's called a stack-smashing attack. It involves corrupting the [call stack](https://en.wikipedia.org/wiki/Call_stack), and overwriting an address to legitimate code to be executed with an address to attacker-supplied code, which gets executed instead.
You can find details at [insecure.org/stf/smashstack.html](http://insecure.org/stf/smashstack.html). |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.