qid
int64
1
74.7M
question
stringlengths
12
33.8k
date
stringlengths
10
10
metadata
list
response_j
stringlengths
0
115k
response_k
stringlengths
2
98.3k
15,447
Oftentimes a bishop will move to b5 or g5 early in the game. [![enter image description here](https://i.stack.imgur.com/4R8sa.png)](https://i.stack.imgur.com/4R8sa.png) I personally never do this because I see a6 coming, and if I move the bishop back to a4, then b5, nearly trapping the bishop, and certainly having made it a waste of time to move it to b5 in the first place. However, when I watch games by top-level players, bishops get moved to the 5 row all the time, and nobody ever fights them away with pawns like I think they will. Why not?
2016/09/22
[ "https://chess.stackexchange.com/questions/15447", "https://chess.stackexchange.com", "https://chess.stackexchange.com/users/725/" ]
Here are a couple of relevant questions and answers that might help to explain why the bishop move happens: * [Morphy Defence in Ruy Lopez](https://chess.stackexchange.com/questions/4733/morphy-defence-in-ruy-lopez/4734#4734) * [Why is the Ruy Lopez such a common opening?](https://chess.stackexchange.com/questions/226/why-is-the-ruy-lopez-such-a-common-opening) In summary the main points seem to be that: * Black loses as much time chasing the bishop as White does retreating it * White can gain a long-term advantage from Black's weaker Queen side pawn structure As you might expect in such a popular opening though it's very finely balanced for both sides. If as you say you are *not* seeing a6 and b5 in the games you're looking at, you might be looking at games where Black plays the Berlin Defence: 3... Nf6. This is quite popular at the moment in very high level games. If you want to see some games where the pawn move a6 is played, "putting the question to the bishop", here's a link: <http://www.chessgames.com/perl/chessopening?eco=C70>
"Spanish Bishop" can become very useful and strong piece, especially when it arrives on the c2 square sometimes.
68,299
I've a question which bothers me since quite some time now. I have several environments and servers. All of them are connected to 1 Gbit ethernet swiches (e.g. Cisco 3560). My understanding is, that a 1 Gbit link should provide 125Mbyte/s - of course this is theory. But at least it should reach ~100Mbyte/s. The problem here is, that one copy process only reaches ~20Mbyte/s. I'm aware of these factors but they don't make any difference: * Source and Destination on the same switch or not * Copy utility: SCP, Rsync, Windows copy, Windows robocopy * SMB/CIFS, NFS, iSCSI * Disk storage: NetApp FAS, locally attached 15k SCSI With all of these configurations, I never get more than ~25Mbyte/s throughput. The thing is, that if I start multiple parallel copy streams, like 3 times an rsync, i almost reach 90Mbyte/s. I also did some IOMeter tests and found out that the "chunksize" makes a huge difference, but is normally not tunable with the given tools (or is it?). Jumbo frames are not enabled but I'm unsure if it would make a difference. TOE is enabled on all NICs. What are the bottlenecks you would think about? Do you have similar experiences? Are these the expected "natural" values? Thanks in advance
2009/09/24
[ "https://serverfault.com/questions/68299", "https://serverfault.com", "https://serverfault.com/users/8085/" ]
Speed. We load jQuery, jQuery UI from [Google AJAX Libraries API](http://code.google.com/apis/ajaxlibs/), which increases the chance there's a cached version of those libraries in any visitor's cache. And Google's infrastructure / CDN is better optimized for serving these kinds of static files than our own web server. Other than that, the bandwidth savings are real - those two libraries are quite heavy together. Con: I don't trust any other host enough to host a library on their site. For any library not listed on Google, we host it ourself.
I think the only real reason is to have always up to date. I'm against linking libs, scripts, etc because I think my traffic stats are a value to be kept at home. Moreover it is quite trivial to have the lib hosted and up to date, a cronjob can do the trick easily, efficiently and safely.
68,299
I've a question which bothers me since quite some time now. I have several environments and servers. All of them are connected to 1 Gbit ethernet swiches (e.g. Cisco 3560). My understanding is, that a 1 Gbit link should provide 125Mbyte/s - of course this is theory. But at least it should reach ~100Mbyte/s. The problem here is, that one copy process only reaches ~20Mbyte/s. I'm aware of these factors but they don't make any difference: * Source and Destination on the same switch or not * Copy utility: SCP, Rsync, Windows copy, Windows robocopy * SMB/CIFS, NFS, iSCSI * Disk storage: NetApp FAS, locally attached 15k SCSI With all of these configurations, I never get more than ~25Mbyte/s throughput. The thing is, that if I start multiple parallel copy streams, like 3 times an rsync, i almost reach 90Mbyte/s. I also did some IOMeter tests and found out that the "chunksize" makes a huge difference, but is normally not tunable with the given tools (or is it?). Jumbo frames are not enabled but I'm unsure if it would make a difference. TOE is enabled on all NICs. What are the bottlenecks you would think about? Do you have similar experiences? Are these the expected "natural" values? Thanks in advance
2009/09/24
[ "https://serverfault.com/questions/68299", "https://serverfault.com", "https://serverfault.com/users/8085/" ]
I think the only real reason is to have always up to date. I'm against linking libs, scripts, etc because I think my traffic stats are a value to be kept at home. Moreover it is quite trivial to have the lib hosted and up to date, a cronjob can do the trick easily, efficiently and safely.
If you link to a resource to save your bandwidth (or to try improve response times to the user in the case of large common libraries) be aware of two potential major problems: 1. The external host may go down at some point, due to accident, DoS or planned maintenance. This may cause your site to break so make sure you account for it (if only by giving the user a useful-ish message like "Unable to load JQuery, some features may be broken. Please refresh this page. If the problem persists please report to ..." 2. When a new version appears it may have some incompatibility with your other code due to a bad assumption (i.e. you have relied on an undocumented or officially undefined behaviour that has changed between versions, or new features bring with them a namespace conflict). This may break your pages until you find and work around the issue. If you hot-link to a script or anything else of importance to the running of your application make sure that you keep a local copy anyway. That way you can switch to using the local copy as an interim measure if either of the above problems happen and don't get fixed quickly.
68,299
I've a question which bothers me since quite some time now. I have several environments and servers. All of them are connected to 1 Gbit ethernet swiches (e.g. Cisco 3560). My understanding is, that a 1 Gbit link should provide 125Mbyte/s - of course this is theory. But at least it should reach ~100Mbyte/s. The problem here is, that one copy process only reaches ~20Mbyte/s. I'm aware of these factors but they don't make any difference: * Source and Destination on the same switch or not * Copy utility: SCP, Rsync, Windows copy, Windows robocopy * SMB/CIFS, NFS, iSCSI * Disk storage: NetApp FAS, locally attached 15k SCSI With all of these configurations, I never get more than ~25Mbyte/s throughput. The thing is, that if I start multiple parallel copy streams, like 3 times an rsync, i almost reach 90Mbyte/s. I also did some IOMeter tests and found out that the "chunksize" makes a huge difference, but is normally not tunable with the given tools (or is it?). Jumbo frames are not enabled but I'm unsure if it would make a difference. TOE is enabled on all NICs. What are the bottlenecks you would think about? Do you have similar experiences? Are these the expected "natural" values? Thanks in advance
2009/09/24
[ "https://serverfault.com/questions/68299", "https://serverfault.com", "https://serverfault.com/users/8085/" ]
Speed. We load jQuery, jQuery UI from [Google AJAX Libraries API](http://code.google.com/apis/ajaxlibs/), which increases the chance there's a cached version of those libraries in any visitor's cache. And Google's infrastructure / CDN is better optimized for serving these kinds of static files than our own web server. Other than that, the bandwidth savings are real - those two libraries are quite heavy together. Con: I don't trust any other host enough to host a library on their site. For any library not listed on Google, we host it ourself.
If you link to a resource to save your bandwidth (or to try improve response times to the user in the case of large common libraries) be aware of two potential major problems: 1. The external host may go down at some point, due to accident, DoS or planned maintenance. This may cause your site to break so make sure you account for it (if only by giving the user a useful-ish message like "Unable to load JQuery, some features may be broken. Please refresh this page. If the problem persists please report to ..." 2. When a new version appears it may have some incompatibility with your other code due to a bad assumption (i.e. you have relied on an undocumented or officially undefined behaviour that has changed between versions, or new features bring with them a namespace conflict). This may break your pages until you find and work around the issue. If you hot-link to a script or anything else of importance to the running of your application make sure that you keep a local copy anyway. That way you can switch to using the local copy as an interim measure if either of the above problems happen and don't get fixed quickly.
565,783
I'm using a linux box as a router: The Box has 2 public ips and local ip, i'm using natting to allow local users to access the web. When a local user access the web, source natting happens here, the packets going through the public interface are they checked through the OUTPUT chain or through the Forward chain ? The same Question for the returned packets of the already established session are they check via the INPUT or forward chain ? One last thing: Consider me as a local user the packets will go like this PC ---> LinuxBox Local Interface ---> Linux Box Public interface ----> External world. When the packets come back: External Word ---->Public interface ----> **Local interface** ----->PC If i want to apply a rule on the packet when they move from the local interface to the PC, should i use the FORWARD chain or the OUTPUT chain ? What do you think ??
2014/01/08
[ "https://serverfault.com/questions/565783", "https://serverfault.com", "https://serverfault.com/users/192415/" ]
Any packets going *through* the router is handled in the FORWARD chain. They will NEVER touch INPUT or OUTPUT. Any packets that originate from the router itself will be handled by OUTPUT. Never FORWARD. Any packets destined to an address that is assigned to one of the routers interfaces, will be handled by INPUT chain. Never FORWARD. The only (kind-of) exception to the INPUT/OUTPUT never being handled by INPUT is if you apply any Destination NAT (as opposed to Source NAT) rules, in which case the destination of a packet originally destined for an address on one of the routers interfaces could be changed to something that is not, in which case it does go to FORWARD because the packet is no longer destined for an address on the local machine.
packets going from your land to the public network are handled in the forward chain. same thing for packets going the other way round. input are for packets for which final destination is the router itself (and are not batted), where ever they come from. and output is for packet originating from the router itself and going to the outside world (be it can or wan) I don't understand your last question but keep in mind you can (and should) configure iptables to be stateful so you don't have to worry about packets from established session, only initial packet.
13,671,909
I'm new to using this scheduler. I will like to know if I need to have a web browser at the page so the scheduler's time always update? Or the scheduler will run so long it have been deployed to tomcat?
2012/12/02
[ "https://Stackoverflow.com/questions/13671909", "https://Stackoverflow.com", "https://Stackoverflow.com/users/903772/" ]
No, that's the whole point. The jobs run in the background while the server is running.
The scheduler requires no active connection to the server to be running (no open browser). In fact, if you are using quartz is mostly because you want to do some tasks on a time basis that does not require user interaction.
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
The method described in [Hallerbach (2004)](http://papers.ssrn.com/sol3/papers.cfm?abstract_id=567721) always worked well for me. > > We derive an estimator for Black-Scholes-Merton implied volatility that, when compared to the familiar Corrado & Miller [JBaF, 1996] estimator, has substantially higher approximation accuracy and extends over a wider region of moneyness. > > >
Jaeckel has a paper "Let's be rational" in which he *"show how Black’s volatility can be implied from option prices with as little as two iterations to maximum attainable precision on standard (64 bit floating point) hardware for all possible inputs."*. I guess it doesn't qualify as closed-form for you, though one might argue that having to apply a deterministic algorithm twice to get accurate answer with machine precision, sort of is. FWIW, I've not tried to implement what Jaeckal did in "Let's be rational" yet, but I have implemented his previous paper "By implication", which always worked well for me (but relies on a root-search without any guarantees on how quickly it converges).
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
The method described in [Hallerbach (2004)](http://papers.ssrn.com/sol3/papers.cfm?abstract_id=567721) always worked well for me. > > We derive an estimator for Black-Scholes-Merton implied volatility that, when compared to the familiar Corrado & Miller [JBaF, 1996] estimator, has substantially higher approximation accuracy and extends over a wider region of moneyness. > > >
Let's Be Rational uses exactly two iterations to give full machine accuracy for all inputs. It can be viewed as a three-stage analytical formula if you like. The code is free to download at www.jaeckel.org. Rgds, Peter
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
The method described in [Hallerbach (2004)](http://papers.ssrn.com/sol3/papers.cfm?abstract_id=567721) always worked well for me. > > We derive an estimator for Black-Scholes-Merton implied volatility that, when compared to the familiar Corrado & Miller [JBaF, 1996] estimator, has substantially higher approximation accuracy and extends over a wider region of moneyness. > > >
There are some other references: * [Li and Lee (2009)](http://dx.doi.org/10.1080/14697680902849361) [[download]](https://ssrn.com/abstract=1027282) An adaptive successive over-relaxation method for computing the Black–Scholes implied volatility * [Stefanica and Radoicic (2017)](https://ssrn.com/abstract=2908494) An Explicit Implied Volatility Formula Related discussions on the implied volatility inversion: * [How can the implied volatility be calculated?](https://quant.stackexchange.com/questions/7761/how-to-compute-implied-volatility-calculation) * [What is an efficient method to find implied volatility?](https://quant.stackexchange.com/questions/15198/what-is-an-efficient-method-to-find-implied-volatility) For the normal (or Bachelier) implied volatility, there's an improvement to Choi et al (2009) [[paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](https://ssrn.com/abstract=990747)] mentioned in the question: * [Fabien Le Floc'h (2016)](https://ssrn.com/abstract=2420757)
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
The method described in [Hallerbach (2004)](http://papers.ssrn.com/sol3/papers.cfm?abstract_id=567721) always worked well for me. > > We derive an estimator for Black-Scholes-Merton implied volatility that, when compared to the familiar Corrado & Miller [JBaF, 1996] estimator, has substantially higher approximation accuracy and extends over a wider region of moneyness. > > >
Peter Jaeckel methods from the papers mentioned are the industry standard used by most practitioners to get IV. In addition in practice the article you mention is probably of very little use because the analytic approximation you refer to in the SSRN paper needs both call and put price to extract the implied vol however usually only one of the 2 instruments is liquid when you are not close to ATM.
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
Let's Be Rational uses exactly two iterations to give full machine accuracy for all inputs. It can be viewed as a three-stage analytical formula if you like. The code is free to download at www.jaeckel.org. Rgds, Peter
Jaeckel has a paper "Let's be rational" in which he *"show how Black’s volatility can be implied from option prices with as little as two iterations to maximum attainable precision on standard (64 bit floating point) hardware for all possible inputs."*. I guess it doesn't qualify as closed-form for you, though one might argue that having to apply a deterministic algorithm twice to get accurate answer with machine precision, sort of is. FWIW, I've not tried to implement what Jaeckal did in "Let's be rational" yet, but I have implemented his previous paper "By implication", which always worked well for me (but relies on a root-search without any guarantees on how quickly it converges).
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
There are some other references: * [Li and Lee (2009)](http://dx.doi.org/10.1080/14697680902849361) [[download]](https://ssrn.com/abstract=1027282) An adaptive successive over-relaxation method for computing the Black–Scholes implied volatility * [Stefanica and Radoicic (2017)](https://ssrn.com/abstract=2908494) An Explicit Implied Volatility Formula Related discussions on the implied volatility inversion: * [How can the implied volatility be calculated?](https://quant.stackexchange.com/questions/7761/how-to-compute-implied-volatility-calculation) * [What is an efficient method to find implied volatility?](https://quant.stackexchange.com/questions/15198/what-is-an-efficient-method-to-find-implied-volatility) For the normal (or Bachelier) implied volatility, there's an improvement to Choi et al (2009) [[paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](https://ssrn.com/abstract=990747)] mentioned in the question: * [Fabien Le Floc'h (2016)](https://ssrn.com/abstract=2420757)
Jaeckel has a paper "Let's be rational" in which he *"show how Black’s volatility can be implied from option prices with as little as two iterations to maximum attainable precision on standard (64 bit floating point) hardware for all possible inputs."*. I guess it doesn't qualify as closed-form for you, though one might argue that having to apply a deterministic algorithm twice to get accurate answer with machine precision, sort of is. FWIW, I've not tried to implement what Jaeckal did in "Let's be rational" yet, but I have implemented his previous paper "By implication", which always worked well for me (but relies on a root-search without any guarantees on how quickly it converges).
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
Let's Be Rational uses exactly two iterations to give full machine accuracy for all inputs. It can be viewed as a three-stage analytical formula if you like. The code is free to download at www.jaeckel.org. Rgds, Peter
Peter Jaeckel methods from the papers mentioned are the industry standard used by most practitioners to get IV. In addition in practice the article you mention is probably of very little use because the analytic approximation you refer to in the SSRN paper needs both call and put price to extract the implied vol however usually only one of the 2 instruments is liquid when you are not close to ATM.
14,706
While the solution for IV can certainly be reached using numerical search methods, I wonder if a high precision closed-form approximation exists. For example, there is a very robust (precise within 10^-12) approximation for Bachelier IV ([paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](http://ssrn.com/abstract=990747)), but is there anything similar for Black'76 and/or Black-Scholes?
2014/09/12
[ "https://quant.stackexchange.com/questions/14706", "https://quant.stackexchange.com", "https://quant.stackexchange.com/users/5086/" ]
There are some other references: * [Li and Lee (2009)](http://dx.doi.org/10.1080/14697680902849361) [[download]](https://ssrn.com/abstract=1027282) An adaptive successive over-relaxation method for computing the Black–Scholes implied volatility * [Stefanica and Radoicic (2017)](https://ssrn.com/abstract=2908494) An Explicit Implied Volatility Formula Related discussions on the implied volatility inversion: * [How can the implied volatility be calculated?](https://quant.stackexchange.com/questions/7761/how-to-compute-implied-volatility-calculation) * [What is an efficient method to find implied volatility?](https://quant.stackexchange.com/questions/15198/what-is-an-efficient-method-to-find-implied-volatility) For the normal (or Bachelier) implied volatility, there's an improvement to Choi et al (2009) [[paper](http://dx.doi.org/10.1080/13504860802583436) / [SSRN](https://ssrn.com/abstract=990747)] mentioned in the question: * [Fabien Le Floc'h (2016)](https://ssrn.com/abstract=2420757)
Peter Jaeckel methods from the papers mentioned are the industry standard used by most practitioners to get IV. In addition in practice the article you mention is probably of very little use because the analytic approximation you refer to in the SSRN paper needs both call and put price to extract the implied vol however usually only one of the 2 instruments is liquid when you are not close to ATM.
35,206
I have a USB device (USB to RS-232) that works on some computers without drivers but not on others. This is a problem because I don't have the drivers disk for it nor any info on who made it. *Are there any tools that will tell me what kind of device it is or whatever info windows (and other OS's) pulls in when it's trying to set it up?*
2009/07/02
[ "https://serverfault.com/questions/35206", "https://serverfault.com", "https://serverfault.com/users/1039/" ]
A list of USB vendor and product IDs is available [here](http://www.linux-usb.org/usb.ids). These programs can give you some or all of the information you need in order to try to track down a source for drivers: * [USBDview](http://www.nirsoft.net/utils/usb_devices_view.html) * [USBView](http://www.ftdichip.com/Resources/Utilities.htm) * [SniffUSB](http://www.pcausa.com/Utilities/UsbSnoop/) * [Bus Hound](http://www.perisoft.net/bushound/) - USB, SCSI, IDE, Firewire, etc.
I've had good luck with Everest <http://www.softpedia.com/get/System/System-Info/Everest-Home-Edition.shtml> It will give you all kinds of info about devices that are present without drivers.
35,206
I have a USB device (USB to RS-232) that works on some computers without drivers but not on others. This is a problem because I don't have the drivers disk for it nor any info on who made it. *Are there any tools that will tell me what kind of device it is or whatever info windows (and other OS's) pulls in when it's trying to set it up?*
2009/07/02
[ "https://serverfault.com/questions/35206", "https://serverfault.com", "https://serverfault.com/users/1039/" ]
Its not a general solution, but a USB to RS-232 interface cable has a high probability of being built out of either a [Silicon Labs CP210x](https://www.silabs.com/products/interface/usbtouart/Pages/default.aspx) chip, or a close relative of the [FTDI FT232R](http://www.ftdichip.com/Products/FT232R.htm) chip. You can get driver packages for both easily on-line, and the odds that one is right are pretty good. Finding VID, PID, REV, CLASS, SUBCLASS, PROTOCOL etc. ----------------------------------------------------- In the general case, you should be able to extract the VID/PID and device class codes from the node in the device manager corresponding to the unknown device. VID and PID can be read from the Details tab, by selecting Hardware Ids, and the Class and Subclass by selecting Compatible Ids. VID/PID/REV ----------- The VID can usually be translated to an actual vendor name since vendors are **supposed** to buy VIDs and register with the [USB-IF](http://www.usb.org/). The [list linked by Dennis](http://www.linux-usb.org/usb.ids) is kept reasonably up to date, and may be the quickest way to just look up a vendor name. Windows will match a driver as specifically as it can, so a vendor can bug-fix a product without using up one of their precious PID codes by modifying the revision code, and produce a driver that compensates for the bug in old devices by matching it to the triple of VID/PID/REV. CLASS/SUB/PROTO --------------- If the device implements one of the standard device classes, then the VID/PID don't matter nearly as much as the class codes. This is why nearly any USB thumb drive "just works"; they all promise to implement the Mass Storage Class, and Windows knows that means they act like a disk and should have a recognized partition table and file system. There is a list of the class codes about 90% of the way down the list linked above. Don't be surprised to find a lot of devices that say CLASS 0xFF, which is the vendor-defined class. Part of your problem is that there isn't a completely obvious standard class for an RS232 bridge device, and it doesn't look Windows has nominated a right answer with a stock driver either. The classes in which you might put a USB to serial translator got standardized relatively late in the game, not all flavors of Windows have a stock driver for it, and not all of the devices claim to implement it.
I've had good luck with Everest <http://www.softpedia.com/get/System/System-Info/Everest-Home-Edition.shtml> It will give you all kinds of info about devices that are present without drivers.
52,379
The phrase "Man will believe anything, as long as it's not in the Bible" is attributed to Napoleon on many sites but there are no further details (date, context,etc.) available, nor a trace of primary sources. Some examples: [AZ Quotes](https://www.azquotes.com/quote/1056429), [Quote Fancy](https://quotefancy.com/quote/870652/Napoleon-Man-will-believe-anything-as-long-as-it-s-not-in-the-bible), [Love Expands](https://loveexpands.com/quotes/napoleon-bonaparte-700294/).
2021/09/23
[ "https://skeptics.stackexchange.com/questions/52379", "https://skeptics.stackexchange.com", "https://skeptics.stackexchange.com/users/17938/" ]
To a bit more history here, there's a 1998 book *Soaring and Settling* by [Rita Gross](https://en.wikipedia.org/wiki/Rita_Gross), which has a different take/version (p. 43): > > I am just as frustrated that many Caucasian Buddhists, reasonably sophisticated in their assessments of certain elements of Christian tradition that are woefully out of date, will believe anything as long as it is said by a Tibetan lama. > > > I suspect it's this passage or something similar that was the basis of the post-2000 broader but similar pronouncements. Regarding Napoleon, he [generally](https://journals.openedition.org/rh19/622) found Catholicism useful, but Protestantism not so much. So it's quite unlikely he would have said something that simplistic about the Bible. On the other hand, Napoleon was/is more often demonized or derided in Protestant circles, which is a more plausible reason why the quote might have been attributed to him in such circles.
There have been similar phrases uttered by many different people. Very often, it is said by religious leaders attacking atheism, especially by creationists attacking Darwinism. Attribution to Napoleon goes back a long way. [Here's a book from 1883](https://www.google.co.uk/books/edition/The_Church_Review/pAPOAAAAMAAJ?hl=en&gbpv=1&dq=Duroc%20%22believe%20anything%22&pg=PA14&printsec=frontcover). Note that the anecdote is uncited, and told by an anti-Darwinist. Very probably apocryphal. > > It is related of Napoleon that one day his friend Duroc was reciting to him a marvelous story. Duroc was a skeptic, indeed more than that - an avowed unbeliever. As he ended the improbable tale Napoleon answered "Some men can believe anything but the Bible." > > > [Here's an earlier example](https://www.google.co.uk/books/edition/Life/mNwXAAAAYAAJ?hl=en&gbpv=1&dq=%22believe%20anything%22%20%22not%20in%20the%20bible%22&pg=PA81&printsec=frontcover),from 1868, attributing the remark to Napoleon speaking of Duroc, without context, and without a cite. The [earliest example](https://www.google.co.uk/books/edition/British_and_Foreign_Medico_chirurgical_R/205YAAAAMAAJ?hl=en&gbpv=1&dq=%22believe%20anything%22%20%22not%20in%20the%20bible%22&pg=PA173&printsec=frontcover) I've been able to find of the phrase itself is in a medical journal dated July 1852. No attribution to Napoleon. > > It is an old observation that complete scepticism in religion is often allied with extraordinary credulity in other matters; and we have a remarkable verification of this in the present time, in the authors of 'Man's Nature and Development' of one of whom it has been not unaptly or untruly remarked, that "she will believe anything that is not in the Bible." > > >
52,379
The phrase "Man will believe anything, as long as it's not in the Bible" is attributed to Napoleon on many sites but there are no further details (date, context,etc.) available, nor a trace of primary sources. Some examples: [AZ Quotes](https://www.azquotes.com/quote/1056429), [Quote Fancy](https://quotefancy.com/quote/870652/Napoleon-Man-will-believe-anything-as-long-as-it-s-not-in-the-bible), [Love Expands](https://loveexpands.com/quotes/napoleon-bonaparte-700294/).
2021/09/23
[ "https://skeptics.stackexchange.com/questions/52379", "https://skeptics.stackexchange.com", "https://skeptics.stackexchange.com/users/17938/" ]
To a bit more history here, there's a 1998 book *Soaring and Settling* by [Rita Gross](https://en.wikipedia.org/wiki/Rita_Gross), which has a different take/version (p. 43): > > I am just as frustrated that many Caucasian Buddhists, reasonably sophisticated in their assessments of certain elements of Christian tradition that are woefully out of date, will believe anything as long as it is said by a Tibetan lama. > > > I suspect it's this passage or something similar that was the basis of the post-2000 broader but similar pronouncements. Regarding Napoleon, he [generally](https://journals.openedition.org/rh19/622) found Catholicism useful, but Protestantism not so much. So it's quite unlikely he would have said something that simplistic about the Bible. On the other hand, Napoleon was/is more often demonized or derided in Protestant circles, which is a more plausible reason why the quote might have been attributed to him in such circles.
Napoleon born: [1769](https://en.wikipedia.org/wiki/Napoleon) First known attestation: [1716](https://books.google.com/books/about/The_Free_holder_Or_Political_Essays.html?id=Yu8Gf39YGgkC&printsec=frontcover&newbks=1&newbks_redir=0&source=gb_mobile_entity#v=onepage&q=Learned%20divine&f=false) Conclusion: NO! (Not the first at least.) It's that easy to prove this negative. But in its place we only find the allegation that it was really [Charles II](https://en.wikipedia.org/wiki/Charles_II_of_England) who said it, referring to [Isaac Vossius](https://en.wikipedia.org/wiki/Isaac_Vossius): > > This Learned Divine, said he, is a very strange Man: He believes every thing but the Bible. > > > ([Alternative link to the quote](https://quod.lib.umich.edu/cgi/t/text/pageviewer-idx?cc=ecco;c=ecco;idno=004856157.0001.000;node=004856157.0001.000:6;seq=45;page=root;view=text))
227,938
Unsurprisingly, the Lightning cable that came with my iPhone 5s ~18 months ago is really worn - to the point where all of the rubber cable has had to come off because it had been left around for a while and had gone all sticky (sooooo gross). The outer metal sheathing is still entirely intact. The resulting cable looks like so. [![It's a *little* bit worn](https://i.stack.imgur.com/L0eL1.jpg)](https://i.stack.imgur.com/L0eL1.jpg) It still charges (and syncs) my iPhone perfectly. Is this at all dangerous to use (I'm charging my phone from my MacBook)? Cheers, Dan
2016/02/17
[ "https://apple.stackexchange.com/questions/227938", "https://apple.stackexchange.com", "https://apple.stackexchange.com/users/158857/" ]
Such Voltage is not 'dangerous' but you should get a new one as soon as possible. Why risk a further damage to your iPhone, your Mac or your charger?
Just underneath the plastic exterior of the cable is a conductive layer called shielding. Shielding is grounded, meaning it is connected to 0 V. It would only be dangerous if you had a serious electrical problem that caused ground in your outlet to be connected to a higher voltage potential. The actual power is transmitted on wires in the core of the cable. While the voltage at the core for iPhone charge cables is not typically dangerous, if the center of your cable becomes exposed, it's time to get a new cable.
743,422
The "classic" high school probability exercises most of the time ends up in a tree of possibilities that deals with all the scenarios (combinations?) that are possible for a given scenario: be it coin tosses, balls in a jar, anagrams, etc. But there is something that I always have a had time deciding: when to add and when to multiply quantities under a certain scenario? I see that for each branch of the tree, I can just multiply, but when does addition enters the picture? ![Image 1](https://i.stack.imgur.com/YCYsw.png)
2014/04/07
[ "https://math.stackexchange.com/questions/743422", "https://math.stackexchange.com", "https://math.stackexchange.com/users/45247/" ]
Add for (disjoint) unions, multiply for (independent) intersections.
This is how I try to think of it from non math perspective. Not sure how accurate this is. If you think that a compounding event will increase your chance compared to the single event, you use the addition rule. Examples: * Probability of rolling 4 or bigger in a fair dice. (The OR generally increase your chance as well). The single event is rolling a number(1/6). The compounding events are rolling 4 or bigger. Logically, you will have more chance. You add. * Probability of getting Number cards from a deck of card. You calculate this using 1/52 probability for each card. Generally you would get a bigger chance if the ask is to drawANY number. Therefore, you add However, if you think compounding events will decrease your chance compared to the single event, use the multiplication rule. Most of the time you will see the word AND indicating 2 events happening together Example: * Probability of getting 2 sixes when rolling a dice 2 times( 6 AND 6). The single event of rolling a six is 1/6. If you need to roll another six the second time, logically, this will be harder. Multiply * Probability of drawing 3 Aces in a row. This is harder than just drawing 1 ACE. multiply.
1,028,679
I have a lot of C# code that uses public fields, and I would like to convert them to properties. I have Resharper, and it will do them one by one, but this will take forever. Does anyone know of an automated refactoring tool that can help with this?
2009/06/22
[ "https://Stackoverflow.com/questions/1028679", "https://Stackoverflow.com", "https://Stackoverflow.com/users/67386/" ]
Resharper does it very quickly, using Alt+PageDown / ALt+Enter (with the default key bindings). If you are at the first field, Alt+PageDown will jump to the next one (since it'll include wrapping public fields as a suggested refactoring), and Alt+Enter will prompt you to wrap it in a property. Since you most likely want to avoid a full blanket wrapping of all properties, this is probably the quickest approach. It's quite fast to do this to a class, since it jumps exactly where you need to go...
If you're in VS.NET when you rename a field, VS prompts you to change all occurrences of the renamed field. So change your public Variable to the property name, tell VS to change all the instances of this variable, then create a private variable to store the value and a public property of the Proper name. Delete the public Variable and you should be good to go.
1,028,679
I have a lot of C# code that uses public fields, and I would like to convert them to properties. I have Resharper, and it will do them one by one, but this will take forever. Does anyone know of an automated refactoring tool that can help with this?
2009/06/22
[ "https://Stackoverflow.com/questions/1028679", "https://Stackoverflow.com", "https://Stackoverflow.com/users/67386/" ]
Resharper does it very quickly, using Alt+PageDown / ALt+Enter (with the default key bindings). If you are at the first field, Alt+PageDown will jump to the next one (since it'll include wrapping public fields as a suggested refactoring), and Alt+Enter will prompt you to wrap it in a property. Since you most likely want to avoid a full blanket wrapping of all properties, this is probably the quickest approach. It's quite fast to do this to a class, since it jumps exactly where you need to go...
Refactor fields to properties (no extensions needed): ===================================================== [![Refactor C# fields to properties](https://i.stack.imgur.com/Bhdd4.gif)](https://i.stack.imgur.com/Bhdd4.gif) **Step 1: Refactor all fields to be encapsulated by properties** **Step 2: Refactor all properties into auto-implemented properties**
1,028,679
I have a lot of C# code that uses public fields, and I would like to convert them to properties. I have Resharper, and it will do them one by one, but this will take forever. Does anyone know of an automated refactoring tool that can help with this?
2009/06/22
[ "https://Stackoverflow.com/questions/1028679", "https://Stackoverflow.com", "https://Stackoverflow.com/users/67386/" ]
Refactor fields to properties (no extensions needed): ===================================================== [![Refactor C# fields to properties](https://i.stack.imgur.com/Bhdd4.gif)](https://i.stack.imgur.com/Bhdd4.gif) **Step 1: Refactor all fields to be encapsulated by properties** **Step 2: Refactor all properties into auto-implemented properties**
If you're in VS.NET when you rename a field, VS prompts you to change all occurrences of the renamed field. So change your public Variable to the property name, tell VS to change all the instances of this variable, then create a private variable to store the value and a public property of the Proper name. Delete the public Variable and you should be good to go.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
To answer simply - an oscilloscope is an essential tool for any electronics lab, whilst an SA is generally not (unless you are an RF engineer, and even then you need a good scope) and for a good quality one much more expensive in comparison (though Rigol have just brought out some pretty powerful SAs at decent scope type prices) The FFT function on your average DSO will do for most work, so unless your frequency range of interest is e.g. > 500MHz or so (if it is let us know), then the DSO is the tool of choice. Basically one does amplitude versus time (scope), and the other does amplitude versus frequency (SA) Scope example: Say you have a digital signal that is intermittently working, you could check on the scope and look for over/undershoot, ringing, noise, gltiches, etc. ![Integrity problems](https://i.stack.imgur.com/1HIa6.gif) (simple) SA example: Say you have a signal and you want to check the harmonic components of it, you can look on the SA screen and check for harmonics (e.g. a pure sine wave should just be one single spike on the screen, at it's frequency, a square wave would be a decreasing series of odd harmonics) Square wave on a Spectrum Analyser: ![SA Square wave](https://i.stack.imgur.com/5Whlb.png) The same signal on a scope would look like this: ![Square wave on scope](https://i.stack.imgur.com/I8AHu.gif)
The difference is that the spectrum analyzer has a mixer frontend allowing it to shift the frequency range it is listening to, while an oscilloscope remains fixed at the lower end. This means that it is possible to see signals at higher frequencies, and at the same time, signals outside of the area being looked at are filtered out, so you can adjust the ADC prescaler for a better resolution. On the other hand, mixers don't like DC at all, so in normal EE work, you won't be able to use a spectrum analyzer in place of an oscilloscope either.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
There were a few correct differences mentioned above, I will try to systemize: 1) Bandwidth (oscilloscope's bandwidth usually wider, but the working band cannot be shifted). I.e. for example Oscilloscope modes are: 0-1kHz, 0-10kHz, 0-50kHz, 0-250kHz, 0-500kHz, 0-2MHz, 0-20MHz,0-100MHz signals, having max sample rate at 500 MSamp/sec. When one looks at FFT, he can see only these 0-100 MHz band. Spectrum Analyzer may have narrower bandwidth, but it can roll across frequency scale: i.e. for example, bandwidth 40 MHz, sampling frequency 200 MSamp/sec, and working frequencies: 0-6.3 GHz. I.e. Spectrum analyzers modes will be: 0-40MHz,10-50MHz, 20-60MHz, 30-70MHz....6260..6300MHz. So one can see, that SA has a tunable band filter instead of anti-aliasing LPF in the oscilloscope. 2) Dynamic range. ADC of a spectrum analyzer has a much better resolution. 3) The spectrum analyzer has a low-noise amplifier, oscilloscope doesn't have it. Low noise amplifier, is a special, radio-frequency amplifier, which works in a big range of frequencies, add very low noise to signal. 4) Oscilloscope and spectrum analyzer have different ways to set up triggers. The oscilloscope is oriented on a shape of signal in the time domain, SA is oriented on capturing certain shapes in the frequency domain. 5) The oscilloscope cannot demodulate signals, a spectrum analyzer usually can (because it is virtually an SDR-receiver). Summarizing: an oscilloscope is an extra-wide band millivoltmeter. spectrum analyzer is a pretty narrow-band receiver, whos main goal to convert radio-waves into the baseband signal (I and Q components) with as low as possible lose and noise.
Another application for a spectrum analyzer is where you'd want to hunt down a source of interference. Latest-gen handhelds make this a lot easier too. For example, in addition to the spectrogram and standard spectrum analyzer measurements, these instruments can make interference-specific measurements such as carrier/noise (C/N) and carrier/interference (C/I). A trace mathematics (diff mode) can help you find, monitor, and characterize interfering signals. Another feature is the ability to record spectrum over a specified time. This allows you to find intermittant faults and frequency variations, over time. Great feature. Personally, I'd go for both: Scope + SA. It just makes your bench more useful, long term.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
To answer simply - an oscilloscope is an essential tool for any electronics lab, whilst an SA is generally not (unless you are an RF engineer, and even then you need a good scope) and for a good quality one much more expensive in comparison (though Rigol have just brought out some pretty powerful SAs at decent scope type prices) The FFT function on your average DSO will do for most work, so unless your frequency range of interest is e.g. > 500MHz or so (if it is let us know), then the DSO is the tool of choice. Basically one does amplitude versus time (scope), and the other does amplitude versus frequency (SA) Scope example: Say you have a digital signal that is intermittently working, you could check on the scope and look for over/undershoot, ringing, noise, gltiches, etc. ![Integrity problems](https://i.stack.imgur.com/1HIa6.gif) (simple) SA example: Say you have a signal and you want to check the harmonic components of it, you can look on the SA screen and check for harmonics (e.g. a pure sine wave should just be one single spike on the screen, at it's frequency, a square wave would be a decreasing series of odd harmonics) Square wave on a Spectrum Analyser: ![SA Square wave](https://i.stack.imgur.com/5Whlb.png) The same signal on a scope would look like this: ![Square wave on scope](https://i.stack.imgur.com/I8AHu.gif)
There were a few correct differences mentioned above, I will try to systemize: 1) Bandwidth (oscilloscope's bandwidth usually wider, but the working band cannot be shifted). I.e. for example Oscilloscope modes are: 0-1kHz, 0-10kHz, 0-50kHz, 0-250kHz, 0-500kHz, 0-2MHz, 0-20MHz,0-100MHz signals, having max sample rate at 500 MSamp/sec. When one looks at FFT, he can see only these 0-100 MHz band. Spectrum Analyzer may have narrower bandwidth, but it can roll across frequency scale: i.e. for example, bandwidth 40 MHz, sampling frequency 200 MSamp/sec, and working frequencies: 0-6.3 GHz. I.e. Spectrum analyzers modes will be: 0-40MHz,10-50MHz, 20-60MHz, 30-70MHz....6260..6300MHz. So one can see, that SA has a tunable band filter instead of anti-aliasing LPF in the oscilloscope. 2) Dynamic range. ADC of a spectrum analyzer has a much better resolution. 3) The spectrum analyzer has a low-noise amplifier, oscilloscope doesn't have it. Low noise amplifier, is a special, radio-frequency amplifier, which works in a big range of frequencies, add very low noise to signal. 4) Oscilloscope and spectrum analyzer have different ways to set up triggers. The oscilloscope is oriented on a shape of signal in the time domain, SA is oriented on capturing certain shapes in the frequency domain. 5) The oscilloscope cannot demodulate signals, a spectrum analyzer usually can (because it is virtually an SDR-receiver). Summarizing: an oscilloscope is an extra-wide band millivoltmeter. spectrum analyzer is a pretty narrow-band receiver, whos main goal to convert radio-waves into the baseband signal (I and Q components) with as low as possible lose and noise.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
To answer simply - an oscilloscope is an essential tool for any electronics lab, whilst an SA is generally not (unless you are an RF engineer, and even then you need a good scope) and for a good quality one much more expensive in comparison (though Rigol have just brought out some pretty powerful SAs at decent scope type prices) The FFT function on your average DSO will do for most work, so unless your frequency range of interest is e.g. > 500MHz or so (if it is let us know), then the DSO is the tool of choice. Basically one does amplitude versus time (scope), and the other does amplitude versus frequency (SA) Scope example: Say you have a digital signal that is intermittently working, you could check on the scope and look for over/undershoot, ringing, noise, gltiches, etc. ![Integrity problems](https://i.stack.imgur.com/1HIa6.gif) (simple) SA example: Say you have a signal and you want to check the harmonic components of it, you can look on the SA screen and check for harmonics (e.g. a pure sine wave should just be one single spike on the screen, at it's frequency, a square wave would be a decreasing series of odd harmonics) Square wave on a Spectrum Analyser: ![SA Square wave](https://i.stack.imgur.com/5Whlb.png) The same signal on a scope would look like this: ![Square wave on scope](https://i.stack.imgur.com/I8AHu.gif)
Another application for a spectrum analyzer is where you'd want to hunt down a source of interference. Latest-gen handhelds make this a lot easier too. For example, in addition to the spectrogram and standard spectrum analyzer measurements, these instruments can make interference-specific measurements such as carrier/noise (C/N) and carrier/interference (C/I). A trace mathematics (diff mode) can help you find, monitor, and characterize interfering signals. Another feature is the ability to record spectrum over a specified time. This allows you to find intermittant faults and frequency variations, over time. Great feature. Personally, I'd go for both: Scope + SA. It just makes your bench more useful, long term.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
An oscilloscope with FFT function uses built in mathematical analysis of the stored waveform to calculate the frequency content and amplitude of the signal. It is displayed on the screen as a frequency vs amplitude graph - just like a spectrum analyser. A 'true' analogue type spectrum analyser, actually measures the amplitude at each frequency ( steps ) from the signal and does not need to do any maths on the measured amplitude other than that required to show the measurement values accurately on the screen. It's true that many oscilloscopes offer a FFT function - but unless you are using a new expensive scope - the resulting display is rather more of a guide than being equivalent to a real spectrum analyser. That said - the newer generation of combined digital instruments do truly offer the same spectrum analysis results and oscilloscope measurements that single task instruments would. They are not cheap however but are useful in that the frequency/analogue content can be synchronised with the digital oscilloscope waveform to identify those signals which are causing RF related problems or EMC.
Another application for a spectrum analyzer is where you'd want to hunt down a source of interference. Latest-gen handhelds make this a lot easier too. For example, in addition to the spectrogram and standard spectrum analyzer measurements, these instruments can make interference-specific measurements such as carrier/noise (C/N) and carrier/interference (C/I). A trace mathematics (diff mode) can help you find, monitor, and characterize interfering signals. Another feature is the ability to record spectrum over a specified time. This allows you to find intermittant faults and frequency variations, over time. Great feature. Personally, I'd go for both: Scope + SA. It just makes your bench more useful, long term.
50,581
Could someone explain to me please which applications demand one or the other and why? As far as I have read it's all about the 'dB'; is that true? And why? At first I can see Digital Storage Oscilloscopes (DSO) with FFT function and Spectrum Analyzers (SA) as being the same thing...they will get a signal from the Time Domain, and convert it to the Frequency Domain and we can check all the harmonics and frequency components of a signal and analyze it in a whole new way.......But since DSOs usually are much cheaper than SA, I keep wondering what functionalities the SA will offer that a DSO can't. Is it about precision, speed of calculation (my DSO FFT is really slow), bandwidth (cheap DSOs usually go only up to 100MHz), or does it just depend on the models and not on being a DSO or a SA? Is there more that I don't know about and you can tell me?
2012/12/05
[ "https://electronics.stackexchange.com/questions/50581", "https://electronics.stackexchange.com", "https://electronics.stackexchange.com/users/15955/" ]
An oscilloscope with FFT function uses built in mathematical analysis of the stored waveform to calculate the frequency content and amplitude of the signal. It is displayed on the screen as a frequency vs amplitude graph - just like a spectrum analyser. A 'true' analogue type spectrum analyser, actually measures the amplitude at each frequency ( steps ) from the signal and does not need to do any maths on the measured amplitude other than that required to show the measurement values accurately on the screen. It's true that many oscilloscopes offer a FFT function - but unless you are using a new expensive scope - the resulting display is rather more of a guide than being equivalent to a real spectrum analyser. That said - the newer generation of combined digital instruments do truly offer the same spectrum analysis results and oscilloscope measurements that single task instruments would. They are not cheap however but are useful in that the frequency/analogue content can be synchronised with the digital oscilloscope waveform to identify those signals which are causing RF related problems or EMC.
Current day spectrum analyzers (SA) are seldom fully sweep tune. Most do FFT and stitch channels together to form a frequency span. Besides a class of modern SA measurement such as Vector Signal Analysis, does not stitch channels, but rather, measure the entire channels base on the IF sampling rate. The analysis bandwidth, which is is usually around [IF sampling rate/1.25] is up to 1 GHz, for the highest end SA -- [Keysight UXA](http://www.keysight.com/en/pdx-x202152-pn-N9040B/uxa-signal-analyzer-multi-touch-3-hz-to-50-ghz?cc=US&lc=eng). Non exhaustive of scope vs spectrum 1. Scope digitize from baseband to the desire frequency range. SA downcovert the RF signals and digitize at IF 2. Being able to digitize at IF allow SA to have better vertical resolution. A scope vertical resolution is mostly 8 bit, while SA is up to 14 bit. (Digitizer designers trade sampling rate with vertical resolution) 3. A scope is useful for time domain analysis. A spectrum is better for frequency domain analysis. SA having a better vertical resolution will have a better performance in S/N ratio, allowing one to see signal at very low power level. While scope having higher sampling rate will allow better time resolution of certain kind of measurement such as rise time. 4. A scope can be more than one port while SA is one port. Hence a scope is able to perform multi channel time domain comparison such as phase, pulse rise time...etc [![](https://i.stack.imgur.com/dBHHP.png)](https://i.stack.imgur.com/dBHHP.png) Above: Scope measuring multi-channel pulses
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
I find whole chickens to be cheaper than the parts (wings, backs, whatever), so I regularly buy a couple whole birds, chop the breasts out (saving for later use), and then make stock with the remaining meat/bones/skin. Place the chicken in a stock pot. Add a couple onions, carrots, celery stalks and some peppercorns. Cover with cold water. Bring to a bare simmer and cook for 2-3 hours. For ease, I usually bring the pot to a boil on the stove top, and then transfer to the oven at 250. Strain the stock, chill, and then freeze for later use.
I serve rotisserie chicken once in a while - I save the bones in the freezer until I have enough to make a big batch of stock. I also save the tops and bottoms of celery stalks and other trimmings when I make veggie trays. The veg may get mushy from freezing but it still has great flavor to add to your stock. I find that for home use freezing the stock in 1-cup increments works great.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
I serve rotisserie chicken once in a while - I save the bones in the freezer until I have enough to make a big batch of stock. I also save the tops and bottoms of celery stalks and other trimmings when I make veggie trays. The veg may get mushy from freezing but it still has great flavor to add to your stock. I find that for home use freezing the stock in 1-cup increments works great.
In general, the parts of the animal to be used in making stock depends on the animal: For chickens, different people will recommend different parts of the animal, but simply chopping up an entire chicken works fairly well. For beef, pretty much any tough piece of meat will work; try looking at shoulder or butt. As for other suggestions, be creative! Almost anybody can make a basic broth by boiling a chicken, but making one that is truly fantastic takes time and work. Try adding different ingredients. For starters, try things that you are familiar with. For example, vegetables and herbs like garlic and basil can improve any broth. From there, try things that are more creative, like sesame oil, miso paste, or perhaps ginger.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
I'd also add that if you have access to a pressure cooker, you can make stock more quickly and have the added benefit of it staying more clear instead of getting cloudy. I also personally think it tastes better.
We start with our older laying hens & roosters. remove Head, dip in boiling water pluck chicken. Next gut. Place aside giblets. Wash in cold well water. Place in large pressure cooker. With salt, onion chopped, garlic, pepper, morening leaf. Feet may be left on. Cook 4 to 6 hours till meat falls of bones. Cool. remove grease from top. Remove all bones. Large chunks of chicken cut up. Taste broth. You may wish to add more spice. Put pot back on stove boil to reduce till you have flavor you wish. Cool. Put in bags. Remove all air. Use masking tape & pen to label bags. Freeze till needed. The secret to good chicken stock is to not use fryers or bakers but your older chickens free range ones.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
***Note that for maximum benefit I am answering the question regarding chicken stock first and at the end have included information on other stocks such as fish and brown chicken and veal stock.*** Properly made stock is made from bones only. If you cut up your own chickens then save the backs and wing tips in the freezer and use for stock. You might ask at your grocery store or butcher and be able to purchase backs from them if you want to make stock but don't have enough. Chicken feet are rich in gelatin and can be added to the pot to help add body to the stock. * Place chicken backs/wing tips in a large stockpot (about a 3-4 gal. stockpot). Cover with cold water and place on stove. Bring to a gentle simmer for about 10 minutes. This is known as "blanching" the bones and is done to help dissolve the blood and other loose proteins from the soft bones. These proteins will form a gray scum on the surface. Once the scum has formed, drain off the water disarding it in the sink. The more clear a stock is the longer the shelflife as it's the additional proteins making a stock cloudy that also cause it to spoil much more quickly. * Return bones to the pot and cover again with cold water. Add a sachet (cheesecloth pouch containing a couple of bay leaves, about a teaspoon of whole peppercorns, a handful of parsley stems, and 8-10 thyme branches or a teaspoon of dried thyme) that's been tied up with some string. Bring the pot to a very gentle simmer once again and allow to simmer for about 3 hours. * After 3 hours add the mirepoix (carrots, onions, celery). Plan on 1-2 lbs. of mirepoix (combined vegetables) per 8 lbs. of bones. I also like to add leeks as they not only add flavor but are supposed to help in clarification of the stock. Continue to simmer the stock for 1-2 hours. (If you're not going to be around to add the vegetables at that time they can be added with the bones but need to be cut in large pieces to avoid cooking down and disintegrating. I would typically start chicken and veal stock at night before leaving the restaurant and allow it to simmer overnight until the next morning.) * Strain the stock through a fine mesh strainer or several layers of cheesecloth. Cool the stock in a metal container (another pot) in an ice bath until cool to the touch and then place in storage containers and refrigerate or freeze. * Once the stock has chilled the fat will have formed a solid layer on top of the stock. Carefully remove and either discard or use for sauteeing (such as duck or chicken fat). **Additional Notes:** The same process would be followed for white veal stock (not frequently used) **If preparing a brown stock (roasted chicken, veal, beef, lamb, duck):\*\*** These stocks are made by first rinsing the bones and then roasting in the oven at about 400 F degrees until medium brown (think iced tea color). Once bones are browned, smear them with a tablespoon or two of tomato paste and roast for another 5 minutes or so. Be careful as the concentrated sugars in the tomato paste can cause it to burn and create a bitter tasting stock. After bones and tomato paste have roasted, place them in the stock pot and add the mirepoix to the roasting pan, tossing to coat with the fat that has rendered out and roast the mirepoix until lightly browned. Add the mirepoix about 2 hours before finishing the stock. Be sure carrots are cut rather large to avoid having them disintegrate and adding a strange excessively sweet taste to the stock. **Recommended cooking times for stocks:** -Fish: 30-45 minutes. Use lean white fish bones and heads. Lobster/shrimp stock is made from the shells and will usually yield more flavor if first roasted. Remove gills and eyes from fish heads to prevent clouding the stock. Do not use carrots in mirepoix or it will color the stock an orange hue. Use onions, celery, leeks, mushrooms, and sliced lemon. -Chicken/Duck: 4-6 hours -Veal/Beef/Lamb: 8-12 hours (knuckle bones of veal will have the greatest amount of gelatin since the bones haven't hardened as much as those of beef) **Stocks should not be seasoned with salt** as one of the primary purposes is for them to be a "stock" item that can be used for a variety of needs: soups, sauces, rice, etc. The salt is added to the final preparation not the stock. If stock is salted and then used for a reduction sauce the resulting sauce will be too salty.
One suggestion I would add is to freeze your stock in ice-cube trays, or in other small portions, so that you can parcel out the amount you need when you need it.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
Darens' recipe sounds lovely, if you are after a proper stock. Here is a quick cheap alternative. Freeze the bones and skin of a roast chicken once you have finished with them. When you need some quick stock, put the bits into a pan and cover with boiling water. Simmer for half an hour or so and then use a sieve to keep the bits behind..
One suggestion I would add is to freeze your stock in ice-cube trays, or in other small portions, so that you can parcel out the amount you need when you need it.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
I find whole chickens to be cheaper than the parts (wings, backs, whatever), so I regularly buy a couple whole birds, chop the breasts out (saving for later use), and then make stock with the remaining meat/bones/skin. Place the chicken in a stock pot. Add a couple onions, carrots, celery stalks and some peppercorns. Cover with cold water. Bring to a bare simmer and cook for 2-3 hours. For ease, I usually bring the pot to a boil on the stove top, and then transfer to the oven at 250. Strain the stock, chill, and then freeze for later use.
In general, the parts of the animal to be used in making stock depends on the animal: For chickens, different people will recommend different parts of the animal, but simply chopping up an entire chicken works fairly well. For beef, pretty much any tough piece of meat will work; try looking at shoulder or butt. As for other suggestions, be creative! Almost anybody can make a basic broth by boiling a chicken, but making one that is truly fantastic takes time and work. Try adding different ingredients. For starters, try things that you are familiar with. For example, vegetables and herbs like garlic and basil can improve any broth. From there, try things that are more creative, like sesame oil, miso paste, or perhaps ginger.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
***Note that for maximum benefit I am answering the question regarding chicken stock first and at the end have included information on other stocks such as fish and brown chicken and veal stock.*** Properly made stock is made from bones only. If you cut up your own chickens then save the backs and wing tips in the freezer and use for stock. You might ask at your grocery store or butcher and be able to purchase backs from them if you want to make stock but don't have enough. Chicken feet are rich in gelatin and can be added to the pot to help add body to the stock. * Place chicken backs/wing tips in a large stockpot (about a 3-4 gal. stockpot). Cover with cold water and place on stove. Bring to a gentle simmer for about 10 minutes. This is known as "blanching" the bones and is done to help dissolve the blood and other loose proteins from the soft bones. These proteins will form a gray scum on the surface. Once the scum has formed, drain off the water disarding it in the sink. The more clear a stock is the longer the shelflife as it's the additional proteins making a stock cloudy that also cause it to spoil much more quickly. * Return bones to the pot and cover again with cold water. Add a sachet (cheesecloth pouch containing a couple of bay leaves, about a teaspoon of whole peppercorns, a handful of parsley stems, and 8-10 thyme branches or a teaspoon of dried thyme) that's been tied up with some string. Bring the pot to a very gentle simmer once again and allow to simmer for about 3 hours. * After 3 hours add the mirepoix (carrots, onions, celery). Plan on 1-2 lbs. of mirepoix (combined vegetables) per 8 lbs. of bones. I also like to add leeks as they not only add flavor but are supposed to help in clarification of the stock. Continue to simmer the stock for 1-2 hours. (If you're not going to be around to add the vegetables at that time they can be added with the bones but need to be cut in large pieces to avoid cooking down and disintegrating. I would typically start chicken and veal stock at night before leaving the restaurant and allow it to simmer overnight until the next morning.) * Strain the stock through a fine mesh strainer or several layers of cheesecloth. Cool the stock in a metal container (another pot) in an ice bath until cool to the touch and then place in storage containers and refrigerate or freeze. * Once the stock has chilled the fat will have formed a solid layer on top of the stock. Carefully remove and either discard or use for sauteeing (such as duck or chicken fat). **Additional Notes:** The same process would be followed for white veal stock (not frequently used) **If preparing a brown stock (roasted chicken, veal, beef, lamb, duck):\*\*** These stocks are made by first rinsing the bones and then roasting in the oven at about 400 F degrees until medium brown (think iced tea color). Once bones are browned, smear them with a tablespoon or two of tomato paste and roast for another 5 minutes or so. Be careful as the concentrated sugars in the tomato paste can cause it to burn and create a bitter tasting stock. After bones and tomato paste have roasted, place them in the stock pot and add the mirepoix to the roasting pan, tossing to coat with the fat that has rendered out and roast the mirepoix until lightly browned. Add the mirepoix about 2 hours before finishing the stock. Be sure carrots are cut rather large to avoid having them disintegrate and adding a strange excessively sweet taste to the stock. **Recommended cooking times for stocks:** -Fish: 30-45 minutes. Use lean white fish bones and heads. Lobster/shrimp stock is made from the shells and will usually yield more flavor if first roasted. Remove gills and eyes from fish heads to prevent clouding the stock. Do not use carrots in mirepoix or it will color the stock an orange hue. Use onions, celery, leeks, mushrooms, and sliced lemon. -Chicken/Duck: 4-6 hours -Veal/Beef/Lamb: 8-12 hours (knuckle bones of veal will have the greatest amount of gelatin since the bones haven't hardened as much as those of beef) **Stocks should not be seasoned with salt** as one of the primary purposes is for them to be a "stock" item that can be used for a variety of needs: soups, sauces, rice, etc. The salt is added to the final preparation not the stock. If stock is salted and then used for a reduction sauce the resulting sauce will be too salty.
We had this problem when my youngest started having food allergies. The quick and dirty method I use, which is easy, but not as fancy as some of the others here, is as follows: buy a 10 lb bag of chicken leg and thigh quarters. put in as big a pot as I can and cover with water. Add 1 tblsp. of poultry seasoning or Mrs. Dash original. Bring to a boil then reduce heat and simmer for about an hour. pull the chicken quarters out reserving the liquid, strip the meat, which should come off easy. Toss the bones back into the pot with the reserved liquid and some of the skin and simmer for another hour or 2. Strain and portion into 1 cup units and toss in the freezer. The chicken I pulled off I use for things like chicken salad, barbeque chicken sandwiches, and for a comfort food like chicken and rice. for the chicken and rice: 1 cup frozen chicken broth(or stock, or whatever), 1 cup rice, enough liquid to make the up the difference in liquid for the rice. When the rice is done, add 1 cup chicken, salt and pepper to taste. You can add a can of cream of chicken soup as well, but It's fine without. Good luck! **edit** side note, I have taken to using a pair of shears to cut the long bones in half before taossing them back into the pot. I think it deepens the flavor
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
***Note that for maximum benefit I am answering the question regarding chicken stock first and at the end have included information on other stocks such as fish and brown chicken and veal stock.*** Properly made stock is made from bones only. If you cut up your own chickens then save the backs and wing tips in the freezer and use for stock. You might ask at your grocery store or butcher and be able to purchase backs from them if you want to make stock but don't have enough. Chicken feet are rich in gelatin and can be added to the pot to help add body to the stock. * Place chicken backs/wing tips in a large stockpot (about a 3-4 gal. stockpot). Cover with cold water and place on stove. Bring to a gentle simmer for about 10 minutes. This is known as "blanching" the bones and is done to help dissolve the blood and other loose proteins from the soft bones. These proteins will form a gray scum on the surface. Once the scum has formed, drain off the water disarding it in the sink. The more clear a stock is the longer the shelflife as it's the additional proteins making a stock cloudy that also cause it to spoil much more quickly. * Return bones to the pot and cover again with cold water. Add a sachet (cheesecloth pouch containing a couple of bay leaves, about a teaspoon of whole peppercorns, a handful of parsley stems, and 8-10 thyme branches or a teaspoon of dried thyme) that's been tied up with some string. Bring the pot to a very gentle simmer once again and allow to simmer for about 3 hours. * After 3 hours add the mirepoix (carrots, onions, celery). Plan on 1-2 lbs. of mirepoix (combined vegetables) per 8 lbs. of bones. I also like to add leeks as they not only add flavor but are supposed to help in clarification of the stock. Continue to simmer the stock for 1-2 hours. (If you're not going to be around to add the vegetables at that time they can be added with the bones but need to be cut in large pieces to avoid cooking down and disintegrating. I would typically start chicken and veal stock at night before leaving the restaurant and allow it to simmer overnight until the next morning.) * Strain the stock through a fine mesh strainer or several layers of cheesecloth. Cool the stock in a metal container (another pot) in an ice bath until cool to the touch and then place in storage containers and refrigerate or freeze. * Once the stock has chilled the fat will have formed a solid layer on top of the stock. Carefully remove and either discard or use for sauteeing (such as duck or chicken fat). **Additional Notes:** The same process would be followed for white veal stock (not frequently used) **If preparing a brown stock (roasted chicken, veal, beef, lamb, duck):\*\*** These stocks are made by first rinsing the bones and then roasting in the oven at about 400 F degrees until medium brown (think iced tea color). Once bones are browned, smear them with a tablespoon or two of tomato paste and roast for another 5 minutes or so. Be careful as the concentrated sugars in the tomato paste can cause it to burn and create a bitter tasting stock. After bones and tomato paste have roasted, place them in the stock pot and add the mirepoix to the roasting pan, tossing to coat with the fat that has rendered out and roast the mirepoix until lightly browned. Add the mirepoix about 2 hours before finishing the stock. Be sure carrots are cut rather large to avoid having them disintegrate and adding a strange excessively sweet taste to the stock. **Recommended cooking times for stocks:** -Fish: 30-45 minutes. Use lean white fish bones and heads. Lobster/shrimp stock is made from the shells and will usually yield more flavor if first roasted. Remove gills and eyes from fish heads to prevent clouding the stock. Do not use carrots in mirepoix or it will color the stock an orange hue. Use onions, celery, leeks, mushrooms, and sliced lemon. -Chicken/Duck: 4-6 hours -Veal/Beef/Lamb: 8-12 hours (knuckle bones of veal will have the greatest amount of gelatin since the bones haven't hardened as much as those of beef) **Stocks should not be seasoned with salt** as one of the primary purposes is for them to be a "stock" item that can be used for a variety of needs: soups, sauces, rice, etc. The salt is added to the final preparation not the stock. If stock is salted and then used for a reduction sauce the resulting sauce will be too salty.
I find whole chickens to be cheaper than the parts (wings, backs, whatever), so I regularly buy a couple whole birds, chop the breasts out (saving for later use), and then make stock with the remaining meat/bones/skin. Place the chicken in a stock pot. Add a couple onions, carrots, celery stalks and some peppercorns. Cover with cold water. Bring to a bare simmer and cook for 2-3 hours. For ease, I usually bring the pot to a boil on the stove top, and then transfer to the oven at 250. Strain the stock, chill, and then freeze for later use.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
In general, the parts of the animal to be used in making stock depends on the animal: For chickens, different people will recommend different parts of the animal, but simply chopping up an entire chicken works fairly well. For beef, pretty much any tough piece of meat will work; try looking at shoulder or butt. As for other suggestions, be creative! Almost anybody can make a basic broth by boiling a chicken, but making one that is truly fantastic takes time and work. Try adding different ingredients. For starters, try things that you are familiar with. For example, vegetables and herbs like garlic and basil can improve any broth. From there, try things that are more creative, like sesame oil, miso paste, or perhaps ginger.
We start with our older laying hens & roosters. remove Head, dip in boiling water pluck chicken. Next gut. Place aside giblets. Wash in cold well water. Place in large pressure cooker. With salt, onion chopped, garlic, pepper, morening leaf. Feet may be left on. Cook 4 to 6 hours till meat falls of bones. Cool. remove grease from top. Remove all bones. Large chunks of chicken cut up. Taste broth. You may wish to add more spice. Put pot back on stove boil to reduce till you have flavor you wish. Cool. Put in bags. Remove all air. Use masking tape & pen to label bags. Freeze till needed. The secret to good chicken stock is to not use fryers or bakers but your older chickens free range ones.
3,226
I am tired of using canned broth/stock and would like to make my own - any suggestions as to the proper technique and parts to use?
2010/07/26
[ "https://cooking.stackexchange.com/questions/3226", "https://cooking.stackexchange.com", "https://cooking.stackexchange.com/users/177/" ]
Darens' recipe sounds lovely, if you are after a proper stock. Here is a quick cheap alternative. Freeze the bones and skin of a roast chicken once you have finished with them. When you need some quick stock, put the bits into a pan and cover with boiling water. Simmer for half an hour or so and then use a sieve to keep the bits behind..
We start with our older laying hens & roosters. remove Head, dip in boiling water pluck chicken. Next gut. Place aside giblets. Wash in cold well water. Place in large pressure cooker. With salt, onion chopped, garlic, pepper, morening leaf. Feet may be left on. Cook 4 to 6 hours till meat falls of bones. Cool. remove grease from top. Remove all bones. Large chunks of chicken cut up. Taste broth. You may wish to add more spice. Put pot back on stove boil to reduce till you have flavor you wish. Cool. Put in bags. Remove all air. Use masking tape & pen to label bags. Freeze till needed. The secret to good chicken stock is to not use fryers or bakers but your older chickens free range ones.
5,818,241
How to send a image or a file using the web service to the server ?
2011/04/28
[ "https://Stackoverflow.com/questions/5818241", "https://Stackoverflow.com", "https://Stackoverflow.com/users/296231/" ]
Here is a good example of how to send a file using NuSOAP: <http://web.imkrisna.com/june/news/simple-file-transfer-using-nusoap/>
The web service's documentation should specify the necessary format, which will probably involve base64 encoding the image data. You would do whatever encoding the web service requires and then pass it to nuSoap in the same way you already do for text.
7,349
I just installed 32-bit Ubuntu Maverick 10 Stable everything is working fine, The system spec's are as below: 3 Ghz Intel DG101 512 MB RAM 80 GB HDD 256 MB Ati Radeon Xpress the only problem is when the Youtube videos are maximized they consume allot of CPU plus video bacomes slower and choppy..... What to do? I have FireFox and installed the FlashPlayer 10 also. But no luck...Any Ideas how to fix? The videos work absolutely good in XP in same computer..... I have tried Google Chrome also but no luck in that also.... Any answer 'll be appreciated.
2011/02/13
[ "https://unix.stackexchange.com/questions/7349", "https://unix.stackexchange.com", "https://unix.stackexchange.com/users/4811/" ]
To find proprietary driver for ATI card head to the [ATI download page](http://support.amd.com/us/gpudownload/Pages/index.aspx) and check if a Linux driver is available for your model. Alternatively, you can use Ubuntu's driver finder by going to System -> Administration -> Hardware Drivers. If installing the driver doesn't help try the newest version of Flash by [getting it from Adobe](http://get.adobe.com/flashplayer/).
He he... You guys are funny . . It s very simple. It will work in 320p when the screen is maximised. Change it from 480p to 320p which is near the volume control when the screen is maximised. All the best watching Youtube!! Regards, Aravindh :)
7,349
I just installed 32-bit Ubuntu Maverick 10 Stable everything is working fine, The system spec's are as below: 3 Ghz Intel DG101 512 MB RAM 80 GB HDD 256 MB Ati Radeon Xpress the only problem is when the Youtube videos are maximized they consume allot of CPU plus video bacomes slower and choppy..... What to do? I have FireFox and installed the FlashPlayer 10 also. But no luck...Any Ideas how to fix? The videos work absolutely good in XP in same computer..... I have tried Google Chrome also but no luck in that also.... Any answer 'll be appreciated.
2011/02/13
[ "https://unix.stackexchange.com/questions/7349", "https://unix.stackexchange.com", "https://unix.stackexchange.com/users/4811/" ]
Finally i made it, all you have to do is to open youtube video any video then right click on video and choose Settings, then a dialogue box will appear, uncheck the "Enable Hardware Acceleration" from that box, reload the video and enjoy!! [SOLVED]
He he... You guys are funny . . It s very simple. It will work in 320p when the screen is maximised. Change it from 480p to 320p which is near the volume control when the screen is maximised. All the best watching Youtube!! Regards, Aravindh :)
21,950
"So Jesus said to them, “Because of your **unbelief**; for assuredly, I say to you, if you have faith as a mustard seed, you will say to this mountain, ‘Move from here to there,’ and it will move; and nothing will be impossible for you" -- Matthew 17:20 (NKJV). To the best of my knowledge, only the KJV and NKJV use the word "unbelief," suggesting a lack of faith. Other versions (NIV, ESV, AMP) use "little faith" instead, which seems to suggest little faith rather than a lack of faith. So, what was Jesus meaning to say: a lack of or a little faith? Thanks.
2016/03/24
[ "https://hermeneutics.stackexchange.com/questions/21950", "https://hermeneutics.stackexchange.com", "https://hermeneutics.stackexchange.com/users/13617/" ]
They translate from the Greek word ὀλιγοπιστίαν ----------------------------------------------- From [3641](http://biblehub.com/greek/3641.htm) /olígos, "little in number, low in quantity" and [4102](http://biblehub.com/greek/4102.htm) /pístis, "faith". Occurs five times in the NT, each time with Jesus rebuking the problem of failing to hear His voice (cf. [Jn 10:3](http://biblehub.com/john/10-3.htm),[4](http://biblehub.com/john/10-4.htm),[27](http://biblehub.com/john/10-27.htm)). "Little-faith" ([3640](http://biblehub.com/greek/3640.htm) /oligópistos) describes someone dull to hearing the Lord's voice, or disinterested in walking intimately with Him. In contrast, the goal of life is to receive (obey) the Lord's gift of faith in each scene of life ([Ro 14:23](http://biblehub.com/romans/14-23.htm); [Heb 11:6](http://biblehub.com/hebrews/11-6.htm)). [Answer Source](http://biblehub.com/greek/3640.htm) Etymology of /[pístis](https://en.wiktionary.org/wiki/%CF%80%CE%AF%CF%83%CF%84%CE%B9%CF%82), "faith" ---------------------------------------------------------------------------------------------------- From πιθ- ‎(pith-), the root of [πείθω](https://en.wiktionary.org/wiki/%CF%80%CE%B5%CE%AF%CE%B8%CF%89#Ancient_Greek) ‎(peíthō), +‎ -[σις](https://en.wiktionary.org/wiki/-%CF%83%CE%B9%CF%82#Ancient_Greek) ‎(-sis), which remained -τις ‎(-tis) after a dental. [Answer Source](https://en.wiktionary.org/wiki/%CF%80%CE%AF%CF%83%CF%84%CE%B9%CF%82) The Usages in the New Testament ------------------------------- **[Matthew 6:30](https://www.biblegateway.com/passage/?search=Matthew+6&version=NKJV)** Here it gets used as "little convinced". Those that understand the power of God do not need to worry about food, drink, or clothing. Those "convinced" know with certainty that God provides. **[Matthew 8:26](https://www.biblegateway.com/passage/?search=Matthew+8&version=NKJV)** Here Jesus rebukes his disciples for if they where convinced, they themselves would have used the word to calm the wind and the seas. However since they were not convinced of the power of what they say, Jesus had to show them again and himself told the wind and seas to calm. **[Matthew 14:31](https://www.biblegateway.com/passage/?search=Matthew+14&version=NKJV)** Here Jesus uses the power of the word and told Peter to "Come" meaning to also come walk on the water. Peter upon seeing the wind began to doubt and started to sink, therefore the rebuke. **[Matthew 16:8](https://www.biblegateway.com/passage/?search=Matthew+16&version=NKJV)** Here the rebuke comes from a general lack of understanding the metaphor "leaven". Perhaps Jesus uses other metaphors? Just something to think about. **[Matthew 17:20](https://www.biblegateway.com/passage/?search=Matthew%2017&version=NKJV)** Here again Jesus tries to explain the power of what a person says. That even telling a mountain to move will move a mountain and that nothings impossible for those that believe in the power of what they say. They still remain unconvinced at the power of what we say, therefore he calls them "hardly convinced". **[Luke 12:28](https://www.biblegateway.com/passage/?search=Luke+12&version=NKJV)** Again used to express how we do not have to worry about food, drink, or clothes. That by simply saying what we need, God will provide.
I believe that what the Lord Jesus Christ wanted to teach us is for us to have faith only, unmixed with unbelief. Because we can have faith and have unbeliefs at the same time. That's why He said, Believe only. All things are possible to those who just believe.
61,835,971
Regarding compatibility between ECMAScript specification and actual implementation; It is fairly easy to check out the data about [browser support for ECMAScript2015 (ES6)](https://caniuse.com/#search=ECMAScript%206), but I found it pretty difficult to have an equivalently clear table for all the following ES versions (ES7+). *By the time this question is asked:* * Mozilla has some info [on their website](https://developer.mozilla.org/en-US/docs/Archive/Web/JavaScript/ECMAScript_Next_support_in_Mozilla): it is possible to see that ES7 and ES8 are fully supported, ES9 still has some problems and ES10 is supported on the latest versions. * I can also guess that IE11 never progressed after ES5. * I did not find anything for the other browsers, just some stolen info here and there. **How can I check what the current browser-support level is?**
2020/05/16
[ "https://Stackoverflow.com/questions/61835971", "https://Stackoverflow.com", "https://Stackoverflow.com/users/4628597/" ]
Browser vendors don't implement specific *versions*, but specific *features*. Almost every modern browser is still missing features from ES2017-ES2020. Hence there is not and won't be a table where you can see an ES version to browser version mapping. But that is not a problem because you as a developer do the same. You use features, not versions of ECMAScript. Caniuse is still a great resource to check for support of individual features. If you are not happy with the data presentation on Caniuse, maybe these [compatibility tables](https://kangax.github.io/compat-table/es2016plus/) are better for you. Additionally, you can use polyfills and Babel for transpiling of newer features to older runtimes.
The simple reason is: they don't support it. Everything above ES6 is still in the works. Since ES6 is still being adopted and not all browsers support everything there's no reason for them to aim for ES7. If you want to use >ES7 features I would suggest looking into Babel, since there are ways to use ES7 and above and compile it back to ES5 so that even IE supports it. Hope that helps, feel free to comment with questions!
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
Also, don't forget you need to have your server secured from current (that is, soon-to-be-past) employees. Several startups were totally wiped due to employee sabotage, e.g. <http://www.geek.com/articles/news/disgruntled-employee-kills-journalspace-with-data-wipe-2009015/>
A few basic "security" measures here that while are more reactive than proactive, are some things to consider. 1) Backup strategy, of course not just for those who hack into your site, but it is nice to restore everything back to pre-hack days if possible, make sure it's reliable and most importantly was tested in a near-live restore drill 2) Mitigation, have plans in place at least on a napkin somewhere for how to react if the server is hacked 3) Insurance, find insurance companies that understand the world of cyber-business and the damages resulting from these things, buy policies 4) Someone already mentioned employee sabotage problems, you're screening your employees beforehand right? Background checks are cheap and do dig up stuff...
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
If security isn't thought of and built into the application and its infrastructure from day one it will be much more difficult to retrofit it in later. Now is the time to build the processes for regular OS/tool patching, upgrades, etc. * What kind of data will users be creating/storing on the site? * What effect will a breach have on your users? * What effect will a breach have on your company? * Will you be able to regain the users' trust after a breach? Since your company is dependent on keeping existing users and attracting new ones, you should present your concerns along the lines of how the users would react to a breach. The higher-ups will understand that the users are your bread and butter.
I agree with Stefan about reputation. You don't want to get hacked because you were lacking on security. Not only will that hurt your site and company, it will look bad on you since you're in charge of that. My personal opinion is to do as much as you can because no matter how much you do there will be vulnerabilities. Unfortunately security like testing and documentation are often afterthoughts. You should really make sure to do risk assessments early in your site/software's life and to keep on doing assessments. I think it is important to patch all software for security holes.
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
Reputation is everything here, especially for a startup. As a startup, you don't have a long history of reliability/security/... - so all depends on users to give you the 'benefit of the doubt' when they start using your app. If your server gets hacked and your users notice that, your reputation is gone. Once it's gone, it doesn't matter whether your app and your features are the 'next new thing' or not. It doesn't matter whether the security breach was minor or not - people won't trust your app/company anymore. So, I would consider security to be the top priority.
Have a look at Mod Security for the various possibilities in the software setup: Do a Google search for "mod\_security howto example" Simple example to start: <http://www.ghacks.net/2009/07/15/install-mod_security-for-better-apache-security/>
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
Also, don't forget you need to have your server secured from current (that is, soon-to-be-past) employees. Several startups were totally wiped due to employee sabotage, e.g. <http://www.geek.com/articles/news/disgruntled-employee-kills-journalspace-with-data-wipe-2009015/>
If you're explicitly trying to attract the sort of users who are inclined to try to crack systems, then you can pretty well bet that your system *will* come under attack. You should suggest to the management that if they're not going to take security seriously, then you should just go ahead and post the company's bank statements and accounting books (in clear text) on the site, with a prominent link from the home page. At least that way, you can tell them, the end result will be about the same, but they're less likely to damage everything else to get what they're looking for. I'd think that the reputation issue might have a slightly different cast with this audience, too -- they may forgive you for being hacked, but they probably *won't* forgive you for being an easy target.
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
A few basic "security" measures here that while are more reactive than proactive, are some things to consider. 1) Backup strategy, of course not just for those who hack into your site, but it is nice to restore everything back to pre-hack days if possible, make sure it's reliable and most importantly was tested in a near-live restore drill 2) Mitigation, have plans in place at least on a napkin somewhere for how to react if the server is hacked 3) Insurance, find insurance companies that understand the world of cyber-business and the damages resulting from these things, buy policies 4) Someone already mentioned employee sabotage problems, you're screening your employees beforehand right? Background checks are cheap and do dig up stuff...
Have a look at Mod Security for the various possibilities in the software setup: Do a Google search for "mod\_security howto example" Simple example to start: <http://www.ghacks.net/2009/07/15/install-mod_security-for-better-apache-security/>
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
These will probably be obvious: * Limit password attempts. * Sanitize your database inputs * Measures to prevent XSS attacks It's also worth mentioning that, as you said, the network architecture should be set up appropriately. You should definitely have a decent firewall that's locked down as much as possible. Some people recommend putting your systems between dual firewalls of different makes so that in the event one of them has a critical vulnerability, the second will most likely not have the same vulnerability and you'll be safe. It all depends on what you can afford since it's a startup.
Have a look at Mod Security for the various possibilities in the software setup: Do a Google search for "mod\_security howto example" Simple example to start: <http://www.ghacks.net/2009/07/15/install-mod_security-for-better-apache-security/>
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
If security isn't thought of and built into the application and its infrastructure from day one it will be much more difficult to retrofit it in later. Now is the time to build the processes for regular OS/tool patching, upgrades, etc. * What kind of data will users be creating/storing on the site? * What effect will a breach have on your users? * What effect will a breach have on your company? * Will you be able to regain the users' trust after a breach? Since your company is dependent on keeping existing users and attracting new ones, you should present your concerns along the lines of how the users would react to a breach. The higher-ups will understand that the users are your bread and butter.
These will probably be obvious: * Limit password attempts. * Sanitize your database inputs * Measures to prevent XSS attacks It's also worth mentioning that, as you said, the network architecture should be set up appropriately. You should definitely have a decent firewall that's locked down as much as possible. Some people recommend putting your systems between dual firewalls of different makes so that in the event one of them has a critical vulnerability, the second will most likely not have the same vulnerability and you'll be safe. It all depends on what you can afford since it's a startup.
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
Make sure you know what version and patch level your servers are running, not just the OS, but all related components and everything that is actually executing the the machine. Then make sure you are never more than a day behind. Not doing so leads to much pain, and you don't hear of most of it - most of my past employers would never publicly admit being hacked as it reflects badly on them, so you can assume systems are getting hacked left and right with pretty serious consequences to companies, you just don't hear about most of these events.
Have a look at Mod Security for the various possibilities in the software setup: Do a Google search for "mod\_security howto example" Simple example to start: <http://www.ghacks.net/2009/07/15/install-mod_security-for-better-apache-security/>
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
Reputation is everything here, especially for a startup. As a startup, you don't have a long history of reliability/security/... - so all depends on users to give you the 'benefit of the doubt' when they start using your app. If your server gets hacked and your users notice that, your reputation is gone. Once it's gone, it doesn't matter whether your app and your features are the 'next new thing' or not. It doesn't matter whether the security breach was minor or not - people won't trust your app/company anymore. So, I would consider security to be the top priority.
A few basic "security" measures here that while are more reactive than proactive, are some things to consider. 1) Backup strategy, of course not just for those who hack into your site, but it is nice to restore everything back to pre-hack days if possible, make sure it's reliable and most importantly was tested in a near-live restore drill 2) Mitigation, have plans in place at least on a napkin somewhere for how to react if the server is hacked 3) Insurance, find insurance companies that understand the world of cyber-business and the damages resulting from these things, buy policies 4) Someone already mentioned employee sabotage problems, you're screening your employees beforehand right? Background checks are cheap and do dig up stuff...
429,352
I'm working with a start-up, mostly doing system administration and I've come across a some security issues that I'm not really comfortable with. I want to judge whether my expectations are accurate, so I'm looking for some insight into what others have done in this situation, and what risks/problems came up. In particular, how critical are measures like placing admin tools behind a vpn, regular security updates (OS and tools), etc. Keep in mind that as this is a start-up, the main goal is to get as many features as possible out the door quickly, so I'll need as much justification as I can get to get the resources for security (i.e. downtime for upgrades, dev time for application security fixes). Background Info: * Application is LAMP as well as a custom java client-server. * Over the next 3 months, I project about 10k anonymous visitors to the site and up to 1000 authenticated users. * Younger audience (16-25) which is guaranteed to have an above average number of black-hats included. Thanks in advance for your responses, and I'll welcome any related advice.
2009/01/09
[ "https://Stackoverflow.com/questions/429352", "https://Stackoverflow.com", "https://Stackoverflow.com/users/2567/" ]
If security isn't thought of and built into the application and its infrastructure from day one it will be much more difficult to retrofit it in later. Now is the time to build the processes for regular OS/tool patching, upgrades, etc. * What kind of data will users be creating/storing on the site? * What effect will a breach have on your users? * What effect will a breach have on your company? * Will you be able to regain the users' trust after a breach? Since your company is dependent on keeping existing users and attracting new ones, you should present your concerns along the lines of how the users would react to a breach. The higher-ups will understand that the users are your bread and butter.
A few basic "security" measures here that while are more reactive than proactive, are some things to consider. 1) Backup strategy, of course not just for those who hack into your site, but it is nice to restore everything back to pre-hack days if possible, make sure it's reliable and most importantly was tested in a near-live restore drill 2) Mitigation, have plans in place at least on a napkin somewhere for how to react if the server is hacked 3) Insurance, find insurance companies that understand the world of cyber-business and the damages resulting from these things, buy policies 4) Someone already mentioned employee sabotage problems, you're screening your employees beforehand right? Background checks are cheap and do dig up stuff...
189,648
**Rationale:** This is intended to address an issue with new users that have *no* reputation. If they are inexperienced, they are very unlikely to gain any reputation very quickly. This results in them very quickly getting blocked from asking questions due to the fact that the questions are often poorly worded. The system of down voting is highly subjective and subject to abuse (originally said: often misused), thus contributing to the problem. Even worse, people with language barriers are likely to have this problem. Overall, the quick blocking of questions without warning seems extreme. **Possible solution options (i.e. choose none, 1 or more):** In order to conform with the automated nature of the routines there are a few alternatives: 1. Provide warnings on the account, (good.) Use the three strikes principle or something similar before the final block. *Automated* 2. Involve a human, (not so good.) Have a moderator review the down voting pattern before approving the block. 3. Kick the question back to the user (very good.) Hide the question until the users has addressed the deficiency. If user's rating continues to drop due to this pseudo deletion, then block them. Hiding would be based on down votes. *Automated* 4. Move the questions to a junk forum, (worse.) Similar to the meta site, set up a rejected questions site. It would be costly but informative. (edit) 5. **The problem lies with different features of the site, make changes in those areas, not with the system of blocking.** 6.[Require a net +1 for questions in high-volume tags to become answerable](https://meta.stackexchange.com/questions/108822/require-a-net-1-for-questions-in-high-volume-tags-to-become-answerable) I would not up vote for number five on this post, it move the issues somewhere else. In terms of my own user experience that is where I see the problem. This is mainly about SO, not meta.SO... i. Exclude down votes of certain types. "Disagreeing with a change" for example is quite different from "the technical content is horrible". ii. How can my post be a duplicate, I didn't find it anywhere? iii. How can my post be a duplicate, the titles are only vaguely related? iv. Being brief got down voted. Adding detail got complaints as well. v. There are not links in the post to similar posts or references. Item (i) is important to some, but a change that is unpopular and involves changes how votes work, also unpopular. Item (iv) is sadly unavoidable and almost off topic except it can block new users. Where down votes, duplicates and closing are concerned, it appears the correct approach is to request reopening and get more feed back by various means. Down votes can be mysterious but a person could ask in the comments for example. I.E. use the system as it exists. **Personally, items ii, iii, and iv as well the need for other change requests I have made really trace back to a problem in the search engine.** Once I figured out the issue I released the duplicates would not have happened. This post would not have happened. I do think item 1 is good and item 4 already exists more or less. Regardless, the importance of my list of changes dropped, particularly in this post. There are some issues on the site where post hi-jacking is concerned what I consider contradictions in the relationship between post subjects vs the requirement to be specific and detailed that is also at the core of items ii and iii. **I will post links to other (probably existing) threads as soon my content is in there. This is a real problem, but I think there are perhaps better more acceptable solutions to it. Thanks.**
2013/07/20
[ "https://meta.stackexchange.com/questions/189648", "https://meta.stackexchange.com", "https://meta.stackexchange.com/users/228676/" ]
I strongly disagree with the whole idea of blocking new and inexperienced users from asking questions. I myself am a new user. A new user learns from his experience and also by asking questions at the meta site. There can be comments guiding him to edit his question by giving the reason for the question being put on hold. While new users are more prone to questions with more down votes, that cannot call for a change in the guidelines and standards of SE Q&A sites. There shouldn't be any limitation in asking or answering questions other than those standards. Regarding points 1 to 6,none of them are practical,as of now, but things may change in the future. Each and every 'new user', but only a new user will face the issues in i to v. SE Chat room is the best place for new users and not SE Q&A sites.
The issue is not that a new user has NO reputation. The issue is that blocked user has a (strongly) NEGATIVE reputation. Many "new" users ask or answer one or two questions, and go for a long time with "no" reputation. (That is, a handful of votes, up or down, with a total score near zero.) These people don't get blocked. The people that get blocked are the ones that have a lot of DOWN votes (and/or deleted posts). That's more likely to happen to a new user, but even an experienced user with a positive reputation who went on a "rampage" and got a lot of downvoted or deleted posts could be blocked, if the previous good record couldn't sufficiently insulate him/her from the recent decline.
122
We currently have a good 5 pages or so of tags at this time. It would help if users were proactive in trying to edit tag Wikis so we can build up quality in our tags. This serves out as a reminder to the community. You do not have to go into too much detail about the subject. Even an excerpt is fine. This will also help us differentiate between useful and useless tags. Finally it should be noted that Wikis discussing the *history* of such concepts are proffered.
2014/11/13
[ "https://hsm.meta.stackexchange.com/questions/122", "https://hsm.meta.stackexchange.com", "https://hsm.meta.stackexchange.com/users/27/" ]
One thing I'd caution against is just copying tag wikis verbatim from other sources. Tag wikis should be written to be directly applicable to this site. Copying the text from another source constitutes plagiarism if you don't cite the source appropriately. Specifically, many users try to copy summaries from Wikipedia, sometimes attributing it to Wikipedia (usually not), but it's impossible for a tag wiki excerpt to satisfy Wikipedia's citation requirements since it is shown truncated and with hyperlinks removed in some places. Generally speaking, the best way to write a good tag wiki is to do it yourself. Tag wiki reviewers should generally check most suggestions via a search engine quickly to make sure they aren't plagiarized. Having been on several beta sites where this wasn't followed from the beginning, it became an enormous headache for us later to try to find the plagiarized ones. For instance, on [Anime & Manga](https://anime.stackexchange.com/tags), at one point we had quite a few users just manually check through all the tag wikis and excerpts to find plagiarized ones, spending quite a few hours in the process. It was a huge mess that could have been avoided just by catching these cases earlier and rejecting them (though there's not much you can do if the plagiarizer has over 4k rep). Hopefully we can control it here and stop it from becoming that rampant. If you find a plagiarized tag wiki, you should rewrite it. Another thing to note is that many tags don't strictly need tag wikis, so if you are going to write them, appropriate your efforts intelligently. They're most useful when the terms are ambiguous or unclear or our use of them here is not standard. If a tag has special policies or it's unclear when to use it, definitely a tag wiki is a good idea. If it's easy to understand, a tag wiki adds more things to maintain without much benefit, so before writing it ask yourself whether it's really needed. Finally, I'll just restate the OP's statement that the tag wikis are best be written in a way that's specific to *this site* (History of Science and Math), not about the concept broadly.
As another note: If you have 1500+ rep (which not enough users do so far, in my opinion), you can approve tag wiki edits. **I strongly recommend that all users who have passed this mark review the Suggested Edits queue to go over them.** Unlike edits to a post, where the poster can approve the edit and not need a second acceptance of the edit, there isn't a person (besides these users) who can approve these tag wiki edits (to my knowledge). If you can review these edits, please do so - and if you can review *anything*, do that, too! Also, users with 1500+ rep may not necessarily make tag wiki edits immediately. They, too, are subject to the peer-review process. Finally, we need to have a unified phrasing for tag wikis. Take a look at tags on other SE sites, and note that their phrasing follows a pattern.
2,527
When creating a bitcoin transaction, you have to choose which coins to use in them. The standard client does this [in a way](https://bitcoin.stackexchange.com/questions/1077/how-does-the-bitcoin-client-determine-which-coins-transactions-wallet-stored-va) to avoid unconfirmed inputs and minimize the number of inputs and amount of change involved. Since [not all coins are created equal](https://bitcoin.stackexchange.com/questions/2496/are-bitcoins-inherently-fungible), I may have other preferences (like spend unconfirmed coins first to get rid of them, or not cross inputs between different trading partners). Is there any way I can have more control over the process?
2012/01/11
[ "https://bitcoin.stackexchange.com/questions/2527", "https://bitcoin.stackexchange.com", "https://bitcoin.stackexchange.com/users/238/" ]
Currently this is not possible with the standard client, other than making separate wallets. There is [a patch for coin selection on github](https://github.com/bitcoin/bitcoin/pull/415), which was very promising. So this feature may make it into a future version.
Check out [Armory](https://bitcointalk.org/index.php?topic=56424.0), a new client that > > uses an algorithm for coin selection which can be optimized for > anonymity or minimal transaction fees. > > >
2,527
When creating a bitcoin transaction, you have to choose which coins to use in them. The standard client does this [in a way](https://bitcoin.stackexchange.com/questions/1077/how-does-the-bitcoin-client-determine-which-coins-transactions-wallet-stored-va) to avoid unconfirmed inputs and minimize the number of inputs and amount of change involved. Since [not all coins are created equal](https://bitcoin.stackexchange.com/questions/2496/are-bitcoins-inherently-fungible), I may have other preferences (like spend unconfirmed coins first to get rid of them, or not cross inputs between different trading partners). Is there any way I can have more control over the process?
2012/01/11
[ "https://bitcoin.stackexchange.com/questions/2527", "https://bitcoin.stackexchange.com", "https://bitcoin.stackexchange.com/users/238/" ]
Old question, so answers are outdated. Anyone reading this now: bitcoin-qt has coin control features that let you choose whichever txin you want to use.
Check out [Armory](https://bitcointalk.org/index.php?topic=56424.0), a new client that > > uses an algorithm for coin selection which can be optimized for > anonymity or minimal transaction fees. > > >
119,918
Is it possible to modify the default fonts to be used from "MS Shell Dlg 2" to another? If yes, where do I have to set it?
2014/10/27
[ "https://gis.stackexchange.com/questions/119918", "https://gis.stackexchange.com", "https://gis.stackexchange.com/users/8419/" ]
Go to settings, options. From here the Applications section has a Font option, unselect default and edit what font you desire. I'm in 2.4, for reference. ![Choosing Font](https://i.stack.imgur.com/y7KaO.png)
You can set the default font for the print composer items in Settings>Options>Layouts: [![enter image description here](https://i.stack.imgur.com/0c67G.jpg)](https://i.stack.imgur.com/0c67G.jpg) But I don't think there is a way to set one for the rest of QGIS, like labels. But the print composer is the most common place for fonts, so that should help a bit.
260,115
I need to forward \*@domain.com to a script. I know the [EXIM way](https://serverfault.com/questions/229964/forward-all-mail-on-a-specified-domain-to-script) and the [PROCMAIL way](https://stackoverflow.com/q/557906/313192). Is there a lighter way? Any experiences? Which one is the fastest if I JUST WANT it to delivery emails to script? (no pop accounts, nothing else) (I knew qmail but I don't want to install "big" software that eats too many resources on the server).
2011/04/15
[ "https://serverfault.com/questions/260115", "https://serverfault.com", "https://serverfault.com/users/75317/" ]
Do you really want the script to just run, regardless of what is passed to it? Or do you want proper SMTP handling? The lightest way might be to use something like Python's Twisted library listening for SMTP or a node.js [SMTP server script](https://github.com/riegel/node.js-SMTP-server), and have it fire of a script on each message required. That way you get full multi-threading, without much of anything else in the way. I would say though that Exim and the others like it will be more than fast enough for 99% of uses.
If you can spare port 25 on a unique IP what about using `netcat` to listen on port 25 ? That's truly zero load and install. A wrapper script to restart it after a reboot/fail should be easy too.
260,115
I need to forward \*@domain.com to a script. I know the [EXIM way](https://serverfault.com/questions/229964/forward-all-mail-on-a-specified-domain-to-script) and the [PROCMAIL way](https://stackoverflow.com/q/557906/313192). Is there a lighter way? Any experiences? Which one is the fastest if I JUST WANT it to delivery emails to script? (no pop accounts, nothing else) (I knew qmail but I don't want to install "big" software that eats too many resources on the server).
2011/04/15
[ "https://serverfault.com/questions/260115", "https://serverfault.com", "https://serverfault.com/users/75317/" ]
Do you really want the script to just run, regardless of what is passed to it? Or do you want proper SMTP handling? The lightest way might be to use something like Python's Twisted library listening for SMTP or a node.js [SMTP server script](https://github.com/riegel/node.js-SMTP-server), and have it fire of a script on each message required. That way you get full multi-threading, without much of anything else in the way. I would say though that Exim and the others like it will be more than fast enough for 99% of uses.
If you just want to recieve E-Mails you can use a "server" like <http://code.google.com/p/subethasmtp/> . You can have a 1 file java program using this library that will accept all emails and execute some code for it. The problem with it is you will need to have java present on the machine.
638,580
I finally crashed my old HDD (backup was made from the first SMART notice). But now the problems is what drive to buy? I had a 500 GB SATA/300 5400 RPM device with 8 MB cache. My machine is a laptop: Intel Code 2 Duo T4500, 4 GB RAM (DDR2), 64-bit machine (tell me what other specificatons could be relevant as I do not remember them now and have to look them up). I have 2 TB of external storage, so storage is not important. I want speed improvement though. 1. What drive should I purchase (skip if the question is too general)? 2. Should I consider an SSD even if I only have SATA II (SATA 300) Interface? 3. If i should stay with HDD, would I benefit from buying a 7200 RPM drive? What other parameters are important for speed gains? Thanks in advnce :)
2013/08/30
[ "https://superuser.com/questions/638580", "https://superuser.com", "https://superuser.com/users/249864/" ]
I haven't worn out my SSD yet (OCS Vertex 4 128Gb), but there is caution: SSD's memory cells can perform only limited number of writes. Therefore, you're advised to minimize the number of writes, on Linux, you should follow [these](https://wiki.archlinux.org/index.php/Solid_State_Drives) instructions. Basically, you should tell the OS not to use swapping very actively (should it really need swapping, it's fast, but OS shouldn't swap when it's not critical). Also, it's good if the SDD/OS/driver supports TRIM command. Notice that SDDs are random-access devices, unlike HDDs. OS may waste time packing adjacent IO requests, so on Linux, for example, you should move to `noop` IO scheduler. Also, google and make sure the SSD you want to buy provides good SMART test data (mine is not very informative). Don't hesitate, HDD is one of the slowest devices in the system, with SSD you'll see the difference even with SATA 2. If you need more space, you may guy hybrid SSD/HDD. P.S. Moreover, SSDs are quite complex devices, so firmware does matter.
Although this is a purchasing recommendation and not really appropriate for Superuser (so don't be surprised if this post is closed) There is no problem using an SSD with a SATA interface and it will give you a massive increase in speed over a standard hard disk. Depending on the age of your computer your sata disk will support a speed of 150 - 600 MB's per second. An SSD will give you speeds of (very roughly) 200 to 500MB's per second, compared to (again, very roughly) 20-70 MB for a typical laptop drive - you should expect a "raw speed" increase of about 3 fold using an SSD - but thats not the whole story. SSD's also have a MUCH lower latency, and this makes everything feel more responsive. I've deliberately stayed clear from recommending a particular brand of SSD but depending on what you are doing you probably want a 128 or 256 Gig SSD to complement your external drive. Just be aware that SSD's are as likely to fail as hard drives, only, when they do they are MUCH less likely to give any warning then hard drives, so make sure you back up anything important on it very regularly. If you do stay with hard drives, 7200 RPM drives are faster then 5400 RPM drives, primarily because (I believe) of their lower latency. If you are going the hard drive root, you might want to explore a hybrid drive - these are typically 5400 rpm drives with a smallish SSD cache which can boost your performance quite dramatically for many desktop-type workloads (moreso then a 7200 RPM drive) - particularly if you tend to read the same data sets frequently. They are much cheaper then pure SSD drives.
62,429
Let's say I'm in an area of a business that is a cost sink: something like customer support. In this support realm there's a KPI that's based on Net Promoter Score. NPS is more or less used in industry to drive growth. Does it make sense to want to use a KPI to drive growth for a cost-sink end of the business? Is there any relevant literature on the subject? Likewise, this is my first post here. Is this a relevant question to ask?
2016/02/21
[ "https://workplace.stackexchange.com/questions/62429", "https://workplace.stackexchange.com", "https://workplace.stackexchange.com/users/47062/" ]
> > Does it make sense to want to use a KPI to drive growth for a > cost-sink end of the business? > > > If you are a believer in KPIs and metrics like Net Promoter Score, then certainly you see value in improving them. And even if you view customer support as a "cost sink", you might concede that customer support is a cost of doing business and growing customer support might be a cost of doing business well. Given the above it makes perfect sense to improve your KPIs by growing this cost sink up to the point where the marginal value of further improvement is outweighed by the cost of attaining it. While I'm not a fan of metrics, I have used similar arguments to get by-in for growing many teams that might fall into what you view as a "cost sink", and what some might view as "making our customers happy".
> > Does it make sense to grow a cost-sink? > > > Sometimes it does, it depends what it is. Services which don't make a visible profit, or actually look like a loss can be essential to a business and give it an edge over competitors. Or they can provide rarely needed redundancy which saves a huge amount in disaster recovery situation. Two examples would be network redundancy and expanded support. In the first I have seen one of my clients experience a power issue that took out 3/4's of their core hardware on a Saturday (servers, switches, backup units), and had them business-as-usual by 6AM Monday morning, most staff didn't even know anything had happened. I have also seen an office burn down and the next week the owners were advertising in the local newspaper that if anyone owed them money, please come forwards (they'd lost all their hard copy and digital records and had no current offsite backup). In the second example some invest heavily in support so that clients have a minimum of frustration when they have issues. This can be THE major selling point for them in terms of customer relations and reputation.
41,846
So tonight I was digging a post hole for a new fence tonight. And as I'm digging the hole I hear a hollow thump, crack. Behold: ![broken ceramic pipe](https://i.stack.imgur.com/qE1JJ.jpg) This pipe is a downspout drain pipe that runs from the front of my house to the back yard. So ideas on how to fix it? Can I just put a piece of PVC pipe on top and cover it back up and call it good? I'm worried the PVC will move on top of it over time. Ideas?
2014/05/10
[ "https://diy.stackexchange.com/questions/41846", "https://diy.stackexchange.com", "https://diy.stackexchange.com/users/8805/" ]
Get a large diameter PVC pipe, cut it in half lengthwise on a bandsaw so it has a sort of shallow C shape cross section. Clean out the broken bits from the clay pipe and use the PVC pipe to cover the hole. Use some adhesive to keep it in place.
So I called a friend who's a retired contractor and he said: 1. Don't replace the pipe. You can't buy those anyway. They're 60 years old. 2. Put a piece of plastic or fabric over it and call it good. You just need dirt to not get in there. 3. More often than not, drain pipes like this aren't even put together tightly because the idea is that they drain at the joints. That is all.
41,846
So tonight I was digging a post hole for a new fence tonight. And as I'm digging the hole I hear a hollow thump, crack. Behold: ![broken ceramic pipe](https://i.stack.imgur.com/qE1JJ.jpg) This pipe is a downspout drain pipe that runs from the front of my house to the back yard. So ideas on how to fix it? Can I just put a piece of PVC pipe on top and cover it back up and call it good? I'm worried the PVC will move on top of it over time. Ideas?
2014/05/10
[ "https://diy.stackexchange.com/questions/41846", "https://diy.stackexchange.com", "https://diy.stackexchange.com/users/8805/" ]
Get a large diameter PVC pipe, cut it in half lengthwise on a bandsaw so it has a sort of shallow C shape cross section. Clean out the broken bits from the clay pipe and use the PVC pipe to cover the hole. Use some adhesive to keep it in place.
Normally I would: 1. Dig around the pipe 2. Cut pipe. 3. By PVC of the same size. 4. Install adjustable rubber gasket to merry the two together. If you don't want to dig out your whole yard (after the break) then you would need a piece of PVC and two gaskets.
41,846
So tonight I was digging a post hole for a new fence tonight. And as I'm digging the hole I hear a hollow thump, crack. Behold: ![broken ceramic pipe](https://i.stack.imgur.com/qE1JJ.jpg) This pipe is a downspout drain pipe that runs from the front of my house to the back yard. So ideas on how to fix it? Can I just put a piece of PVC pipe on top and cover it back up and call it good? I'm worried the PVC will move on top of it over time. Ideas?
2014/05/10
[ "https://diy.stackexchange.com/questions/41846", "https://diy.stackexchange.com", "https://diy.stackexchange.com/users/8805/" ]
Get a large diameter PVC pipe, cut it in half lengthwise on a bandsaw so it has a sort of shallow C shape cross section. Clean out the broken bits from the clay pipe and use the PVC pipe to cover the hole. Use some adhesive to keep it in place.
Before you dig, it's a great idea to call Dig Alert. Set up an appointment with them and have them come in and mark the existing utilities. Just in case you run into an old pipe. It's safe to know that there's a utility present and not assume it's just an old abandoned pipe. Once that is done, proceed in digging around carefully enough making space to work around broken clay pipe, and clean loose debris. Make a deep pothole 14 inches long x 14 inches wide x and 12 inches deep, right in the center below the cracked pipe, so that when you open the clay pipe...the dirty water or sewer gray water goes down below the pipe in the deep pothole and prevent yourself from getting your feet dirty. Cut the roots that caused the break in the first place and or remove palm tree if necessary to prevent future sewer breaks. Dig into the dirt back a few inches away from the broken section, and leave enough space to measure and cut that clay pipe and install pvc sdr 35 or ABS sewer black pipe same outside diameter (O.D.) Never glue parts in other projects and assume its going in perfectly. Always dry fit first, so you have an idea of how all dry fittings will fit. Never gun it, without knowing how the actual connections may fit, or make the mistake in cutting without measuring the pipe to short. Or not having enough pipe to do another cut, or do a second run to the hardware store and it's getting dark. Grab your measuring tape and measure the end of one side to the center of the no hub stainless steel bands coupling rubber fittings, transfer those measurements to the ABS or SRD 35 pvc pipe and to the previously square cut clay pipe. Check pitch fall with level and make sure all marks align with the rubber couplings, then[![enter image description here](https://i.stack.imgur.com/iEABx.jpg)](https://i.stack.imgur.com/iEABx.jpg) ...torque to specs. Let the water run for a bit and and check for leaks. After all is checked OK! Backfill 12,14 inches with clean damp sand and carefully compact around the pipe little by little. Then the rest of the dirt can be covered with clean soil compact it to top of grade and leave a few inches to put the green turf back how it was. Always make a note and take measurements where the crack happened and make a small drawing or sketch for the near future, in case of replacing the whole sewer or making other repairs to the sewer system. Very simple job! There's nothing to it! Best regards Juan Good luck! :)
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
In most disciplines and in the larger more developed countries it's perfectly straightforward to provide real-time transcription where the speech is transcribed into text on a big screen (or streamed to a delegate's laptop or tablet). > > Many deaf and hearing-impaired people do not have sign language, or > fluent sign language, and as with translation between any two > languages, nuance can be lost and ambiguities can be introduced in the > translation from English to Sign. Sign interpretation also requires > that a student maintain continuous eye contact, usually necessitating > additional notetakers. Students using real-time transcription can take > their own notes; the last dozen or so lines of the lecture are > displayed on the transcription screen for easy reference. > > > (adapted from [Mirabai Knight Stenography](http://stenoknight.com/FAQ.html#cartname)) Using computer-assisted realtime transcription (CART) is useful not just for people with hearing impairment; people whose first language is not the same as the speaker's, or people with some specific learning difficulties, may find it useful to have a real-time transcript. > > **In the USA, one in seven people have hearing loss. For people over 65, > that rate goes up to one in three.** Events at most conferences seat > hundreds of people, so statistically it’s a sure bet that at least > some of those people would benefit from captioning. Even people with > mild hearing loss, who do quite well in one-on-one social situations > by using a combination of residual hearing, lip reading, and context > clues, often have trouble with conference audio, which can be > distorted in the amplification process, and which puts the speaker so > far away from the audience that lipreading becomes impossible. There’s > also the benefits that captioning can offer people without hearing > loss, who may be more comfortable reading written English than > understanding spoken English (very common when English isn’t a > person’s first language), or who may have central audio processing > issues (very common in Aspergers and autism) or attention deficit > issues such as ADHD. > > > <https://ccacblog.wordpress.com/2013/02/11/conference-captioning-ask-for-it/> And the transcript forms a useful record of the conference, and can be used by people who couldn't attend the conference because of cost or distance, or who have disabilities making travel impossible or difficult, or who are unable to obtain visas. The transcript can also be translated (or machine-translated, which may be good enough for getting the gist or deciding whether ot order a paid-for translation). All this widens access to and participation in the conference, and means that the cost of providing the transcript can be covered by the general budget rather than being a specific cost for one or two delegates. An example of a conference video being transcribed <https://livestream.com/accounts/686369/events/4375102/videos/100329035> And a picture of a conference with, it appears, both transcription and sign. [![delegates on a conference stage, with a presentation video screen and a transcription video screen behind them; the transcription screen is showing the words spoken by a delegate ](https://i.stack.imgur.com/t5mYg.jpg)](https://i.stack.imgur.com/t5mYg.jpg)
COVID briefings are aimed at the general public (and many Deaf people have poor English skills due to a history of bad Deaf education) and are low on jargon (hence fairly straightforward to interpret). Interpreting into the local sign language is the right choice for maximum accessibility. For a high-jargon scientific speech to an audience of scientists, transcription may be the better option. You would need a specialist interpreter to manage such a speech; and it would be tricky for the Deaf audience member to view both the interpreter and the slides at the same time.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
In what was a very large astronomy conference I have seen a sign language interpreter (actually they had two, who round swap every few minutes during the talk) in certain sessions (presumably going to the sessions which the deaf scientist(s) was attending). I haven't seen this at other conferences, but whether that is because of their smaller sizes or because deaf scientists didn't attend I don't know.
They do not work. Most conferences in most fields of science do not work well for people with any sensory/communication disabilities. A few online conferences offer automatically generated captions. These help some but they are not very accurate. Edit: The fact that conferences are ableist is not because they are intentionally ableist, but rather because of ignorance. Many conference organizers would be willing to adjust their conferences if they knew how to do so and had the funds.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
I am neither a deaf scientist nor an organizer for conferences. However, I was a student and staff member at a university with a significant deaf population so I'll speak from that perspective. The prevalence of interpreters and other accommodations for those with disabilities varied significantly. It was a given that American Sign Language (ASL) interpreters would be present at large scale events and presentations aimed at the community as a whole. Additionally, interpreters would be provided for deaf students taking classes taught in spoken English. Beyond that, it was usually the responsibility of the deaf person to arrange for an interpreter; though many interpreters were employed by the university, there was no guarantee one would be available on short notice. This may explain why I seldom saw interpreters at club meetings and other smaller events, especially those that emphasized spontaneous direct communication. For example, interpreters were present at the anime club where much of the activities involved watching Japanese animation with subtitles, but I did not see any at the club for tabletop role playing games where interpersonal interaction was constant. In short, students and staff had the accommodations necessary to fulfill their primary goal(s) but often not enough for secondary ones, such as socialization. I suspect a similar dynamic applies to scientific conferences. Conferences that regularly have deaf people attending likely organize interpreters themselves but the rest will only provide accommodations when requested. (In other words, at a typical conference, you're not going to see an ASL interpreter unless there's at least one deaf person present!) The effectiveness of such accommodations are apt to suffer if the organizers lack experience and/or resources. This would discourage deaf people from attending unless the conference was essential - for example, a topic of particular interest is being discussed or a colleague is in need of material or moral support - or had a positive reputation of proper accommodations. In summary, the attendance of deaf scientists at conferences is presumably dictated by the following: * The importance of the conference to the individual deaf scientist * The presence and quality of accommodations that enable deaf scientists to be productive * The number of people with whom the deaf persons can readily communicate with * Each deaf scientist's personal comfort level with attending conferences designed primarily for hearing people
In spite of what Anonymous Physicist says, conference can work for deaf scientists. This is the case in my own area of work, digital accessibility. There are two main methods to make conferences accessible for deaf attendants: 1. **Sign language interpretation**: this means that a sign language interpreter translates the spoken words into sign language. If the conference is a local or national event, the sign language interpreter will translate into the national sign language, e.g. American Sign Language (ASL), British Sign Language (unrelated to ASL), German Sign Language, etc. However, if the conference is an international event, there won't be sign language interpreters for each of the nationalities represented in the audience. In that case, sign language interpreters will translate into [International Sign](https://en.wikipedia.org/wiki/International_Sign). Sing language interpreters take turns every 20-30 minutes. It is also helpful if they know in advance what the talk will be about, so they can prepare for it. 2. **Live transcription into text**: the talk is transcribed live and the transcription is displayed on a large screen at the front of the conference room. The transcriber can be physically at the conference or connected remotely. The transcription is in the same language as the talk. For congenitally deaf persons, sign language translation is preferable, since that represents translation into their native language (except if the translation is into International Sign), whereas written language is essentially a foreign language. Research into sign language avatars has been ongoing for quite some time, but I have never seen them in action at conferences. The above is about deaf scientists in the audience. What about deaf scientists giving a talk? In 19 years of research in accessibility, I have seen this only once. It was a congenitally deaf researcher from Finland who gave a talk in English. His pronunciation was only marginally harder to understand than that of many other non-native speakers of English, i.e. something one could get used to within minutes. The only issue was that he did not know how loud he was talking, so he asked he wasn't talking to loud. After feedback from the audience, he started talking less loudly. (The conferences I am referring to include [ICCHP / International Conference on Computers Helping People with Special Needs](https://www.icchp.org/), where translation was into International Sign, the e-Accessibility Forums in Paris (transcription), events organised by the Swedish organisation Funka/FunkaNu (transcription) and conferences organised by EU-funded projects in the domain of digital accessibility.)
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
I am neither a deaf scientist nor an organizer for conferences. However, I was a student and staff member at a university with a significant deaf population so I'll speak from that perspective. The prevalence of interpreters and other accommodations for those with disabilities varied significantly. It was a given that American Sign Language (ASL) interpreters would be present at large scale events and presentations aimed at the community as a whole. Additionally, interpreters would be provided for deaf students taking classes taught in spoken English. Beyond that, it was usually the responsibility of the deaf person to arrange for an interpreter; though many interpreters were employed by the university, there was no guarantee one would be available on short notice. This may explain why I seldom saw interpreters at club meetings and other smaller events, especially those that emphasized spontaneous direct communication. For example, interpreters were present at the anime club where much of the activities involved watching Japanese animation with subtitles, but I did not see any at the club for tabletop role playing games where interpersonal interaction was constant. In short, students and staff had the accommodations necessary to fulfill their primary goal(s) but often not enough for secondary ones, such as socialization. I suspect a similar dynamic applies to scientific conferences. Conferences that regularly have deaf people attending likely organize interpreters themselves but the rest will only provide accommodations when requested. (In other words, at a typical conference, you're not going to see an ASL interpreter unless there's at least one deaf person present!) The effectiveness of such accommodations are apt to suffer if the organizers lack experience and/or resources. This would discourage deaf people from attending unless the conference was essential - for example, a topic of particular interest is being discussed or a colleague is in need of material or moral support - or had a positive reputation of proper accommodations. In summary, the attendance of deaf scientists at conferences is presumably dictated by the following: * The importance of the conference to the individual deaf scientist * The presence and quality of accommodations that enable deaf scientists to be productive * The number of people with whom the deaf persons can readily communicate with * Each deaf scientist's personal comfort level with attending conferences designed primarily for hearing people
COVID briefings are aimed at the general public (and many Deaf people have poor English skills due to a history of bad Deaf education) and are low on jargon (hence fairly straightforward to interpret). Interpreting into the local sign language is the right choice for maximum accessibility. For a high-jargon scientific speech to an audience of scientists, transcription may be the better option. You would need a specialist interpreter to manage such a speech; and it would be tricky for the Deaf audience member to view both the interpreter and the slides at the same time.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
In what was a very large astronomy conference I have seen a sign language interpreter (actually they had two, who round swap every few minutes during the talk) in certain sessions (presumably going to the sessions which the deaf scientist(s) was attending). I haven't seen this at other conferences, but whether that is because of their smaller sizes or because deaf scientists didn't attend I don't know.
In most disciplines and in the larger more developed countries it's perfectly straightforward to provide real-time transcription where the speech is transcribed into text on a big screen (or streamed to a delegate's laptop or tablet). > > Many deaf and hearing-impaired people do not have sign language, or > fluent sign language, and as with translation between any two > languages, nuance can be lost and ambiguities can be introduced in the > translation from English to Sign. Sign interpretation also requires > that a student maintain continuous eye contact, usually necessitating > additional notetakers. Students using real-time transcription can take > their own notes; the last dozen or so lines of the lecture are > displayed on the transcription screen for easy reference. > > > (adapted from [Mirabai Knight Stenography](http://stenoknight.com/FAQ.html#cartname)) Using computer-assisted realtime transcription (CART) is useful not just for people with hearing impairment; people whose first language is not the same as the speaker's, or people with some specific learning difficulties, may find it useful to have a real-time transcript. > > **In the USA, one in seven people have hearing loss. For people over 65, > that rate goes up to one in three.** Events at most conferences seat > hundreds of people, so statistically it’s a sure bet that at least > some of those people would benefit from captioning. Even people with > mild hearing loss, who do quite well in one-on-one social situations > by using a combination of residual hearing, lip reading, and context > clues, often have trouble with conference audio, which can be > distorted in the amplification process, and which puts the speaker so > far away from the audience that lipreading becomes impossible. There’s > also the benefits that captioning can offer people without hearing > loss, who may be more comfortable reading written English than > understanding spoken English (very common when English isn’t a > person’s first language), or who may have central audio processing > issues (very common in Aspergers and autism) or attention deficit > issues such as ADHD. > > > <https://ccacblog.wordpress.com/2013/02/11/conference-captioning-ask-for-it/> And the transcript forms a useful record of the conference, and can be used by people who couldn't attend the conference because of cost or distance, or who have disabilities making travel impossible or difficult, or who are unable to obtain visas. The transcript can also be translated (or machine-translated, which may be good enough for getting the gist or deciding whether ot order a paid-for translation). All this widens access to and participation in the conference, and means that the cost of providing the transcript can be covered by the general budget rather than being a specific cost for one or two delegates. An example of a conference video being transcribed <https://livestream.com/accounts/686369/events/4375102/videos/100329035> And a picture of a conference with, it appears, both transcription and sign. [![delegates on a conference stage, with a presentation video screen and a transcription video screen behind them; the transcription screen is showing the words spoken by a delegate ](https://i.stack.imgur.com/t5mYg.jpg)](https://i.stack.imgur.com/t5mYg.jpg)
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
In what was a very large astronomy conference I have seen a sign language interpreter (actually they had two, who round swap every few minutes during the talk) in certain sessions (presumably going to the sessions which the deaf scientist(s) was attending). I haven't seen this at other conferences, but whether that is because of their smaller sizes or because deaf scientists didn't attend I don't know.
In spite of what Anonymous Physicist says, conference can work for deaf scientists. This is the case in my own area of work, digital accessibility. There are two main methods to make conferences accessible for deaf attendants: 1. **Sign language interpretation**: this means that a sign language interpreter translates the spoken words into sign language. If the conference is a local or national event, the sign language interpreter will translate into the national sign language, e.g. American Sign Language (ASL), British Sign Language (unrelated to ASL), German Sign Language, etc. However, if the conference is an international event, there won't be sign language interpreters for each of the nationalities represented in the audience. In that case, sign language interpreters will translate into [International Sign](https://en.wikipedia.org/wiki/International_Sign). Sing language interpreters take turns every 20-30 minutes. It is also helpful if they know in advance what the talk will be about, so they can prepare for it. 2. **Live transcription into text**: the talk is transcribed live and the transcription is displayed on a large screen at the front of the conference room. The transcriber can be physically at the conference or connected remotely. The transcription is in the same language as the talk. For congenitally deaf persons, sign language translation is preferable, since that represents translation into their native language (except if the translation is into International Sign), whereas written language is essentially a foreign language. Research into sign language avatars has been ongoing for quite some time, but I have never seen them in action at conferences. The above is about deaf scientists in the audience. What about deaf scientists giving a talk? In 19 years of research in accessibility, I have seen this only once. It was a congenitally deaf researcher from Finland who gave a talk in English. His pronunciation was only marginally harder to understand than that of many other non-native speakers of English, i.e. something one could get used to within minutes. The only issue was that he did not know how loud he was talking, so he asked he wasn't talking to loud. After feedback from the audience, he started talking less loudly. (The conferences I am referring to include [ICCHP / International Conference on Computers Helping People with Special Needs](https://www.icchp.org/), where translation was into International Sign, the e-Accessibility Forums in Paris (transcription), events organised by the Swedish organisation Funka/FunkaNu (transcription) and conferences organised by EU-funded projects in the domain of digital accessibility.)
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
They do not work. Most conferences in most fields of science do not work well for people with any sensory/communication disabilities. A few online conferences offer automatically generated captions. These help some but they are not very accurate. Edit: The fact that conferences are ableist is not because they are intentionally ableist, but rather because of ignorance. Many conference organizers would be willing to adjust their conferences if they knew how to do so and had the funds.
COVID briefings are aimed at the general public (and many Deaf people have poor English skills due to a history of bad Deaf education) and are low on jargon (hence fairly straightforward to interpret). Interpreting into the local sign language is the right choice for maximum accessibility. For a high-jargon scientific speech to an audience of scientists, transcription may be the better option. You would need a specialist interpreter to manage such a speech; and it would be tricky for the Deaf audience member to view both the interpreter and the slides at the same time.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
I am neither a deaf scientist nor an organizer for conferences. However, I was a student and staff member at a university with a significant deaf population so I'll speak from that perspective. The prevalence of interpreters and other accommodations for those with disabilities varied significantly. It was a given that American Sign Language (ASL) interpreters would be present at large scale events and presentations aimed at the community as a whole. Additionally, interpreters would be provided for deaf students taking classes taught in spoken English. Beyond that, it was usually the responsibility of the deaf person to arrange for an interpreter; though many interpreters were employed by the university, there was no guarantee one would be available on short notice. This may explain why I seldom saw interpreters at club meetings and other smaller events, especially those that emphasized spontaneous direct communication. For example, interpreters were present at the anime club where much of the activities involved watching Japanese animation with subtitles, but I did not see any at the club for tabletop role playing games where interpersonal interaction was constant. In short, students and staff had the accommodations necessary to fulfill their primary goal(s) but often not enough for secondary ones, such as socialization. I suspect a similar dynamic applies to scientific conferences. Conferences that regularly have deaf people attending likely organize interpreters themselves but the rest will only provide accommodations when requested. (In other words, at a typical conference, you're not going to see an ASL interpreter unless there's at least one deaf person present!) The effectiveness of such accommodations are apt to suffer if the organizers lack experience and/or resources. This would discourage deaf people from attending unless the conference was essential - for example, a topic of particular interest is being discussed or a colleague is in need of material or moral support - or had a positive reputation of proper accommodations. In summary, the attendance of deaf scientists at conferences is presumably dictated by the following: * The importance of the conference to the individual deaf scientist * The presence and quality of accommodations that enable deaf scientists to be productive * The number of people with whom the deaf persons can readily communicate with * Each deaf scientist's personal comfort level with attending conferences designed primarily for hearing people
In what was a very large astronomy conference I have seen a sign language interpreter (actually they had two, who round swap every few minutes during the talk) in certain sessions (presumably going to the sessions which the deaf scientist(s) was attending). I haven't seen this at other conferences, but whether that is because of their smaller sizes or because deaf scientists didn't attend I don't know.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
I am neither a deaf scientist nor an organizer for conferences. However, I was a student and staff member at a university with a significant deaf population so I'll speak from that perspective. The prevalence of interpreters and other accommodations for those with disabilities varied significantly. It was a given that American Sign Language (ASL) interpreters would be present at large scale events and presentations aimed at the community as a whole. Additionally, interpreters would be provided for deaf students taking classes taught in spoken English. Beyond that, it was usually the responsibility of the deaf person to arrange for an interpreter; though many interpreters were employed by the university, there was no guarantee one would be available on short notice. This may explain why I seldom saw interpreters at club meetings and other smaller events, especially those that emphasized spontaneous direct communication. For example, interpreters were present at the anime club where much of the activities involved watching Japanese animation with subtitles, but I did not see any at the club for tabletop role playing games where interpersonal interaction was constant. In short, students and staff had the accommodations necessary to fulfill their primary goal(s) but often not enough for secondary ones, such as socialization. I suspect a similar dynamic applies to scientific conferences. Conferences that regularly have deaf people attending likely organize interpreters themselves but the rest will only provide accommodations when requested. (In other words, at a typical conference, you're not going to see an ASL interpreter unless there's at least one deaf person present!) The effectiveness of such accommodations are apt to suffer if the organizers lack experience and/or resources. This would discourage deaf people from attending unless the conference was essential - for example, a topic of particular interest is being discussed or a colleague is in need of material or moral support - or had a positive reputation of proper accommodations. In summary, the attendance of deaf scientists at conferences is presumably dictated by the following: * The importance of the conference to the individual deaf scientist * The presence and quality of accommodations that enable deaf scientists to be productive * The number of people with whom the deaf persons can readily communicate with * Each deaf scientist's personal comfort level with attending conferences designed primarily for hearing people
They do not work. Most conferences in most fields of science do not work well for people with any sensory/communication disabilities. A few online conferences offer automatically generated captions. These help some but they are not very accurate. Edit: The fact that conferences are ableist is not because they are intentionally ableist, but rather because of ignorance. Many conference organizers would be willing to adjust their conferences if they knew how to do so and had the funds.
154,710
Watching various COVID-related briefings, I notice there is always a sign language interpreter. This is clearly necessary, yet thinking about all the conferences I've attended in the past, literally none of them have had a sign language interpreter. How do conferences work for deaf scientists? Do deaf scientists simply not attend conferences? If so, how does one even work in fields such as computer science where conferences are an integral part of the academic experience?
2020/09/02
[ "https://academia.stackexchange.com/questions/154710", "https://academia.stackexchange.com", "https://academia.stackexchange.com/users/84834/" ]
In what was a very large astronomy conference I have seen a sign language interpreter (actually they had two, who round swap every few minutes during the talk) in certain sessions (presumably going to the sessions which the deaf scientist(s) was attending). I haven't seen this at other conferences, but whether that is because of their smaller sizes or because deaf scientists didn't attend I don't know.
COVID briefings are aimed at the general public (and many Deaf people have poor English skills due to a history of bad Deaf education) and are low on jargon (hence fairly straightforward to interpret). Interpreting into the local sign language is the right choice for maximum accessibility. For a high-jargon scientific speech to an audience of scientists, transcription may be the better option. You would need a specialist interpreter to manage such a speech; and it would be tricky for the Deaf audience member to view both the interpreter and the slides at the same time.
111,486
My Ubuntu 11.10 desktop has an option to suspend when inactive (under System Settings > Power) but no option to hibernate. How can I hibernate it after a period of inactivity?
2012/03/09
[ "https://askubuntu.com/questions/111486", "https://askubuntu.com", "https://askubuntu.com/users/3813/" ]
Corel as well as Illustrator can recognize and read SVG format. But the problem is that sometimes they don't open SVG properly. Generally some publishers can print from pdf. I think, the main issue is to get along with the publishing house about the file prepared for print.
You can Use [Inkscape](http://inkscape.org/) you can import the ilustrator 9.0 version file, and corel 7 X14 version files but i don't recommend to you those formats because the conversion is not so good. Normally wen i have to deal with files from others programs, i use PDF or SVG the standart file of Inkscape. I hope i have helped. <http://inkscape.org/>
10,836
I was reviewing a bit and ran into a bad answer. Usually I open the question to see it within context (I know this is not the supposed course of action) and so I did. The bad answer I was reviewing wasn't there anymore, it apparently had been deleted. I went back to the review page and tried to edit the answer to make sure it was deleted, and I got a blank page saying it was a deleted post. In view of all this I clicked 'no action needed' as no action was needed. And I got the following message: > > STOP! Look and Listen. > > > This was an audit, designed to see if you were paying attention. You didn't pass. This post was of very poor quality, and needed significant improvements to be useful – passing over such posts hurts the ability of others to find and answer good questions. Please try to fix such posts by editing, downvoting, closing, or flagging as "very low quality". > > > Don't worry, we've already handled this post appropriately – but please take a minute to look it over closely, keeping in mind the guidance above. > > > Am I incurring any sort of penalty with what just happened? In my opinion I reviewed correctly. No action was needed. I find this a bit offensive, to be honest. I am reasonably careful while reviewing and most of my reviews are actually 'skip'.
2013/08/31
[ "https://math.meta.stackexchange.com/questions/10836", "https://math.meta.stackexchange.com", "https://math.meta.stackexchange.com/users/55235/" ]
Regardless of the practical consequences (or lack thereof) of your actions, there's one part of your complaint that I agree with: > > I find this a bit offensive, to be honest. > > > It is! It's an unnecessarily sanctimonious phrasing in a message from a system that is known to have false positives on a semi-regular basis. The context in which this message is presented is basically, "that thing I just asked you to do for the good of the community out of a mutual trust and understanding? Yeah, that thing was actually a complete waste of your time, and not only that, *you wasted your time incorrectly*". It wouldn't do any harm to phrase the subsequent complaint on the presumption of innocence: the guilty users *know* they are guilty, and the innocent users already have the timewasting and mistrust to be affronted by without also heaping injustice and hurt pride on top of that. (Sorry that this is a bit off-topic, but I felt the need to help legitimise your personal response, because it does kind of sound like you've been insulted here.)
The only consequences of failed audits can be a temporary and short-lived review ban if this happens multiple times within a short duration. Even if this does happen, you can still vote to close questions you find elsewhere. As you have just noticed, the system isn't perfect hence why the penalty is so small. It's just to prevent people from mindlessly racing through the audits to get a badge. It's really nothing to worry about and it happens to lots of people all the time for various reasons.
10,836
I was reviewing a bit and ran into a bad answer. Usually I open the question to see it within context (I know this is not the supposed course of action) and so I did. The bad answer I was reviewing wasn't there anymore, it apparently had been deleted. I went back to the review page and tried to edit the answer to make sure it was deleted, and I got a blank page saying it was a deleted post. In view of all this I clicked 'no action needed' as no action was needed. And I got the following message: > > STOP! Look and Listen. > > > This was an audit, designed to see if you were paying attention. You didn't pass. This post was of very poor quality, and needed significant improvements to be useful – passing over such posts hurts the ability of others to find and answer good questions. Please try to fix such posts by editing, downvoting, closing, or flagging as "very low quality". > > > Don't worry, we've already handled this post appropriately – but please take a minute to look it over closely, keeping in mind the guidance above. > > > Am I incurring any sort of penalty with what just happened? In my opinion I reviewed correctly. No action was needed. I find this a bit offensive, to be honest. I am reasonably careful while reviewing and most of my reviews are actually 'skip'.
2013/08/31
[ "https://math.meta.stackexchange.com/questions/10836", "https://math.meta.stackexchange.com", "https://math.meta.stackexchange.com/users/55235/" ]
The only consequences of failed audits can be a temporary and short-lived review ban if this happens multiple times within a short duration. Even if this does happen, you can still vote to close questions you find elsewhere. As you have just noticed, the system isn't perfect hence why the penalty is so small. It's just to prevent people from mindlessly racing through the audits to get a badge. It's really nothing to worry about and it happens to lots of people all the time for various reasons.
I have seen one such test edit. It was so obvious to reject it. May be they can one day train their SW to send less and less obviously bad edits? May be the software can eventually do good edits? [When that happens!](http://xkcd.com/810/)
10,836
I was reviewing a bit and ran into a bad answer. Usually I open the question to see it within context (I know this is not the supposed course of action) and so I did. The bad answer I was reviewing wasn't there anymore, it apparently had been deleted. I went back to the review page and tried to edit the answer to make sure it was deleted, and I got a blank page saying it was a deleted post. In view of all this I clicked 'no action needed' as no action was needed. And I got the following message: > > STOP! Look and Listen. > > > This was an audit, designed to see if you were paying attention. You didn't pass. This post was of very poor quality, and needed significant improvements to be useful – passing over such posts hurts the ability of others to find and answer good questions. Please try to fix such posts by editing, downvoting, closing, or flagging as "very low quality". > > > Don't worry, we've already handled this post appropriately – but please take a minute to look it over closely, keeping in mind the guidance above. > > > Am I incurring any sort of penalty with what just happened? In my opinion I reviewed correctly. No action was needed. I find this a bit offensive, to be honest. I am reasonably careful while reviewing and most of my reviews are actually 'skip'.
2013/08/31
[ "https://math.meta.stackexchange.com/questions/10836", "https://math.meta.stackexchange.com", "https://math.meta.stackexchange.com/users/55235/" ]
Regardless of the practical consequences (or lack thereof) of your actions, there's one part of your complaint that I agree with: > > I find this a bit offensive, to be honest. > > > It is! It's an unnecessarily sanctimonious phrasing in a message from a system that is known to have false positives on a semi-regular basis. The context in which this message is presented is basically, "that thing I just asked you to do for the good of the community out of a mutual trust and understanding? Yeah, that thing was actually a complete waste of your time, and not only that, *you wasted your time incorrectly*". It wouldn't do any harm to phrase the subsequent complaint on the presumption of innocence: the guilty users *know* they are guilty, and the innocent users already have the timewasting and mistrust to be affronted by without also heaping injustice and hurt pride on top of that. (Sorry that this is a bit off-topic, but I felt the need to help legitimise your personal response, because it does kind of sound like you've been insulted here.)
I have seen one such test edit. It was so obvious to reject it. May be they can one day train their SW to send less and less obviously bad edits? May be the software can eventually do good edits? [When that happens!](http://xkcd.com/810/)
3,184
Are there any extruders and printers that do a good job with ceramic clay? I assume the clay has to be fine grained and with enough water added to be able to be pushed out through the extruder's nozzle. I also assume the print speed would have to be arther low for the already printed layers to gain stability before printing the next layer. Perhaps some extra equipment on the printer to spped up this drying process? At the same time, the print speed can't be too low: if the bottom starts to dry enough to start shrinking, the object will crack. Also, some solutions I've seen tend to result in rather low resolution/thick layers, which is negative. So, what equipment is "out there"?
2016/12/13
[ "https://3dprinting.stackexchange.com/questions/3184", "https://3dprinting.stackexchange.com", "https://3dprinting.stackexchange.com/users/5224/" ]
What you are looking for is the Frostruder style extruders. Here is a link to a [thingiverse file for it](http://www.thingiverse.com/thing:1143). You honestly cannot expect fantastic results but it will get the job done! There are many other types. Quality will depend on the size of the point. However with something as high viscosity as clay I would not expect you to be able to push the quality very high.
I cant speak for the quality or type of clay, but there are several designs out there that print in clay. Here is the first one that popped up on my google search. <http://www.wasproject.it/w/en/wasp-launches-the-new-professional-clay-extruder/>
3,184
Are there any extruders and printers that do a good job with ceramic clay? I assume the clay has to be fine grained and with enough water added to be able to be pushed out through the extruder's nozzle. I also assume the print speed would have to be arther low for the already printed layers to gain stability before printing the next layer. Perhaps some extra equipment on the printer to spped up this drying process? At the same time, the print speed can't be too low: if the bottom starts to dry enough to start shrinking, the object will crack. Also, some solutions I've seen tend to result in rather low resolution/thick layers, which is negative. So, what equipment is "out there"?
2016/12/13
[ "https://3dprinting.stackexchange.com/questions/3184", "https://3dprinting.stackexchange.com", "https://3dprinting.stackexchange.com/users/5224/" ]
Here is an open source clay extruder from standard parts, that i designed <https://www.stoneflower3d.com/projects/clay-extruder-prototype/> Different designs of 3D printers, pumps, extruders etc. for clay are posted and discussed in the Google Community "Make Your Own Ceramic 3D Printer", driven by Jonathan Keep (<https://plus.google.com/u/0/communities/109375785524125994679>).
I cant speak for the quality or type of clay, but there are several designs out there that print in clay. Here is the first one that popped up on my google search. <http://www.wasproject.it/w/en/wasp-launches-the-new-professional-clay-extruder/>
178,412
> > ![Signature and text at the bottom of the image with a signature from the author](https://i.stack.imgur.com/oTkwR.jpg)[Click to preview in new tab](https://i.stack.imgur.com/oTkwR.jpg) > > > > > ![Naked lady posing with a planet-like object in the background](https://i.stack.imgur.com/gveEd.jpg)[Click to preview in new tab](https://i.stack.imgur.com/gveEd.jpg) > > > I have this original art and am trying to find who the creator is.. It seems familiar.. Has a 1980's feel.
2018/01/08
[ "https://scifi.stackexchange.com/questions/178412", "https://scifi.stackexchange.com", "https://scifi.stackexchange.com/users/94822/" ]
The same painting, *down to the numbering (3/12)*, and one other from same artist, was found on a [Goodwill auction site](https://www.shopgoodwill.com/Item/45902954), but they don't give the artist name. I would think that given the materials of the framing these are both from before the 1980's (the yellowing of tape on the back of the frame makes me think they glue is quite old, probably from before the 80's). [![Goodwill images](https://i.stack.imgur.com/mxu1W.jpg)](https://i.stack.imgur.com/mxu1W.jpg) From here, it would probably be best to try to find a signature match somewhere else, but I have not idea what it says: [![Artists sig](https://i.stack.imgur.com/HFZt6.jpg)](https://i.stack.imgur.com/HFZt6.jpg)
The title is given in image: "pictorial ode to a torn woman." I recognized the style (with the peculiar musculature), and I was pretty sure it was somebody who had done some early illustrations for Dungeons & Dragons. However, I couldn't quite figure the artists's name, so I had to go through lots of Googling to figure it out. It appears to be [Jim Roslof](https://en.wikipedia.org/wiki/Jim_Roslof). I suspect this may be very early work, since it is fairly crude and does not have one of the later standard signatures that Roslof used. Note, however, that the handwritten name is clearly consistent with "Jim Roslof," although it's not possible to be sure that's what it says. Searching over images shows that Roslof used many different signatures styles over his career. Here is a clearly legible version of his name. [![Written name](https://i.stack.imgur.com/r4xVD.jpg)](https://i.stack.imgur.com/r4xVD.jpg) Obviously, this name bears scarce resemblance to the signature in the above artwork. Note, however, the last letters of the words "Jim" and "Roslof." The terminal "m" trails off just like the written "n" at the end of "woman" in the picture title. Even more clearly, the quite distinctive "f" at the end of "Roslof" (apart from the underline that crosses it) is practically identical to the letter at the end of "Artists proof." The most objectively similar picture by Roslof that I could find with a bit of Googling is this image of the goddess Hecate from the *Deities & Demigods Cyclopedia*. [![Hecate](https://i.stack.imgur.com/tgDDg.jpg)](https://i.stack.imgur.com/tgDDg.jpg) Note the similar angularity in the figure. The women's feet and bellies are especially similar. The same style of hatched shading is visible in the moon behind Hecate and on some of the flames at the unknown figure's feet. There is also the similarity imagery of the (occluded?) moon.
313,239
Let's say my girlfriend has made me/us dinner and I want to compliment her for her cooking (while we're eating). Would all these be equally correct and idiomatic? It tastes wonderful, honey. It tastes wonderfully, honey. It's tasting wonderful, honey. It's tasting wonderfully, honey.
2022/04/12
[ "https://ell.stackexchange.com/questions/313239", "https://ell.stackexchange.com", "https://ell.stackexchange.com/users/149951/" ]
> > It tastes wonderful. > > > The verb "taste" is one of the verbs that can act as a linking verb, to connect a subject and an adjective. In this linking form, you should normally use the simple present tense (unless you particularly want to emphasise that that it will soon not taste wonderful) and use an adjective, not an adverb. The word "wonderful" describes "it". It doesn't describe the action of "taste", and so you would not use an adverb.
Take a look at the following examples: I'm tasting the food. It tastes good / wonderful. I'm smelling the flower. It smells good / wonderful. It feels wonderful. It looks wonderful.
5,201,874
I am going to develop the web application in asp.net. where as in this application there is 4 user perspectives. all are independent all togather. but I want to develop under the single web solution. also want to create the number web applications for each user perspective, under this solution. so can i use common web.config file for this solution? what will impact ? if can not then how to manage the config files so that I need to change configuration at once (if need in future)
2011/03/05
[ "https://Stackoverflow.com/questions/5201874", "https://Stackoverflow.com", "https://Stackoverflow.com/users/165309/" ]
Yes, you can use the common web.config file. Config files are hierarchical -- with some caveats that don't concern you now because you will use only one config file. If, in the future, you decide each app needs a slightly different config, then you can create a web.config in each app, containing only the config deltas. Having said all this, I have never developed an app this way because * my apps all need a different config * the common web.config is in a different location than my apps, which makes version control more complicated
Web.config files are "cascading" in nature. Config values in parent folders are applied first. If there are config files in sub folders they will override the parent values where applicable. Unless you have a high degree of reuse of custom configuration settings it may make more sense to simply copy & paste the values between files on the four projects. If the configurations really are identical, you could also create a build task to copy a common web.config file into each of the projects.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
The number of processor cores is not coupled with the number of agents. It's MSBuild that takes the advantage of multi-core processors. So the right answer here is #1: The default number of agents is one. Having multiple agents on one build machine will allow you to parallelize building different platform/configuration, or running more than one builds on a controller at a time, though.
According to <http://msdn.microsoft.com/en-us/library/ee259687.aspx> it looks like the default number of build agents is one. That is a weird question for them to be asking.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
According to <http://msdn.microsoft.com/en-us/library/ee259687.aspx> it looks like the default number of build agents is one. That is a weird question for them to be asking.
In the Hashimi/Bartholomew MSPress book "Using MSBuild and Team Foundation Build" the authors recommend NOT having Build Agents on the same machine as the Team Foundation Server. The reasons they give are: * compilation is resource intensive and could degrade TFS performance * build scripts and tests running on the TFS could compromise security, since these are usually written by non-admins * build scripts and projects often require 3rd party software or libraries, and installing these on the TFS could likewise compromise security. The only situation in which they say its OK is for demo or testing purposes where it isn't practical to have these on different machines or virtuals.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
Thanks for the work. It turns out that the correct answer was answer number 2. We found "If you installed Team Foundation Build Service on its own server, the default setting for **Number of build agents to run on this build machine is equal to the number of processor cores** on this server." at <http://msdn.microsoft.com/en-us/library/ee259683.aspx> And this was apparently the correct answer.
According to <http://msdn.microsoft.com/en-us/library/ee259687.aspx> it looks like the default number of build agents is one. That is a weird question for them to be asking.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
The number of processor cores is not coupled with the number of agents. It's MSBuild that takes the advantage of multi-core processors. So the right answer here is #1: The default number of agents is one. Having multiple agents on one build machine will allow you to parallelize building different platform/configuration, or running more than one builds on a controller at a time, though.
In the Hashimi/Bartholomew MSPress book "Using MSBuild and Team Foundation Build" the authors recommend NOT having Build Agents on the same machine as the Team Foundation Server. The reasons they give are: * compilation is resource intensive and could degrade TFS performance * build scripts and tests running on the TFS could compromise security, since these are usually written by non-admins * build scripts and projects often require 3rd party software or libraries, and installing these on the TFS could likewise compromise security. The only situation in which they say its OK is for demo or testing purposes where it isn't practical to have these on different machines or virtuals.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
Thanks for the work. It turns out that the correct answer was answer number 2. We found "If you installed Team Foundation Build Service on its own server, the default setting for **Number of build agents to run on this build machine is equal to the number of processor cores** on this server." at <http://msdn.microsoft.com/en-us/library/ee259683.aspx> And this was apparently the correct answer.
The number of processor cores is not coupled with the number of agents. It's MSBuild that takes the advantage of multi-core processors. So the right answer here is #1: The default number of agents is one. Having multiple agents on one build machine will allow you to parallelize building different platform/configuration, or running more than one builds on a controller at a time, though.
5,963,634
I am attending a competition where I am supposed to answer a couple of engineering and technical related questions. I hope that some of you guys could help me with this one. I've already searched the web for a couple of hours, but I can't seem to make out the difference in the names used. I saw this article <http://msdn.microsoft.com/en-us/library/dd793166.aspx> on msdn, but I still can't make out the answer. **The question is:** In Team Fitting Solutions we use Team Foundation Server 2010 to manage and build our C# and C++ code. Team Foundation Server is able to build the code using a number of build agents and build services. What is the default number of build agents that can run on a build service when the build service is installed on its own dedicated server? 1. One build agent per build service 2. The default number of build agents is defined by the number of processor cores on the server hosting the build service 3. None, because a build agent cannot run on a server hosting the build service; it needs its own dedicated server
2011/05/11
[ "https://Stackoverflow.com/questions/5963634", "https://Stackoverflow.com", "https://Stackoverflow.com/users/503899/" ]
Thanks for the work. It turns out that the correct answer was answer number 2. We found "If you installed Team Foundation Build Service on its own server, the default setting for **Number of build agents to run on this build machine is equal to the number of processor cores** on this server." at <http://msdn.microsoft.com/en-us/library/ee259683.aspx> And this was apparently the correct answer.
In the Hashimi/Bartholomew MSPress book "Using MSBuild and Team Foundation Build" the authors recommend NOT having Build Agents on the same machine as the Team Foundation Server. The reasons they give are: * compilation is resource intensive and could degrade TFS performance * build scripts and tests running on the TFS could compromise security, since these are usually written by non-admins * build scripts and projects often require 3rd party software or libraries, and installing these on the TFS could likewise compromise security. The only situation in which they say its OK is for demo or testing purposes where it isn't practical to have these on different machines or virtuals.
10,366,401
I'm working on a system that have multiple threads and one shared object. There is a number of threads that do read operations very often, but write operations are rare, maybe 3 to 5 per day. I'm using rwlock for synchronization but the lock acquisition operation it's not fast enough since it happens all the time. So, I'm looking for a faster way of doing it. Maybe a way of making the write function atomic or looking all threads during the write. Portability it's not a hard requirement, I'm using Linux with GCC 4.6.
2012/04/28
[ "https://Stackoverflow.com/questions/10366401", "https://Stackoverflow.com", "https://Stackoverflow.com/users/1183970/" ]
You might want to use a multiple objects rather than a single one. Instead of actually sharing the object, create an object that holds the object and an atomic count, then share a *pointer* to this structure among the threads. [Assuming that there is only one writer] Each reader will get the pointer, then atomically increment the counter and use the object, after reading, atomically decrement the counter. The writer will create a new object that holds a copy of the original and modify it. Then perform an atomic swap of the two pointers. Now the problem is releasing the old object, which is why you need the count of the readers. The writer needs to continue checking the count of the old object until all readers have completed the work at which point you can delete the old object. If there are multiple writers (i.e. there can be more than one thread updating the variable) you can follow the same approach but with writers would need to do a compare-and-swap exchange of the pointer. If the pointer from which the updated copy has changed, then the writer restarts the process (deletes it's new object, copies again from the pointer and retries the CAS)
Maybe you could use a spinlock, the threads will busy wait until unlocked. If the threads aren't locked for long it can be much more efficent than mutexes since the locking and unlocking is completed with less instructions. spinlock is a part of POSIX pthread although optional so I don't know if it's implemented on your system. I used them in a C program on ubuntu but had to compile with -std=gnu99 instead of c99.
10,366,401
I'm working on a system that have multiple threads and one shared object. There is a number of threads that do read operations very often, but write operations are rare, maybe 3 to 5 per day. I'm using rwlock for synchronization but the lock acquisition operation it's not fast enough since it happens all the time. So, I'm looking for a faster way of doing it. Maybe a way of making the write function atomic or looking all threads during the write. Portability it's not a hard requirement, I'm using Linux with GCC 4.6.
2012/04/28
[ "https://Stackoverflow.com/questions/10366401", "https://Stackoverflow.com", "https://Stackoverflow.com/users/1183970/" ]
Have you considered using read-copy-update with [liburcu](http://lttng.org/urcu)? This lets you avoid atomic operations and locking *entirely* on the read path, at the expense of making writes quite a bit slower. Note that some readers might see stale data for a short time, though; if you need the update to take effect immediately, it may not be the best option for you.
Maybe you could use a spinlock, the threads will busy wait until unlocked. If the threads aren't locked for long it can be much more efficent than mutexes since the locking and unlocking is completed with less instructions. spinlock is a part of POSIX pthread although optional so I don't know if it's implemented on your system. I used them in a C program on ubuntu but had to compile with -std=gnu99 instead of c99.
56,589
In a few different English versions of the Bible, Isaiah 14:12 begins with something like: > > How art thou fallen from heaven, O Lucifer, son of the morning! (KJV) > > > I thought, however, that Jesus was son of the morning. Does anyone think this Isaiah passage was translated incorrectly, and that maybe it should be "son of *mourning*" instead of "son of the morning"? What about Satanail (or *Satanial*?), or Samuel? Is either one of *these* more properly Satan's name?
2017/04/29
[ "https://christianity.stackexchange.com/questions/56589", "https://christianity.stackexchange.com", "https://christianity.stackexchange.com/users/34945/" ]
From the CSI website comes the following (from the drop-down menu "About Us"): > > The Church of South India is the result of the union of churches of varying traditions Anglican, Methodist, Congregational, Presbyterian, and Reformed. It was inaugurated in September 1947, after protracted negotiation among the churches concerned. Organized into 24 dioceses, each under the spiritual supervision of a bishop, the church as a whole is governed by a synod, which elects a moderator (presiding bishop) every 2 years. Episcopacy is thus combined with Synodical government, and the church explicitly recognizes that Episcopal, Presbyterian, and congregational elements are all necessary for the church's life. The Scriptures are the ultimate standard of faith and practice. The historic creeds are accepted as interpreting the biblical faith, and the sacraments of baptism and the Lord's Supper are recognized as of binding obligation. > > > Based on this excerpt from CSI's website, I'd say you would be safe in visiting churches with the following words in their names: > > * Anglican > * Methodist > * Congregational > * Presbyterian > * Reformed > > > From the CSI's statement of faith, as quoted by [gracelightcity](https://gracelightcity.wordpress.com/2011/04/03/doctrinal-beliefs-of-the-church-of-south-india/): > > “. . . the Scriptures are the ultimate standard of faith and practice. The historic creeds are accepted as interpreting the biblical faith, and the sacraments of baptism and the Lord’s Supper are recognized as of binding obligation.” > > > Given the Church of South India's emphasis on the Word of God being the final word in both faith and practice, I'd stick with churches which have a similar commitment to the Bible.[![enter image description here](https://i.stack.imgur.com/cw0MM.png)](https://i.stack.imgur.com/cw0MM.png) That you are "struggling to find the right church" is a good sign. I recommend you pick out a church with one of the five key words in it (as listed above) and see if that church has a website to which you can go for further information. Does something you read about a given church resonate strongly within you? If so, visit that church. See if you feel comfortable there. Ask yourself if the church members there make you feel welcomed, accepted, and loved. If so, perhaps the Holy Spirit is indicating that that church is for you. If not, try another church. In conclusion, the matter of choosing a church to be your home church is very important. Church "shopping" can sometimes be a daunting task, but the criteria for choosing should be pretty much the same for any Christian. More specifically, the church you choose to be your home church should be a place where you are fed spiritually, experience true fellowship in the Holy Spirit, and are "put to work" in a ministry in which your spiritual gift-mix can be used to edify that local assembly of believers and bring glory to God. I wish you well in the process!
Well there are not dozens of denominations here in the US, there are thousands. The CSI is interesting in that it blends some Protestant traditions that have classically been opposed to each other here in the US. I will leave out Roman Catholic and Eastern Orthodox from the initial discussion as they are not Protestant as your home church was. I have visited many different types of churches here and generally like to break them down on a set of spectrums (forgive the simplifications): * Calvinist to Arminian (or Augustinian to Cassian for historical folks, or Reformed to Free Will) * Liturgical to Evangelical (bit of a misnomer as no church isn't evangelical, Evangelical here refers to a more "seeker-friendly" style of worship and soteriology) * Fundamental to Liberal **Liturgical** Liturgical vs Evangelical is the worship style and generally also encapsulates whether the church uses a more sacramental or scholastic/individualistic soteriology. It is also usually a good indication of whether the church is high church or low church (explanation to follow). A liturgical worship service is focused around the participation in communion and most such churches teach that such participation confers saving grace upon the communicants (ie - it is sacramental). Sacramental also means that they tend to teach that the church is indeed essential to a person's salvation (ie - communal not individualistic soteriology). Such churches will also tend to be high church meaning that there is a hierarchical order of authority that terminates with a high ranking bishop. There is usually a communion of bishops that together direct the entire communion. The worship services at a liturgical church tend to include formal vestments for the clergy, traditional music with pianos and organs, and service of communion from a single chalice served by the presbyter/priest/pastor and generally more traditionally beautiful buildings. Anglican and Lutheran fall into this category with Methodists and Presbyterian often taking up some middle ground. Note that some Anglican and Lutheran communions have moved towards the Evangelical side recently. **Evangelical** Evangelical covers a broad range of stuff, but for the purposes of this post we'll say that it covers the individualistic soteriology and seeker-friendly worship style (although technically there are Evangelical Liturgical Lutherans, whatever I'm not a scholar). At these churches the worship emphasizes reaching a contemporary audience with contemporary music and a well crafted sermon. This can range from a single acoustic guitar to an entire rock band with flashing lights, stage fog, and skinny jeans. But the main focus is usually the sermon from what is hopefully an engaging pastor. These churches tend to teach that salvation is mainly a personal relationship stemming from a personal commitment/conversion in which the church simply serves as a place of fellowship. The sacraments don't confer grace so much as they are a reminder. The governance of such churches is usually centralized at the local level with a looser affiliation with other members of the communion. The church governance is usually more democratic, but a local pastor can also turn into a mini-pope. Reformed, Baptist, Congregationalists, community churches, and most any mega-church around you will fall into this category. **Calvinist vs Arminian** This is a pretty simple difference. Arminians believe that the free will of man must be exercised for salvation and Calvinists don't. Presbyterians and Reformed will be more hard core Calvinist and Lutherans to a lesser degree. Methodists are more explicitly Arminian and not sure about Anglicans (CS Lewis was certainly not Calvinist). Low churches are usually split on the issue, so you'd have to ask the pastor. Most mega-churches avoid the topic as they see it as something that just brings division. **Fundamental to Liberal** Fundamentalism arose in response to modern skepticism. A fundamental church confesses that you must believe in some basic tenants of the Christian faith to be Christian like the divine inspiration of the Bible, the virgin birth and incarnation, the resurrection of Christ, etc. Liberals generally like to question fundamental doctrines, like criticizing whether the Bible is authentic and questioning whether homosexuality is an acceptable Christian lifestyle. In general there are deep ends on both sides of this pool. Liberalism tends to be more prominent in high church communions and fundamentalism in low church (There is a ton of detail here, but I can't go into it right now, for instance I believe there is an unhealthy form of ultra fundamentalism that involves calling people heretics when they don't believe in the literal 6 day creation and stuff like that, but that's for another post). Now if I were you I would skip this whole Protestant mess and just return to the ancient church and become Eastern Orthodox :).
56,589
In a few different English versions of the Bible, Isaiah 14:12 begins with something like: > > How art thou fallen from heaven, O Lucifer, son of the morning! (KJV) > > > I thought, however, that Jesus was son of the morning. Does anyone think this Isaiah passage was translated incorrectly, and that maybe it should be "son of *mourning*" instead of "son of the morning"? What about Satanail (or *Satanial*?), or Samuel? Is either one of *these* more properly Satan's name?
2017/04/29
[ "https://christianity.stackexchange.com/questions/56589", "https://christianity.stackexchange.com", "https://christianity.stackexchange.com/users/34945/" ]
From the CSI website comes the following (from the drop-down menu "About Us"): > > The Church of South India is the result of the union of churches of varying traditions Anglican, Methodist, Congregational, Presbyterian, and Reformed. It was inaugurated in September 1947, after protracted negotiation among the churches concerned. Organized into 24 dioceses, each under the spiritual supervision of a bishop, the church as a whole is governed by a synod, which elects a moderator (presiding bishop) every 2 years. Episcopacy is thus combined with Synodical government, and the church explicitly recognizes that Episcopal, Presbyterian, and congregational elements are all necessary for the church's life. The Scriptures are the ultimate standard of faith and practice. The historic creeds are accepted as interpreting the biblical faith, and the sacraments of baptism and the Lord's Supper are recognized as of binding obligation. > > > Based on this excerpt from CSI's website, I'd say you would be safe in visiting churches with the following words in their names: > > * Anglican > * Methodist > * Congregational > * Presbyterian > * Reformed > > > From the CSI's statement of faith, as quoted by [gracelightcity](https://gracelightcity.wordpress.com/2011/04/03/doctrinal-beliefs-of-the-church-of-south-india/): > > “. . . the Scriptures are the ultimate standard of faith and practice. The historic creeds are accepted as interpreting the biblical faith, and the sacraments of baptism and the Lord’s Supper are recognized as of binding obligation.” > > > Given the Church of South India's emphasis on the Word of God being the final word in both faith and practice, I'd stick with churches which have a similar commitment to the Bible.[![enter image description here](https://i.stack.imgur.com/cw0MM.png)](https://i.stack.imgur.com/cw0MM.png) That you are "struggling to find the right church" is a good sign. I recommend you pick out a church with one of the five key words in it (as listed above) and see if that church has a website to which you can go for further information. Does something you read about a given church resonate strongly within you? If so, visit that church. See if you feel comfortable there. Ask yourself if the church members there make you feel welcomed, accepted, and loved. If so, perhaps the Holy Spirit is indicating that that church is for you. If not, try another church. In conclusion, the matter of choosing a church to be your home church is very important. Church "shopping" can sometimes be a daunting task, but the criteria for choosing should be pretty much the same for any Christian. More specifically, the church you choose to be your home church should be a place where you are fed spiritually, experience true fellowship in the Holy Spirit, and are "put to work" in a ministry in which your spiritual gift-mix can be used to edify that local assembly of believers and bring glory to God. I wish you well in the process!
**The Church of South India currently aligns under Anglican Communion** [Wikipedia]. It formed in 1947 and is the second largest Christian denomination in Southern India behind Roman Catholic, currently claiming 4 million parishioners, 11,000 ministers, 100 hospitals, 2000 schools, and 130 colleges. It absorbed Calvinist denominations at some point and maintains a certain degree of dogma suggesting such. However, it's sacraments, liturgy, and hierarchy are Anglican in origin [this is all Wikipedia]. Socially, it's course has become more liberal in the last 10 years, ordaining female ministers and being (perhaps reluctantly) supportive of the LGBTQ community. **An Anglican church seems like your best first choice.** But I liked the comment mentioning any church is better than staying at home!
306,069
A gentleman on youtube called Tom Campbel, PhD, claims the following. If we perform the double slit experiment with 2 kinds of detectors, one which records the *which path* information about the particle and another that detects the pattern formed by the particles in the screen, something very unusual would happen: If we looked at the screen detector alone, we would see a clump pattern, because there exists data, information, on the *which path* part of the experiment. If we looked at the *which path* device alone, it would show data regarding which path each particle took. If we looked at none, the result would remain uncertain. If we would let the *which path* recording device sealed away for 100 years, never examining it's data, and in the mean time we'd never look at the screen pattern device, and if, 100 years later, we were to completely destroy the information recorded in the *which path* device so it could never, ever be recovered, when we looked at the daata f the screen pattern it would show an interference pattern, instead of a clump pattern, even though we tood the information on *which path*. but since we completely and utterly erased that information without first observing it, the *which way* measurement would be like it never existed in the first place, because there isn't data anywhere in the universe of what it recorded. Is there any truth to this? Could any of you conduct such an experiment (no need for the 100 years, but absolute need for the complete destruction of the information in the *which path* measurement device. Is that even possible? to completely destroy information?)
2017/01/19
[ "https://physics.stackexchange.com/questions/306069", "https://physics.stackexchange.com", "https://physics.stackexchange.com/users/123779/" ]
In short, hiding the data is no different than seeing the data. **Quantum limits how good our meters are but we can always build a bad meter** An "observation" happens any time a Q (a "small" quantum system) interacts with a C (a "large" classical system) enough to perturb the Q. In the double slit the Q is the photon, and the C's are any detector device. If a C gives us some information, it must count as a measurement, so any C that can figure out which slit the photon passed through will destroy the interference pattern. However, we always do worse: a C that erases it's own data after a set amount of time will still count as an observation (destroy the pattern) even it said device is worthless to us. Even putting a brick in front of both slits counts as "observing" which slit each photon went through! **No time travel/faster than light** We can't build a C that will allow someone i.e. a lightyear away and a year in the future to choose whether or not the photons going through the slit are observed. It is "observed" whether or not they intercept the i.e beams that tell them the results.
The question in focus here is known as that of "quantum erasing" and relates to the big interpretation issues of quantum mechanics. One may look e.g. at the Wikipedia article: <https://en.wikipedia.org/wiki/Quantum_eraser_experiment> or at Preskill's lectures notes [http://www.theory.caltech.edu/~preskill/ph219/chap2\_15.pdf](http://www.theory.caltech.edu/%7Epreskill/ph219/chap2_15.pdf) (section 2.5.4). Basically, the answer to your question ("is there any truth to this?") is yes, although much depends on what one accepts as a bona fide "recording" of the data. If it means intrication with a quantum object (here both for the which path device and the screen), then it can indeed be "erased" and interference recovered. If recording means something like a quantum measurement implying Born postulate ("wave packet collapse" involving a "macroscopic apparatus") then it cannot be erased and interference will not be recovered. What makes the difference (and is there really one) between a quantum object and a "macroscopic apparatus" has been a hot question for ages. On practical grounds, interaction with the environment will cause "decoherence" making "recording" look very much like "measurement", but this leaves open Bell's criticism about "For All Practical Purpose" arguments.
306,069
A gentleman on youtube called Tom Campbel, PhD, claims the following. If we perform the double slit experiment with 2 kinds of detectors, one which records the *which path* information about the particle and another that detects the pattern formed by the particles in the screen, something very unusual would happen: If we looked at the screen detector alone, we would see a clump pattern, because there exists data, information, on the *which path* part of the experiment. If we looked at the *which path* device alone, it would show data regarding which path each particle took. If we looked at none, the result would remain uncertain. If we would let the *which path* recording device sealed away for 100 years, never examining it's data, and in the mean time we'd never look at the screen pattern device, and if, 100 years later, we were to completely destroy the information recorded in the *which path* device so it could never, ever be recovered, when we looked at the daata f the screen pattern it would show an interference pattern, instead of a clump pattern, even though we tood the information on *which path*. but since we completely and utterly erased that information without first observing it, the *which way* measurement would be like it never existed in the first place, because there isn't data anywhere in the universe of what it recorded. Is there any truth to this? Could any of you conduct such an experiment (no need for the 100 years, but absolute need for the complete destruction of the information in the *which path* measurement device. Is that even possible? to completely destroy information?)
2017/01/19
[ "https://physics.stackexchange.com/questions/306069", "https://physics.stackexchange.com", "https://physics.stackexchange.com/users/123779/" ]
> > but since we completely and utterly erased that information without first observing it, the which way measurement would be like it never existed in the first place, > > > This is absolutely wrong. I blame Schrodinger's cat thought experiment for this type of misunderstanding of how quantum mechanics experiments work. There exist various physics frameworks, and in these frameworks there are elegant theories which have been tested over and over again and work beautifully within their framework. Classical mechanics works for macroscopic observations, classical electrodynamics for light, thermodynamics, an emergent theory from the many body problems of classical mechanics, also is a self contained framework etc. Quantum mechanics is the framework of the very small , a self consistent framework when the Heisenberg uncertainty has to be obeyed , with its own differential equations and mathematical formulations. It can be shown mathematically that it is the underlying framework for most other frameworks which emerge naturally from the laws of the microcosm.( except gravity, and most physicists hope that will be shown to be quantized in the future). Paradoxes arise when one mixes frameworks without keeping track of the hypotheses entering the problem. Our measurements and measuring devices are in the classical frameworks. We explore the underlying quantum mechanical framework by using interactions that obey quantum mechanical rules and studying the result in our macroscopic detectors. As an example, the whole experiment, poison and all, of Schrodinger's cat is a macroscopic detector of what happens at the quantum mechanical level. The quantum mechanical equations enter in calculating the probability of decay, and the dead or alive cat is the *measurement* of whether a decay happened or not. The "dead or alive" cat is as unknown as the "is there a measurement on the counter". The measurement will be there whether I look at it or not, if a decay happened, because measurement of counters is in the classical framework. In the case of the double slit , any measurements, photos... are in the classical framework and will be there whether looked at or not. The quantum mechanical uncertainties enter at the level of the QM equations with the boundary conditions "particle impinging on two slits of certain width and distance' and "particle impinging on two slits of certain width and distance + a detector at one of the slits". The interference pattern observed ( or not) is in the probability distribution from the complex conjugate square of the wavefunction describing the experiment, the quantum mechanical framework. The measurement in the detecting devices is in the classical framework. P.S. an interesting experiment that shows origin of the difference in the two QM distributions [is here](https://phys.org/news/2011-01-which-way-detector-mystery-double-slit.html).
The question in focus here is known as that of "quantum erasing" and relates to the big interpretation issues of quantum mechanics. One may look e.g. at the Wikipedia article: <https://en.wikipedia.org/wiki/Quantum_eraser_experiment> or at Preskill's lectures notes [http://www.theory.caltech.edu/~preskill/ph219/chap2\_15.pdf](http://www.theory.caltech.edu/%7Epreskill/ph219/chap2_15.pdf) (section 2.5.4). Basically, the answer to your question ("is there any truth to this?") is yes, although much depends on what one accepts as a bona fide "recording" of the data. If it means intrication with a quantum object (here both for the which path device and the screen), then it can indeed be "erased" and interference recovered. If recording means something like a quantum measurement implying Born postulate ("wave packet collapse" involving a "macroscopic apparatus") then it cannot be erased and interference will not be recovered. What makes the difference (and is there really one) between a quantum object and a "macroscopic apparatus" has been a hot question for ages. On practical grounds, interaction with the environment will cause "decoherence" making "recording" look very much like "measurement", but this leaves open Bell's criticism about "For All Practical Purpose" arguments.
9,133,862
My partner and I are working on a project using TFS. For some reason when he adds files to the project then checks them in, I don't get them when doing an get latest version. The same thing happens when I check in files, he doesn't see them. We can both go in to source control and see the files are there, and even get latest version, but they are never added to the solution explorer. Any ideas what is happening?
2012/02/03
[ "https://Stackoverflow.com/questions/9133862", "https://Stackoverflow.com", "https://Stackoverflow.com/users/623719/" ]
If he, or you, has checked in the files, but not checked in the project file, then TFS will not download them if you execute a "get latest" by right clicking on the project or solution in Solution Explorer. You WILL however get the files downloaded correctly if you right click and "get latest" on the folder in the source control window, however, they won't get loaded into your solution, because you haven't told TFS that they belong there.
Can you verify that the .csproj (or .vbproj, etc. depending on language) has been updated to actually include the new file(s)? Is the project file also being committed along with the new files?
2,916,044
I'm wondering if there is any news about a "Android Dev Phone 3"? Is an updated hardware of the "Android Dev Phone 2" that is coming soon? (probably an unlocked version of the Nexus One)
2010/05/26
[ "https://Stackoverflow.com/questions/2916044", "https://Stackoverflow.com", "https://Stackoverflow.com/users/1406/" ]
The Nexus One IS the unlocked version of the Nexus One! It's sim unlocked and the bootloader is unlocked.
There is no news or suggestion by Google, nor are there any leaks as of today.
2,916,044
I'm wondering if there is any news about a "Android Dev Phone 3"? Is an updated hardware of the "Android Dev Phone 2" that is coming soon? (probably an unlocked version of the Nexus One)
2010/05/26
[ "https://Stackoverflow.com/questions/2916044", "https://Stackoverflow.com", "https://Stackoverflow.com/users/1406/" ]
There is no news or suggestion by Google, nor are there any leaks as of today.
"The Nexus One is no longer available for direct purchase from Google, but will be made available through a partner for sale to registered developers. Please check back for more updates regarding developer availability." <http://www.google.com/support/android/bin/answer.py?hl=en&answer=166508>