qid int64 1 74.7M | question stringlengths 12 33.8k | date stringlengths 10 10 | metadata list | response_j stringlengths 0 115k | response_k stringlengths 2 98.3k |
|---|---|---|---|---|---|
47,029 | I am a newly appointed editor to a top journal. I have received my first manuscript assignment. I see in the journal system that the authors have provided preferences for reviewers for their paper.
I was wondering what is the norm like with respect to this. Do editors normally go by author's preference or do they ignore it? What factors should I take into account before considering author preference of reviewers?
On one hand this makes my task of searching appropriate reviewers easy but I suspect this might also give an unfair edge to the authors if the reviewers have some/any kind of bias. | 2015/06/11 | [
"https://academia.stackexchange.com/questions/47029",
"https://academia.stackexchange.com",
"https://academia.stackexchange.com/users/6103/"
] | The way the preferred reviewers are used varies. Some go by these suggestions whole-heartedly while others do not. I lean towards the latter since my experience with some preferred names is less than favourable.
In my experience names listed can be good. I usually double check to see if persons seem affiliated in some way and if they do I avoid appointing them. As a rule, however, I try to find persons independently and based on my own experience. I tend to use the preferred names as back-ups unless my preference and the authors coincide.
The reason for my slight aversion towards the preferred is that some authors tend to list friends and other persons who are obviously close to the authors. I have seen many low quality reviews come out from such reviewers an clearly at a rate very different from independently chosen reviewers. Judging what is too close is not easy and sometimes it may be justified if, for example, the topic is such that local knowledge comes into play. For the reason of uncertainty I therefore try to at least mix them up so that one is chosen by me independently and the other is selected from the authors suggestions.
So, try to assess the quality of the preferred reviewers and at least try to find some to complement a preferred reviewer will be my advice.
It is also common that authors list non-preferred reviewers. I always stay clear of such reviewers since I do not know what lies beneath the sentiment. | Be extra careful when following authors' suggestions for suitable peer reviewers. There has been a recent case of authors suggesting fabricated contacts as "reviewers", as described on <http://publicationethics.org/news/cope-statement-inappropriate-manipulation-peer-review-processes>. That case led some publishers to stop asking for reviewer suggestions explicitly within their submission processes.
Measures of caution that I find useful include:
* Only choose an author-suggested reviewer if you can verify independently that this person is suited as reviewer, ideally from your own prior knowledge.
* Don't use the contact address provided by the authors, but use a contact address that you can obtain independently, for example from the reviewer's university web page.
* Verify very carefully that there's no conflict of interests for that reviewer, for example joint publications, same affiliations also in the recent past, or similar.
* Don't make a decision if you only have reviews from author-suggested reviewers, but have at least one independently chosen reviewer. |
47,029 | I am a newly appointed editor to a top journal. I have received my first manuscript assignment. I see in the journal system that the authors have provided preferences for reviewers for their paper.
I was wondering what is the norm like with respect to this. Do editors normally go by author's preference or do they ignore it? What factors should I take into account before considering author preference of reviewers?
On one hand this makes my task of searching appropriate reviewers easy but I suspect this might also give an unfair edge to the authors if the reviewers have some/any kind of bias. | 2015/06/11 | [
"https://academia.stackexchange.com/questions/47029",
"https://academia.stackexchange.com",
"https://academia.stackexchange.com/users/6103/"
] | I assume you're one new member of an established editorial board, with an Editor in Chief and other members of the board fully involved. Why not ask them what the convention is for this particular journal? | Be extra careful when following authors' suggestions for suitable peer reviewers. There has been a recent case of authors suggesting fabricated contacts as "reviewers", as described on <http://publicationethics.org/news/cope-statement-inappropriate-manipulation-peer-review-processes>. That case led some publishers to stop asking for reviewer suggestions explicitly within their submission processes.
Measures of caution that I find useful include:
* Only choose an author-suggested reviewer if you can verify independently that this person is suited as reviewer, ideally from your own prior knowledge.
* Don't use the contact address provided by the authors, but use a contact address that you can obtain independently, for example from the reviewer's university web page.
* Verify very carefully that there's no conflict of interests for that reviewer, for example joint publications, same affiliations also in the recent past, or similar.
* Don't make a decision if you only have reviews from author-suggested reviewers, but have at least one independently chosen reviewer. |
16,211 | In a game where where player 1 is on black and player 2 still has balls remaining on the table. Given the scenarios where player 2 commits a foul giving player 1 a free shot, then on that free shot player 1 first hits one of player 2's remaining balls before sinking the black on the same shot. Does player 1 win in this case? If not, what is the correct flow of play following this?
The WPA rule book states that the "Wrong Ball First" rule is suspended on free shots. However, does this still apply on the black given that you are essentially winning off a foul shot?
Using this version of the rule book for reference: <http://wpapool.co.za/wp-content/uploads/2017/03/WPA_New_Rules.pdf> | 2017/05/29 | [
"https://sports.stackexchange.com/questions/16211",
"https://sports.stackexchange.com",
"https://sports.stackexchange.com/users/13435/"
] | The shooter wins the rack if they sink the black ball, and do not commit a foul under the rules.
As it is a free shot, the shooter cannot be called foul under 6.2 Wrong Ball First.
They may of course shoot the cue ball into the opponent ball, so that it contacts the black ball and sinks it.
As this meets the conditions of the introduction to part 5 for winning the rack, they have won, subject to not making an illegal shot and breaching rule 5.14, part (a) for a Loss of Rack foul. | You lose as shot was not called in that manner, i.e. white off opponent's ball first, if called that way is all good. Must call black pocket and if not obvious and easy, how as well, especially if hitting opponent's ball first.
Imagine my black over hole being covered with opponent's ball and I have ball in hand. I must call opponent's ball to pot black for win as unable to hit black directly with white. |
16,211 | In a game where where player 1 is on black and player 2 still has balls remaining on the table. Given the scenarios where player 2 commits a foul giving player 1 a free shot, then on that free shot player 1 first hits one of player 2's remaining balls before sinking the black on the same shot. Does player 1 win in this case? If not, what is the correct flow of play following this?
The WPA rule book states that the "Wrong Ball First" rule is suspended on free shots. However, does this still apply on the black given that you are essentially winning off a foul shot?
Using this version of the rule book for reference: <http://wpapool.co.za/wp-content/uploads/2017/03/WPA_New_Rules.pdf> | 2017/05/29 | [
"https://sports.stackexchange.com/questions/16211",
"https://sports.stackexchange.com",
"https://sports.stackexchange.com/users/13435/"
] | The shooter wins the rack if they sink the black ball, and do not commit a foul under the rules.
As it is a free shot, the shooter cannot be called foul under 6.2 Wrong Ball First.
They may of course shoot the cue ball into the opponent ball, so that it contacts the black ball and sinks it.
As this meets the conditions of the introduction to part 5 for winning the rack, they have won, subject to not making an illegal shot and breaching rule 5.14, part (a) for a Loss of Rack foul. | **Player 1 loses in this game**. Why?:
* **8-ball made on an illegal shot**: As player 1 has called 8-ball and the pocket, he cannot make contact with any other ball first. |
23,063,022 | I am trying to write the following function in Excel.
if P2 is greater than or equal to 3 and AD2 is 0
OR
if P2 is greater than or equal to 2 and AD2 is greater than or equal to 1
OR
if P2 is greater than or equal to 1 and AD2 is 2
Then do the following:
(H2+V2)/(P2+AD2),-999)
I have had a go at writing the following to no avail.
=IF(AND(P2>=3,AD2>=0),OR(AND(P2>=2,AD2>=1)),OR(AND(P2>=1,AD2>=2)),(H2+V2)/(P2+AD2),-999)
Any pointers in the right direction would be much appreciated as I am a novice at Excel functions.
Many thanks,
Ash | 2014/04/14 | [
"https://Stackoverflow.com/questions/23063022",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2879612/"
] | >
> Quicksort gained widespread adoption, appearing, for example, in Unix as the default library sort function, whence it lent its name to the C standard library function qsort and in the reference implementation of Java.
>
>
> Merge sort type algorithms allowed large data sets to be sorted on early computers that had small random access memories by modern standards. Records were stored on magnetic tape and processed on banks of magnetic tape drives . merge sort is implemented with disk drives.
>
>
>
For more read : <http://en.wikipedia.org/wiki/Merge_sort>
<http://en.wikipedia.org/wiki/Quicksort#References> | Their applications vary; it depends on the size and contents of what you're sorting.
Quicksort is one of the fastest sorting algorithms, so it is commonly used in commercial applications.
Merge sort is used mostly under size constraints because it isn't as hefty an algorithm as Quicksort. |
23,063,022 | I am trying to write the following function in Excel.
if P2 is greater than or equal to 3 and AD2 is 0
OR
if P2 is greater than or equal to 2 and AD2 is greater than or equal to 1
OR
if P2 is greater than or equal to 1 and AD2 is 2
Then do the following:
(H2+V2)/(P2+AD2),-999)
I have had a go at writing the following to no avail.
=IF(AND(P2>=3,AD2>=0),OR(AND(P2>=2,AD2>=1)),OR(AND(P2>=1,AD2>=2)),(H2+V2)/(P2+AD2),-999)
Any pointers in the right direction would be much appreciated as I am a novice at Excel functions.
Many thanks,
Ash | 2014/04/14 | [
"https://Stackoverflow.com/questions/23063022",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/2879612/"
] | >
> Quicksort gained widespread adoption, appearing, for example, in Unix as the default library sort function, whence it lent its name to the C standard library function qsort and in the reference implementation of Java.
>
>
> Merge sort type algorithms allowed large data sets to be sorted on early computers that had small random access memories by modern standards. Records were stored on magnetic tape and processed on banks of magnetic tape drives . merge sort is implemented with disk drives.
>
>
>
For more read : <http://en.wikipedia.org/wiki/Merge_sort>
<http://en.wikipedia.org/wiki/Quicksort#References> | Both algorithms can be used in distributed systems for sorting data.
Merge sort is a typical example used in online sorting.
On the other hand quick sort also has some usage in graphic/game development where u need to sort data points and cluster them out. |
5,909 | I just got a used Fender Starcaster S1 to compliment my existing Behringer strat.
Upon inspecting it, I noticed that it has different backs on it's tuning machines.
What are the differences between these besides appearance?
 | 2012/04/09 | [
"https://music.stackexchange.com/questions/5909",
"https://music.stackexchange.com",
"https://music.stackexchange.com/users/2060/"
] | Inside them both you have a cog and a worm gear, which works as per this gif from Wikipedia:

Worm gears have a very useful property - they can cope with high tension without slipping.
The casing differences are purely cosmetic. | There are some easily noticeable differences:
1. Tuners on Behringers are attached with screws on the back side, while the ones on Fenders have a nut on the front.
2. There might be differences in the string attachment itself (not pictured here).
Some differences are not as easily spotted, but are a huge deciding factor:
1. For example, **differences in gear ratio.** Generally, higher-end tuning machines have a **high ratio** for easy, minute adjustments in pitch.
2. In addition, **differences in build quality.** Lower-end tuners are less likely to hold their pitch, and may be more susceptible to damage from use. |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | >
> In the field of Data Mining, is there a specific sub-discipline called 'Similarity'?
>
>
>
Yes. There is a specific subfield in data mining and machine learning called metric learning, which aims to learn a better distance metric among data instances.
Do you know any of the following concepts?
[Euclidean distance](http://en.wikipedia.org/wiki/Euclidean_distance)
[Mahalanobis distance](http://en.wikipedia.org/wiki/Mahalanobis_distance)
[Pearson correlation](http://en.wikipedia.org/wiki/Correlation_and_dependence)
[Cosine similarity](https://stackoverflow.com/questions/2859970/about-cosine-similarity) and [here](https://stackoverflow.com/questions/520241/cosine-similarity-of-vectors)
Kernel functions
After you know these, you will know what is 'similarity'.
>
> I would like the community opinion on how closely related Data Mining and Artificial Intelligence are.
>
>
>
It is very hard to distinguish what is data mining, what is AI. Don't discuss this question when you are new in the field. When you have learned 10 algorithms in data mining and read some AI books, you will know the difference and the relation. | Appropriate definitions of 'similarity' (which features you extract, what you do with them afterwards) are almost the definition of clustering, and clustering is a fairly wide sub-field of data mining.
If you make the standard cynical definition of AI as the set of problems we can't solve well (indeed, that we can't specify well enough to start solving), data mining shades into it once the space in which you're looking for correlations starts to be larger than your algorithms can handle. |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | Appropriate definitions of 'similarity' (which features you extract, what you do with them afterwards) are almost the definition of clustering, and clustering is a fairly wide sub-field of data mining.
If you make the standard cynical definition of AI as the set of problems we can't solve well (indeed, that we can't specify well enough to start solving), data mining shades into it once the space in which you're looking for correlations starts to be larger than your algorithms can handle. | Similarity is a concept that is used in several data mining tasks such as clustering, classification. Dependings on what kind of data you have, you may used different similarity measures such as cosine similarity for text documents, euclidian distance, etc |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | Appropriate definitions of 'similarity' (which features you extract, what you do with them afterwards) are almost the definition of clustering, and clustering is a fairly wide sub-field of data mining.
If you make the standard cynical definition of AI as the set of problems we can't solve well (indeed, that we can't specify well enough to start solving), data mining shades into it once the space in which you're looking for correlations starts to be larger than your algorithms can handle. | There are lots of similarity measurement used in data mining. for text mining, to find similarity in texts, cosine similarity, jaccard similarity widely used
For reference, you can see raghavan and amnnings information retrieval book |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | >
> In the field of Data Mining, is there a specific sub-discipline called 'Similarity'?
>
>
>
Yes. There is a specific subfield in data mining and machine learning called metric learning, which aims to learn a better distance metric among data instances.
Do you know any of the following concepts?
[Euclidean distance](http://en.wikipedia.org/wiki/Euclidean_distance)
[Mahalanobis distance](http://en.wikipedia.org/wiki/Mahalanobis_distance)
[Pearson correlation](http://en.wikipedia.org/wiki/Correlation_and_dependence)
[Cosine similarity](https://stackoverflow.com/questions/2859970/about-cosine-similarity) and [here](https://stackoverflow.com/questions/520241/cosine-similarity-of-vectors)
Kernel functions
After you know these, you will know what is 'similarity'.
>
> I would like the community opinion on how closely related Data Mining and Artificial Intelligence are.
>
>
>
It is very hard to distinguish what is data mining, what is AI. Don't discuss this question when you are new in the field. When you have learned 10 algorithms in data mining and read some AI books, you will know the difference and the relation. | Just to stress the importance of the "similarity" concept.
Data mining (AI, machine learning, modelling etc) is about bringing some function to either it's maximum or minimum value. Take the best optimization/learning/mining algorithm and a wrong function and you get a complete garbage. Note that we use "value" and not "valueS". That's because there is no (to my best knowledge) algorithm (computational or other) that is capable of optimizing more than one value. However, in our Universe, complex optimizations are more frequent than one-dimensional ones (we want to be rich AND young AND healthy). That is why there a plethora of similarity and other scoring functions exists. And that is why none of them is "the right one" |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | >
> In the field of Data Mining, is there a specific sub-discipline called 'Similarity'?
>
>
>
Yes. There is a specific subfield in data mining and machine learning called metric learning, which aims to learn a better distance metric among data instances.
Do you know any of the following concepts?
[Euclidean distance](http://en.wikipedia.org/wiki/Euclidean_distance)
[Mahalanobis distance](http://en.wikipedia.org/wiki/Mahalanobis_distance)
[Pearson correlation](http://en.wikipedia.org/wiki/Correlation_and_dependence)
[Cosine similarity](https://stackoverflow.com/questions/2859970/about-cosine-similarity) and [here](https://stackoverflow.com/questions/520241/cosine-similarity-of-vectors)
Kernel functions
After you know these, you will know what is 'similarity'.
>
> I would like the community opinion on how closely related Data Mining and Artificial Intelligence are.
>
>
>
It is very hard to distinguish what is data mining, what is AI. Don't discuss this question when you are new in the field. When you have learned 10 algorithms in data mining and read some AI books, you will know the difference and the relation. | Similarity is a concept that is used in several data mining tasks such as clustering, classification. Dependings on what kind of data you have, you may used different similarity measures such as cosine similarity for text documents, euclidian distance, etc |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | >
> In the field of Data Mining, is there a specific sub-discipline called 'Similarity'?
>
>
>
Yes. There is a specific subfield in data mining and machine learning called metric learning, which aims to learn a better distance metric among data instances.
Do you know any of the following concepts?
[Euclidean distance](http://en.wikipedia.org/wiki/Euclidean_distance)
[Mahalanobis distance](http://en.wikipedia.org/wiki/Mahalanobis_distance)
[Pearson correlation](http://en.wikipedia.org/wiki/Correlation_and_dependence)
[Cosine similarity](https://stackoverflow.com/questions/2859970/about-cosine-similarity) and [here](https://stackoverflow.com/questions/520241/cosine-similarity-of-vectors)
Kernel functions
After you know these, you will know what is 'similarity'.
>
> I would like the community opinion on how closely related Data Mining and Artificial Intelligence are.
>
>
>
It is very hard to distinguish what is data mining, what is AI. Don't discuss this question when you are new in the field. When you have learned 10 algorithms in data mining and read some AI books, you will know the difference and the relation. | There are lots of similarity measurement used in data mining. for text mining, to find similarity in texts, cosine similarity, jaccard similarity widely used
For reference, you can see raghavan and amnnings information retrieval book |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | Just to stress the importance of the "similarity" concept.
Data mining (AI, machine learning, modelling etc) is about bringing some function to either it's maximum or minimum value. Take the best optimization/learning/mining algorithm and a wrong function and you get a complete garbage. Note that we use "value" and not "valueS". That's because there is no (to my best knowledge) algorithm (computational or other) that is capable of optimizing more than one value. However, in our Universe, complex optimizations are more frequent than one-dimensional ones (we want to be rich AND young AND healthy). That is why there a plethora of similarity and other scoring functions exists. And that is why none of them is "the right one" | Similarity is a concept that is used in several data mining tasks such as clustering, classification. Dependings on what kind of data you have, you may used different similarity measures such as cosine similarity for text documents, euclidian distance, etc |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | Just to stress the importance of the "similarity" concept.
Data mining (AI, machine learning, modelling etc) is about bringing some function to either it's maximum or minimum value. Take the best optimization/learning/mining algorithm and a wrong function and you get a complete garbage. Note that we use "value" and not "valueS". That's because there is no (to my best knowledge) algorithm (computational or other) that is capable of optimizing more than one value. However, in our Universe, complex optimizations are more frequent than one-dimensional ones (we want to be rich AND young AND healthy). That is why there a plethora of similarity and other scoring functions exists. And that is why none of them is "the right one" | There are lots of similarity measurement used in data mining. for text mining, to find similarity in texts, cosine similarity, jaccard similarity widely used
For reference, you can see raghavan and amnnings information retrieval book |
2,887,541 | In the field of Data Mining, is there a specific sub-discipline called 'Similarity'? If yes, what does it deal with. Any examples, links, references will be helpful.
Also, being new to the field, I would like the community opinion on how closely related Data Mining and Artificial Intelligence are. Are they synonyms, is one the subset of the other?
Thanks in advance for sharing your knowledge. | 2010/05/22 | [
"https://Stackoverflow.com/questions/2887541",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/243355/"
] | Similarity is a concept that is used in several data mining tasks such as clustering, classification. Dependings on what kind of data you have, you may used different similarity measures such as cosine similarity for text documents, euclidian distance, etc | There are lots of similarity measurement used in data mining. for text mining, to find similarity in texts, cosine similarity, jaccard similarity widely used
For reference, you can see raghavan and amnnings information retrieval book |
8,502 | What are we supposed to do when a question we ask receives an answer that doesn't help answer the question?
There are multiple ideas on the proper response, which end up summing together to the rule that if someone doesn't understand your question, then you're straight out of luck and aren't allowed to clarify the question in any way.
Is this the intention? Or is there supposed to be some way to do this that actually lets you get your questions answered? | 2021/10/22 | [
"https://worldbuilding.meta.stackexchange.com/questions/8502",
"https://worldbuilding.meta.stackexchange.com",
"https://worldbuilding.meta.stackexchange.com/users/75161/"
] | >
> What are we supposed to do when a question we ask receives an answer that doesn't help answer the question?
>
>
>
A good question! The very first thing you should always try is writing your questions in such a way that they are clear as to what you want and have sufficient background to encourage a knowledgeable person to work out an answer for you!
>
> There are multiple ideas on the proper response, which end up summing together to the rule that if someone doesn't understand your question, then you're straight out of luck and aren't allowed to clarify the question in any way
>
>
>
This is false. You are not only allowed to clarify your question, you are encouraged to do so! You might be referring to your most recent query: that got reverted to its original state because there was already an answer that your clarifying edit would invalidate. That's not good. You'll notice, though, that the respondent altered the answer to suit your edit, so your revision got restored. All's swell that ends well.
>
> Is this the intention? Or is there supposed to be some way to do this that actually lets you get your questions answered?
>
>
>
The ideal intention, really, is to encourage you to write high quality questions in the first place, so we don't have to go through all this nonsense asking for clarification, editing, reverting and rolling back. That gets old.
I can tell you this plain and simple: I love weird anatomy questions; and I get anatomy. I'd ***absolutely love to work on all your weird anatomy questions!*** But as I said in comments elsewhere, until you get your question writing skills on par with your creativity, I'm not going to bother!
The reason for that is simple: I don't want to waste time working on an assumption about what you really mean but didn't or couldn't express only to have you complain that my hard work doesn't even answer the question. Well? Ask the right question then! | There are 2 chief reasons for not getting the answers you want/expect:
1. Your writing is not clear enough.
2. Your target audience is not familiar with the topic and unable to provide an answer.
The first reason is something that you can work on. For example:
* Take a look at thematically similar questions and see how they are asked.
* Read the answers and see what words, reasoning, and assumptions they use (assumptions are rarely stated openly but they can be inferred).
* Try to use vocabulary and delivery formats familiar to your audience so they can focus on your question rather than words and post's form.
* It also might be helpful to ask people directly what causes them to misunderstand your question (please be warned that not everyone will be open to this kind of discussion and some people may even get offended at the suggestion that they misunderstood you or your intentions).
The second reason (people are not familiar with the topic) is something that you cannot control. If this is the case it might be better to ask your question elsewhere. Each community has its area of expertise and the WB.SE is no exception.
There is also another, albeit not so common, reason for misunderstandings: Failure of imagination. Sometimes people are so used to the way things are that they cannot imagine anything else. It is often the case with cultural beliefs and attitudes (e.g. something that is common in one's native culture is perceived as the only possible and natural way to do things). You may attempt to talk to people but it is never certain that you will get any positive outcome. Biases are not easy to overcome. It might be better to look for another community to ask your question. |
350,581 | You may have seen this situation a lot of times:
A guy answers the question in a comment, but comments are easy to become forgotten. Thus another guy takes the original author's solution and posts it as an answer, which is great for a lot of people, but it is originally not his solution.
Wouldn't it be cool, for the second guy to offer the credits to @OriginalPoster?
Example: [Glassfish DeploymentException: Error in linking security policy for](https://stackoverflow.com/questions/7322476/glassfish-deploymentexception-error-in-linking-security-policy-for/18699520#18699520) | 2017/06/12 | [
"https://meta.stackoverflow.com/questions/350581",
"https://meta.stackoverflow.com",
"https://meta.stackoverflow.com/users/2492569/"
] | If they wanted the reputation (or honestly, even if they didn't) they shouldn't have posted an answer in the comments; they should have posted an answer. | Adding a comment isn't answering.
Notice that comment has max length and answer havs minimum length and that because they have different purpose.
When you comment you don't take any risk (mainly of down voting). The person who answer takes risk and should check and append to the solution described in the comment. |
350,581 | You may have seen this situation a lot of times:
A guy answers the question in a comment, but comments are easy to become forgotten. Thus another guy takes the original author's solution and posts it as an answer, which is great for a lot of people, but it is originally not his solution.
Wouldn't it be cool, for the second guy to offer the credits to @OriginalPoster?
Example: [Glassfish DeploymentException: Error in linking security policy for](https://stackoverflow.com/questions/7322476/glassfish-deploymentexception-error-in-linking-security-policy-for/18699520#18699520) | 2017/06/12 | [
"https://meta.stackoverflow.com/questions/350581",
"https://meta.stackoverflow.com",
"https://meta.stackoverflow.com/users/2492569/"
] | If they wanted the reputation (or honestly, even if they didn't) they shouldn't have posted an answer in the comments; they should have posted an answer. | No, "answers" in comments do not need reputation as it is explicit (also probably misguided) choice by author to not create real answer. The very common reason to post answer in comments is to explicitly avoid reputation loss on partial/unrelated answer for unclear questions.
If actual answer is 100% copied from the comment it should be marked community wiki and mention author of the original source. If answer just happen to have the same idea (whether inspired by the comment or independently arrived to the same solution) there is no additional requirements to cite the comment.
See also [Comment Poaching](https://meta.stackoverflow.com/questions/269913/comment-poaching) |
471,594 | I've noticed that in English there are several words which describe light or radiance remaining in the sky after the sun *has set*.
For instance, there is an "afterglow" which, in my opinion, refers to a more effulgent kind of sunlight that is scattered in the sky after sunset. There is "gloaming" which refers to a dim and faint light. There is "dusk" which is a darker stage of twilight. I can also think of crepuscular light, but I think it has a bit [different meaning](https://en.wikipedia.org/wiki/Crepuscular_rays) with a subtle twilight connotation and also looks a tad esoteric.
As you might have already noticed, there are several words which denote sunset colours, but I wasn't able to find anything more precise, which would imply the same glow but with a reference to a sunrise.
If we check a definition of an "afterglow" in the [Oxford Dictionary](https://en.oxforddictionaries.com/definition/afterglow), we can see that it is explicitly stated:
>
> after the sun has set.
>
>
>
In case of other words such as "gloaming", "twilight" or "crepuscular" this is not mentioned. Can a native speaker confirm that I may use these words in the context of a sunrise glow?
The closest I've found so far is [twilight](https://www.merriam-webster.com/dictionary/twilight)
and the word combination "sunrise twilight" can be found in different text corpora. But I am eminently willing to know whether there is a more precise word.
P.S. "dawn", "cockcrow" and "daybreak" denote mostly a time instance, hence I think they are irrelevant here. | 2018/11/05 | [
"https://english.stackexchange.com/questions/471594",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/323014/"
] | Consulting the Oxford English Dictionary reveals that while the first definition for 'gloaming' refers specifically to sunset
>
> a. Evening twilight.
>
>
>
the OED also admits of a second meaning
>
> b. Said occasionally of morning twilight.
>
>
>
from which we can also note that 'twilight' itself is not reserved to sunset, as confirmed by the OED definition:
>
> 1. The light diffused by the reflection of the sun's rays from the atmosphere before sunrise, and after sunset; the period during which this prevails between daylight and darkness.
>
>
>
'Crepuscular' is also related to the ambiguous 'twilight'
>
> 1. Of or pertaining to twilight.
>
>
>
but also specifically to morning
>
> b. esp. Resembling or likened to the morning twilight as preceding the full light of day; characterized by (as yet) imperfect enlightenment.
>
>
>
So certainly there is nothing in any of those definitions to prevent you using them to describe morning light. Indeed, even 'effulgent' which you claim as a quality more pertinent to sunset is defined as:
>
> Shining forth brilliantly; sending forth intense light; resplendent, radiant
>
>
>
Which cannot be more applicable to the going down of the sun than its rising.
However, I think I do understand what you are getting at. The growing dawn light is perceived as perhaps paler and cooler than sunset, and you seem to be looking for a term relating to the sky itself rather than general light levels, as such I would suggest that '**Brightening**' may be appropriate:
>
> Brightening n.
> 1. The action of making or becoming bright; illumination.
>
>
>
While the term is not exclusive to this meaning, it is used and understood in context, as per this example from [Photoreview.com.au](https://www.photoreview.com.au/tips/shooting/ten-tips-for-better-landscape-photography/)
>
> The 'golden hours' for landscape photography occur between pre-dawn brightening in the sky and about an hour-and-a-half after sunrise in the morning
>
>
>
and this from a [random blog](https://jeannelombardo.com/2017/06/desert-dawn-greeting-arizona-summer-days/)
>
> I crack the window. I blink, keeping my eyes closed too long. There's a brightening in the sky. I step on it. I arrive home with the dawn, relieved.
>
>
>
\*all definitions per OED | ***[First light](https://dictionary.cambridge.org/dictionary/english/first-light)*** is used to refer to the luminescence of dawn:
>
> the time when light is first seen in the morning : dawn
>
>
> * She was up at first light.
>
>
>
(M-W) |
471,594 | I've noticed that in English there are several words which describe light or radiance remaining in the sky after the sun *has set*.
For instance, there is an "afterglow" which, in my opinion, refers to a more effulgent kind of sunlight that is scattered in the sky after sunset. There is "gloaming" which refers to a dim and faint light. There is "dusk" which is a darker stage of twilight. I can also think of crepuscular light, but I think it has a bit [different meaning](https://en.wikipedia.org/wiki/Crepuscular_rays) with a subtle twilight connotation and also looks a tad esoteric.
As you might have already noticed, there are several words which denote sunset colours, but I wasn't able to find anything more precise, which would imply the same glow but with a reference to a sunrise.
If we check a definition of an "afterglow" in the [Oxford Dictionary](https://en.oxforddictionaries.com/definition/afterglow), we can see that it is explicitly stated:
>
> after the sun has set.
>
>
>
In case of other words such as "gloaming", "twilight" or "crepuscular" this is not mentioned. Can a native speaker confirm that I may use these words in the context of a sunrise glow?
The closest I've found so far is [twilight](https://www.merriam-webster.com/dictionary/twilight)
and the word combination "sunrise twilight" can be found in different text corpora. But I am eminently willing to know whether there is a more precise word.
P.S. "dawn", "cockcrow" and "daybreak" denote mostly a time instance, hence I think they are irrelevant here. | 2018/11/05 | [
"https://english.stackexchange.com/questions/471594",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/323014/"
] | Merriam-Webster's (the preferred American reference dictionary) suggests "alpenglow," which arrived in English via German at some point in the late-ish 19th century. Though I think the arguments made in Answer 1 are thorough and well researched, they don't quite convince me. Words such as crepuscular connote, if not dennote, post-sunset to most readers. Using them may have the support of the OED (a most peculiar dictionary, as it is based entirely on historical usage), but will be confusing to most readers. The phrase "pre-dawn light" is dull and not evoctive, but it has the benefit of being unambiguous. | ***[First light](https://dictionary.cambridge.org/dictionary/english/first-light)*** is used to refer to the luminescence of dawn:
>
> the time when light is first seen in the morning : dawn
>
>
> * She was up at first light.
>
>
>
(M-W) |
471,594 | I've noticed that in English there are several words which describe light or radiance remaining in the sky after the sun *has set*.
For instance, there is an "afterglow" which, in my opinion, refers to a more effulgent kind of sunlight that is scattered in the sky after sunset. There is "gloaming" which refers to a dim and faint light. There is "dusk" which is a darker stage of twilight. I can also think of crepuscular light, but I think it has a bit [different meaning](https://en.wikipedia.org/wiki/Crepuscular_rays) with a subtle twilight connotation and also looks a tad esoteric.
As you might have already noticed, there are several words which denote sunset colours, but I wasn't able to find anything more precise, which would imply the same glow but with a reference to a sunrise.
If we check a definition of an "afterglow" in the [Oxford Dictionary](https://en.oxforddictionaries.com/definition/afterglow), we can see that it is explicitly stated:
>
> after the sun has set.
>
>
>
In case of other words such as "gloaming", "twilight" or "crepuscular" this is not mentioned. Can a native speaker confirm that I may use these words in the context of a sunrise glow?
The closest I've found so far is [twilight](https://www.merriam-webster.com/dictionary/twilight)
and the word combination "sunrise twilight" can be found in different text corpora. But I am eminently willing to know whether there is a more precise word.
P.S. "dawn", "cockcrow" and "daybreak" denote mostly a time instance, hence I think they are irrelevant here. | 2018/11/05 | [
"https://english.stackexchange.com/questions/471594",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/323014/"
] | Consulting the Oxford English Dictionary reveals that while the first definition for 'gloaming' refers specifically to sunset
>
> a. Evening twilight.
>
>
>
the OED also admits of a second meaning
>
> b. Said occasionally of morning twilight.
>
>
>
from which we can also note that 'twilight' itself is not reserved to sunset, as confirmed by the OED definition:
>
> 1. The light diffused by the reflection of the sun's rays from the atmosphere before sunrise, and after sunset; the period during which this prevails between daylight and darkness.
>
>
>
'Crepuscular' is also related to the ambiguous 'twilight'
>
> 1. Of or pertaining to twilight.
>
>
>
but also specifically to morning
>
> b. esp. Resembling or likened to the morning twilight as preceding the full light of day; characterized by (as yet) imperfect enlightenment.
>
>
>
So certainly there is nothing in any of those definitions to prevent you using them to describe morning light. Indeed, even 'effulgent' which you claim as a quality more pertinent to sunset is defined as:
>
> Shining forth brilliantly; sending forth intense light; resplendent, radiant
>
>
>
Which cannot be more applicable to the going down of the sun than its rising.
However, I think I do understand what you are getting at. The growing dawn light is perceived as perhaps paler and cooler than sunset, and you seem to be looking for a term relating to the sky itself rather than general light levels, as such I would suggest that '**Brightening**' may be appropriate:
>
> Brightening n.
> 1. The action of making or becoming bright; illumination.
>
>
>
While the term is not exclusive to this meaning, it is used and understood in context, as per this example from [Photoreview.com.au](https://www.photoreview.com.au/tips/shooting/ten-tips-for-better-landscape-photography/)
>
> The 'golden hours' for landscape photography occur between pre-dawn brightening in the sky and about an hour-and-a-half after sunrise in the morning
>
>
>
and this from a [random blog](https://jeannelombardo.com/2017/06/desert-dawn-greeting-arizona-summer-days/)
>
> I crack the window. I blink, keeping my eyes closed too long. There's a brightening in the sky. I step on it. I arrive home with the dawn, relieved.
>
>
>
\*all definitions per OED | Merriam-Webster's (the preferred American reference dictionary) suggests "alpenglow," which arrived in English via German at some point in the late-ish 19th century. Though I think the arguments made in Answer 1 are thorough and well researched, they don't quite convince me. Words such as crepuscular connote, if not dennote, post-sunset to most readers. Using them may have the support of the OED (a most peculiar dictionary, as it is based entirely on historical usage), but will be confusing to most readers. The phrase "pre-dawn light" is dull and not evoctive, but it has the benefit of being unambiguous. |
63,047,524 | I'm developing a simple app for Android TV using Flutter.
I would like to add authenticated users.
Android TV would display something like:
To login go to [www.domain.com/activate](http://www.domain.com/activate) on any device and enter this code: ABC123XYZ
Once logged in on a phone or computer and the code is entered Android TV app is authenticated and ready for future use.
Just wondering how is this implemented, any help is appreciated. | 2020/07/23 | [
"https://Stackoverflow.com/questions/63047524",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1068975/"
] | Multiple-way you can perform this
1. Websocket
2. API services
First let's check for API Web Services, create a page for Android TV users and send the code they have entered from remote to the server let the user do the computation and get back that user is authenticated.
Although I personally do not like this feature as typing on tv is rather a bad user experience with the remote keypad.
Let's go with the WebSocket, create a QR code with UUID information on tv, connect tv app with WebSocket to receive authentication updates, then scan QR code via a mobile app the code that you have generated on tv is UUID of that tv and you got unique token after scanning, now send this UUID to the server and server will fire socket message to the TV that you have been authenticated.
also, it would be best if you go to your backend developer regarding this query is regarding flow and that can be determined by any backend engineer. | I would suggest implementing
<https://oauth.net/2/grant-types/device-code/>
Your app would periodically poll the server asking if authenticated
once the user has gone to said website and put in said code, the next poll you would return a token etc. for the app to use and then let your app do its thing. |
793,570 | My daughter has a new Apple laptop. Her old laptop is an older Dell Inspiron. It runs Windows XP. The Dell has pictures and videos of her son (my grandson), we want to transfer to an external hard drive. My question is, can the data that is transferred to the external hard drive from Windows XP be transferred to her new Apple laptop which runs Mac OS/X? | 2014/08/07 | [
"https://superuser.com/questions/793570",
"https://superuser.com",
"https://superuser.com/users/354273/"
] | Well the answer turned out to be rather simple for me. I live in an apartment complex, and at this time I share a public IP with all who live in the complex. I have to request my own public IP to get this working. However I hope others will benefit from the other comments and answers given!
EDIT: I suppose a solution for this is posted [here](https://superuser.com/questions/85285/ssh-to-computer-with-shared-public-ip-and-no-admin-rights-over-the-router?rq=1) | Verify if your ISP is blocking the port 22 (it's easy to search).
Also, be sure that your port forwarding is correct:
* IP source must be empty
* Source port is 22
* IP Destination must point to your computer local ip (which is given by your router and looks like 192.168.X.X [warning: 192.168.X.1 is your router])
* Destination port is 22
If your ISP does block port 22, a little trick that could work is setting the source port to 443 (SSL) [keep the destination one to 22], which is more often open that SSH. Be sure if you try to connect to your server from outside that you specified the port 443. |
561,995 | Since the internet connection in our house does break down from time to time I set up a little experiment:
For the last two month, one of my machines is pinging google.com on an half-hourly basis. One measurement consists of 50 pings.
I now calculated the mean percentage of packets lost for each hour of the day:

My questions:
1. Could this peak in the evening be caused by choosing google.com as the ping destination?
2. Would you recommend using an other destination and which?
3. Does this indicate that something is wrong with my connection?
4. What would be a better strategy to measure where exactly the problem in our internet connection is? Our ISP tells us it is working fine so I try to aggregate some proof...
Regards!
Edit: I forgot to mention that the machine is directly connected to the router (no WiFi). And the router is pinged as well, with no packet loss at all. | 2013/03/07 | [
"https://superuser.com/questions/561995",
"https://superuser.com",
"https://superuser.com/users/151077/"
] | Unfortunately you really have not provided enough information to work out where the problem is. To answer as best as I possibly can with the limited information provided:
1. If my experiences are anything to go by, pinging Google is normally quite a good bet, as they design their network to be as fast as possible. Also as ICMP is prioritized, evening peak probably does not make a significant difference - particularly in terms of packet loss - which I'd argue should be 0.
2. Google is a good destination, but to get a better idea of what is going on
you may want to additionally try pinging your gateway, and if they allow it,
your providers DNS, mail or web server. This will help show where the packet
loss is creeping in. Realistically though, at the level of packet loss
you are seeing, look at downloading MTR (or WinMTR) and running that at peek
to get a better idea of where packet loss is coming in.
3. Subjectively, 5% packet loss is at the top end of acceptable for a Wifi
based network - assuming you are not saturating your network. On the flip
side, I get upset about 0.5% packet loss on my fibre connections - as a
point of reference, loosely speaking, less then 1% is OK for VOIP, above
that not so much. If you expect to be able to use Skype or Viber or
what have you on your connection then 5% packet loss is not OK. For
just Web browsing it may just about suffice.
4. As an ISP, I want to see the results of an MTR, which show the latencies
and packet loss between the destination - this helps me look where the
bottleneck might be and is a good first step. I would also like to know
when the test was done so I can correlate it with the customers other
usage and what is going on on the system. The packet loss graphs you
have done are useful as well, but not in isolation.
As a client, my ISP has been unable to excuse away my graphs, which plot
packet loss (I do it for 250 pings, once a second over 5 minute intervals,
combined with minimimum, average and maximum latencies for those pings). I
also have a set of graphs showing my utilization of the link, and have
sets of graphs showing local (ie very close to me), and to another POP
they own of specific interest a few hundred KM's away.
Other observations:
It looks like your latency increases in the afternoons - which means the first places I would be looking are if the problem is WIFI when everyone around me is using it. After ruling that out I'd start questioning my ISP about oversubscribing the connection. | This more than likely is a result of congestion somewhere along the line. It could be your router but more likely an upstream provider.
You don't state how you're doing the 50 pings e.g. what time interval, are you waiting for one to fail/succeed before the next or firing 50 off all at once (flood pinging).
Such a loss during periods of high congestion is not unusual in my experience. It can be down to lower prioritisation for ICMP traffic but more likely is happening for the same percentage of all connections - it's just TCP will gracefully resend and reorder packets so you are less likely to notice.
To gain a better picture of the situation I would suggest you implement the following:
1. Increase the interval between your pings
2. Ping an IP address for google not the domain - google.com will return a number of A records and it may be you're using different end IPs (and hence routing differently) without knowing it
3. Record the mean time to respond; see if this correlates with loss - if it does so you see higher ping roundtrip times and higher loss then it indicates congestion. You could then investigate by storing traceroute logs instead and see if there is a likely bottleneck somewhere where you're seeing suddenly increased times
4. Try pinging more than google. When I've benchmarked network performance in the past I've done it using 4 or 5 good end-points (again with the IP address not the hostname) so you can rule out congestion or a specific problem within just google's network causing you to question your entire connection |
561,995 | Since the internet connection in our house does break down from time to time I set up a little experiment:
For the last two month, one of my machines is pinging google.com on an half-hourly basis. One measurement consists of 50 pings.
I now calculated the mean percentage of packets lost for each hour of the day:

My questions:
1. Could this peak in the evening be caused by choosing google.com as the ping destination?
2. Would you recommend using an other destination and which?
3. Does this indicate that something is wrong with my connection?
4. What would be a better strategy to measure where exactly the problem in our internet connection is? Our ISP tells us it is working fine so I try to aggregate some proof...
Regards!
Edit: I forgot to mention that the machine is directly connected to the router (no WiFi). And the router is pinged as well, with no packet loss at all. | 2013/03/07 | [
"https://superuser.com/questions/561995",
"https://superuser.com",
"https://superuser.com/users/151077/"
] | This more than likely is a result of congestion somewhere along the line. It could be your router but more likely an upstream provider.
You don't state how you're doing the 50 pings e.g. what time interval, are you waiting for one to fail/succeed before the next or firing 50 off all at once (flood pinging).
Such a loss during periods of high congestion is not unusual in my experience. It can be down to lower prioritisation for ICMP traffic but more likely is happening for the same percentage of all connections - it's just TCP will gracefully resend and reorder packets so you are less likely to notice.
To gain a better picture of the situation I would suggest you implement the following:
1. Increase the interval between your pings
2. Ping an IP address for google not the domain - google.com will return a number of A records and it may be you're using different end IPs (and hence routing differently) without knowing it
3. Record the mean time to respond; see if this correlates with loss - if it does so you see higher ping roundtrip times and higher loss then it indicates congestion. You could then investigate by storing traceroute logs instead and see if there is a likely bottleneck somewhere where you're seeing suddenly increased times
4. Try pinging more than google. When I've benchmarked network performance in the past I've done it using 4 or 5 good end-points (again with the IP address not the hostname) so you can rule out congestion or a specific problem within just google's network causing you to question your entire connection | This is typical of most residential ISP accounts. You're seeing a peak due to network congestion as people get home after work, then go on-line for the whole evening. This sort of evening peak is especially pronounced in high-tech communities with lots of on-line gamers (like where I live, here in Redmond, home of Microsoft.) |
561,995 | Since the internet connection in our house does break down from time to time I set up a little experiment:
For the last two month, one of my machines is pinging google.com on an half-hourly basis. One measurement consists of 50 pings.
I now calculated the mean percentage of packets lost for each hour of the day:

My questions:
1. Could this peak in the evening be caused by choosing google.com as the ping destination?
2. Would you recommend using an other destination and which?
3. Does this indicate that something is wrong with my connection?
4. What would be a better strategy to measure where exactly the problem in our internet connection is? Our ISP tells us it is working fine so I try to aggregate some proof...
Regards!
Edit: I forgot to mention that the machine is directly connected to the router (no WiFi). And the router is pinged as well, with no packet loss at all. | 2013/03/07 | [
"https://superuser.com/questions/561995",
"https://superuser.com",
"https://superuser.com/users/151077/"
] | Unfortunately you really have not provided enough information to work out where the problem is. To answer as best as I possibly can with the limited information provided:
1. If my experiences are anything to go by, pinging Google is normally quite a good bet, as they design their network to be as fast as possible. Also as ICMP is prioritized, evening peak probably does not make a significant difference - particularly in terms of packet loss - which I'd argue should be 0.
2. Google is a good destination, but to get a better idea of what is going on
you may want to additionally try pinging your gateway, and if they allow it,
your providers DNS, mail or web server. This will help show where the packet
loss is creeping in. Realistically though, at the level of packet loss
you are seeing, look at downloading MTR (or WinMTR) and running that at peek
to get a better idea of where packet loss is coming in.
3. Subjectively, 5% packet loss is at the top end of acceptable for a Wifi
based network - assuming you are not saturating your network. On the flip
side, I get upset about 0.5% packet loss on my fibre connections - as a
point of reference, loosely speaking, less then 1% is OK for VOIP, above
that not so much. If you expect to be able to use Skype or Viber or
what have you on your connection then 5% packet loss is not OK. For
just Web browsing it may just about suffice.
4. As an ISP, I want to see the results of an MTR, which show the latencies
and packet loss between the destination - this helps me look where the
bottleneck might be and is a good first step. I would also like to know
when the test was done so I can correlate it with the customers other
usage and what is going on on the system. The packet loss graphs you
have done are useful as well, but not in isolation.
As a client, my ISP has been unable to excuse away my graphs, which plot
packet loss (I do it for 250 pings, once a second over 5 minute intervals,
combined with minimimum, average and maximum latencies for those pings). I
also have a set of graphs showing my utilization of the link, and have
sets of graphs showing local (ie very close to me), and to another POP
they own of specific interest a few hundred KM's away.
Other observations:
It looks like your latency increases in the afternoons - which means the first places I would be looking are if the problem is WIFI when everyone around me is using it. After ruling that out I'd start questioning my ISP about oversubscribing the connection. | This is typical of most residential ISP accounts. You're seeing a peak due to network congestion as people get home after work, then go on-line for the whole evening. This sort of evening peak is especially pronounced in high-tech communities with lots of on-line gamers (like where I live, here in Redmond, home of Microsoft.) |
8,331,175 | Does such program exist?
I have to study Java SE and diagram with all classes and interfaces from given package will be immensely helpful.
For example I want to plot all relations between subclasses of types Collection and Map.
I know there are a lot of images with core package structure already, but don't really trust them because they are proven to be incomplete, outdated or even inaccurate. | 2011/11/30 | [
"https://Stackoverflow.com/questions/8331175",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/362721/"
] | Whilst I don't know of any tools for generating UML class diagrams from JavaDocs, there are many tools available that can generate UML class diagrams from source code, and there are already many [questions on StackOverflow](https://stackoverflow.com/search?q=generate%20uml%20class%20diagram%20from%20java%20code) that suggest various tools that can do this
[This question](https://stackoverflow.com/questions/1687323/from-java-code-to-uml-diagram) seems to have a good range of options that you could explore. I've used [ArgoUML](http://argouml.tigris.org/) myself, but I suggest looking at the options to find the one what appeals to you most.
You can download the [Java SE source code](http://download.java.net/jdk6/source/) and generate the class diagrams for the sections you're interested in. | You can create RCP application using [ZEST](http://eclipse.org/gef/zest). This is pretty cool |
8,331,175 | Does such program exist?
I have to study Java SE and diagram with all classes and interfaces from given package will be immensely helpful.
For example I want to plot all relations between subclasses of types Collection and Map.
I know there are a lot of images with core package structure already, but don't really trust them because they are proven to be incomplete, outdated or even inaccurate. | 2011/11/30 | [
"https://Stackoverflow.com/questions/8331175",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/362721/"
] | You can create RCP application using [ZEST](http://eclipse.org/gef/zest). This is pretty cool | You can not create diagrams from JavaDoc because there is no official implementation. Only Java code could be reversed and displayed as class or sequence diagrams.
I played to reverse the full Java language with EclipseUML Omondo. It was really interesting to get all dependencies, inheritances, associations at package but also at project level, or even at multi-projects level.
The trick Omondo use is to convert manually each Java Id to a single UML Id. You get after a 30 mn process a huge model composed by hundred of thousands UML Ids saved in an XMI format respecting java project structure. I mean you get project > packages > class> iner classes > attrbites etc...
From this model then you create views by drag and drop or directly selecting the entire package. You can show hide relations, navigate in the code at graphical level etc.....
Really amazing technology !! |
8,331,175 | Does such program exist?
I have to study Java SE and diagram with all classes and interfaces from given package will be immensely helpful.
For example I want to plot all relations between subclasses of types Collection and Map.
I know there are a lot of images with core package structure already, but don't really trust them because they are proven to be incomplete, outdated or even inaccurate. | 2011/11/30 | [
"https://Stackoverflow.com/questions/8331175",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/362721/"
] | Whilst I don't know of any tools for generating UML class diagrams from JavaDocs, there are many tools available that can generate UML class diagrams from source code, and there are already many [questions on StackOverflow](https://stackoverflow.com/search?q=generate%20uml%20class%20diagram%20from%20java%20code) that suggest various tools that can do this
[This question](https://stackoverflow.com/questions/1687323/from-java-code-to-uml-diagram) seems to have a good range of options that you could explore. I've used [ArgoUML](http://argouml.tigris.org/) myself, but I suggest looking at the options to find the one what appeals to you most.
You can download the [Java SE source code](http://download.java.net/jdk6/source/) and generate the class diagrams for the sections you're interested in. | You can not create diagrams from JavaDoc because there is no official implementation. Only Java code could be reversed and displayed as class or sequence diagrams.
I played to reverse the full Java language with EclipseUML Omondo. It was really interesting to get all dependencies, inheritances, associations at package but also at project level, or even at multi-projects level.
The trick Omondo use is to convert manually each Java Id to a single UML Id. You get after a 30 mn process a huge model composed by hundred of thousands UML Ids saved in an XMI format respecting java project structure. I mean you get project > packages > class> iner classes > attrbites etc...
From this model then you create views by drag and drop or directly selecting the entire package. You can show hide relations, navigate in the code at graphical level etc.....
Really amazing technology !! |
86,001 | I've gotten a Raspberry Pi 2B as a gift, but without a case. I'm trying to buy me a case, but I mostly find Pi 3 cases being offered.
So, I'm asking the opposite of this question: [Is the Raspberry Pi 2 Model B or B+ case compatible with the Raspberry Pi 3?](https://raspberrypi.stackexchange.com/q/44446/88531)
I understand that there might be some issue with the LEDs, so I should prefer a transparent case. Can I be sure there are no other incompatibilities? | 2018/07/13 | [
"https://raspberrypi.stackexchange.com/questions/86001",
"https://raspberrypi.stackexchange.com",
"https://raspberrypi.stackexchange.com/users/88531/"
] | The only significant physical difference in the positioning of the ACT/PWR LEDs.
I have cases with LED cutouts on both sides. | Yes, screw holes & dimensions are the same. I know this from experience. |
21,323,452 | I'm using gae for a project and I have a cron script to keep it alive so that the handful of users I have so far wont have to wait +5 sec on their first query.
Does anyone know if Google will enforce some limit on my app if the logs reflect that there are significantly more hits from the cron script than actuall users?
Thanks | 2014/01/24 | [
"https://Stackoverflow.com/questions/21323452",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1764080/"
] | To address the wait time that you are facing, App Engine has a feature called Warmup requests. Warmup requests are a specific type of loading request and their task is to load the application initialisation code into an instance in advance before the standard requests from your users hit the application.
Please look at [Warmup Requests](https://developers.google.com/appengine/docs/java/config/appconfig?csw=1#Java_appengine_web_xml_Warmup_requests) documentation and a [tutorial](http://googcloudlabs.appspot.com/codelabexercise9.html). | so far google has no such criteria but better option is enable billing with automatic scaling
it really helped me to pace my site's speed.
if you are really worried on pace increase min idle instance and decrease pending latencies you can find them on your dashboard if your app has billing enabled. |
21,323,452 | I'm using gae for a project and I have a cron script to keep it alive so that the handful of users I have so far wont have to wait +5 sec on their first query.
Does anyone know if Google will enforce some limit on my app if the logs reflect that there are significantly more hits from the cron script than actuall users?
Thanks | 2014/01/24 | [
"https://Stackoverflow.com/questions/21323452",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/1764080/"
] | I don't know if Google would enforce a limit or not.
Instead of using a cron job, you can [set the min idle instances to 1(or more)](https://developers.google.com/appengine/docs/adminconsole/performancesettings#Setting_the_Number_of_Idle_Instances) if you're worried about response time.
Enable billing to do this. | so far google has no such criteria but better option is enable billing with automatic scaling
it really helped me to pace my site's speed.
if you are really worried on pace increase min idle instance and decrease pending latencies you can find them on your dashboard if your app has billing enabled. |
239,116 | How do I create a second version a node add page?
For example if have node/add and I want node/add-quick?
Some fields of the required fields form node/add would be missing or not required on node/add-quick.
I know node add fields can be hidden based on user role or other variables but how do I create an entirely new page with a select set of fields. | 2017/06/22 | [
"https://drupal.stackexchange.com/questions/239116",
"https://drupal.stackexchange.com",
"https://drupal.stackexchange.com/users/46625/"
] | This is a classic case of trying to do something in the non-Drupal way -- conceptually, defining a custom route for your home page route makes sense, and it's possible in other frameworks, but Drupal has a different way of doing things that makes this approach completely unintuitive.
Here's how to do it:
1. Define your route as something other than '/' (i.e. '/students/overview')
2. In /admin/config/system/site-information, set your "Default Front Page" to whatever your route was:
[](https://i.stack.imgur.com/lwPS5.png)
Voila! | Also looking at your code you presented above, you have a typo.
'acess content' should be 'access content' |
11,324 | Here is the situation that has come up often in our group:
1. A creature is immobilized(save ends) at the start of its turn. As no enemies are adjacent, its options are limited.
2. The creature readies a charge
3. The trigger chosen is "When I am not immobilized"
* Note that Ready requires an action to trigger off of. For a while, we ruled that the trigger had to be something like, "When Bob tells me I'm not Immobilized." Talking is a free action, so that satisfied our inner rules lawyer. It got silly fast, and we went with this shorthand.
4. The creature processes its end of turn actions, saving half the time\* and immediately charges an opponent.
Both characters and monsters have been doing this awhile. However it doesn't sit well with me, and some others in the group as it does nerf the Immobilized condition. We've agreed to play this way as we think it is how we interpret the rules.
Is this legal? Are we reading the rules correctly? Are we missing a rules clarification somewhere?
My group and I are very much "Rules as Written" players, so we aren't very interested in house rules to change it. Thanks for the suggestions though.
\* ok, not exactly half, but that isn't the point here, nitpickers :) | 2011/12/14 | [
"https://rpg.stackexchange.com/questions/11324",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/460/"
] | So my question would be how is the immobilizing being put onto the target? Is it something that is "until the end of the target's next turn" or "save ends". I personally don't count your turn over with until all of your actions have preformed. Since a readied action moves your place in the initiative order, I treat it just like a delay.
The rules for [delay](http://www.wizards.com/dndinsider/compendium/glossary.aspx?id=327) specify that harmful effects end when you ACT, while beneficial effects end when you DELAY. I think that because readied actions and delayed turns are very similar, you should apply that logic to the player's readied actions.
If the immobilizing effect was put onto the target from an power that said, "until the end of the caster's next turn", then I think you would have an argument for allowing what is happening. That is because the end of the effect isn't based upon the person who is readying the action, but instead is based upon who used the power in the first place.
In the end, I tend to just follow the rule of "is it too good to be true". These players were effectively making immobilizing useless in most cases, as the player was able to still move and attack and basically ignore the condition. | There are two issues:
1. Unlike Delay, it doesn't explicitly say when (if) you get your End of Turn saving throw, but it makes sense to follow Delay's example and have it be *Make Saving Throws after You Act*:
>
> After you return to the initiative order and take your actions, you make saving throws against effects on you.
>
>
>
2. The action triggered by Ready an Action is an immediate action, which means you can't take it on your turn. |
11,324 | Here is the situation that has come up often in our group:
1. A creature is immobilized(save ends) at the start of its turn. As no enemies are adjacent, its options are limited.
2. The creature readies a charge
3. The trigger chosen is "When I am not immobilized"
* Note that Ready requires an action to trigger off of. For a while, we ruled that the trigger had to be something like, "When Bob tells me I'm not Immobilized." Talking is a free action, so that satisfied our inner rules lawyer. It got silly fast, and we went with this shorthand.
4. The creature processes its end of turn actions, saving half the time\* and immediately charges an opponent.
Both characters and monsters have been doing this awhile. However it doesn't sit well with me, and some others in the group as it does nerf the Immobilized condition. We've agreed to play this way as we think it is how we interpret the rules.
Is this legal? Are we reading the rules correctly? Are we missing a rules clarification somewhere?
My group and I are very much "Rules as Written" players, so we aren't very interested in house rules to change it. Thanks for the suggestions though.
\* ok, not exactly half, but that isn't the point here, nitpickers :) | 2011/12/14 | [
"https://rpg.stackexchange.com/questions/11324",
"https://rpg.stackexchange.com",
"https://rpg.stackexchange.com/users/460/"
] | **Your reading is correct**
Rules as written, there is nothing preventing your groups actions.
As others have noted, it is a small stretch of a house rule to give ready and delay the same treatment with respect to AEOT effects, but though you could consider it "preemptive errata" with the thinking that WotC just has not seen fit to disco yet, it would be a house rule.
I would also note that it doesn't completely nerf *immobilized*, as you still can't use your move action to move, and you may not have your full selection of attack powers available to you on the charge. | There are two issues:
1. Unlike Delay, it doesn't explicitly say when (if) you get your End of Turn saving throw, but it makes sense to follow Delay's example and have it be *Make Saving Throws after You Act*:
>
> After you return to the initiative order and take your actions, you make saving throws against effects on you.
>
>
>
2. The action triggered by Ready an Action is an immediate action, which means you can't take it on your turn. |
41,695,601 | I've been thinking about the concept of controller in the MVC pattern, but there's something i would like to discuss, should we create a controller per entity (let's say Product) or per view when using angularJS (since angularJS follows the singleton concept) or is it right to have a controller for all the views related to "products" for example?
I'll give a situation:
Let's say we have a module called purchases and under that module we've got three submodules: products, suppliers, & stock-moves, each submodule has views for the CRUD operations, let's say that under products we have: list-products.view.html, create-product.view.html, edit-product.view.html; what would be the best approach for the controllers? a single controller per view, or a controller for all views?
Thanks for your help | 2017/01/17 | [
"https://Stackoverflow.com/questions/41695601",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/7224642/"
] | I always go for a controller per view, regardless of whether that results in having multiple controllers per entity or not. The reasoning for this is simple. Whenever a view is malfunctioning I want to look at the smallest subset of code to figure out what went wrong.
Also, if you think about it the controller's job is to respond to commands from view and communicate model changes to views. So, in essence your controller is not supposed to manage entities but it's used to manage your view.
This approach has served me in 98% of the cases without any issues. Obviously, in software there are no absolute answers so there might be some scenarios where a per entity controller might fit in more seamlessly. But for me, from a practical as well as a philosophical perspective a controller per view makes more sense. | I usually use one controller and route for each view, this is the most used approach. It helps to logically structure the application especially when it begins to grow.
There is another case when I use multiple controllers per view to govern different areas of the view. In such a scenario each controller manages a fragment of the overall page. |
1,191 | I read on [Wikipedia](http://en.wikipedia.org/wiki/Milky_Way#Appearance):
>
> depending on the time of night and the year, the arc of Milky Way can
> appear relatively low or relatively high in the sky. For observers
> from about 65 degrees north to 65 degrees south on the Earth's surface
> the Milky Way passes directly overhead twice a day
>
>
>
What happens and how does the Milky Way look like out of that range?
If it matters, consider also the all-summer light and the dark winter. | 2013/12/17 | [
"https://astronomy.stackexchange.com/questions/1191",
"https://astronomy.stackexchange.com",
"https://astronomy.stackexchange.com/users/623/"
] | On midnight, right around this time of year, the Milky Way will be in the zenith. Here is an XEphem rendering for the north of finland (65th latitude) for yesterday midnight (the brown outline marks the Milky Way):

You can also see this on photographs by finnish photographer [Tommy Eliassen](http://www.tommyeliassen.com/#!/index/15). He has many examples of this on his website. I won't put any in here, because I guess they are copyrighted, and not free to use.
This is an animated 24 hour version of the above sky map, which shows the rotation of the milkyway around the zenith:
 | Outside of this range you are close to the poles of our planet. The milky way will be low on the horizon and hard to see. |
29,242 | I'm running Visual Studio Team System Development Edition 2008 on Vista Business 64bit along with Resharper 4.5, Telerik Reporting 2009/Q2 and GhostDoc 2.5.9166.0 and it keeps crashing randomly. It typically happens when I start entering text into a .cs or text file. The event log gets an application error entry:
>
> Faulting application devenv.exe,
> version 9.0.30729.1, time stamp
> 0x488f2b50, faulting module ntdll.dll,
> version 6.0.6002.18005, time stamp
> 0x49e03824, exception code 0xc0000374,
> fault offset 0x000ab0bf, process id
> 0xb0, application start time
> 0x01ca255374e4ec56.
>
>
>
Any thoughts? | 2009/08/25 | [
"https://superuser.com/questions/29242",
"https://superuser.com",
"https://superuser.com/users/-1/"
] | Try removing the extensisons (ReSharper, Telerik & GhostDoc) one by one and see if the problem goes away - it might be in one of those or due to an interaction between them. If so reinstall and see if the problem goes away.
If that doesn't work repair/reinstall Visual Studo itself. | I'd guess it has to do with the extensions you have installed. To disable them temporarily, start VS and click on Tools -> Add-in manager. Deselect all "Start" entries and restart Visual Studio. |
29,242 | I'm running Visual Studio Team System Development Edition 2008 on Vista Business 64bit along with Resharper 4.5, Telerik Reporting 2009/Q2 and GhostDoc 2.5.9166.0 and it keeps crashing randomly. It typically happens when I start entering text into a .cs or text file. The event log gets an application error entry:
>
> Faulting application devenv.exe,
> version 9.0.30729.1, time stamp
> 0x488f2b50, faulting module ntdll.dll,
> version 6.0.6002.18005, time stamp
> 0x49e03824, exception code 0xc0000374,
> fault offset 0x000ab0bf, process id
> 0xb0, application start time
> 0x01ca255374e4ec56.
>
>
>
Any thoughts? | 2009/08/25 | [
"https://superuser.com/questions/29242",
"https://superuser.com",
"https://superuser.com/users/-1/"
] | Try removing the extensisons (ReSharper, Telerik & GhostDoc) one by one and see if the problem goes away - it might be in one of those or due to an interaction between them. If so reinstall and see if the problem goes away.
If that doesn't work repair/reinstall Visual Studo itself. | Ensure you are running the latest version of ReSharper. There was a known issue with one of the beta release that would cause a random crash when trying to edit anything related to HTML or CSS or when using the `<` or `>` keys. I would suggest disabling ReSharper first. |
29,242 | I'm running Visual Studio Team System Development Edition 2008 on Vista Business 64bit along with Resharper 4.5, Telerik Reporting 2009/Q2 and GhostDoc 2.5.9166.0 and it keeps crashing randomly. It typically happens when I start entering text into a .cs or text file. The event log gets an application error entry:
>
> Faulting application devenv.exe,
> version 9.0.30729.1, time stamp
> 0x488f2b50, faulting module ntdll.dll,
> version 6.0.6002.18005, time stamp
> 0x49e03824, exception code 0xc0000374,
> fault offset 0x000ab0bf, process id
> 0xb0, application start time
> 0x01ca255374e4ec56.
>
>
>
Any thoughts? | 2009/08/25 | [
"https://superuser.com/questions/29242",
"https://superuser.com",
"https://superuser.com/users/-1/"
] | Try removing the extensisons (ReSharper, Telerik & GhostDoc) one by one and see if the problem goes away - it might be in one of those or due to an interaction between them. If so reinstall and see if the problem goes away.
If that doesn't work repair/reinstall Visual Studo itself. | Make sure that you have SP1 for VS2008 installed. and Latest version of ReSharper that supports SP1. |
29,242 | I'm running Visual Studio Team System Development Edition 2008 on Vista Business 64bit along with Resharper 4.5, Telerik Reporting 2009/Q2 and GhostDoc 2.5.9166.0 and it keeps crashing randomly. It typically happens when I start entering text into a .cs or text file. The event log gets an application error entry:
>
> Faulting application devenv.exe,
> version 9.0.30729.1, time stamp
> 0x488f2b50, faulting module ntdll.dll,
> version 6.0.6002.18005, time stamp
> 0x49e03824, exception code 0xc0000374,
> fault offset 0x000ab0bf, process id
> 0xb0, application start time
> 0x01ca255374e4ec56.
>
>
>
Any thoughts? | 2009/08/25 | [
"https://superuser.com/questions/29242",
"https://superuser.com",
"https://superuser.com/users/-1/"
] | I'd guess it has to do with the extensions you have installed. To disable them temporarily, start VS and click on Tools -> Add-in manager. Deselect all "Start" entries and restart Visual Studio. | Ensure you are running the latest version of ReSharper. There was a known issue with one of the beta release that would cause a random crash when trying to edit anything related to HTML or CSS or when using the `<` or `>` keys. I would suggest disabling ReSharper first. |
29,242 | I'm running Visual Studio Team System Development Edition 2008 on Vista Business 64bit along with Resharper 4.5, Telerik Reporting 2009/Q2 and GhostDoc 2.5.9166.0 and it keeps crashing randomly. It typically happens when I start entering text into a .cs or text file. The event log gets an application error entry:
>
> Faulting application devenv.exe,
> version 9.0.30729.1, time stamp
> 0x488f2b50, faulting module ntdll.dll,
> version 6.0.6002.18005, time stamp
> 0x49e03824, exception code 0xc0000374,
> fault offset 0x000ab0bf, process id
> 0xb0, application start time
> 0x01ca255374e4ec56.
>
>
>
Any thoughts? | 2009/08/25 | [
"https://superuser.com/questions/29242",
"https://superuser.com",
"https://superuser.com/users/-1/"
] | I'd guess it has to do with the extensions you have installed. To disable them temporarily, start VS and click on Tools -> Add-in manager. Deselect all "Start" entries and restart Visual Studio. | Make sure that you have SP1 for VS2008 installed. and Latest version of ReSharper that supports SP1. |
134,392 | If you were the mentor of someone new to SharePoint where would you start to teach, developer, the world of SharePoint.
Where would you guide him first, in order for him to know what Sharepoint is all about?
And what tutorials would you recommend him to start developing his first Apps and workflows? | 2015/03/06 | [
"https://sharepoint.stackexchange.com/questions/134392",
"https://sharepoint.stackexchange.com",
"https://sharepoint.stackexchange.com/users/40279/"
] | For me, I would prefer reading books, they make you more focused and they go in depth with you more than videos. For sure PluralSight are great, but you can get easily distracted and not knowing what they're talking about, plus reading a book on your laptop makes it easy for you to copy code snippets and paste them here so people can help you more. And it's not about which site to watch videos from.. there was a large number of videos on Youtube before the idea of PluralSight, but it's a bit hard to keep up watching videos.
If you're a beginner and would like to focus on the development side of SharePoint, then make sure to master HTML5 and JavaScript, anyway this is the first exam you need to take to become MCSD SharePoint Certified. After you become familiar with these technologies, get into SharePoint from end user perspective, you should know how to use it without thinking as a developer, understand what it offers, try to make stuff with it without writing code. Then it's good to know a little bit of **Administration**, know how to configure some service applications like Search and User Profiles, and understand their implications in your site collections. Then you should take it one level up and use SharePoint designer to create some workflows and understand this concept and what SharePoint Designer can do for you.
After all of this, try with applying some master pages for SharePoint, understand the themes, and when to use master pages, try to apply one for SharePoint and understand differences in design between all SharePoint Editions (Server VS Foundation). After you know this area, get into development and use what you know in SharePoint OOB, what you learned from JavaScript and HTML, what you knew SharePoint Designer can't do, and what designs you created for SharePoint and try to implement that in a Visual Studio project, like packaging your design, creating SharePoint columns, content types, list definitions, so on.
After you know that area, learn the App model, and see how things are transitioned from doing stuff the regular way in SharePoint 2010 to doing them in App model and how to provision everything using an App. That's how I would mentor someone to learn SharePoint Development. | Most of my learning as a beginner was from the courses on Lynda.com. These are the ones I started with:
[SharePoint Server 2013 Essential Training](http://www.lynda.com/Office-tutorials/SharePoint-Server-2013-Essential-Training/121679-2.html)
[SharePoint Designer 2013: Custom Workflows](http://www.lynda.com/SharePoint-tutorials/SharePoint-Designer-2013-Custom-Workflows/144025-2.html)
[SharePoint Designer 2013: Creating Data-Driven Sites](http://www.lynda.com/SharePoint-Designer-tutorials/SharePoint-Designer-2013-Creating-Data-Driven-Sites/180862-2.html)
[SharePoint Designer 2013: Branding SharePoint Sites](http://www.lynda.com/SharePoint-Designer-tutorials/SharePoint-Designer-2013-Branding-SharePoint-Sites/155976-2.html)
These should give you a solid grounding in what SharePoint is and how to start building solutions with it.
Now that I'm at a more intermediate level and moving to advanced, I've benefited greatly from the courses on PluralSight.com.
Trust me when I say the subscription cost to these sites is well worth it. |
121,667 | The 3-SAT problem is NP-complete, meaning that no known algorithm can provide an exact solution in polynomial time, while a solution can be tested very quickly in polynomial time.
My question is, if asking for an algorithm that only provides yes/no, or solvable/not solvable, without providing an exact solution of the 3-SAT formula, will this result in another complexity class?
The reason I have in mind of asking this, is a similar situation when looking for prime numbers and factorizations. Where it is known that it's much easier to test if a number is a prime number yes/no, than to find an exact prime factorization of a number. | 2020/03/11 | [
"https://cs.stackexchange.com/questions/121667",
"https://cs.stackexchange.com",
"https://cs.stackexchange.com/users/88261/"
] | Check Belare and Goldwasser, "The complexity of decision versus search", SIAM J. Of Computing, 23:1 (feb 1994), pp. 97-119. Belare has [notes for a class](https://cseweb.ucsd.edu/~mihir/cse200/decision-search.pdf). | Short answer: if you can solve the decision (yes/no) problem, calling that tells you if it has no solution; if there is a solution, pick a variable and set it to true, see if the result can be satisfied; if not, it has to be false. This way, with one call to the oracle per variable you get a "solution" (set of values of the variables making the expression true, it here are any). Thus the cost of the search problem isn't "too much more" (multiply by the size of the input --number of variables-- is polynomial). |
6,153 | A friend of mine, actually my best friend whom I've known for 5 years now is a very emotionally driven person. He is 19 years old, attends Engineering school in which he really struggles, and he even had to repeat last year.
So now to the problem: He smokes weed. Normally not a problem, especially not for me as I smoked it as well a couple of times and had a great time mostly. But my problem with his consumption is that he takes it to a completely different level as I for example ever did. He smokes *every day*. Every evening when he comes home from school he blasts himself away. Every weekend he goes out with a friend of ours and they blast themselves away every time. I sadly cannot be there and try to help him currently as I now live in another town 1,5 hours away. I am really concerned about him because, as stated by him, he smokes to forget.
When he is high he doesn't have to deal with his troubles, his problems and so on. So that means that his mental state contributes a great deal to that.
So to wrap this up, I am really concerned about the psychological aspect to all that.
Is there any way I can show him that I care for him and his mental health and well-being and would like him to stop smoking pot and rather focus on important things that *really* make someone happy, like quality time with friends or finding some meaningful hobby? I'm no psychologist of course and maybe some advice from someone having knowledge in this part may be of tremendous help!
I never told him "STOP SMOKING POT!!" as I know that won't help. I was sometimes a bit suggestive to steer him more to healthier things, but I know that brute force wouldn't change a thing. | 2017/11/02 | [
"https://interpersonal.stackexchange.com/questions/6153",
"https://interpersonal.stackexchange.com",
"https://interpersonal.stackexchange.com/users/7301/"
] | Depression isn't simply solved by saying "just go do something that makes you happy". When you're depressed, if you are truly feeling miserable, it is very hard to push yourself to do something, anything. Finding a meaningful hobby is very much harder than you think if there seems to be so little that brings you joy.
Can you see the downwards spiral here?
1. Feel miserable
2. This makes it hard to find joy in doing things
3. This makes it hard to move yourself to do those things, to do anything
4. Do nothing instead, which is something you feel guilty about, making you even more miserable.
5. Repeat
It is really really really hard to break this spiral. Especially when looking at it from the outside, because it seems so simple, just go do something, right?
Instead, this spiral makes it easy to slip into a habit that makes you feel slightly better, even if only for a while. This can easily lead to an addictive habit like gaming and/or substance abuse.
---
So how could you help?
This is really hard. I don't have *the* answer. Honestly, if *the* answer existed, depression wouldn't be that big of a problem. Instead, it is important to get help from someone who is trained to deal with these situations.
This only really works if your friend cooperates, so they have to see the need (and commit the time) for therapy, else it is just a waste of time and money. This means you need to convince your friend that:
1. They have a problem. This is hard enough on its own. The important part when trying to convince them of this is to be careful that you are not blaming them. They can't help their depression, they are a victim.
2. That this problem is substantially big, that they can't solve it on their own, that you can't solve it for them either, that there is help available, and most important of all that **it is not a bad thing to ask for help**!
And honestly, if he thinks he doesn't have time for it during his studies, if there is no immediate risk (selfharm for example), it is not that bad to wait a while.
---
So is there nothing that you can do?
Why don't you invite him along on your activities every now and then. While it might be hard for him to get moving on his own, he might get some motivation or help from just doing things together.
But be sure to take the initiative, don't sit back thinking "if he wants my help, he can ask for it". This is one of the big reasons so many people are depressed without telling anyone about it. "If I feel terrible about myself, I sure as hell shouldn't bother others with my problems."
---
In your question you edited:
>
> AS tinkingerbell said it seems like i want to force a lifestyle on him he does not want. THe problem is that he DOES ENJOY stuff like this, again the example of going bouldering with me. We did that a couple of times, and had a blast. It was really fun for both of us. THe problem is he has no self drive and would never go alone with his initiative, only if someone drags him to do stuff. Then he really anjoys it again. I actually was the same at his age. At that time it was my brother that dragged me everytime because i always enjoyed it. Now i have drive on my own. He sadly just doesnt have that.
>
>
>
This is good, he enjoys something, but clearly he lacks the motivation to do it himself. This is something I recognize from myself. For a while I just went jogging a few times a week, to wake up before my driving lessons. But once those were finished, the big stick that kept me jogging disappeared, and it again becomes more and more easy to fall back to those old habits of gaming etc.
The place where you could break the cycle in the top of the answer would be point 3. They have trouble moving themselves to do things, so they need you to drag them there (not literally). Be that stick, give them those few hours that he enjoys, and make sure he keeps doing it. Don't rely on him to take the initiative, but be proactive in your help. | You are on the right track. Telling someone to stop a behaviour they know is not ideal and typically causes them to be defensive and on the opposite side from the person who was trying to help. This can mean pushing you (the supportive friend) away and that is bad for everyone. Instead, looking at alternative things to help as you have already started thinking of is the best way to actively help your friend.
This does a few things, most of which you have already mentioned. It occupies his mind which he is currently depending on weed to do, this would reduce his need reliance on the drug. It encourages being more social (other than getting together to smoke) interacting socially can help a lot with depression. It encourages him to be active which produces endorphines that make you feel good, and it physically takes up his time which is a very efficient yet subtle way to cut into the time spent alone, feeling depressed where he then turns to weed.
Next is to just decide what you think would be best suited for him to take part in and how to actually get him to participate. This is trickier, will probably require some effort on your part and depends a lot more on what he likes and such. Keep in mind its best if you do this activity with him, participating with a group helps give more of a push to get out of the house on the days you are not feeling like it, and this way you can continue helping in other ways too. You mentioned you are now 1.5 hours away, so if you absolutely can not do it with him then clue-ing in some other friends is the next best thing.
Here are a few ideas, dont take any of them too seriously please, remember it is up to you.
**Hiking** - can take multiple days if you are enthusiastic and can be a good excuse to travel the distance.
**Mountain Biking** - similar to hiking
**Table Tennis tournaments** - There are plenty of beginner ones of these and its a sport with not much required strength but tons of growth room.
**LARPing** - if you dont know what it is already then maybe not
**Going to semi-local bands/convensions/events** - less frequent but less likely to be rejected
It may be quite a challenge to convince him but since you are so far away I would say there is not too much else you could do that would help in a significant way |
8,869 | I am a 24 year old who has always aspired to become a pilot. I've decided to eventually get a private pilot's license so I can fly just for fun. However, a few months ago I was diagnosed with brain tumor which I had removed. There were no complications, I have never had any seizures, I'm not epileptic, etc. As far as I am concerned (and my doctors as well), I am now in perfect health.
With this in mind, is it still possible for me to obtain a PPL? | 2014/10/01 | [
"https://aviation.stackexchange.com/questions/8869",
"https://aviation.stackexchange.com",
"https://aviation.stackexchange.com/users/3720/"
] | According to [AOPA](http://www.aopa.org/)'s medical guru on their discussion board (members only, or I would link to an example) it is possible to get a medical certificate following a tumor removal but only after a 5-year wait and a battery of tests.
But having said that, medical issues are very individual and the FAA's [AME guidance](http://www.faa.gov/about/office_org/headquarters_offices/avs/offices/aam/ame/guide/app_process/exam_tech/item46/amd/cd/) suggests that one year is enough in some cases. In the end the only way to get a definite answer is to find a good AME who can review your case in detail and advise you. Note that standard advice is to schedule a consultation first, rather than go straight for the actual flight medical.
Disclaimer: I'm not a doctor or AME and I have no idea how brains work :-) | If you are in the United States and you have never been denied an FAA medical certificate or had your medical revoked and you have a driver's license, then you can fly as a sport pilot in light sport aircraft. There are some limitations compared to a PPL, the major ones being that you are limited to one passenger and daytime VFR operations.
If you decide to later move up to a PPL, then you will have to get a medical. There is some risk, though. If you attempt to get a medical and are denied, you will also lose your sport pilot priviliges. |
8,869 | I am a 24 year old who has always aspired to become a pilot. I've decided to eventually get a private pilot's license so I can fly just for fun. However, a few months ago I was diagnosed with brain tumor which I had removed. There were no complications, I have never had any seizures, I'm not epileptic, etc. As far as I am concerned (and my doctors as well), I am now in perfect health.
With this in mind, is it still possible for me to obtain a PPL? | 2014/10/01 | [
"https://aviation.stackexchange.com/questions/8869",
"https://aviation.stackexchange.com",
"https://aviation.stackexchange.com/users/3720/"
] | According to [AOPA](http://www.aopa.org/)'s medical guru on their discussion board (members only, or I would link to an example) it is possible to get a medical certificate following a tumor removal but only after a 5-year wait and a battery of tests.
But having said that, medical issues are very individual and the FAA's [AME guidance](http://www.faa.gov/about/office_org/headquarters_offices/avs/offices/aam/ame/guide/app_process/exam_tech/item46/amd/cd/) suggests that one year is enough in some cases. In the end the only way to get a definite answer is to find a good AME who can review your case in detail and advise you. Note that standard advice is to schedule a consultation first, rather than go straight for the actual flight medical.
Disclaimer: I'm not a doctor or AME and I have no idea how brains work :-) | I have a PPL license and I was on the way to obtain a CPL license when I had a seizure.
After many tests it came out that I had a cavernous angioma in my brain on the left temporal lobe anterior. It had bleed outside causing the seizure.
After removal of the cavernoma with brain incision, I had no other seizures and no complications and I went to see the AME.
I had lots of EEGs, MRIs and the AMEs decision is giving back my Class I certificate for at least 1 year after the surgery.
Neurology is decreasing the medication now and after medication is finished I will have few other checks.
If everything goes OK, I will have my Class I medical license back with an OML restriction (most probably).
This may give you lots of information about your situation : <https://www.icao.int/publications/Documents/8984_cons_en.pdf>
I hope this helps and don't give up. :)
Note : I have my Class 1 with OML and CPL(A) now. |
37,886 | I was wondering if there are established rules of thumb (or algorithms) that, given a set of observations can help:
1. choose an initial number of class intervals.
2. refine that choice to a better number.
I could find talk of using square-root(N), where N is the number of observations as an initial guess of the number of class intervals.
Thanks in advance. | 2012/09/24 | [
"https://stats.stackexchange.com/questions/37886",
"https://stats.stackexchange.com",
"https://stats.stackexchange.com/users/14163/"
] | The help of the R command `hist` <http://stat.ethz.ch/R-manual/R-patched/library/grDevices/html/nclass.html> has some references to algorithms for computing the number of the bins:
Sturges, H. A. (1926) The choice of a class interval. Journal of the American Statistical Association 21, 65–66.
Scott, D. W. (1979) On optimal and data-based histograms. Biometrika 66, 605–610.
Freedman, D. and Diaconis, P. (1981) On the histogram as a density estimator: L\_2 theory. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete 57, 453–476. | See also
[HOGG, David W. Data analysis recipes: Choosing the binning for a histogram. arXiv preprint arXiv:0807.4820, 2008.](http://arxiv.org/pdf/0807.4820v1.pdf)
The abstract:
>
> Data points are placed in bins when a histogram is created, but there
> is always a decision to be made about the number or width of the bins.
> This decision is often made arbitrarily or subjectively, but it need
> not be. A jackknife or leave-one-out cross-validation likelihood is
> defined and employed as a scalar objective function for optimization
> of the locations and widths of the bins. The objective is justified as
> being related to the histogram’s usefulness for predicting future
> data. The method works for data or histograms of any dimensionality.
>
>
> |
170,749 | A month ago I sent my application for a job for X company for a full-stack job. They didn't move forward with my application because I am not senior level. I found a new DevOps job. I tried to call the HR person to apply for the job but he did not respond to my call or email. I use my nickname, alternate email and my home phone number to apply for the job.
I want to ask if changing my name to a nickname or alternate email is viewed as a scam, can I get into trouble if someone found out? | 2021/03/23 | [
"https://workplace.stackexchange.com/questions/170749",
"https://workplace.stackexchange.com",
"https://workplace.stackexchange.com/users/11202/"
] | >
> I want to ask if changing my name to a nickname or alternate email is scam can I get into trouble if someone found out?
>
>
>
"Scam" is probably too strong a word - but ultimately you're asking whether attempting to circumvent the company's prior knowledge of you via obfuscating who you are is okay. You aren't *lying* (yet) but you aren't exactly being 100% honest either.
Let's take this a bit further - say they bring you in for an interview under the "new" details (nickname etc) and during that interview they ask you if you are the same person who applied for the other job. What are you going to say? Do you tell the truth? If you do and they then follow up with asking why you used different details this time around what are you going to say?
If you persist in the pretense that you are a *different* person to the original application how long do you plan on keeping that up? How long do you think you *can* keep that up? Getting rejected for lacking skills isn't a big deal, it happens all the time and is rarely personal. Getting discovered engaging in shennaigans and then getting canned with prejudice **is** a big deal.
All of this based on an *assumption* that you are being stonewalled unfairly for the DevOps role based on you prior application. You say they rejected that based on lack of senior-level skills/experience, how do you know they aren't just dismissing you for similar reasons this time around?
You're acting as if this HR person is some evil mustache-twirling villain keeping you from a job there for, well, *reasons* and if the plucky hero (you) can just outfox them you'll get to work there and everything will be sunshine and light while they stand and the side impotently shaking their fist muttering "next time, Gadget!". But life isn't a movie or a cartoon - you've sent your details in for a a couple of jobs and they didn't pan out. It sucks, but it's something that happens day in day out so don't over-invest - keep your head down and keep applying for different jobs at different companies and something **will** pan out for you, and then, sooner than you think, you'll have forgotten all about X company. Good luck! | >
> I want to ask if changing my name to a nickname or alternate email is scam can I get into trouble if someone found out?
>
>
>
**No it can't.**
But eventually, you will have to give them your real full legal name for a background check. If your issue is a single HR person, then this will work. If you've been blocked by the company and HR is enforcing it, you'll be found out at the background check stage. |
437,022 | **Can one boot M1 MacBook from a *unauthorized* bootable external drive**: i.e. the external drive that was prepared on some other Mac, and the user not having admin password of the MacBook he is trying to boot into?
Case in question: suppose my Mac (2021 Pro) is lost or stolen; its internal drive is FileVault-encrypted; MacBook is kept offline and is not aware that "Lost Mode" is activated for it; still someone tries to use it with his own external drive that I never authorized to boot from on my MacBook. | 2022/02/15 | [
"https://apple.stackexchange.com/questions/437022",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/54644/"
] | I reached out to Howard Oakley, author of [EclecticLight blog](https://eclecticlight.co) which goes quite deep into Mac security.
TLDR of his answer: **booting from external drive is only possible if that drive was previously authorized using login and password of Mac's internal-drive admin**.
Full answer below:
>
> This is straightforward.
>
>
> You can't boot an M1 Mac from an external disk without it having a LocalPolicy on its internal storage. So the only way to try that would be to boot in Recovery, there select the external disk, and try to boot from it.
>
>
> Recovery won't let you do that without giving it a LocalPolicy, which then prompts you to enter the username and password of someone who already has ownership of its internal SSD - in normal circumstances, the admin user. Without that, LocalPolicy can't be created, and the M1 won't boot from the external disk.
>
>
> What the 'thief' can do is put the M1 into DFU mode and restore it to factory settings, which trashes the keys to the internal SSD safeguarding any data there. And when it starts up, it then goes through the 'new M1 Mac' routine of connecting to Apple before it can be set up. If that Mac has been reported as missing, then Apple could refuse to set it up at that stage.
>
>
> The relevant article on my blog is at
> <https://eclecticlight.co/2021/07/21/owners-and-users-primary-and-secondary-systems-on-m1-macs/>
>
>
> | This piece from Apple Platform Security Guide ([HTML](https://support.apple.com/nl-nl/guide/security/sec7d92dc49f/web), [PDF](https://manuals.info.apple.com/MANUALS/1000/MA1902/en_US/apple-platform-security-guide.pdf)) *seems to suggest* that only users authenticated on this particular Mac can boot it from an external drive:
>
> A Mac with Apple silicon doesn’t require or support a specific media boot policy, because technically all boots are performed locally. If a user chooses to boot from external media, that operating system version must first be personalized using an authenticated reboot from recoveryOS. This reboot creates a LocalPolicy file on the internal drive that’s used to perform a trusted boot from the operating system stored on the external media.
>
>
> This means the configuration of booting from external media is always explicitly enabled on a per operating system basis, and already requires user authorization, so no additional secure configuration is necessary.
>
>
>
---
Another piece: as of M1, [volumes have ownership](https://support.apple.com/guide/deployment/use-secure-and-bootstrap-tokens-dep24dbdcf9e/web#dep409564e88). When creating a second boot drive, [a consent is required from a user on the default boot drive to hand off Ownership to the users](https://support.apple.com/guide/security/contents-a-localpolicy-file-mac-apple-silicon-secc745a0845/web#secb7cb6686c) on the second boot drive.
Which *suggests* that users who "own" an external boot drive *could be* not allowed to boot someone else's Mac as they are not on the same hierarchy with users on that Mac.
---
But I would very welcome someone more knowledgeable to comment on that. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | No. Please, *please*, no!
If someone gives a clear answer, supports it with references to authoritative sources, provides easy-to-follow examples, and patiently answers follow-up questions... Then that's enough, *even if it's the only answer they've ever provided on the site, and googling their name turns up nothing but a lone blog post from 2001... about cats.*
If someone gives a brief, unclear answer, with no supporting references, missing or unhelpful examples, and steadfastly ignores follow-up questions... Then it's **not** enough, *even if they have a high reputation number on SO, are heavily awarded by various industry organizations, have written several books on the topic, and currently hold the top spot on Google for [the best programmer ever](http://www.google.com/search?q=the+best+programmer+ever).*
Judge people by their answers. Or better yet, **don't**. Just judge the answers, ignore the people. SO is no place to be resting on your laurels... We're here to answer questions, not fawn over celebrities. | Put the information in your profile.
From what I have seen, most people are pretty forthcoming about their information and I haven't seen much in the way of posers and fakers. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | No. Please, *please*, no!
If someone gives a clear answer, supports it with references to authoritative sources, provides easy-to-follow examples, and patiently answers follow-up questions... Then that's enough, *even if it's the only answer they've ever provided on the site, and googling their name turns up nothing but a lone blog post from 2001... about cats.*
If someone gives a brief, unclear answer, with no supporting references, missing or unhelpful examples, and steadfastly ignores follow-up questions... Then it's **not** enough, *even if they have a high reputation number on SO, are heavily awarded by various industry organizations, have written several books on the topic, and currently hold the top spot on Google for [the best programmer ever](http://www.google.com/search?q=the+best+programmer+ever).*
Judge people by their answers. Or better yet, **don't**. Just judge the answers, ignore the people. SO is no place to be resting on your laurels... We're here to answer questions, not fawn over celebrities. | My attitude on forums like these has always been that I'm a guy with an opinion and an ISP. The quality of my answer depends a whole lot more about what I say than that I have >10K of StackOverflow rep and a C++ badge. (The rep and badges are for personal gloating when I'm not asking or answering questions.)
However, if we are going to put rep and badge count next to our answers, we probably want to note relevant tag badges. We are apparently trying to indicate answers from more trustworthy sources, and I'm a lot more trustworthy answering C++ questions than Visual Basic questions. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | MVPs are awarded by Microsoft for those that help spread the word to the community more than it is to experts. The more you know...
Regardless, this isn't a Microsoft (or any other company's) site, and the only currency on here is reputation. We vote up answers so that you can quickly distinguish between random morons and decent responses. | Give a MVP a badges and letting them start with 100 reps may be good if it gets more of them to use the site. Otherwise let everyone live by their rep.
*Who will be the first person to get a MVP due to their ansers on StackOverflow?* |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | MVPs are awarded by Microsoft for those that help spread the word to the community more than it is to experts. The more you know...
Regardless, this isn't a Microsoft (or any other company's) site, and the only currency on here is reputation. We vote up answers so that you can quickly distinguish between random morons and decent responses. | I have no interest in being identified as an MVP on this site, anywhere except in my profile. That's more than enough. I primarily put that information there (translated to acceptable HTML from a much prettier signature I use), so that, when needed, I can say, "go look at my profile and see if you see any reason I might be right about this". I think this has happened once in the five months I've been here.
Besides, my reputation here, especially as it distributes over tags, is a better indicator of the likely quality of my answers than my MVP-hood. I'm an MVP in the Connected Systems area (roughly, Web Services, etc). But my experience ranges over ASP.NET, SQL Server, SSIS, multi-threading issues, etc. If you only looked at my MVP award, you might wonder what I'd have to say about SSIS. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | Put the information in your profile.
From what I have seen, most people are pretty forthcoming about their information and I haven't seen much in the way of posers and fakers. | My attitude on forums like these has always been that I'm a guy with an opinion and an ISP. The quality of my answer depends a whole lot more about what I say than that I have >10K of StackOverflow rep and a C++ badge. (The rep and badges are for personal gloating when I'm not asking or answering questions.)
However, if we are going to put rep and badge count next to our answers, we probably want to note relevant tag badges. We are apparently trying to indicate answers from more trustworthy sources, and I'm a lot more trustworthy answering C++ questions than Visual Basic questions. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | No. Please, *please*, no!
If someone gives a clear answer, supports it with references to authoritative sources, provides easy-to-follow examples, and patiently answers follow-up questions... Then that's enough, *even if it's the only answer they've ever provided on the site, and googling their name turns up nothing but a lone blog post from 2001... about cats.*
If someone gives a brief, unclear answer, with no supporting references, missing or unhelpful examples, and steadfastly ignores follow-up questions... Then it's **not** enough, *even if they have a high reputation number on SO, are heavily awarded by various industry organizations, have written several books on the topic, and currently hold the top spot on Google for [the best programmer ever](http://www.google.com/search?q=the+best+programmer+ever).*
Judge people by their answers. Or better yet, **don't**. Just judge the answers, ignore the people. SO is no place to be resting on your laurels... We're here to answer questions, not fawn over celebrities. | MVPs are awarded by Microsoft for those that help spread the word to the community more than it is to experts. The more you know...
Regardless, this isn't a Microsoft (or any other company's) site, and the only currency on here is reputation. We vote up answers so that you can quickly distinguish between random morons and decent responses. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | Why can't these users simply put this in their profile page?
These things shouldn't make a user absolutely trusted by anyone anyway. Voting should be done based on the quality of the post, not the user. | No. Please, *please*, no!
If someone gives a clear answer, supports it with references to authoritative sources, provides easy-to-follow examples, and patiently answers follow-up questions... Then that's enough, *even if it's the only answer they've ever provided on the site, and googling their name turns up nothing but a lone blog post from 2001... about cats.*
If someone gives a brief, unclear answer, with no supporting references, missing or unhelpful examples, and steadfastly ignores follow-up questions... Then it's **not** enough, *even if they have a high reputation number on SO, are heavily awarded by various industry organizations, have written several books on the topic, and currently hold the top spot on Google for [the best programmer ever](http://www.google.com/search?q=the+best+programmer+ever).*
Judge people by their answers. Or better yet, **don't**. Just judge the answers, ignore the people. SO is no place to be resting on your laurels... We're here to answer questions, not fawn over celebrities. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | Why can't these users simply put this in their profile page?
These things shouldn't make a user absolutely trusted by anyone anyway. Voting should be done based on the quality of the post, not the user. | Put the information in your profile.
From what I have seen, most people are pretty forthcoming about their information and I haven't seen much in the way of posers and fakers. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | Why can't these users simply put this in their profile page?
These things shouldn't make a user absolutely trusted by anyone anyway. Voting should be done based on the quality of the post, not the user. | My attitude on forums like these has always been that I'm a guy with an opinion and an ISP. The quality of my answer depends a whole lot more about what I say than that I have >10K of StackOverflow rep and a C++ badge. (The rep and badges are for personal gloating when I'm not asking or answering questions.)
However, if we are going to put rep and badge count next to our answers, we probably want to note relevant tag badges. We are apparently trying to indicate answers from more trustworthy sources, and I'm a lot more trustworthy answering C++ questions than Visual Basic questions. |
12,210 | Has there been any discussion about denoting the experts of certain topics, for example MVPs or Microsoft employees in the small profile icon?
This way users can quickly distinguish between random internet morons like myself, and the ones accredited with their chosen speciality? | 2009/08/04 | [
"https://meta.stackexchange.com/questions/12210",
"https://meta.stackexchange.com",
"https://meta.stackexchange.com/users/130507/"
] | Why can't these users simply put this in their profile page?
These things shouldn't make a user absolutely trusted by anyone anyway. Voting should be done based on the quality of the post, not the user. | Reputation and badges are, roughly speaking, SO's way of marking people as possible experts. Certainly you can't put too much weight into someone's reputation without doing a little research into how they achieved their rep. But regardless, I cannot see SO doing anything to provide "external" approval of a person. If you want to make your credentials known, you can put them in your profile. |
6,411,203 | My java application calls C++ code using JNI calls. I want to check the program for memory leaks in the C++ code. Tools like jconsole are only for analyzing java objects heap. What can I do? | 2011/06/20 | [
"https://Stackoverflow.com/questions/6411203",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/474323/"
] | My favourite native heap leak detection tool for Windows is [umdh.exe](http://support.microsoft.com/kb/268343). However, this will also display the current Java GC heap memory footprint, in whatever native form the JVM uses.
You should still be able to identify memory attributable to your C++ code since it will (provided symbols are set up properly according to the UMDH instructions) have a callstack matching C++ code. So having Java code coresident will muddy the waters but should not make it impossible to track C++/native memory usage. | You need a native heap debugging tool. There are many available depending on your platform and what compiler was used for the native component. |
27,107,029 | I am using solr 4.10 setup on a **3 node cluster(solrcloud)** with zookeeper and **RF = 1.** Total 3 shards.
The problem here is 2 phase.
1. I added **50 million** records into the index which has a uuid field(**user\_id**) as a unique key. The uuid field was generated by the application and not solr. The records were indexed in batches of 5k and committed. After indexing i saw there were a lot of duplicates in the index. Even though uuid was specified as a unique key. **Somehow 2 different shards had the same document.**
When i tried to return the counts for a particular query, the shards returned the counts including duplicates. It returned **numFound=60,686** .
But when i tried to retrieve all the 60,686 records **it returned only 52,760** which was the actual records expected.
2. It looked like there were 7926 duplicates, **but there were only 74 duplicates** when i checked
select?q=tag%3A123&rows=0&wt=json&indent=true&facet=true&facet.field=user\_id&facet.mincount=2
i manually deleted the duplicates and the count went down to 60538 which is still wrong.
I see each shard is contributing to this false count when i turned on debugQuery.
Is there a facet cache or something that needs to be cleared ? what are these remaining extra counts ? My application is count dependent so this is a big issue i'm facing.
Thanks in advance. | 2014/11/24 | [
"https://Stackoverflow.com/questions/27107029",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/4125141/"
] | The problem seemed to be with the document router. By default solr uses "implicit" router, i changed it to "compositeId" and a document was always sent to the same shard.
Ref: <https://lucidworks.com/blog/2013/06/13/solr-cloud-document-routing/> | If you had deleted or updated any records, they are still contributing to the facet counts until Solr does automatic merge or you do manual optimize trigger (very expensive).
That's the price of the fact that Lucene does not actually allow deleting documents in place, so it is marked gone at the higher level. Those documents never show up in the search, but they do contribute to the counts.
So, if this is a test system, try *optimize* call and see if your problem goes away. But this may not be a solution for production, just a way to confirm the hypothesis. |
14,940 | If Linear Cryptanalysis exploits the fact that the plaintext and ciphertext are not completely unrelated, is the attack possible without having access to the plaintext? | 2014/03/11 | [
"https://crypto.stackexchange.com/questions/14940",
"https://crypto.stackexchange.com",
"https://crypto.stackexchange.com/users/12420/"
] | Yes, linear cryptanalysis may still be possible, depending upon the distribution on the plaintexts and the specifics of the block cipher.
For instance, suppose we know that the plaintext is English encoded in ASCII. Then we know that the high bit of each 8-bit byte is zero. We may also know some additional linear approximations that hold on the plaintext that hold with non-zero bias. This may be enough to mount linear cryptanalysis, e.g., if you can find a linear characteristic for the whole cipher that only involves those bits of the plaintext (only the high bits of each byte) and not any other bits of the plaintext.
As I recall, there has been results showing that ciphertext-only linear cryptanalysis of DES is possible, assuming that the plaintexts are ASCII-encoded English text. This result might even have been included in Matsui's original paper on linear cryptanalysis; I cannot remember. | To apply this kind if cryptoanalysis you need to have at least few pair of plaintext-ciphertext. Without knowing plaintext you could not build correct linear expression, because you will not know where to start.
Quote from [Tutorial on Linear and Differential Cryptanalysis](http://www.engr.mun.ca/~howard/PAPERS/ldc_tutorial.pdf):
>
> Linear cryptanalysis tries to take advantage of high probability
> occurrences of linear expressions involving plaintext bits,
> "ciphertext" bits, and subkey bits. It is a known plaintext attack:
> that is, it is premised on the attacker having information on a set of
> plaintexts and the corresponding ciphertexts.
>
>
> |
13,459 | What is the difference between radiation doses of a medical scanner and airport security scanner (X-Ray full body scan)? Is it the same kind of radiation? Does it pose any danger for people who fly often? | 2013/11/12 | [
"https://biology.stackexchange.com/questions/13459",
"https://biology.stackexchange.com",
"https://biology.stackexchange.com/users/1373/"
] | The type of radiation is quite different in a medical X-ray vs. an airport scanner.
Medical X-rays are high frequency (beyond ultraviolet) radiation, typically on a wavelength of a few angstroms. While I would emphasize that @Ram is right to point out that there [is not very much radiation](http://hps.org/physicians/documents/Doses_from_Medical_X-Ray_Procedures.pdf) in a medical X-ray since electronic detectors have been in place over film, the radiation itself is capable of penetrating the entire body and causes ionization. The body can tolerate a certain amount of this, but that's why we use X-rays - they go through just about anything but bone/minerals and metals.
Airport security scanners emit [terahertz radiation](http://en.wikipedia.org/wiki/Terahertz_radiation), which has a wavelength between microwave and infrared. This is a very low level of energy per photon. Terahertz radiation will not penetrate more than a centimeter of light material, which is more suitable to find hidden metal objects (which reflect terawaves). Relative to X-rays it causes practically no radiation damage. The intensity of the radiation is also pretty low since these scans are also detected by digital cameras, which are quite sensitive.
Compare the machinery. X-ray technicians, who might have to take dozens of X-ray exposures a day, usually go behind a protective screen to take a medical x-ray. The airport body scanners are open to the air and the terahertz radiation spills out into the crowd and the workers stand by it every day. It has practically no expectation of being hazardous. This device is pretty good as frequent fliers can have as many scans per year as are needed and health concerns are not significant.
Its always possible we might find some issues with Terahertz radiation in the future, but its hard to imagine it - its safer than standing 2 feet away from a working, closed oven. | As per the American Association of Physicists in Medicine the radiation exposure from full body airport scanners is equivalent to what an individual receives every 1.8 minutes on the ground from natural background radiation or equivalent to every 12 seconds during an airplane flight.
<http://www.aapm.org/pubs/reports/RPT_217.pdf>
The Back scatter full body scanners at airports use X-rays. |
28,309 | I just installed Odin on an older Windows 8 vintage Dell laptop. I have totally forgotten how to troubleshoot the chipset and get the kernel sources installed correctly. I’ve gotten as far as installing the BCMWL Kernel source package from AppCenter, but I am totally lost. | 2021/08/10 | [
"https://elementaryos.stackexchange.com/questions/28309",
"https://elementaryos.stackexchange.com",
"https://elementaryos.stackexchange.com/users/6173/"
] | Have been struggling with this, too. You need to install Kernel Headers first. See this:
<https://github.com/elementary/os/issues/526#issue-965454894>
you might need to
sudo apt-get install --reinstall bcmwl-kernel-source | Unfortunately I cannot comment: use the checkmark left to my answer, below the down/upvote arrows to mark it as solved |
422,581 | I refer to a 2N3055 power transistor, the spec. sheet gives the following:
Max. current 15A, Max. Voltage 50V, Power rating is 115 Watts.
* What is the power rating calculated on?
* If it is E\*I then the power would be 750 Watts, this is of course incorrect, so how is it calculated please.? | 2019/02/16 | [
"https://electronics.stackexchange.com/questions/422581",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/130169/"
] | The power rating is based on the thermal design of the TO-3 transistor package, how easy or hard it is to extract heat from the transistor die inside. The MJE3055 is basically the same die in a smaller TO-220 package, and is rated for only 75 W. The thermal path from the die to the mounting surface is better than the old TO-3 package, but that mounting surface is significantly smaller. | Total Power Dissipation @ Tc = 25°C
Derate PDmax=115W Above 25°C by 0.657 W/°C for case temp, Tc so **Rjc= 1/0.657= 1.52'C/W**
This implies the max junction temp, **Tj = 175'C+25'C=200'C** but this accelerates failure rate significantly, so a **good design has Tj<< 100'C**.
That assumes an infinite heatsink which is not possible.
So to compute Tc , you need Tja heatsink rating like 0.1 'C/W for a super CPU cooler with Fan or 1'C/W for a large passive heatsink then add internal ambient rise above room temp (or outside temp)
Then Temp rise of junction is Rja\*W= Rjc+Rca(total) \* Watts = Tj rise |
422,581 | I refer to a 2N3055 power transistor, the spec. sheet gives the following:
Max. current 15A, Max. Voltage 50V, Power rating is 115 Watts.
* What is the power rating calculated on?
* If it is E\*I then the power would be 750 Watts, this is of course incorrect, so how is it calculated please.? | 2019/02/16 | [
"https://electronics.stackexchange.com/questions/422581",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/130169/"
] | You can have maximum current through the transistor and you can have maximum voltage across it but you can't have both at the same time. This is specified in the "safe operating area" power curve.
[](https://i.stack.imgur.com/yAQU1.png)
*Figure 1. The safe operating area of a BDV66C transistor. Source: Wikipedia [Safe operating area](https://en.wikipedia.org/wiki/Safe_operating_area).*
In Figure 1 we can see the following:
1. 16 A at about 12 V. Pmax = 192 W.
2. 5 A at 50 V. Pmax = 250 W.
3. 0.2 A at 120 V. Pmax = 30 W.
Taking the two maxima from the graph one might have expected that the device could handle 16 A at 120 V = 1,920 W which is not the case.
SOA is defined as the voltage and current conditions over which the device can be expected to operate without self-damage. The SOA figures are, of course, subject to other constraints such as maximum junction temperature and so the power dissipated as heat, even when inside the SOA, must be conducted away to maintain a safe operating temperature. There is more information in the linked wiki article. | Total Power Dissipation @ Tc = 25°C
Derate PDmax=115W Above 25°C by 0.657 W/°C for case temp, Tc so **Rjc= 1/0.657= 1.52'C/W**
This implies the max junction temp, **Tj = 175'C+25'C=200'C** but this accelerates failure rate significantly, so a **good design has Tj<< 100'C**.
That assumes an infinite heatsink which is not possible.
So to compute Tc , you need Tja heatsink rating like 0.1 'C/W for a super CPU cooler with Fan or 1'C/W for a large passive heatsink then add internal ambient rise above room temp (or outside temp)
Then Temp rise of junction is Rja\*W= Rjc+Rca(total) \* Watts = Tj rise |
354,670 | I managed to get myself in a two-factor authentication pickle. My Apple id is associated with a Mac (currently running OS X Yosemite), an iPhone and an iPad.
The iPad just came back from repair and I needed to restore it from the cloud. Unfortunately before doing that I lost my iPhone! I now have nothing that can receive the two-factor authentication codes. (Didn't see that one coming!)
My Mac is tied to my iCloud account and Apple Id, and I see that Sierra can receive these codes and that I can upgrade from Yosemite which is what's currently running on the Macbook Pro.
If I upgrade to Sierra will that mean my Mac can then also receive the two factor codes (allowing me then to get my iPad back up and running)?
Or will I be out of luck as I may need to use two-factor to get the laptop properly set up with Sierra? If that's the case then I do not want to upgrade as it means I may lose the limited iCloud access that I have via the Mac.
Help!
N.B. Yes, have a full trustworthy backup of the Mac on Time Machine. | 2019/03/24 | [
"https://apple.stackexchange.com/questions/354670",
"https://apple.stackexchange.com",
"https://apple.stackexchange.com/users/325347/"
] | You can request the two factor code to be sent to your phone number instead. Click the "Didn't get a code" on iPad, where you're asked to enter the code. Then you can get it sent to your trusted phone number.
The trusted phone number doesn't have to be an iPhone, so you can use any dumb phone that can receive SMS text messages. | I am fairly confident that the answer to this is “No, the device is not yet trusted and updating the operating system will not change that.”
The reason being that having regained my phone and therefore the ability to get two factor codes I see my devices described as “trusted” with the exception of the laptop.
It seems unlikely that the laptop could be considered trusted until it is explicitly signed in by one of the other trusted devices.
If I feel brave one day I may upgrade the OS and see if that is correct.
(Oh, and yes, lesson learned - I have added a whole bunch of extra backup numbers so I can get verification codes through multiple avenues!)
**Update**
The answer is indeed “no”. Having updated to Sierra I cannot be fully integrated with iCloud without getting it in to the trust circle. |
1,510,047 | Despite thorough research, I could not find out the form factor of the Fujitsu TX1310 M1 server's mainboard, which is a D3219-Axx.
I have found the manuals here: <http://manuals.ts.fujitsu.com/index.php?id=5406-5635-5814-16661-18262>
Unfortunately, none of the manuals tells the mainboard's form factor. Does anybody know if I can mount that board into a standard mATX, ATX or eATX case? | 2019/12/14 | [
"https://superuser.com/questions/1510047",
"https://superuser.com",
"https://superuser.com/users/650572/"
] | >
> In my opinion, it is an ATX form factor motherboard.
>
>
>
I looked through a number of spec sheets, websites, and even some videos. However, I could not find anything definitively stating what form factor it was. That being said, it definitely *appears* to be an ATX form factor motherboard in what *appears* to be an ATX form factor case. It probably is, but there is no guarantee.
>
> One important thing to note:
>
>
>
This machine does ***NOT*** use a standard ATX power supply. It is a non-standard/custom power supply. It might not mount properly in the case you were planning to put it in. Additionally, even if it does mount, the cables running to the motherboard are very short and do not use the standard ATX connector, so extenders will not work. It also apparently uses non standard voltages (11v) on some if the pins. | I have seen some of these in real life a couple of months ago.
The motherbaord is a non-standard ATX form-factor. It has the ATX size, but it is just different enough that it won't easily fit in a standard case.
Some of the mounting holes are in the wrong spot. You can only mount it with 2 or 3 screws depending on the layout of a normal case.
The backplane insert for the connectors is fixed to the original case, so you can't take that out to put in another case.
I could be wrong but it looks as if the backplane connectors are also somewhat offset from a normal ATX baord which may mean they could potentially interfere with the case.
The power-supply is totally non-standard. Not easily replaceable by a standard PU.
The various front-panel headers for case-lights, on/off switch, HD led, audio, etc. are non-standard. |
39,643 | What is another way of saying, "raise the roof"? This slang phrase means something like, "get noisy and have a good time at a party," but it doesn't sound correct for some reasons.
Why is that? What would be better? | 2011/08/27 | [
"https://english.stackexchange.com/questions/39643",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/5877/"
] | A few of the idioms i've heard:
"Tear the house down"
"Rock the house"
"Heat this place up"
There are others. | It seems "lame" because it's antiquated -- the only thing that spreads as quickly is slang is the knowledge that a slang term has become "uncool."
"Raise the roof" is a dance move, usually used (the key word here is USED -- it's completely out of favor) in hip hop, in which you push the palm of your hands towards the ceiling as if you're raising and then lowering the roof over and over.
"Raise the roof" is also a phrase that expresses the sentiment that a group intends to "go out". It can also just be an expression of excitement. Depending on which meaning you're using, there are many ways to say it, and you probably don't need to use slang regardless of what group you're addressing. One alternative would be "Let's hit the club!" |
39,643 | What is another way of saying, "raise the roof"? This slang phrase means something like, "get noisy and have a good time at a party," but it doesn't sound correct for some reasons.
Why is that? What would be better? | 2011/08/27 | [
"https://english.stackexchange.com/questions/39643",
"https://english.stackexchange.com",
"https://english.stackexchange.com/users/5877/"
] | You might [get down](http://en.wiktionary.org/wiki/get_down), especially if you plan to dance. Or, if you want to make a lot of noise, you might intend to [rock the house](http://en.wiktionary.org/wiki/rock_the_house).
Slang terms for partying tend to change along with pop culture, which means that while suggesting you get [slizzard](http://slizzard.urbanup.com/5134275) now is popular due to recent mentions in music, it'll be uncool soon enough; calling something [crunk](http://onlineslangdictionary.com/meaning-of/crunk) probably sounds outdated; and suggesting you [get the party started](http://en.wikipedia.org/wiki/Get_the_Party_Started) might get you an eye-roll unless your friends are fans of Pink.
If you're looking for something more timeless, try the simple version: *let's party!* | It seems "lame" because it's antiquated -- the only thing that spreads as quickly is slang is the knowledge that a slang term has become "uncool."
"Raise the roof" is a dance move, usually used (the key word here is USED -- it's completely out of favor) in hip hop, in which you push the palm of your hands towards the ceiling as if you're raising and then lowering the roof over and over.
"Raise the roof" is also a phrase that expresses the sentiment that a group intends to "go out". It can also just be an expression of excitement. Depending on which meaning you're using, there are many ways to say it, and you probably don't need to use slang regardless of what group you're addressing. One alternative would be "Let's hit the club!" |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | Even **"Assalam o Alaikum"** could be used for women, it is more convenient to use **"Assalam o Alaikun"** ( without the appended -a ); unless you want to say another thing else without stopping, then you can use **"Assalam o Alikuna** (oh mothers/sisters or something else)". I wish it helps. | بسم الله الرحمن الرحیم
Kum (کم) means plural you (‘men’ or 'men & women')
Kuna (کنَّ) means plural you (only women)
Yet in conversations it is common to say Alaikum, however if we are to speak with eloquence to an only women group, then say Alaikuna is more correct yet uncommon, especially in non-Arab dialogues. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | In regards to greeting in Arabic (for females):
According to Arabic grammar, KOM is used for men, and Kon (Konna) is used for women. As a result if you’d like to use the correct grammatical shape of it, you ought to say “Assalam o Alaikon (Konna) which is written like the following phrase:
>
> السلام علیکن
>
>
>
---
Of course, there is another nice point, and it is related to singular and plural nouns. Actually “Kon” is related to plural noun (according to the correct Arabic grammar).
---
But is the singular noun it is writes:
>
> السلام علیک / Assalam o Alaike (For female)
>
>
> | بسم الله الرحمن الرحیم
Kum (کم) means plural you (‘men’ or 'men & women')
Kuna (کنَّ) means plural you (only women)
Yet in conversations it is common to say Alaikum, however if we are to speak with eloquence to an only women group, then say Alaikuna is more correct yet uncommon, especially in non-Arab dialogues. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | If you want to get it really right, then you need to say:
* To one woman: as-salāmu ʻalayki (or ʻalayk)
* To two women: as-salāmu ʻalaykumā
* To three or more women: as-salāmu ʻalaykunna (or ʻalaykunn)
* To a mixed group (men and women): as-salāmu ʻalaykum | بسم الله الرحمن الرحیم
Kum (کم) means plural you (‘men’ or 'men & women')
Kuna (کنَّ) means plural you (only women)
Yet in conversations it is common to say Alaikum, however if we are to speak with eloquence to an only women group, then say Alaikuna is more correct yet uncommon, especially in non-Arab dialogues. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | Even **"Assalam o Alaikum"** could be used for women, it is more convenient to use **"Assalam o Alaikun"** ( without the appended -a ); unless you want to say another thing else without stopping, then you can use **"Assalam o Alikuna** (oh mothers/sisters or something else)". I wish it helps. | As much as I see, it can use for both men and women in informal form. But if you are looking for the formal form of the greeting or saying salam for women, you should be aware that if there are more than 2 women, you should say Assalamo Alaikon.
There is another important matter that you can note, and that is about two women. In fact if there are two women, you should say Assalamo alaikoma in the form of formal and in fact according to the true grammar of Arabic. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | If you want to get it really right, then you need to say:
* To one woman: as-salāmu ʻalayki (or ʻalayk)
* To two women: as-salāmu ʻalaykumā
* To three or more women: as-salāmu ʻalaykunna (or ʻalaykunn)
* To a mixed group (men and women): as-salāmu ʻalaykum | Even **"Assalam o Alaikum"** could be used for women, it is more convenient to use **"Assalam o Alaikun"** ( without the appended -a ); unless you want to say another thing else without stopping, then you can use **"Assalam o Alikuna** (oh mothers/sisters or something else)". I wish it helps. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | In regards to greeting in Arabic (for females):
According to Arabic grammar, KOM is used for men, and Kon (Konna) is used for women. As a result if you’d like to use the correct grammatical shape of it, you ought to say “Assalam o Alaikon (Konna) which is written like the following phrase:
>
> السلام علیکن
>
>
>
---
Of course, there is another nice point, and it is related to singular and plural nouns. Actually “Kon” is related to plural noun (according to the correct Arabic grammar).
---
But is the singular noun it is writes:
>
> السلام علیک / Assalam o Alaike (For female)
>
>
> | As much as I see, it can use for both men and women in informal form. But if you are looking for the formal form of the greeting or saying salam for women, you should be aware that if there are more than 2 women, you should say Assalamo Alaikon.
There is another important matter that you can note, and that is about two women. In fact if there are two women, you should say Assalamo alaikoma in the form of formal and in fact according to the true grammar of Arabic. |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | If you want to get it really right, then you need to say:
* To one woman: as-salāmu ʻalayki (or ʻalayk)
* To two women: as-salāmu ʻalaykumā
* To three or more women: as-salāmu ʻalaykunna (or ʻalaykunn)
* To a mixed group (men and women): as-salāmu ʻalaykum | In regards to greeting in Arabic (for females):
According to Arabic grammar, KOM is used for men, and Kon (Konna) is used for women. As a result if you’d like to use the correct grammatical shape of it, you ought to say “Assalam o Alaikon (Konna) which is written like the following phrase:
>
> السلام علیکن
>
>
>
---
Of course, there is another nice point, and it is related to singular and plural nouns. Actually “Kon” is related to plural noun (according to the correct Arabic grammar).
---
But is the singular noun it is writes:
>
> السلام علیک / Assalam o Alaike (For female)
>
>
> |
24,593 | Normally we say "Assalam o Alaikum", but my sir said that for women we need to say "Assalam o Alaikuna". Is that right? | 2015/05/30 | [
"https://islam.stackexchange.com/questions/24593",
"https://islam.stackexchange.com",
"https://islam.stackexchange.com/users/12586/"
] | If you want to get it really right, then you need to say:
* To one woman: as-salāmu ʻalayki (or ʻalayk)
* To two women: as-salāmu ʻalaykumā
* To three or more women: as-salāmu ʻalaykunna (or ʻalaykunn)
* To a mixed group (men and women): as-salāmu ʻalaykum | As much as I see, it can use for both men and women in informal form. But if you are looking for the formal form of the greeting or saying salam for women, you should be aware that if there are more than 2 women, you should say Assalamo Alaikon.
There is another important matter that you can note, and that is about two women. In fact if there are two women, you should say Assalamo alaikoma in the form of formal and in fact according to the true grammar of Arabic. |
15,984 | We are building a web service(SOAP, .Net) which would be talking to (mostly) native clients (windows, C++) and we are wondering what is the best way to communicate errors to the client (e.g. SomethingBadHappened like login service not available or something like user not found) and haven't been able to decide between throwing exception to the client or using some kind of error code model to do the above.
What you would prefer on the handling on the client side: receiving a error code or handling a ServerFault exception which contains the reason for the error?
1) Why are we thinking exception: Because it would make server side code a lot more uniform
2) Why are we thinking error codes: Because we *think* it makes more sense from the client side perspective.
If 2) is really true we would probably want to go for error codes than exceptions? Is that the case here?
Also, would the answer change if we were talking to managed clients instead of native clients? | 2010/11/01 | [
"https://softwareengineering.stackexchange.com/questions/15984",
"https://softwareengineering.stackexchange.com",
"https://softwareengineering.stackexchange.com/users/3755/"
] | SOAP has a concept of [faults](http://www.w3schools.com/soap/soap_fault.asp), you can convert an exception to a fault on the server side and on the client proxy the fault can again be converted back to an exception. This works remarkably well in WCF and Java metro stack, cannot comment on native C++ clients.
As regards to SOA best practice define one generic fault and few specific faults only if the client need to handle a certain type of error differently. Never send a exception stack trace to client in production deployment. This is because in theory the server trace has no meaning for the client and for security reasons as well. Log the full error and stacktrace on the server and send a unique reference to the log in the fault. In WCF I use the Microsoft Exception Handling block from Enterprise Library to generate a guid and also convert a exception to SOAP fault.
Check the guidance at [Microsoft Patterns and Practices](http://msdn.microsoft.com/en-us/practices/bb969054.aspx#web_svrcs). | I recently did a web service with the Java 6 libraries, which can report an exception back to the caller (I haven't looked into how as it is done automatically).
The ability to have the client provide a stack trace in the error report to the developer has been very useful (opposed to getting an approximate timestamp and then have to look it up in your logs, if you happen to log it).
So, seen from a developers point of view, use Exceptions. |
162,334 | I'm looking to attach three posts the the side of my house and I'm wondering if a ledger board is needed?
Looks like this:
[](https://i.stack.imgur.com/wCjBS.jpg) | 2019/04/18 | [
"https://diy.stackexchange.com/questions/162334",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/99657/"
] | Wood siding on a house is there for weather protection and looks, it has little structural strength. The ledger board is used to connect the patio cover beams to the internal structural members of the house walls for strength against wind shear forces and (possibly) snow loads that will want to pull your cover away from the house, ***spreading that force across as many internal structural members (studs) as possible.*** If you just attach the beams directly to the studs, you are attaching only to one stud and essentially concentrating the mechanical forces onto fewer points, increasing the stresses they will see. If they fail, it becomes major damage to your house walls. | If your connection points line up with the floor assembly for the floor above you can fasten to the rim joist.
If not you need someway to support half the weight of the structure at the house wall side. There are different approaches. Ledger is one. Posts for independent structure is another. If you have windows with headers that match the height of the connection point this can also work.
In the picture you've posted it is probable that there is a header over the sliding glass door. If they had this overhang in mind from the beginning they might have put blocking at the level required that spans between studs for the other connection points. Often studs will have fire blocking between them which tends to make them act as a structural system instead of just a single stud which is also tied to the top and bottom plates. |
5,582 | [This comment](http://buddhismnow.com/2012/08/17/compassion-without-drowning/#comment-3274) says,
>
> True the world becomes more SPACIOUS when we focus on the moment and thus see that to save one life is to save the world. This is why it is important to translate spaciousness as spaciousness and not emptiness.
>
>
>
Are "spacious" and "empty" words that are confused or mis-translated?
Is "spacious" a word that has a use in Buddhism?
A search for "spacious" on Access to Insight suggests that word is used, but is not used often and not used canonically.
I think "empty" usually means "empty of self", am I right? I.e. when people say "empty" do they mean [Anatta](http://en.wikipedia.org/wiki/Three_marks_of_existence#Anatta)? | 2014/11/20 | [
"https://buddhism.stackexchange.com/questions/5582",
"https://buddhism.stackexchange.com",
"https://buddhism.stackexchange.com/users/254/"
] | More like shunyata instead of anatta. The former has a philosophical as well as an experiential component. It speaks to the inherent emptiness of all things as well as the spacious feelings that often accompany the direct realization of that emptiness. The etymology of the word shunyata also conveys a notion of "hollowness" and has actually been translated as "emptiness", "voidness", and "spaciousness".
Side note - I'd be careful not to confuse the perception of shunyata with the experiences of the formless attainments (namely the base of infinite space and the base of nothingness). Those are very different things; even the Pali/Sanskrit roots of the word are vastly different. | Simplicity is considered virtuous.
Spaciousness is the accompaniment of emptiness.
But all are perceptive; objectively, neither spaciousness nor emptiness exists, because we are filled with space, or we are filled with something else. There is a dimensional zero as much as there is a dimensional one. But in the abstract sense, both are numbers, and therefore the same.
Truly accomplishing yourself means that you understand that the perceptions are illusions; but the illusions are what we have to work with, and so are the only truth we have.
Hope that makes sense. |
23,947 | I am using Gmail.
I know how to forward an email received from A to B, by simply choosing the forward option for the email to be forwarded, and putting B's address there.
But I wonder how to forward an email received from A to B, within a reply to another email received from B? Or what do people usually do alternatively in such a case? | 2012/02/16 | [
"https://webapps.stackexchange.com/questions/23947",
"https://webapps.stackexchange.com",
"https://webapps.stackexchange.com/users/4079/"
] | I think copy and paste is your only option here. There is (to my knowledge) no way of attaching another email to a new one in gmail. | Best way I found is to hit 'print all' in the email you want to forward and save as PDF. Then simply attach the PDF to your reply. |
23,947 | I am using Gmail.
I know how to forward an email received from A to B, by simply choosing the forward option for the email to be forwarded, and putting B's address there.
But I wonder how to forward an email received from A to B, within a reply to another email received from B? Or what do people usually do alternatively in such a case? | 2012/02/16 | [
"https://webapps.stackexchange.com/questions/23947",
"https://webapps.stackexchange.com",
"https://webapps.stackexchange.com/users/4079/"
] | I think copy and paste is your only option here. There is (to my knowledge) no way of attaching another email to a new one in gmail. | Email clients usually group emails by title (and other stuff).
So you are able to forward an email in Gmail, **as a reply in another thread**, if you **edit the subject** to the same exact subject on the other thread (this will maintain images and other attachments properly!).
In this case, your are actually **moving** your email, so it won't show as a "forwarded email" in the original thread.
Just confirmed it works, so it might help other people. It should work with other clients such as Outlook too.
*I hate receiving emails > with attached emails > with attached other stuff* |
106,426 | Two-row connector blocks are ubiquitous in small-run and hobbyist PCB designs. Pitches of 0.1" or 2mm are the most common, and there are all sorts of options for board-to-board or board-to-wire connectors. The problem is that pin numbering schemes vary wildly.
Pin numbering schemes I've seen include:
1. DIP-style counterclockwise: 
2. Zig-zag between the two rows: 
3. Down one row then the other, but NOT counterclockwise: 
In my experience, getting the pin numbers wrong is one of the most common sources of PCB heartbreak. I've done a lot of wire patching to correct 6-pin header layout mistakes... and I've thrown away boards with 70-pin header errors.
My question is: **Are there industry standard terms for these (or other) 2-row pin numbering schemes?**
I would love to have a short, reliable way to specify these schemes, to allow saying things like "We're using a 14-pin 0.1" header with *zig-zag* numbering" (or whatever).
Although standardized names with the broadest possible user base are (obviously) better, I realize that there probably isn't one true answer to this question. I'd appreciate answers about how these schemes are named even within smaller subsets of the industry, as long as you're clear about the scope of use (e.g. "At company X we always called the second scheme 'shoelacing'"). | 2014/04/13 | [
"https://electronics.stackexchange.com/questions/106426",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/40307/"
] | If not the connector is made to a certain specification, say IEC60603-2, there is nothing that says that the manufacturer can't make it anyway he like.
And I have not seen a "default" standard yet. | If You are using/connecting flat/ribbon cable than pin numbering must be zig-zag, because of cable construction.
It is also depending on software you are using for develop pcb.
In my experience in repairing boards, more or less 1st pin is somehow visibile, other wires mark as for your convenience.
Happy HW connecting. |
106,426 | Two-row connector blocks are ubiquitous in small-run and hobbyist PCB designs. Pitches of 0.1" or 2mm are the most common, and there are all sorts of options for board-to-board or board-to-wire connectors. The problem is that pin numbering schemes vary wildly.
Pin numbering schemes I've seen include:
1. DIP-style counterclockwise: 
2. Zig-zag between the two rows: 
3. Down one row then the other, but NOT counterclockwise: 
In my experience, getting the pin numbers wrong is one of the most common sources of PCB heartbreak. I've done a lot of wire patching to correct 6-pin header layout mistakes... and I've thrown away boards with 70-pin header errors.
My question is: **Are there industry standard terms for these (or other) 2-row pin numbering schemes?**
I would love to have a short, reliable way to specify these schemes, to allow saying things like "We're using a 14-pin 0.1" header with *zig-zag* numbering" (or whatever).
Although standardized names with the broadest possible user base are (obviously) better, I realize that there probably isn't one true answer to this question. I'd appreciate answers about how these schemes are named even within smaller subsets of the industry, as long as you're clear about the scope of use (e.g. "At company X we always called the second scheme 'shoelacing'"). | 2014/04/13 | [
"https://electronics.stackexchange.com/questions/106426",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/40307/"
] | If not the connector is made to a certain specification, say IEC60603-2, there is nothing that says that the manufacturer can't make it anyway he like.
And I have not seen a "default" standard yet. | Pin 1 generally gets a square pad on the PCB as an index marker, whereas those of the other pins are generally round. |
106,426 | Two-row connector blocks are ubiquitous in small-run and hobbyist PCB designs. Pitches of 0.1" or 2mm are the most common, and there are all sorts of options for board-to-board or board-to-wire connectors. The problem is that pin numbering schemes vary wildly.
Pin numbering schemes I've seen include:
1. DIP-style counterclockwise: 
2. Zig-zag between the two rows: 
3. Down one row then the other, but NOT counterclockwise: 
In my experience, getting the pin numbers wrong is one of the most common sources of PCB heartbreak. I've done a lot of wire patching to correct 6-pin header layout mistakes... and I've thrown away boards with 70-pin header errors.
My question is: **Are there industry standard terms for these (or other) 2-row pin numbering schemes?**
I would love to have a short, reliable way to specify these schemes, to allow saying things like "We're using a 14-pin 0.1" header with *zig-zag* numbering" (or whatever).
Although standardized names with the broadest possible user base are (obviously) better, I realize that there probably isn't one true answer to this question. I'd appreciate answers about how these schemes are named even within smaller subsets of the industry, as long as you're clear about the scope of use (e.g. "At company X we always called the second scheme 'shoelacing'"). | 2014/04/13 | [
"https://electronics.stackexchange.com/questions/106426",
"https://electronics.stackexchange.com",
"https://electronics.stackexchange.com/users/40307/"
] | If You are using/connecting flat/ribbon cable than pin numbering must be zig-zag, because of cable construction.
It is also depending on software you are using for develop pcb.
In my experience in repairing boards, more or less 1st pin is somehow visibile, other wires mark as for your convenience.
Happy HW connecting. | Pin 1 generally gets a square pad on the PCB as an index marker, whereas those of the other pins are generally round. |
3,542,988 | I am working on CRC64 Reversing algorithm on C# but unable to code it. If anyone can help me out, it will be nice of him. I am desperate to complete the coding. | 2010/08/22 | [
"https://Stackoverflow.com/questions/3542988",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/-1/"
] | That's not possible. CRC is a [hash function](http://en.wikipedia.org/wiki/Hash_function). Hash functions **cannot** be reversed. The best you can do is try and find collisions for a given hash, but they are *designed* to be hard to find such collisions.. | You definitely won't be able to reverse it, since you lose critical data doing the "forward" computation that would be necessary to the "backward" computation. You can try to brute-force it if you have some idea of what the input might have been, but because a CRC is weak there might be a lot of collisions (meaning you might find lots of possible inputs that could have generated your result) in which case you have no way of knowing which one is correct. |
3,542,988 | I am working on CRC64 Reversing algorithm on C# but unable to code it. If anyone can help me out, it will be nice of him. I am desperate to complete the coding. | 2010/08/22 | [
"https://Stackoverflow.com/questions/3542988",
"https://Stackoverflow.com",
"https://Stackoverflow.com/users/-1/"
] | That's not possible. CRC is a [hash function](http://en.wikipedia.org/wiki/Hash_function). Hash functions **cannot** be reversed. The best you can do is try and find collisions for a given hash, but they are *designed* to be hard to find such collisions.. | Pretty sure he's asking for a 'reflected' algorithm.
The 'standard' and 'reflected' algorithms are an exact mirror
image of each other, including the polynomial used in each.
A message created/checked on one-endian system should not be
reversed on other-endian system as this can harm the checks
ability to identify burst errors. An 'mirror' algorithm on
the other-endian system will produce a check that exactly
matches the transmitted check value. Not sure if this code
is still needed? I could probably do it for you with the
original code to be reflected. |
873,146 | Does anyone know how to boot Linux "Ubuntu" from a Usb Stick when there is no Legacy boot option In the Bios I have Microsoft 8.1 and want to replace it. | 2015/02/03 | [
"https://superuser.com/questions/873146",
"https://superuser.com",
"https://superuser.com/users/415157/"
] | Actually, the 32-bit UEFI is not that big a problem. Here is how I set up my Asus EeeX205T to dual-boot Ubuntu 14.10. Note that the process does require a couple of steps in Windows:
PRE-INSTALLATION
Turn off BitLocker (this will allow Ubuntu to mount and/or resize the primary data drive):
In Windows, click on PC Settings/PC and devices/PC info; scroll to the bottom of the screen and click to turn off BitLocker/decrypt drive
Prepare Ubuntu Live bootable USB memory stick:
Download .iso file for 64-bit Ubuntu 14.10
Create a bootable memory stick:
In Windows, download and run Rufus
In Linux, sudo apt-get install unetbootin; run
After the memory stick is set up, copy bootia32.efi to the /EFI/BOOT directory on the memory stick
Fetch bootia32.efi from github (?) or from one of the links listed at the end of this post.
Alternately, instructions for generating this file can be found here on github -- look for JF Well's site on the t100.
Boot Ubuntu Live:
Insert memory stick in Asus EEE USB port
In Windows, hold Shift key and click on Restart; choose Advanced Options/UEFI Firmware to enter BIOS
Turn off SecureBoot
Select USB stick as boot device
Save and exit
When grub menu loads, select Try Ubuntu
Optional but highly recommended: make a backup of the complete SSHD image
Plug in a USB external hard disk with at least 32GB of free space
Right-click on the OS and Recovery icons in the Unity launcher and choose Unmount for each
Open a terminal and run the following: dd if=/dev/mmcblk0 of=/media/ubuntu/HD/FILENAME bs=1M
(Replace HD with the name by which the external drive was mounted; replace FILENAME with the filename you wish to use for the resulting image (or set another path as desired).)
It will take about 20 minutes for the command to complete; do NOT interrupt it, or you will have to start over!
Note: to restore the X205T to its original state, you can run this command “in reverse”: dd if=/media/ubuntu/HD/FILENAME of=/dev/mmcblk0 bs=1M -- and yes, I have done this to verify that it works!
Optional: resize Recovery partition (note that the Recovery partition is required for booting Windows on this machine, so it cannot be removed if you want dual-boot, but it can be reduced in size to allow some additional room):
Right-click on the OS and Recovery icons in the Unity launcher and make sure each is unmounted (choose Unmount for each if needed)
Run gparted (note: this runs very slowly at this point; be patient!)
Shrink the Recovery partition as small as it will go (a little less than 6 GB)
Expand the OS partition to take up the extra room
INSTALLATION (from Ubuntu Live)
Set up a wireless connection (to download updates during installation):
Insert a USB based wifi adapter (Ubuntu does not yet recognize the internal wifi)
Choose a wireless connection and connect
Double-click on Install Ubuntu
Select language; continue
Click to install additional drivers and restricted software; continue
Choose Install Ubuntu alongside Windows; continue
Set relative sizes of Ubuntu/Windows partitions as desired (I set mine up with 8 GB for Windows and 13 GB for Ubuntu); continue
Allow the installation to complete, but do not (yet) allow it to reboot
Open a terminal and enter the following commands:
sudo -s // sets up a root shell
mount /dev/mmcblk0p5 /mnt // make sure that this is the partition to which
// Ubuntu was installed
mount /dev/mmcblk0p1 /mnt/boot/efi // make sure this is the EFI partition
for I in /dev /dev/pts /proc /sys /run; do sudo mount -B $i /mnt$i; done
// transfer key elements of the current (Ubuntu
// Live) file system to the newly mounted
// Ubuntu partition on /mnt
mount -o bind /etc/resolv.conf /mnt/etc/resolv.conf
// likewise make the current resolv.conf available
// via the mounted Ubuntu partition on /mnt
chroot /mnt /bin/bash // set up /mnt as the new root and run bash
apt-get remove grub-efi-amd64 // remove the 64-bit grub …
apt-get install grub-efi-ia32 // and install the 32-bit grub
Reboot; the machine should boot to the grub menu, allowing a choice between Ubuntu and Windows. If so, congratulations – you have succeeded! If not, and if you made the recommended backup, see the note above for how to restore to the original state. If you did not make the backup … you may have a brick on your hands.
FINALIZE INSTALLATION
Clean up Windows:
Reboot to grub menu and choose Windows
Allow Windows to check and “repair” drive C: (basically Windows is responding to the fact that the drive has been re-sized)
When finished, it will automatically reboot; choose Windows again to verify that the Windows installation is still functioning properly
Set up SDHC card reader:
Reboot to grub menu and choose Ubuntu
Open a terminal
sudo gedit /etc/modprobe.d/sdhci.conf (note: this file may not yet exist; if not, create it)
Add the following line: options sdhci debug\_quirks=0x8000
Save and exit
sudo update-initramfs -u -k all
Reboot; SDHC card reader should now work
At this point, you should have a mostly working dual-boot machine. What doesn’t work yet (hopefully a new kernel down the road will fix these issues):
Internal wifi (have to use a USB wifi adapter for now)
Bluetooth (it sort-of sees the bluetooth, but not completely)
Sound (volume control appears to work, but it is a “dummy” sound system)
ACPI (battery monitor, power-saving, suspend)
What does work at this point (mostly):
USB ports
SDHC card reader (if you do the tweak above)
Screen brightness (from System settings, but not from the keyboard)
Screen turn-off after specified time; lock when screen turns off
Note that sometimes after the screen has turned off / system has locked, it will not return to the log-in screen when a key is pressed. In this case, it is possible to call up a terminal screen (Ctrl-Alt-F1), log in, and re-start the X server (sudo restart lightdm), but any open programs will be closed and any unsaved data may be lost
Built-in camera
Optional: set up micro-SD card as /home
Insert a micro-SD card and format it to ext4
One way to do this is to install gparted (sudo apt-get install gparted); note that you will need a working wifi (see above)!
Note that you can also partition the micro-SD card and install one partition as /home, and another partition as something else
Copy all of the contents of /home to the SD card (or desired partition)
Determine the UUID of the SD card or the desired partition (sudo blkid)
sudo gedit /etc/fstab and add the following line:
UUID=XXXX /home ext4 defaults 0 2 (replace XXXX with the UUID noted above)
Reboot
Optional: make the Recovery partition non-mountable
Use sudo blkid to determine the UUID of the Recovery partition
sudo mkdir /mnt/Recovery
sudo gedit /etc/fstab and add the following line:
UUID=XXXX /mnt/Recovery ntfs noauto,umask=222 0 2 (replace XXXX with the UUID determined above)
Reboot
Continue setup of Ubuntu as desired!
HELPFUL REFERENCES
<http://www.jfwhome.com/2014/03/07/perfect-ubuntu-or-other-linux-on-the-asus-transformer-book-t100/>
<http://www.linuxforen.de/forums/showthread.php?276316-bootbaren-USB-Stick-f%FCr-32bit-UEFI-erstellen&p=1820224&viewfull=1#post1820224> | In general, the Asus x205T isn't Linux compatible due to its 32-bit UEFI implementation.
This question over at the AskUbuntu Stack Exchange site covers it:
[Creating a boot disc and Installing ubuntu ASUS X205T(A)](https://askubuntu.com/questions/560741/creating-a-boot-disc-and-installing-ubuntu-asus-x205ta)
Highlights:
>
> The short answer is, you can't do it. Your laptop isn't the only one. I recall reading the reason is an incompatibility with the linux kernel and the combination of a 64bit processor but a 32bit uefi like these laptops have.
>
>
>
.
>
> the only way of getting linux to boot on one of these is to use a 64bit distro with a 32bit boot loader. so, long story short, you're stuck with windoze, which is about the only os that will boot.
>
>
>
.
>
> I was able to get ubuntu to boot using a 32 bit grub with 64 bit operating system. The x205t isnt linux incompatible. Linux boots, but with a few bugs.
>
>
>
There are also links provided there to other sites with information on how to do the 32-bit GRUB with a 64-bit Linux install.
This one seems to have the meat of it: <http://www.jfwhome.com/2014/03/07/perfect-ubuntu-or-other-linux-on-the-asus-transformer-book-t100/> |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | Sometimes you can knock them out by striking the wood furniture next to the hole, either with your hand or with a hammer. If using a hammer, put a small piece of wood (like, a chunk of 2x4) against your furniture, then strike this wood. That way it won't mess up the finish on your furniture.
Just a few taps. Don't hit it so hard that you break your other connections! :)
Sometimes it works to hit the same face of the wood which has the hole drilled into it (right next to the hole). Other times it works if works if you tap the furniture directly behind the hole where the nut is buried. Give them both a try. | You can try a really strong magnet. If that doesn't work you can try drilling it. If it starts spinning then reduce the speed of the drill and try to insert your screw while its spinning. If it doesn't spin and you drill through to the open area then you should be able to pull it out with a coat hanger with the end bent. |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | You can try a really strong magnet. If that doesn't work you can try drilling it. If it starts spinning then reduce the speed of the drill and try to insert your screw while its spinning. If it doesn't spin and you drill through to the open area then you should be able to pull it out with a coat hanger with the end bent. | Stick a piece of double-sided mounting tape on the end of a dowel or large-headed nail and press it against the cam lock. Pull it out. If you don't have ready access to mounting tape, use a small dot of hot melt glue in the center of the dowel end (if the cam lock is buried inside a hole, hot melt glue could spread out when you press the dowel, and glue in the dowel, so not the best solution in that case). |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | You can try a really strong magnet. If that doesn't work you can try drilling it. If it starts spinning then reduce the speed of the drill and try to insert your screw while its spinning. If it doesn't spin and you drill through to the open area then you should be able to pull it out with a coat hanger with the end bent. | Use a Dremel moto-tool with a cutting disc to cut a slot in the blank face.
DONE.
Now the cam lock is double faced. |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | Sometimes you can knock them out by striking the wood furniture next to the hole, either with your hand or with a hammer. If using a hammer, put a small piece of wood (like, a chunk of 2x4) against your furniture, then strike this wood. That way it won't mess up the finish on your furniture.
Just a few taps. Don't hit it so hard that you break your other connections! :)
Sometimes it works to hit the same face of the wood which has the hole drilled into it (right next to the hole). Other times it works if works if you tap the furniture directly behind the hole where the nut is buried. Give them both a try. | I'd start with Platinum Goose's drilling suggestion. Otherwise, a drop of super glue on the end of a steel dowel with a very flat surface would do well. |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | Sometimes you can knock them out by striking the wood furniture next to the hole, either with your hand or with a hammer. If using a hammer, put a small piece of wood (like, a chunk of 2x4) against your furniture, then strike this wood. That way it won't mess up the finish on your furniture.
Just a few taps. Don't hit it so hard that you break your other connections! :)
Sometimes it works to hit the same face of the wood which has the hole drilled into it (right next to the hole). Other times it works if works if you tap the furniture directly behind the hole where the nut is buried. Give them both a try. | Stick a piece of double-sided mounting tape on the end of a dowel or large-headed nail and press it against the cam lock. Pull it out. If you don't have ready access to mounting tape, use a small dot of hot melt glue in the center of the dowel end (if the cam lock is buried inside a hole, hot melt glue could spread out when you press the dowel, and glue in the dowel, so not the best solution in that case). |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | Sometimes you can knock them out by striking the wood furniture next to the hole, either with your hand or with a hammer. If using a hammer, put a small piece of wood (like, a chunk of 2x4) against your furniture, then strike this wood. That way it won't mess up the finish on your furniture.
Just a few taps. Don't hit it so hard that you break your other connections! :)
Sometimes it works to hit the same face of the wood which has the hole drilled into it (right next to the hole). Other times it works if works if you tap the furniture directly behind the hole where the nut is buried. Give them both a try. | Use a Dremel moto-tool with a cutting disc to cut a slot in the blank face.
DONE.
Now the cam lock is double faced. |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | I'd start with Platinum Goose's drilling suggestion. Otherwise, a drop of super glue on the end of a steel dowel with a very flat surface would do well. | Stick a piece of double-sided mounting tape on the end of a dowel or large-headed nail and press it against the cam lock. Pull it out. If you don't have ready access to mounting tape, use a small dot of hot melt glue in the center of the dowel end (if the cam lock is buried inside a hole, hot melt glue could spread out when you press the dowel, and glue in the dowel, so not the best solution in that case). |
139,305 | I messed up trying to assemble an IKEA bed. I accidentally pushed the cam lock with the smooth side facing out. Now I have no way to rotate it to tighten the cam lock and bolt to secure the two pieces of wood. How do I remove it? I have attached some images.[](https://i.stack.imgur.com/INAzB.png)
The part numbered 104895 is facing the wrong way at one location. Is there a way I can extract it? | 2018/05/21 | [
"https://diy.stackexchange.com/questions/139305",
"https://diy.stackexchange.com",
"https://diy.stackexchange.com/users/86025/"
] | I'd start with Platinum Goose's drilling suggestion. Otherwise, a drop of super glue on the end of a steel dowel with a very flat surface would do well. | Use a Dremel moto-tool with a cutting disc to cut a slot in the blank face.
DONE.
Now the cam lock is double faced. |
26,239 | Wikimedia Commons has two images of [Atilla the Hun](https://commons.wikimedia.org/wiki/Attila_the_Hun) from the Chronicon Pictum, dated around 1360. Is there anything (coin, portrait, bust, statue) older than that? | 2015/11/05 | [
"https://history.stackexchange.com/questions/26239",
"https://history.stackexchange.com",
"https://history.stackexchange.com/users/8103/"
] | It is not surprising that you ask about this topic for it is a very little studied phase of the complicated relationship between Nazi Germany and the Soviet Union. As you mentioned, many communists were gravely troubled by the rapprochement between Hitler and Stalin, leading to widespread disaffection in many of the Western Communist Parties. I would highly commend to you Roger Moorehouse’s “The Devils’ Alliance” (Hitler’s Pact with Stalin, 1939-1941; 2014) as one of the few references dealing with the 22 month pre-Barbarosa collaboration between these ideological enemies.
Prior to the Ribbentrop Molotov pact, the KPD had been
“… outlawed since the Nazi seizure of power in 1933, it had been forced underground, its members subject to arrest and persecution, and given only limited succour via the often tortuous lines of communication with their superiors Moscow. The fate of its leader, Ernst Thalmann, was indicative of how far the party had fallen. Once the giant of the political scene who had contested the presidency in 1925 and 1932, Thalmann was arrested by the Gestapo barely a month after Hitler came to power. Kept in solitary confinement, he was repeatedly questioned, abused and beaten-losing 4 teeth in one interrogation-but never granted the dignity of a trial. He simply disappeared, shunted between a succession of prisons and concentration camps from which he would never reemerge.”
“By 1939, the German Communists have been reduced to an underground fringe movement, isolated and largely swimming against the tide of public opinion, with its lines of command fractured, compromised and unreliable. Little wonder that the Nazi Soviet Pact was viewed with utter bewilderment in German Communists circles. Officially, at least, it was greeted as potential lifeline with the party announcing its approval of the pact as a “blow for peace” and expressing the hope that further, similar pacts would follow. Some Communists went further, speculating that the pact would signal an end to the persecution with the expectation that they would soon be able to hold their meetings without hindrance and that Thalmann and other prisoners would be released.”
From the tenor of your question, it appears that you reasonably suspect that this situation might have changed for the better, at least from the Communists’ point of view. After all the “line emanating from Moscow… came perilously close to advocating a political truce with Nazism, with communist energies instead to be focused on attacking the Western Powers as the true enemies of world revolution.” From Moscow, Walter Ulbricht, who later achieved considerable success in East German politics, “… blamed the war squarely on capitalism and “big business”, and branded British imperialism as more reactionary and more dangerous than Nazi imperialism, indeed as the “most reactionary force in the world”. Similarly, Izvestia ridiculed the West’s “war on Hitlerism” while the KPD’s official newspaper explained the rapid conquest of France and the Low Countries as the result of “the baleful politics of the ruling classes in England and France and their social democratic lackeys…”
However, Hitler was not to be mollified, his hatred of Communism never flagged, and “the Gestapo’s attention had scarcely lessened…” while most Communists lapsed into inaction with confiscations of communist leaflets declining from a monthly average of 1000 prior to the pact down to 82 by spring of 1940 and arrests of communists, declining from over 950 in 1937 to a mere 70 in April 1940; so that in June 1940 the SS could “no longer speak of organized resistance from Communist and Marxist circles”.
Accordingly, it seems that the short answer to your question is that the Nazi state brutally suppressed the KPD from the moment they acquired power, this continued through the 22 months of the pact, and that the only reason for the decline in arrests and confiscations of literature during the 22 months of the pact was the utter passivity of the KPD in that time frame leading “… one prominent historian of the period… [to describe]… the German Communists of that era as ‘the most shameful of Hitler’s accomplices’”. | In a now-deleted answer @Alex has correctly pointed out that Stalin handed over to the Nazis many German communists who had sought asylum in the USSR.
Other users have asked to name some people who have been handed over. Well, for example:
* [Margarete Buber-Neumann](http://Margarete%20Buber-Neumann)
* [Fritz Houtermans](https://en.wikipedia.org/wiki/Fritz_Houtermans)
* [Hans Walter David](https://de.wikipedia.org/wiki/Hans_Walter_David)
Some more names and details are given [here](http://www.e-reading.club/chapter.php/1015793/39/Chernaya_kniga_kommunizma_Prestupleniya._Terror._Repressii.html) and [here](http://trst.narod.ru/rogovin/t7/ii_vii.htm).
All in all, it seems that about 500 communists were handed over by the USSR more or less directly to the care of the Gestapo (or 1000, it is possible that 500 refers to a first large tranche, I haven't got time to research it more fully right now). |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.