text
stringlengths
0
30.5k
title
stringclasses
1 value
embeddings
listlengths
768
768
all costs. I've long since completed the project which prompted this question, but recently I've had another project come along with very minor data requirements, so I spent some more time experimenting with this. I had assumed that Sql Server Express required licensing fees to deploy, but this is not in fact the case. According to Microsoft's website, you are free to use it with certain restrictions: * Maximum database size: 4 GB * Maximum memory used: 1 GB * Maximum CPUs used: 1 (complete procs, not cores) Sql Server Compact is a bad idea for web applications because it requires a hack to make it
[ -0.13365596532821655, 0.0842062309384346, 0.5224185585975647, 0.30801481008529663, 0.2057100087404251, -0.4530908763408661, -0.05619672313332558, 0.15620917081832886, -0.3882986903190613, -0.475042462348938, -0.25134825706481934, 0.46699678897857666, -0.16287700831890106, 0.253125607967376...
work, and it isn't built for the concurrent access you'd need for the web. But if your application can fit within the modest limitations of Sql Server Express, it works pretty well. And since it speaks regular T-SQL like its larger siblings, you can use Linq to SQL with it. I hear that [Linq to Sql support is now in the Mono trunk](http://twitter.com/migueldeicaza/status/2368465000) for the 2.6 release, so L2S' tight-coupling to Sql Server will likely be a moot point in the near future. I will either end up porting my code to use Mono's superior Linq to Sql implementation on the
[ 0.3814506530761719, -0.05997323989868164, 0.40432208776474, 0.027815409004688263, -0.009489637799561024, -0.215229794383049, 0.20338426530361176, -0.18082459270954132, -0.13173392415046692, -0.47200655937194824, -0.03801874816417694, 0.7069669365882874, -0.2384594976902008, -0.193236365914...
db of my choice, or go [another route entirely](http://subsonicproject.com/) (SubSonic has improved by leaps and bounds since I last tried it). But for the time being, Sql Server Express is a valid choice for very small database-driven apps.
[ -0.008391265757381916, -0.0665629506111145, 0.1674550622701645, 0.12447694689035416, -0.46768537163734436, -0.3002796769142151, 0.6237261295318604, 0.028857480734586716, 0.3107801377773285, -0.5055274963378906, 0.06745836138725281, 0.21928691864013672, -0.25134819746017456, -0.042926289141...
I want to escape a DOS filename so I can use it with sed. I have a DOS batch file something like this: ``` set FILENAME=%~f1 sed 's/Some Pattern/%FILENAME%/' inputfile ``` (Note: `%~f1` - expands `%1` to a Fully qualified path name - `C:\utils\MyFile.txt`) I found that the backslashes in `%FILENAME%` are just escaping the next letter. How can I double them up so that they are escaped? (I have cygwin installed so feel free to use any other \*nix commands) Solution ======== Combining Jeremy and Alexandru Nedelcu's suggestions, and using | for the delimiter in the sed command I have ``` set FILENAME=%~f1 cygpath "s|Some Pattern|%FILENAME%|" >sedcmd.tmp sed -f sedcmd.tmp inputfile del /q sedcmd.tmp ``` This will
[ 0.07283664494752884, -0.0021323959808796644, 0.2522534132003784, -0.14767718315124512, 0.16904275119304657, 0.4121309518814087, 0.5940120816230774, -0.7334264516830444, -0.0873609110713005, -0.3830704987049103, -0.21589694917201996, 0.1700345128774643, -0.6457915902137756, 0.22044572234153...
work. It's messy because in BAT files you can't use set var=`cmd` like you can in unix. The fact that echo doesn't understand quotes is also messy, and could lead to trouble if `Some Pattern` contains shell meta characters. ``` set FILENAME=%~f1 echo s/Some Pattern/%FILENAME%/ | sed -e "s/\\/\\\\/g" >sedcmd.tmp sed -f sedcmd.tmp inputfile del /q sedcmd.tmp ``` [Edited]: I am suprised that it didn't work for you. I just tested it, and it worked on my machine. I am using sed from <http://sourceforge.net/projects/unxutils> and using cmd.exe to run those commands in a bat file.
[ 0.6169731020927429, 0.12682019174098969, 0.056705113500356674, -0.2055199146270752, -0.27925145626068115, -0.07888144999742508, 0.5437144637107849, -0.5733827352523804, -0.04852641001343727, -0.5955603122711182, -0.054381273686885834, 0.47428667545318604, -0.4459330439567566, -0.2411962896...
I have always used `||` (two pipes) in OR expressions, both in C# and PHP. Occasionally I see a single pipe used: `|`. What is the difference between those two usages? Are there any caveats when using one over the other or are they interchangeable? Just like the `&` and `&&` operator, the double Operator is a "short-circuit" operator. For example: ``` if(condition1 || condition2 || condition3) ``` If condition1 is true, condition 2 and 3 will NOT be checked. ``` if(condition1 | condition2 | condition3) ``` This will check conditions 2 and 3, even if 1 is already true. As your conditions can be quite expensive functions, you can
[ 0.42909592390060425, 0.01493152603507042, 0.07516171783208847, -0.22050729393959045, -0.07003770768642426, -0.034011151641607285, 0.45759162306785583, -0.1820264458656311, 0.14533638954162598, -0.5084085464477539, -0.07595547288656235, 0.6779773235321045, -0.30789512395858765, -0.078138194...
get a good performance boost by using them. There is one big caveat, NullReferences or similar problems. For example: ``` if(class != null && class.someVar < 20) ``` If class is null, the if-statement will stop after `class != null` is false. If you only use &, it will try to check `class.someVar` and you get a nice `NullReferenceException`. With the Or-Operator that may not be that much of a trap as it's unlikely that you trigger something bad, but it's something to keep in mind. No one ever uses the single `&` or `|` operators though, unless you have a design where each condition is
[ 0.33725669980049133, -0.2205759882926941, 0.08367104083299637, -0.02610628865659237, -0.12103965878486633, -0.43093568086624146, 0.6490790843963623, -0.17828218638896942, -0.39662888646125793, -0.21655845642089844, -0.10840637981891632, 0.8589480519294739, -0.2833051383495331, 0.0252695344...
a function that HAS to be executed. Sounds like a design smell, but sometimes (rarely) it's a clean way to do stuff. The `&` operator does "run these 3 functions, and if one of them returns false, execute the else block", while the `|` does "only run the else block if none return false" - can be useful, but as said, often it's a design smell. There is a Second use of the `|` and `&` operator though: [Bitwise Operations](http://www.c-sharpcorner.com/UploadFile/chandrahundigam/BitWiserOpsInCS11082005050940AM/BitWiserOpsInCS.aspx).
[ 0.37468528747558594, -0.13106808066368103, -0.1138540655374527, -0.10422079265117645, -0.05729229748249054, -0.2059444636106491, 0.4828737676143646, -0.3990095257759094, 0.05669662728905678, -0.28726181387901306, -0.18249592185020447, 0.6101307272911072, -0.48058509826660156, 0.05992031097...
What's the best way to combine results sets from disparate data sources in SSRS? In my particular example, I need to write a report that pulls data from SQL Server and combines it with another set of data that comes from a DB2 database. In the end, I need to join these separate data sets together so I have one combined dataset with data from both sources combined on to the same rows. (Like an inner join if both tables were coming from the same SQL DB). I know that you can't do this "out of the box" in SSRS 2005.
[ 0.12370815128087997, -0.1959114521741867, -0.21623095870018005, 0.041945621371269226, -0.2015015333890915, -0.15697717666625977, 0.011931678280234337, -0.4008030891418457, -0.2999834716320038, -0.4214949607849121, 0.26699098944664, 0.4154151976108551, -0.485024631023407, 0.1625586003065109...
I'm not excited about having to pull the data into a temporary table on my SQL box because users need to be able to run this report on demand and it seems like having to use SSIS to get the data into the table on demand will be slow and hard to manage with multiple users trying to get at the report simultaneously. Are there any other, more elegant solutions out there? I know that the linked server solution mentioned below would technically work, however, for some reason our DBAs will simply not allow us to use linked servers. I know that
[ 0.49812769889831543, 0.19826284050941467, 0.25043338537216187, 0.31971171498298645, 0.21886125206947327, -0.17923566699028015, -0.054054390639066696, 0.03592047840356827, -0.3590814769268036, -0.6247971653938293, 0.4013223648071289, 0.2882193624973297, -0.1710197478532791, 0.28776958584785...
you can add two different data sets to a report, however, I need to be able to join them together. Anybody have any ideas on how to best accomplish this? We had to do something similar (i.e. inner join 2 data sources from different servers). I believe the best way is to write your own custom Data Extension. It's not very difficult and it would give you the ability to do this and more.
[ 0.8428751826286316, -0.045711833983659744, 0.1530800610780716, 0.32094311714172363, -0.08799314498901367, -0.1824001520872116, -0.051051948219537735, 0.13229387998580933, -0.5450578331947327, -0.5784635543823242, 0.4430656433105469, 0.3746507167816162, -0.04369097575545311, 0.0910471677780...
Let X be the set of all sets that do not contain themselves. Is X a member of X? In **ZFC**, either the axiom of foundation [as mentioned] or the axiom (scheme) of comprehension will prohibit this. The first, for obvious reasons; the second, since it basically says that for given *z* and first-order property *P*, you can construct { *x* ∈ *z* : *P*(*x*) }, but to generate the Russell set, you would need *z* = *V* (the class of all sets), which is not a set (i.e. cannot be generated from any of the given axioms). In New Foundations (**NF**),
[ 0.25050967931747437, 0.09659115225076675, -0.15101145207881927, 0.08759001642465591, -0.2055511623620987, -0.306924432516098, 0.13426616787910461, -0.1841334104537964, -0.15722674131393433, -0.4065827429294586, -0.6720808148384094, 0.27964481711387634, -0.5686317682266235, 0.33001533150672...
"*x* ∉ *x*" is not a stratified formula, and so again we cannot define the Russell set. Somewhat amusingly, however, *V* *is* a set in **NF**. In von Neumann--Bernays--Gödel set theory (**NBG**), the class *R* = { *x* : *x* is a set and *x* ∉ *x* } is definable. We then ask whether *R* ∈ *R*; if so, then also *R* ∉ *R*, giving a contradiction. Thus we must have *R* ∉ *R*. But there is no contradiction here, since for any given class *A*, *A* ∉ *R* implies either *A* ∈ *A* or *A* is a proper class. Since
[ -0.08943057805299759, 0.011560313403606415, -0.2385598123073578, -0.11205817759037018, -0.2905932366847992, 0.18826507031917572, 0.18676066398620605, -0.22889341413974762, -0.17774799466133118, -0.431421160697937, -0.5021392107009888, 0.30638086795806885, -0.46385571360588074, 0.3430123925...
*R* ∉ *R*, we must simply have that *R* is a proper class. Of course, the class *R* = { *x* : *x* ∉ *x* }, without the restriction, is simply not definable in **NBG**. Also of note is that the above procedure is formally constructable as a proof in **NBG**, whereas in **ZFC** one has to resort to meta-reasoning.
[ -0.10796855390071869, 0.2744528353214264, -0.16498708724975586, -0.07672441750764847, -0.058701153844594955, -0.014893083833158016, 0.4132276773452759, -0.36624258756637573, 0.04613348841667175, -0.28194448351860046, -0.3676031529903412, 0.7300816774368286, -0.28208744525909424, 0.22262188...
I am a developer. An architect on good days. Somehow I find myself also being the DBA for my small company. My background is fair in the DB arts but I have never been a full fledged DBA. My question is what do I have to do to ensure a realiable and reasonably functional database environment with as little actual effort as possible? I am sure that I need to make sure that backups are being performed and that is being done. That is an easy one. What else should I be doing on a consistant basis? Who else is involved in
[ 0.696641743183136, 0.3305697441101074, 0.07248696684837341, -0.016802771016955376, -0.01664958894252777, -0.13088786602020264, 0.3550909757614136, 0.08498664200305939, -0.059018783271312714, -0.42052343487739563, 0.06390424817800522, 0.6122197508811951, 0.27295228838920593, 0.1664750874042...
the database? Are you the only person making schema changes (creating new objects, releasing new stored procedures, permissioning new users)? * Make sure that the number of users doing anything that could impact performance is reduced to as close to zero as possible, ideally including you. * Make sure that you're testing your backups - ideally run a DEV box that is recreating the production environment periodically, 1. a DEV box is a good idea, 2. a backup is only useful if you can restore from it. * Create groups for the various apps that connect to your database, so when a
[ 0.5560792088508606, 0.003932047635316849, 0.04496864229440689, 0.4429807662963867, 0.12464425712823868, -0.2260020524263382, 0.317960649728775, -0.04727554693818092, -0.3480381667613983, -0.409853458404541, 0.003925221506506205, 0.8328981399536133, -0.19324132800102234, -0.3386701345443725...
new user comes along you don't guess what permissions they need, just add them to the group, meanwhile permission the database objects to only the groups that need them * Use indices, primary keys, foreign keys, constraints, stats and whatever other tools your database supports. Normalise. * Optimise the most common code against your box - bad stored procedures/data access code will kill you.
[ 0.09873537719249725, -0.1690419465303421, 0.3335793912410736, 0.3443557322025299, 0.14652962982654572, -0.27212682366371155, 0.46719077229499817, -0.104104183614254, -0.19097138941287994, -0.7551416754722595, -0.1438596248626709, 0.45733869075775146, -0.482023149728775, 0.132501482963562, ...
I am trying to build an website for my college's magazine. I used the "views" module to show a block of static content I created on the front page. My question is: how can I edit the theme's css so it changes the way that block of static content is displayed? For reference, [here's the link](http://www.historia.uff.br/aroda/) to the site (in portuguese, and with almost zero content for now). The main css file that drives your content is the styles.css file located in your currently selected theme. In your case that means that most of your site styling is driven by this file: /aroda/roda/themes/garland/style.css
[ 0.6595236659049988, 0.12276560068130493, 0.6625710725784302, 0.056761275976896286, -0.14509591460227966, -0.04208621755242348, -0.33550333976745605, 0.13369712233543396, -0.18687456846237183, -0.24427573382854462, 0.11516021192073822, 0.2920692265033722, 0.1327170431613922, 0.3307811915874...
with basic coloring effects handled by this file: /aroda/roda/files/color/garland-d3985506/style.css You're currently using Garland, the default Drupal theme included with the core download, so for best practices you shouldn't edit the included style.css file directly. Instead, you should, as Daniel James said, create a subdirectory in /sites/all called "themes". If you're using Drupal 6, I'd follow Daniel James directions from there. If you're using Drupal 5, I'd go ahead and copy the garland directory into the themes directory and rename it for something specific to your site (aroda\_v1) so you would have something like /sites/all/themes/aroda\_v1 which would contain styles.css. At that point, you can
[ 0.33848485350608826, -0.12185706198215485, 0.5431274175643921, -0.012583479285240173, 0.030839117243885994, 0.07454443722963333, 0.08213990926742554, -0.13222159445285797, -0.2275538146495819, -0.6462455987930298, -0.2535346746444702, 0.17634718120098114, -0.5114243030548096, 0.01971684955...
edit the styles.css file directly to make any changes you see fit. Hope that helps!
[ 0.9200524091720581, -0.3545306622982025, 0.37680351734161377, 0.16416612267494202, -0.09491246193647385, -0.38294297456741333, 0.03125036880373955, 0.4297442138195038, -0.16608299314975739, -0.8050894737243652, -0.15334975719451904, 0.1915673464536667, -0.11531663686037064, -0.002100684447...
I have a solution consisting of five projects, each of which compile to separate assemblies. Right now I'm code-signing them, but I'm pretty sure I'm doing it wrong. What's the best practice here? * Sign each with a different key; make sure the passwords are different * Sign each with a different key; use the same password if you want * Sign each with the same key * Something else entirely Basically I'm not quite sure what "signing" does to them, or what the best practices are here, so a more generally discussion would be good. All I really know is that [FxCop](http://en.wikipedia.org/wiki/FxCop) yelled at
[ 1.0046579837799072, 0.39137133955955505, 0.038996681571006775, -0.15490539371967316, -0.0257420614361763, 0.04652643948793411, 0.2607420086860657, -0.17287836968898773, -0.31496840715408325, -0.5303229689598083, 0.2267037332057953, 0.48384442925453186, -0.09525937587022781, -0.082705020904...
me, and it was easy to fix by clicking the "Sign this assembly" checkbox and generating a .pfx file using Visual Studio (2008). If your only objective is to stop FxCop from yelling at you, then you have found the best practice. The best practice for signing your assemblies is something that is completely dependent on your objectives and needs. We would need more information like your intended deployment: * For personal use * For use on corporate network PC's as a client application * Running on a web server * Running in SQL Server * Downloaded over the internet * Sold on a CD in shrink wrap *
[ 0.9273072481155396, 0.34442493319511414, 0.04371248930692673, -0.242679163813591, -0.3771943151950836, -0.01579197682440281, 0.33508020639419556, -0.4088735282421112, -0.21099895238876343, -0.6005414128303528, 0.0855209156870842, 0.4969354569911957, -0.4264165759086609, -0.1728659272193908...
Uploaded straight into a cybernetic brain * Etc. Generally you use code signing to verify that the Assemblies came from a specific trusted source and have not been modified. ***So each with the same key is fine.*** Now how that trust and identity is determined is another story. **UPDATE:** How this benefits your end users when you are deploying over the web is if you have obtained a [software signing certificate from a certificate authority](http://www.thawte.com/code-signing/index.html?click=main-nav-products-codesigning). Then when they download your assemblies they can verify they came from *Domenic's Software Emporium*, and they haven't been modified or corrupted along the way. You will also
[ 1.032723069190979, -0.024746134877204895, 0.22456209361553192, 0.33540159463882446, 0.07910890877246857, -0.045998405665159225, 0.2010267823934555, -0.036005809903144836, -0.3535926938056946, -0.5387208461761475, -0.26633137464523315, 0.3388808071613312, -0.04098076373338699, -0.0474599450...
want to sign the installer when it is downloaded. This prevents the warning that some browsers display that it has been obtained from an unknown source. Note, you will pay for a software signing certificate. What you get is the certificate authority become the trusted 3rd party who verifies you are who you say you are. This works because of a web of trust that traces its way back to a root certificate that is installed in their operating system. There are a few certificate authorities to choose from, but you will want to make sure they are supported by the
[ 0.8541359305381775, 0.11982816457748413, 0.346933513879776, 0.08485899865627289, 0.31367629766464233, -0.6151953339576721, 0.4865628182888031, 0.03334253653883934, -0.09531812369823456, -0.46445706486701965, -0.5087171196937561, 0.6264830231666565, 0.03785817325115204, 0.004545079078525305...
root certificates on the target operating system.
[ -0.16744579374790192, 0.0032233886886388063, 0.0058777108788490295, 0.32275208830833435, 0.061480335891246796, -0.17150403559207916, 0.3067289888858795, 0.28211790323257446, -0.08922011405229568, -0.3909953832626343, -0.5524369478225708, 0.12572883069515228, 0.04075801372528076, 0.40587985...
When using Resharper to encapsulate a class's properties, is there a way to get it to do more than one property at a time? You might or might not already know this (R# does suffer from a lack of discoverability, unless you get the one-page key-shortcut page printed out), but ALT-INS opens a box which can at least mass-generate properties for fields. Not sure if that's any use - it's not the same as a retrospective encapsulation.
[ 0.3130739629268646, -0.220133438706398, -0.13535961508750916, 0.363252192735672, -0.25641152262687683, -0.24790161848068237, 0.43453431129455566, -0.26303812861442566, -0.12104180455207825, -0.4398028254508972, -0.2856309115886688, 0.5287548899650574, 0.13621054589748383, 0.163204565644264...
Graphics and audio editing and processing software often contain functions called "High-Pass Filter" and "Low-Pass Filter". Exactly what do these do, and what are the algorithms for implementing them? Wikipedia: * [High-pass filter](http://en.wikipedia.org/wiki/High-pass_filter) * [Low-pass filter](http://en.wikipedia.org/wiki/Low-pass_filter) * [Band-pass filter](http://en.wikipedia.org/wiki/Band-pass_filter) These "high", "low", and "band" terms refer to *frequencies*. In high-pass, you try to remove low frequencies. In low-pass, you try to remove high. In band pass, you only allow a continuous frequency range to remain. Choosing the cut-off frequency depends upon your application. Coding these filters can either be done by simulating RC circuits or by playing around with Fourier transforms of your time-based data. See
[ 0.38350653648376465, -0.08377435058355331, 0.08645778894424438, 0.07903784513473511, -0.2066308706998825, -0.15220074355602264, -0.06771808862686157, -0.31089985370635986, -0.16127696633338928, -0.6536720991134644, -0.21621477603912354, 0.6255617737770081, -0.42416051030158997, -0.17263618...
the wikipedia articles for code examples.
[ 0.5142500400543213, 0.419913113117218, -0.31511688232421875, 0.3684902489185333, -0.0022637834772467613, -0.26795849204063416, 0.48472344875335693, 0.1409507691860199, -0.29116731882095337, -0.47115975618362427, -0.3585122227668762, 0.13883095979690552, -0.21099448204040527, 0.234335392713...
Doug McCune had created something that was exactly what I needed (<http://dougmccune.com/blog/2007/05/10/analyze-your-actionscript-code-with-this-apollo-app/>) but alas - it was for AIR beta 2. I just would like some tool that I can run that would provide some decent metrics...any idea's? There is a Code Metrics Explorer in the Enterprise Flex Plug-in below: <http://www.deitte.com/archives/2008/09/flex_builder_pl.htm>
[ 0.42154550552368164, 0.2800614535808563, 0.0725613608956337, 0.2994972765445709, -0.10995979607105255, 0.002785187680274248, -0.11952604353427887, 0.0299996268004179, -0.275683730840683, -0.3543725907802582, 0.19854703545570374, 0.55120849609375, 0.07396867126226425, -0.3750332295894623, ...
In a project I'm working on FxCop shows me lots of (and I mean more than 400) errors on the InitializeComponent() methods generated by the Windows Forms designer. Most of those errors are just the assignment of the Text property of labels. I'd like to suppress those methods in source, so I copied the suppression code generated by FxCop into AssemblyInfo.cs, but it doesn't work. This is the attribute that FxCop copied to the clipboard. ``` [module: SuppressMessage("Microsoft.Globalization", "CA1303:DoNotPassLiteralsAsLocalizedParameters", Scope = "member", Target = "WindowsClient.MainForm.InitializeComponent():System.Void", MessageId = "System.Windows.Forms.Control.set_Text(System.String)")] ``` Anyone knows the correct attribute
[ 0.1733311265707016, 0.03526601567864418, 0.40875276923179626, -0.13952796161174774, 0.07249926030635834, -0.12220893055200577, 0.4468627870082855, -0.3781788945198059, -0.23154203593730927, -0.44504186511039734, -0.08384454250335693, 0.5296754240989685, -0.5976277589797974, -0.052062667906...
to suppress this messages? PS: I'm using Visual Studio 2005, C#, FxCop 1.36 beta. You've probably got the right code, but you also need to add CODE\_ANALYSIS as a precompiler defined symbol in the project properties. I think those SuppressMessage attributes are only left in the compiled binaries if CODE\_ANALYSIS is defined.
[ 0.3804524838924408, -0.05379466712474823, 0.16510440409183502, -0.0014123549917712808, -0.13727673888206482, -0.20352394878864288, 0.45178666710853577, 0.09267175197601318, -0.06439799815416336, -0.3834550082683563, -0.24993333220481873, 0.438623309135437, -0.46025076508522034, -0.39467459...
I've been using Window Home Server for my backups here at home for most of a year now, and I'm really pleased with it. It's far better than the software I was using previously (Acronis). I'm thinking about a backup strategy for my work machine and I'd like to know how WHS compares with Vista's built-in backup and restore features. The plan is to do a full backup to a local external hard drive and backup the documents folder to a network drive on the server. Anyone have experience using the Vista backup feature like this? Chris, They're different beasts. WHS
[ 0.4213356673717499, 0.1492677927017212, 0.4601035714149475, 0.058083340525627136, 0.11644000560045242, -0.2362554371356964, 0.22082406282424927, 0.17882798612117767, -0.46220824122428894, -0.6897254586219788, 0.08575615286827087, 0.669670820236206, 0.2505973279476166, 0.269344687461853, ...
backup is pretty much automatic and uses deltas - Vista's is manual and I don't believe offers incremental updates. While your solution (Vista + network copy) would preserve your data it has two problems I an see; 1. Your documents will only have the latest revision. If you find something was corrupted a month ago it could be very awkward to recover it. Vista's shadow copies may help though. 2. As soon as you install a program/patch/config your Vista backup is out of date and needs remade, or these repeated if you reinstall. These might not be dealbreakers and indeed Vista's backup is pretty
[ -0.003228818066418171, 0.0008080218685790896, 0.566616952419281, 0.3255164325237274, 0.3409728407859802, -0.5195698738098145, 0.5167843103408813, 0.0259314626455307, -0.4153376817703247, -0.5925925970077515, -0.10369708389043808, 0.8773996233940125, -0.20398543775081635, -0.068835087120532...
decent, it's just nowhere near as good as WHS. In my opinion WHS leaves almost everything else standing, you can be sure this tech will be in the "big brother" server versions shortly.
[ 0.4871000349521637, 0.09593547135591507, 0.5166513323783875, 0.422883003950119, 0.13153749704360962, -0.5172170996665955, 0.40474286675453186, 0.3291521370410919, 0.17013531923294067, -0.4385690987110138, 0.27691832184791565, 0.19267968833446503, 0.4623968005180359, 0.18064022064208984, ...
I am ready to start using SVN, but I have NO (as in the money required for free beer) experience with source control. I have installed subversion on my server (that was easy, 'apt-get install subversion') but now I don't know what to do, how to configure it, or how to use it. What suggestions do you have, and where can I find good resources to learn to start using it? Update: O.K. So the feedback has been great and I have read through a bit of it but I want to clarify my question by saying
[ 0.7632178664207458, 0.29095274209976196, 0.33936750888824463, -0.29908138513565063, -0.33155322074890137, -0.5289150476455688, 0.28602898120880127, 0.3106768727302551, -0.03621574491262436, -0.5482961535453796, 0.11683178693056107, 0.4566280245780945, 0.039639540016651154, 0.10138857364654...
that I am looking for more information on how to actually go about setting my up my repositories, clients, server, etc. I know that I could do a quick Google search and find dozens (or more) resources but I'm hoping that someone whom has experience with subversion and a client(I have installed tortoise) could suggest a good reference that will be reliable, and have quality content. Eric Sink has an [excellent series](http://www.ericsink.com/scm/source_control.html) on source code control aimed at beginners. For Subversion specifics, including setting up and administering a server, the [Subversion book](http://svnbook.red-bean.com/) is a great resource, and includes a section with
[ 0.4303358495235443, -0.04470185935497284, -0.13521337509155273, 0.17895756661891937, -0.13787804543972015, -0.1418769657611847, 0.33493176102638245, 0.013716437853872776, -0.2860076427459717, -0.29214900732040405, 0.08288570493459702, 0.2911296486854553, 0.04027663543820381, -0.04030738398...
examples of a typical session with Subversion (checkout, commit, merging and updating basics). **Update:** I forgot to mention that for beginners, I'd also recommend messing around in a graphical client, which removes the command-line hassle from the learning experience. [RapidSVN](http://rapidsvn.tigris.org/) is a reasonable cross-platform client. You'll also find that common IDEs either come with Subversion support, or have plugins which can be installed, which allow most version control operations to be performed within that environment. **@John Millikin:** While setting up a Subversion server can be complicated, depending on one's general admin experience, don't forget that you don't need to do that just
[ 0.3530169129371643, -0.25272393226623535, -0.2363361120223999, 0.15604665875434875, -0.24262481927871704, -0.47409844398498535, 0.7478890419006348, 0.3014693558216095, -0.5664746761322021, -0.6682339310646057, -0.08473403751850128, 0.43293043971061707, -0.3796462416648865, -0.2894557416439...
to mess about with a repository and get to grips with the basics - the client can interact with a repository in the local filesystem.
[ 0.4128305912017822, 0.022199276834726334, -0.020766502246260643, 0.548896849155426, 0.0486588217318058, -0.37586724758148193, 0.07754682749509811, -0.016591429710388184, -0.35138368606567383, -0.626950204372406, -0.2983989119529724, 0.6787951588630676, 0.08458039164543152, 0.05363748595118...
With a distributed application, where you have lots of clients and one main server, should you: * Make the clients dumb and the server smart: clients are fast and non-invasive. Business rules are needed in only 1 place * Make the clients smart and the server dumb: take as much load as possible off of the server Additional info: * Clients collect tons of data about the computer they are on. The server must analyze all of this info to determine the health of these computers * The owners of the client computers are temperamental and will shut down the clients if the client starts
[ 0.3109530210494995, -0.13188426196575165, 0.1819072961807251, 0.292221337556839, 0.1484237164258957, -0.3636642098426819, 0.2788248062133789, 0.141009122133255, -0.599729061126709, -0.8128507733345032, 0.0663367286324501, 0.30773651599884033, -0.4714311361312866, 0.299649715423584, 0.063...
to consume too many resources (thus negating the purpose of the distributed app in helping diagnose problems) You should do as much client-side processing as possible. This will enable your application to scale better than doing processing server-side. To solve your temperamental user problem, you could look into making your client processes run at a very low priority so there's no noticeable decrease in performance on the part of the user.
[ -0.029721124097704887, -0.0857340395450592, 0.11946222186088562, 0.4440927505493164, 0.11695574969053268, 0.004933598916977644, 0.40121474862098694, 0.13080443441867828, -0.2138337343931198, -0.7201182246208191, 0.2143757939338684, 0.6349725723266602, -0.33868086338043213, -0.0654431879520...
I have taken over a large code base and would like to get an overview how and where certain classes and their methods are used. Is there any good tool that can somehow visualize the dependencies and draw a nice call tree or something similar? The code is in C++ in Visual Studio if that helps narrow down any selection. Here are a few options: * [CodeDrawer](http://www.codedrawer.com/index.html) * [CC-RIDER](http://www.westernwares.com/) * [Doxygen](http://www.doxygen.nl/index.html) The last one, doxygen, is more of an automatic documentation tool, but it is capable of generating dependency graphs and inheritance diagrams. It's also licensed under the GPL, unlike the first two which are not free.
[ 0.8052533864974976, -0.06270133703947067, -0.0358516126871109, 0.12862692773342133, -0.18719325959682465, -0.621889054775238, -0.03694529086351395, -0.26818424463272095, 0.03264051675796509, -0.7530359625816345, 0.1758614182472229, 0.5990352630615234, -0.11344657838344574, 0.10735587775707...
I would like to create events for certain resources that are used across various processes and access these events by name. The problem seems to be that the names of the events must be known to all applications referring to them. Is there maybe a way to get a list of names events in the system? I am aware that I might use some standard names, but it seems rather inflexible with regard to future extensibility (all application would require a recompile). --- I'm afraid, I can't even consider ZwOpenDirectoryObject, because it is described as needing Windows XP or higher, so it is out
[ 0.2377304583787918, 0.15027134120464325, 0.07203318178653717, 0.2019154578447342, 0.011650379747152328, -0.4480748772621155, 0.43891745805740356, 0.1690661758184433, -0.2221849113702774, -0.48146453499794006, -0.011394030414521694, 0.6169289946556091, -0.21782143414020538, 0.20174176990985...
of question. Thanks for the suggestion though. I am a little unsure about shared memory, because I haven't tried it so far. Might do some reading in that area I guess. Configuration files and registry are a slight problem, because they do tend to fail with Vista due to access problems. I am a bit afraid, that shared memory will have the same problem. The idea with ProcessExplorer sounds promising. Does anyone know an API that could be used for listing events for a process? And, does it work without administrative rights? --- Thank you for the clarification. There is not really a master process.
[ 0.47526323795318604, 0.04449194297194481, -0.0607350692152977, 0.31092536449432373, -0.1458197981119156, -0.10883060097694397, 0.16536755859851837, 0.2134198099374771, -0.4045464098453522, -0.7003456354141235, 0.5307945609092712, 0.5347014665603638, 0.07727319747209549, 0.44454237818717957...
It is more of a driver dll that is used from different processes and the events would be used to "lock" resources used by these processes. I am thinking about setting up a central service that has sufficient access rights even under Vista. It will certainly complicate things, but it might be the only thing left facing the problems with security. Do not mix up the user mode ZwOpenDirectoryObject with the kernel mode ZwOpenDirectoryObject -- the kernel mode API (<http://msdn.microsoft.com/en-us/library/ms800966.aspx>) indeed seems to available as of XP only, but the user mode version should be available at least since NT 4. Anyway,
[ -0.07963002473115921, 0.10140689462423325, 0.49492672085762024, 0.15818624198436737, -0.07929395139217377, -0.4717417359352112, 0.26239073276519775, 0.17921718955039978, -0.03860688954591751, -0.655908465385437, -0.16273748874664307, 0.6992182731628418, -0.3142755627632141, 0.2112743109464...
I would not recommend using ZwOpenDirectoryObject. Why should configuration files and registry keys fail on Vista? Of course, you have to get the security settings right -- but you would have to do that for your named events as well -- so there should not be a big difference here. Maybe you should tell us some more details about the nature of your processes -- do they all run within the same logon session or do they run as different users even? And is there some master process or who creates the events in the first place? Frankly, I tend to find
[ 0.546880841255188, 0.0816701129078865, -0.035197269171476364, 0.3158782720565796, -0.12867513298988342, -0.45384806394577026, 0.20677568018436432, 0.17102766036987305, -0.3503836393356323, -0.5157872438430786, -0.0840430036187172, 0.5036795735359192, -0.39413949847221375, 0.268161356449127...
the Process Explorer idea to be not a very good one. Despite the fact that you probably will not be able to accomplish that without using undocumented APIs and/or a device driver, I do not think that a process should be spelunking around in the handle table of another process just to find out the names of some kernel objects. And, of course, the same security issues apply again.
[ 0.14999143779277802, 0.2460888922214508, -0.02345842681825161, 0.2635605037212372, -0.08307819068431854, -0.3148881793022156, 0.20616842806339264, 0.3909004330635071, -0.24371929466724396, -0.5166658163070679, 0.10194012522697449, 0.4945119619369507, -0.06731293350458145, 0.462529867887496...
I'd like to have a link in my ASP.NET web site that authenticated users click to download a windows app that is already pre-configured with their client ID and some site config data. My goal is no typing required for the user during the client app install, both for the user friendliness, and to avoid config errors from mis-typed technical bits. Ideally I'd like the web server-side code to run as part of the ASP.NET app. FogBugz seems to do something like this. There is a menu option within the web app to download a screenshot tool, and when you
[ 0.4810709059238434, 0.10150011628866196, 0.7283770442008972, 0.11730192601680756, 0.15125268697738647, -0.17850416898727417, 0.41586899757385254, 0.27747100591659546, -0.31174302101135254, -0.5361416339874268, -0.03606743365526199, 0.5444738268852234, -0.1570555865764618, -0.03467944636940...
download and run the installer, it knows your particular FogBugz web address so it can send screenshots there. (Hey Joel, looking for a question to answer? *hint—hint*) The way the FogBugz screenshot setup tool does this is that it appends a 256 byte block at the end of the setup program at the moment it is downloaded. In other words, the download script spits out all the bytes from setup.exe and then an extra 256 with the url for the FogBugz server, plus any padding. Windows ignores these extra bytes when the .exe is run (provided you turned off the CRC check
[ 0.4632272720336914, 0.12155280262231827, 0.469586044549942, 0.3562594950199127, -0.15255507826805115, -0.2761881649494171, 0.09762203693389893, 0.2844310998916626, -0.22227509319782257, -0.5492319464683533, -0.23377203941345215, 0.7693706750869751, -0.2861552834510803, 0.006931781303137541...
for your setup installer - we're using [InnoSetup](http://www.innosetup.com)). After installation, we run the Screenshot program with a command line switch that tells it where the setup installer is. It looks at the end of the setup.exe and finds it's info, and then writes that to the registry so the user doesn't have to know it.
[ 0.7292323112487793, -0.10442107915878296, 0.26820212602615356, -0.024484191089868546, 0.5779571533203125, -0.49828359484672546, -0.09338418394327164, 0.16585850715637207, -0.2295437455177307, -0.9327363967895508, -0.12701082229614258, 1.0242599248886108, -0.18870048224925995, -0.2006082683...
How do I get the id of my Java process? I know there are several platform-dependent hacks, but I would prefer a more generic solution. There exists no platform-independent way that can be guaranteed to work in all jvm implementations. `ManagementFactory.getRuntimeMXBean().getName()` looks like the best (closest) solution, and typically includes the PID. It's short, and *probably* works in every implementation in wide use. On linux+windows it returns a value like `"12345@hostname"` (`12345` being the process id). Beware though that [according to the docs](http://docs.oracle.com/javase/6/docs/api/java/lang/management/RuntimeMXBean.html#getName%28%29), there are no guarantees about this value: > Returns the name representing the running Java virtual machine. The > returned name string
[ 0.19111758470535278, -0.03200430050492287, 0.41624486446380615, -0.019563229754567146, 0.05242924764752388, -0.21888761222362518, 0.19566048681735992, -0.27131474018096924, -0.03268745541572571, -0.8157516717910767, 0.025427449494600296, 0.5148626565933228, -0.1640196144580841, 0.162545099...
can be any arbitrary string and a Java virtual > machine implementation can choose to embed platform-specific useful > information in the returned name string. Each running virtual machine > could have a different name. **In Java 9** the new [process API](https://docs.oracle.com/javase/9/docs/api/java/lang/ProcessHandle.html) can be used: ``` long pid = ProcessHandle.current().pid(); ```
[ 0.2630217671394348, -0.0662582591176033, 0.5149475336074829, -0.13310344517230988, 0.23240926861763, 0.2890307605266571, -0.11579987406730652, -0.25579047203063965, -0.38372892141342163, -0.5694534778594971, -0.31391260027885437, 0.06332138925790787, -0.4120637774467468, 0.3644683361053467...
I have a list of more than 15 thousand latitude and longitude coordinates. Given any X,Y coordinates, what is the fastest way to find the closest coordinates on the list? You will want to use a geometric construction called a [Voronoi diagram](http://mathworld.wolfram.com/VoronoiDiagram.html). This divides up the plane into a number of areas, one for each point, that encompass all the points that are closest to each of your given points. The code for the exact algorithms to create the Voronoi diagram and arrange the data structure lookups are too large to fit in this little edit box. :) @Linor: That's essentially what you
[ 0.38574790954589844, 0.19388365745544434, 0.754233181476593, 0.40582409501075745, 0.13720887899398804, 0.334119588136673, 0.0688602551817894, -0.05244215577840805, -0.2241480052471161, -0.7083340883255005, 0.10757850110530853, -0.02976631559431553, -0.05756109207868576, 0.27200135588645935...
would do after creating a Voronoi diagram. But instead of making a rectangular grid, you can choose dividing lines that closely match the lines of the Voronoi diagram (this way you will get fewer areas that cross dividing lines). If you recursively divide your Voronoi diagram in half along the best dividing line for each subdiagram, you can then do a tree search for each point you want to look up. This requires a bit of work up front but saves time later. Each lookup would be on the order of log N where N is the number of points.
[ 0.05069354176521301, -0.2802121639251709, 0.18485161662101746, 0.041655004024505615, -0.07269261032342911, 0.21614447236061096, 0.11901450902223587, -0.5751736760139465, -0.11899610608816147, -0.7225070595741272, -0.1149587482213974, 0.11608126759529114, -0.21973036229610443, 0.21886588633...
16 comparisons is a lot better than 15,000!
[ 0.36709603667259216, 0.21824951469898224, 0.351223886013031, 0.29776281118392944, -0.5268000364303589, 0.3098871409893036, 0.3151454031467438, -0.2049623429775238, -0.23768842220306396, -0.485862672328949, 0.33256906270980835, 0.4148518443107605, 0.15265874564647675, -0.004864242859184742,...
I want to keep logs of some things that people do in my app, in some cases so that it can be undone if needed. Is it best to store such logs in a file or a database? I'm completely at a loss as to what the pros and cons are except that it's another table to setup. Is there a third (or fourth etc) option that I'm not aware of that I should look into and learn about? You will almost certainly want to use a database for flexible, record based access and to take advantage of the database's ability to handle
[ 0.7148544788360596, 0.1851387917995453, 0.06439048796892166, 0.2650921046733856, 0.39623016119003296, -0.08173271268606186, 0.17452099919319153, 0.06426475197076797, -0.5807285904884338, -0.41743791103363037, 0.2327578067779541, 0.4521866738796234, -0.018478915095329285, -0.037740308791399...
concurrent data access. If you need to track information that may need to be undone, having it in a structured format is a benefit, as is having the ability to update a row indicating when and by whom a given transaction has been undone. You likely only want to write to a file if very high performance is an issue, or if you have very unstructured or large amounts of data per record that might be unweidly to store in a database. Note that Unless your application has a very large number of transactions database speed is unlikely to be
[ 0.0827459767460823, 0.08272495120763779, 0.28523293137550354, 0.5751766562461853, -0.20138199627399445, -0.18289922177791595, 0.2923610806465149, 0.070316381752491, -0.9251365661621094, -0.4567526578903198, 0.14700573682785034, 0.3689526319503784, -0.31439724564552307, 0.33558517694473267,...
an issue. Also note that if you are working with a file you'll need to handle concurrent access (read / write / locking) very carefully which is likely not something you want to have to deal with.
[ 0.0734388530254364, 0.06814529746770859, 0.16820861399173737, 0.23323610424995422, 0.00021114121773280203, -0.3541625440120697, 0.0004025925591122359, -0.027184609323740005, -0.5839765071868896, -0.4873030483722687, -0.12410574406385422, 0.4293401539325714, -0.15992288291454315, 0.35269540...
I need an associative container that makes me index a certain object through a string, but that also keeps the order of insertion, so I can look for a specific object by its name or just iterate on it and retrieve objects in the same order I inserted them. I think this [hybrid of linked list and hash map](http://www.ddj.com/cpp/184406207) should do the job, but before I tried to use `std::tr1::unordered_map` thinking that it was working in that way I described, but it wasn't. So could someone explain me the meaning and behavior of `unordered_map`? --- @wesc: I'm sure std::map is implemented by STL,
[ 0.13549302518367767, 0.01000653300434351, 0.1226021945476532, 0.1294630765914917, -0.1179296225309372, -0.19742503762245178, 0.023715922608971596, 0.13475912809371948, -0.39669883251190186, -0.7720608711242676, 0.049553923308849335, 0.34202587604522705, -0.15553626418113708, 0.184259042143...
while I'm sure std::hash\_map is NOT in the STL (I think older version of Visual Studio put it in a namespace called stdext). @cristopher: so, if I get it right, the difference is in the implementation (and thus performances), not in the way it behaves externally. [Boost documentation of unordered containers](http://www.boost.org/doc/libs/1_36_0/doc/html/unordered.html) The difference is in the method of how you generate the look up. In the map/set containers the `operator<` is used to generate an ordered tree. In the unordered containers, an `operator( key ) => index` is used. See hashing for a description of how that works.
[ 0.07779791951179504, 0.00001963909926416818, 0.17321224510669708, -0.2841469943523407, -0.2863028049468994, 0.1642172485589981, 0.018728207796812057, -0.3491040766239166, -0.3233841359615326, -0.49772006273269653, -0.2512654960155487, 0.7503988146781921, -0.40678587555885315, -0.5496873259...
What is the best approach to define additional data for typedef enums in C? Example: ``` typedef enum { kVizsla = 0, kTerrier = 3, kYellowLab = 10 } DogType; ``` Now I would like to define names for each, for example `kVizsla` should be "vizsla". I currently use a function that returns a string using a large switch block. @dmckee: I think the suggested solution is good, but for simple data (e.g. if only the name is needed) it could be augmented with auto-generated code. While there are lots of ways to auto-generate code, for something as simple as this I believe you could
[ 0.17458368837833405, -0.10021143406629562, 0.37882494926452637, 0.003830939531326294, -0.11040379852056503, -0.10742893815040588, 0.18278656899929047, -0.3258216679096222, -0.22491444647312164, -0.507804811000824, -0.07746273279190063, 0.4207359552383423, -0.2118043750524521, 0.19774673879...
write a simple XSLT that takes in an XML representation of the enum and outputs the code file. The XML would be of the form: ``` <EnumsDefinition> <Enum name="DogType"> <Value name="Vizsla" value="0" /> <Value name="Terrier" value="3" /> <Value name="YellowLab" value="10" /> </Enum> </EnumsDefinition> ``` and the resulting code would be something similar to what dmckee suggested in his solution. For information of how to write such an XSLT try [here](http://www.w3schools.com/xsl/) or just search it up in google and find a
[ 0.18581682443618774, -0.04550644010305405, 0.3436163365840912, -0.033522479236125946, -0.20279376208782196, -0.05897170677781105, 0.17773665487766266, -0.30340149998664856, 0.2237284630537033, -0.6267383694648743, -0.22815930843353271, 0.23465193808078766, -0.6538358330726624, -0.140109330...
tutorial that fits. Writing XSLT is not much fun IMO, but it's not that bad either, at least for relatively simple tasks such as these.
[ -0.11497997492551804, -0.14136986434459686, 0.08273027092218399, 0.5901539921760559, -0.28302955627441406, -0.25655245780944824, 0.2161255031824112, 0.33770114183425903, 0.05239028111100197, -0.5424282550811768, -0.11565554887056351, 0.43750661611557007, 0.04342876747250557, -0.59670388698...
I'm working on an email solution in SQL Server ONLY that will use Database Mail to send out HTML formatted emails. The catch is that the images in the HTML need to be embedded in the outgoing email. This wouldn't be a problem if I were using a .net app to generate & send the emails but, unfortunately, all I have is SQL Server. Is it possible for SQL Server to embed images on its own? Yes, what you need to do is include the images as attachments and then they can be referenced within the HTML. Use the `@file_attachment` parameter of `sp_send_dbmail`
[ 0.6172770857810974, 0.2248876690864563, 0.5680629014968872, -0.013712641783058643, -0.3306739926338196, -0.12291314452886581, -0.09098217636346817, 0.017302067950367928, -0.3877533972263336, -0.8356764912605286, 0.3146745264530182, 0.3667268753051758, -0.39419490098953247, 0.14899426698684...
Has anyone used the [Sphinx](http://cmusphinx.sourceforge.net/html/cmusphinx.php) speech recognition stack to build IVR applications? I am looking for open source alternatives to the expensive and somewhat limiting choices from MSFT and others. I have not been able to find a comprehensive package that ties open source speech/voip applications together. Last I looked at Sphinx, it had issues with 8khz audio which resulted in really poor performance. There's not a lot of people talking about successful deployments of Sphinx in real environments, but you might be able to get it to work with some trailblazing effort. See here for more info: [<http://www.voip-info.org/wiki-Sphinx>](http://www.voip-info.org/wiki-Sphinx) The closest thing to open-source that really works
[ 0.19658631086349487, 0.24645134806632996, -0.11494376510381699, 0.13629195094108582, -0.30918675661087036, -0.02552076429128647, 0.4930112957954407, 0.06359301507472992, -0.31401368975639343, -0.46791765093803406, -0.056802038103342056, 0.39604365825653076, -0.19099535048007965, 0.12489792...
is using LumenVox with Asterisk. Asterisk is the open-source PBX that you can use to integrate with a VoIP service or gateway, or even the PSTN. LumenVox is a commercial speech engine that integrates with Asterisk: [<http://www.asterisk.org>](http://www.asterisk.org) [<http://www.lumenvox.com>](http://www.lumenvox.com) [<http://www.lumenvox.com/partners/digium/Asterisk.aspx>](http://www.lumenvox.com/partners/digium/Asterisk.aspx) There's lots of people successfully using LumenVox with Asterisk.
[ 0.2680237293243408, 0.0016068513505160809, 0.568690299987793, 0.11487758159637451, -0.03856106102466583, 0.08136006444692612, -0.05876915529370308, -0.08760346472263336, -0.3719867467880249, -0.6507208347320557, -0.04107373580336571, 0.3228321969509125, -0.4616684913635254, -0.092896580696...
This might seem like a stupid question I admit. But I'm in a small shop me plus two designers. Our backups are getting out of hand because they just copy/paste files if they need to make a change (version). I was all set to try Subversion to handle all of our files my text (code) files and their photoshop/illustrator and asset files. That is until I noticed there was a new version of Adobe Version Cue v3. We've tried previously to use version cue but it got complicated and the designers quickly stopped using it. Looking for anyone that has some
[ 0.5613125562667847, 0.012798868119716644, 0.5043517351150513, 0.037862684577703476, -0.09832030534744263, -0.01245911791920662, -0.14717905223369598, 0.0878361165523529, -0.4817298352718353, -0.2602628469467163, 0.23290368914604187, 0.5205605626106262, -0.2040184736251831, 0.06778546422719...
experience with version 3 of Version Cue. Thanks for the great feedback. Maybe I should have asked what's the best tool to use for Versioning Photoshop and related files. I did notice the binary file issue and was worried about trying to explain it and keep it "working". I signed up for the beta at Gridiron thanks for that! [Here](https://stackoverflow.com/questions/36186/best-versioning-tools-to-use-for-photoshopillustrator-and-related-binary-files) is the other question related to this one. I have used Subversion for this exact thing, and Theo is right, you have to remember to lock your files. I am on CS2 and so have not used Version Cue, but I have not
[ 0.19987721741199493, 0.062178321182727814, 0.36861947178840637, -0.07055504620075226, -0.39884838461875916, -0.363193541765213, 0.15395495295524597, -0.3166263997554779, -0.2528364360332489, -0.24212345480918884, 0.3384074568748474, 0.5261093974113464, -0.12132276594638824, -0.283060878515...
been able to find a whole lot online about other folks using it either, for some reason. The other problem I had using Subversion related to disk space. Subversion stores an alternate "shadow copy" of your files in your working directory. For Photoshop and Illustrator this is normally not that big of a deal, but I was using Premiere and After Effects as well, and the disk space required for the shadow copies doubled my disk usage. You might also check out Gridiron's new Gridiron Flow product, which [John Nack raves about](http://blogs.adobe.com/jnack/2008/01/gridiron_flow_r.html). I would love to use it - it's
[ 0.5871350169181824, -0.29049810767173767, 0.3831639587879181, 0.2077063024044037, -0.1321195662021637, -0.34354615211486816, -0.48127037286758423, -0.18645834922790527, -0.2619428038597107, -0.47150278091430664, 0.39673784375190735, 0.42584845423698425, -0.19586792588233948, -0.16158807277...
due out about now, and it will likely run in the several-hundred dollar range, I think... Update 3/6/2009: Gridiron Flow is out, and it does versioning on a single machine, but it's not clear from their demos whether it does collaborative versioning. Also, I just stumbled across this [very good comparison](http://joshcarter.com/productivity/svn_hg_git_for_home_directory) of subversion, git & mercurial for managing a home directory - including various versions of large Photoshop files.
[ 0.20161880552768707, -0.17562812566757202, 1.0310072898864746, -0.021358979865908623, 0.09200669080018997, -0.4770716726779938, -0.0974503606557846, 0.2623049318790436, -0.32944101095199585, -0.5616488456726074, 0.31594544649124146, 0.4190235137939453, 0.14520058035850525, -0.3047337532043...
Some WPF controls (like the `Button`) seem to happily consume all the available space in its' container if you don't specify the height it is to have. And some, like the ones I need to use right now, the (multiline) `TextBox` and the `ListBox` seem more worried about just taking the space necessary to fit their contents, and no more. If you put these guys in a cell in a `UniformGrid`, they will expand to fit the available space. However, `UniformGrid` instances are not right for all situations. What if you have a grid with some rows set to a \*
[ 0.23921151459217072, -0.20937447249889374, 0.4587292969226837, 0.2810186445713043, 0.148552805185318, -0.20491226017475128, -0.0695280060172081, -0.140823632478714, -0.5645706653594971, -0.8985393643379211, 0.10956946015357971, 0.0036184447817504406, -0.28892138600349426, 0.119905389845371...
height to divide the height between itself and other \* rows? What if you have a `StackPanel` and you have a `Label`, a `List` and a `Button`, how can you get the list to take up all the space not eaten by the label and the button? I would think this would really be a basic layout requirement, but I can't figure out how to get them to fill the space that they could (putting them in a `DockPanel` and setting it to fill also doesn't work, it seems, since the `DockPanel` only takes up the space needed by its' subcontrols). A
[ 0.1700359731912613, 0.11280369758605957, 0.5027514100074768, -0.1428089737892151, 0.2569856345653534, 0.06961137056350708, -0.39765608310699463, -0.3338776230812073, -0.3421209156513214, -0.56302410364151, -0.05408092215657234, 0.5535860061645508, 0.1228012964129448, 0.06304353475570679, ...
resizable GUI would be quite horrible if you had to play with `Height`, `Width`, `MinHeight`, `MinWidth` etc. Can you bind your `Height` and `Width` properties to the grid cell you occupy? Or is there another way to do this? Each control deriving from **`Panel`** implements distinct layout logic performed in `Measure()` and `Arrange()`: * `Measure()` determines the size of the panel and each of its children * `Arrange()` determines the rectangle where each control renders The last child of the **`DockPanel`** fills the remaining space. You can disable this behavior by setting the `LastChild` property to `false`. The **`StackPanel`** asks each child for its desired
[ -0.3230830729007721, -0.312086284160614, 0.3341132700443268, 0.18873505294322968, 0.3018215000629425, 0.4935142993927002, -0.21223494410514832, -0.2681944668292999, -0.21893227100372314, -0.8900392055511475, -0.12136202305555344, 0.5208004117012024, 0.022765086963772774, 0.0589646026492118...
size and then stacks them. The stack panel calls `Measure()` on each child, with an available size of `Infinity` and then uses the child's desired size. A **`Grid`** occupies all available space, however, it will set each child to their desired size and then center them in the cell. You can implement your own layout logic by deriving from `Panel` and then overriding `MeasureOverride()` and `ArrangeOverride()`. See [this article](https://learn.microsoft.com/en-us/dotnet/framework/wpf/controls/how-to-create-a-custom-panel-element) for a simple example.
[ 0.28903326392173767, -0.2365724742412567, 0.37378454208374023, -0.06566935777664185, 0.3247804343700409, 0.492289274930954, -0.24267664551734924, -0.5097830891609192, -0.5092546939849854, -0.8626841306686401, -0.10153432935476303, -0.10497693717479706, 0.032287195324897766, 0.1333958804607...
The following shell script takes a list of arguments, turns Unix paths into WINE/Windows paths and invokes the given executable under WINE. ``` #! /bin/sh if [ "${1+set}" != "set" ] then echo "Usage; winewrap EXEC [ARGS...]" exit 1 fi EXEC="$1" shift ARGS="" for p in "$@"; do if [ -e "$p" ] then p=$(winepath -w $p) fi ARGS="$ARGS '$p'" done CMD="wine '$EXEC' $ARGS" echo $CMD $CMD ``` However, there's something wrong with the quotation of command-line arguments. ``` $ winewrap '/home/chris/.wine/drive_c/Program Files/Microsoft Research/Z3-1.3.6/bin/z3.exe' -smt /tmp/smtlib3cee8b.smt Executing: wine '/home/chris/.wine/drive_c/Program Files/Microsoft Research/Z3-1.3.6/bin/z3.exe' '-smt' 'Z: mp\smtlib3cee8b.smt' wine: cannot find ''/home/chris/.wine/drive_c/Program' ``` Note that: 1. The path to the executable is being chopped off at the
[ -0.24389687180519104, 0.29141974449157715, 0.38979628682136536, -0.47331497073173523, 0.17580635845661163, 0.11494408547878265, 0.3919249176979065, -0.31593188643455505, 0.004535801708698273, -0.4113084077835083, -0.450625479221344, 0.8467750549316406, -0.18478187918663025, 0.0309743229299...
first space, even though it is single-quoted. 2. The literal "\t" in the last path is being transformed into a tab character. Obviously, the quotations aren't being parsed the way I intended by the shell. How can I avoid these errors? EDIT: The "\t" is being expanded through two levels of indirection: first, `"$p"` (and/or `"$ARGS"`) is being expanded into `Z:\tmp\smtlib3cee8b.smt`; then, `\t` is being expanded into the tab character. This is (seemingly) equivalent to ``` Y='y\ty' Z="z${Y}z" echo $Z ``` which yields ``` zy\tyz ``` and *not* ``` zy yz ``` UPDATE: `eval "$CMD"` does the trick. The "`\t`" problem seems to be echo's fault: "If the first operand is -n, or if any
[ -0.07086089998483658, 0.0749443769454956, 0.36767348647117615, -0.1942373663187027, -0.05693471431732178, 0.08934033662080765, 0.30513614416122437, -0.277860552072525, -0.15504691004753113, -0.44012516736984253, -0.47465255856513977, 0.5838708281517029, -0.368580162525177, -0.0596327856183...
of the operands contain a backslash ( '\' ) character, the results are implementation-defined." ([POSIX specification of `echo`](http://www.opengroup.org/onlinepubs/009695399/utilities/echo.html)) I you do want to have the assignment to CMD you should use `eval $CMD` instead of just `$CMD` in the last line of your script. This should solve your problem with spaces in the paths, I don't know what to do about the "\t" problem.
[ 0.14154011011123657, -0.03189358860254288, 0.008063127286732197, -0.04138997569680214, 0.1182696744799614, -0.02974507212638855, 0.2498481571674347, -0.2432096004486084, -0.03809738531708717, -0.20185218751430511, -0.1692313700914383, 0.64153653383255, -0.3414980173110962, -0.0956839919090...
Are there any VC++ settings I should know about to generate better PDB files that contain more information? I have a crash dump analysis system in place based on the project [crashrpt](http://code.google.com/p/crashrpt/). Also, my production build server has the source code installed on the D:\, but my development machine has the source code on the C:\. I entered the source path in the VC++ settings, but when looking through the call stack of a crash, it doesn't automatically jump to my source code. I believe if I had my dev machine's source code on the D:\ it would work. > "Are there
[ 0.37608495354652405, 0.3853040933609009, 0.13849100470542908, 0.18384622037410736, -0.01871349848806858, -0.061446670442819595, 0.010705514810979366, -0.11761954426765442, -0.19979339838027954, -0.6764542460441589, -0.1493546962738037, 0.8477954268455505, -0.057326268404722214, -0.05504326...
any VC++ settings I should know about" Make sure you turn off Frame pointer ommision. Larry osterman's blog has [the historical details](http://blogs.msdn.com/larryosterman/archive/2007/03/12/fpo.aspx) about fpo and the issues it causes with debugging. > Symbols are loaded successfully. It shows the callstack, but double clicking on an entry doesn't bring me to the source code. What version of VS are you using? (Or are you using Windbg?) ... in VS it should defintely prompt for source the first time if it doesn't find the location. However it also keeps a list of source that was 'not found' so it doesn't ask you for it every
[ 0.13659338653087616, -0.17187020182609558, 0.5627474188804626, 0.09547632932662964, 0.04794463515281677, -0.04468485340476036, 0.4902022182941437, -0.43732136487960815, -0.2583470344543457, -0.6081164479255676, -0.024808576330542564, 0.6222515106201172, -0.24428017437458038, 0.023572472855...
time. Sometimes the don't look list is a pain ... to get the prompt back up you need to go to solution explorer/solution node/properties/debug properties and edit the file list in the lower pane. Finally you might be using 'stripped symbols'. These are pdb files generated to provide debug info for walking the callstack past FPO, but with source locations stripped out (along with other data). The public symbols for windows OS components are stripped pdbs. For your own code these simply cause pain and are not worth it unless you are providing your pdbs to externals. How would you have
[ 0.27491894364356995, -0.10348111391067505, 0.32806190848350525, 0.04166087880730629, 0.12518948316574097, 0.11900431662797928, 0.42611876130104065, -0.08295056968927383, -0.1978258639574051, -0.930885910987854, -0.3423565626144409, 0.4028477966785431, -0.2536623477935791, 0.052910368889570...
one of these horrible stripped pdbs? You might have them if you use "binplace" with the -a command. Good luck! A proper mini dump story is a godsend for production debugging.
[ 0.6078797578811646, -0.09067641943693161, 0.04297058656811714, 0.5445538759231567, 0.10515457391738892, -0.3356248736381531, 0.21479962766170502, 0.6907371282577515, -0.1586487740278244, -0.11610722541809082, -0.16181734204292297, 0.11464756727218628, 0.0897815003991127, 0.2091421186923980...
I'm trying to write a regex function that will identify and replace a single instance of a match within a string without affecting the other instances. For example, I have this string: ``` 12||34||56 ``` I want to replace the second set of pipes with ampersands to get this string: ``` 12||34&&56 ``` The regex function needs to be able to handle x amount of pipes and allow me to replace the nth set of pipes, so I could use the same function to make these replacements: ``` 23||45||45||56||67 -> 23&&45||45||56||67 23||34||98||87 -> 23||34||98&&87 ``` I know that I could just split/replace/concat the string at the pipes, and I also know that I
[ 0.14267894625663757, 0.04672131687402725, 0.21424271166324615, -0.014384377747774124, -0.24072766304016113, 0.15456640720367432, 0.2969447374343872, -0.5479193925857544, 0.22579623758792877, -0.6257569789886475, 0.29898691177368164, 0.5751294493675232, -0.526542603969574, 0.265143245458602...
can match on `/\|\|/` and iterate through the resulting array, but I'm interested to know if it's possible to write a single expression that can do this. Note that this would be for Javascript, so it's possible to generate a regex at runtime using `eval()`, but it's not possible to use any Perl-specific regex instructions. here's something that works: ``` "23||45||45||56||67".replace(/^((?:[0-9]+\|\|){n})([0-9]+)\|\|/,"$1$2&&") ``` where n is the one less than the nth pipe, (of course you don't need that first subexpression if n = 0) And if you'd like a function to do this: ``` function pipe_replace(str,n) { var RE = new RegExp("^((?:[0-9]+\\|\\|){" + (n-1) + "})([0-9]+)\|\|");
[ 0.11998040229082108, -0.24343593418598175, 0.7198656797409058, -0.28534847497940063, 0.006300690583884716, 0.012999852187931538, 0.3090163767337799, -0.6297779083251953, 0.24528321623802185, -0.6019365787506104, 0.060073770582675934, 0.8508888483047485, -0.48336362838745117, -0.05532000958...
return str.replace(RE,"$1$2&&"); } ```
[ 0.043465543538331985, 0.31390297412872314, 0.323597252368927, -0.40825143456459045, 0.08519838005304337, 0.07584035396575928, 0.4294182360172272, -0.20100870728492737, 0.042108260095119476, -0.3030843436717987, -0.44659489393234253, 0.9762457609176636, -0.49108025431632996, 0.0972084105014...
I would like to think that some of the software I'm writing today will be used in 30 years. But I am also aware that a lot of it is based upon the UNIX tradition of exposing time as the number of seconds since 1970. ```c #include <stdio.h> #include <time.h> #include <limits.h> void print(time_t rt) { struct tm * t = gmtime(&rt); puts(asctime(t)); } int main() { print(0); print(time(0)); print(LONG_MAX); print(LONG_MAX+1); } ``` Execution results in: * Thu Jan 1 00:00:00 1970 * Sat Aug 30 18:37:08 2008 * Tue Jan 19 03:14:07 **2038** * Fri
[ 0.2366451472043991, 0.01438059937208891, 0.5764334201812744, -0.24633625149726868, 0.127583846449852, -0.01365380547940731, 0.32410240173339844, -0.2932661175727844, -0.47820764780044556, -0.5685475468635559, -0.04028257727622986, 0.6640697121620178, -0.09813877940177917, 0.407091349363327...
Dec 13 20:45:52 **1901** > > The functions ctime(), gmtime(), and localtime() all take as an argument a time value representing the time in seconds since the Epoch (00:00:00 UTC, January 1, 1970; see time(3) ). I wonder if there is anything proactive to do in this area as a programmer, or are we to trust that all software systems (aka Operating Systems) will some how be magically upgraded in the future? **Update** It would seem that indeed 64-bit systems are safe from this: ```java import java.util.*; class TimeTest { public static void main(String[] args) {
[ 0.12403954565525055, 0.20208007097244263, 0.4208131432533264, 0.01022123172879219, 0.13117092847824097, 0.04861494526267052, 0.37126198410987854, 0.03407200425863266, -0.30085840821266174, -0.67798912525177, -0.09253575652837753, 0.38803407549858093, -0.10165850073099136, 0.612606883049011...
print(0); print(System.currentTimeMillis()); print(Long.MAX_VALUE); print(Long.MAX_VALUE + 1); } static void print(long l) { System.out.println(new Date(l)); } } ``` * Wed Dec 31 16:00:00 PST 1969 * Sat Aug 30 12:02:40 PDT 2008 * Sat Aug 16 23:12:55 PST **292278994** * Sun Dec 02 08:47:04 PST **292269055** But what about the year 292278994? I have written portable replacement for time.h (currently just localtime(), gmtime(), mktime() and timegm()) which uses 64 bit time even
[ 0.019100002944469452, -0.06790728121995926, 0.6767523288726807, -0.07310229539871216, 0.3999727964401245, 0.29999950528144836, 0.04555528238415718, 0.07038324326276779, -0.4255577325820923, -0.6755651831626892, 0.016622651368379593, 0.3797381520271301, -0.13671042025089264, 0.5038349032402...
on 32 bit machines. It is intended to be dropped into C projects as a replacement for time.h. It is being used in Perl and I intend to fix Ruby and Python's 2038 problems with it as well. This gives you a safe range of +/- 292 million years. You can find the code [at the y2038 project](https://github.com/schwern/y2038). Please feel free to post any questions to the [issue tracker](https://github.com/schwern/y2038/issues). As to the "this isn't going to be a problem for another 29 years", peruse this [list of standard answers](https://github.com/schwern/y2038/wiki/Why-Bother%3F) to that. In short, stuff happens in the future and sometimes you need
[ 0.3850005865097046, 0.10627981275320053, 0.2982668876647949, 0.06848045438528061, 0.2711108922958374, -0.3181508183479309, 0.30974671244621277, 0.22125466167926788, -0.3863498568534851, -0.5412355065345764, -0.12480013072490692, 0.2872544825077057, -0.2773727774620056, 0.07815064489841461,...
to know when. I also have [a presentation on the problem, what is not a solution, and what is](http://www.slideshare.net/schwern/whos-afraid-of-2038-presentation). Oh, and don't forget that many time systems don't handle dates before 1970. Stuff happened before 1970, sometimes you need to know when.
[ 0.046932563185691833, -0.27497556805610657, 0.434909850358963, 0.2830487787723541, 0.4903240501880646, -0.3338366746902466, 0.10967360436916351, 0.24497383832931519, -0.5546598434448242, -0.45598337054252625, 0.4092678427696228, 0.1002504974603653, 0.2740755081176758, 0.27445322275161743, ...
I'm trying to "install SGML::Parser::OpenSP" from the cpan shell, but it fails on the first "make test". I also get the same error if I go into the build directory and run make test. I believe this bit of the output below is the relevant part. Note the Symbol not found when perl gets to the "use" line for the new library. The file listed there exists and is readable. When I run the unix command "nm" it *does* show the symbol. I don't know what to make of the symbol not found error. I'm not running as admin/root if that matters.
[ 0.27355092763900757, 0.32315370440483093, 0.2638205587863922, -0.3038002550601959, 0.1955706924200058, -0.31377309560775757, 0.5351234674453735, 0.023405928164720535, -0.13348670303821564, -0.5555871725082397, 0.033643703907728195, 0.7720773816108704, -0.19755065441131592, 0.33744966983795...
This is on a mac, 10.4.11 My googling turned up some hints that this can happen if gcc is called instead of g++, but I believe that is set up correctly. What else could it be, and how can I try to fix? Here's the excerpt from running make test: ``` PERL_DL_NONLAZY=1 /usr/bin/perl "-MExtUtils::Command::MM" "-e" "test_harness(0, 'blib/lib', 'blib/arch')" t/*.t t/01basic...........1/4 # Failed test 'use SGML::Parser::OpenSP;' # at t/01basic.t line 14. # Tried to use 'SGML::Parser::OpenSP'. # Error: Can't load '/Users/joshgold/.cpan/build/SGML-Parser-OpenSP-0.994/blib/arch/auto/SGML/Parser/OpenSP/OpenSP.bundle' for module SGML::Parser::OpenSP: dlopen(/Users/joshgold/.cpan/build/SGML-Parser-OpenSP-0.994/blib/arch/auto/SGML/Parser/OpenSP/OpenSP.bundle, 2): Symbol not found: __ZTI15SGMLApplication # Referenced from: /Users/joshgold/.cpan/build/SGML-Parser-OpenSP-0.994/blib/arch/auto/SGML/Parser/OpenSP/OpenSP.bundle #
[ 0.37474575638771057, -0.10329718887805939, 0.4889772832393646, -0.01797478459775448, 0.20388248562812805, -0.1646679788827896, 0.3755396604537964, -0.4628482460975647, -0.05266272649168968, -0.42051956057548523, -0.13182446360588074, 0.5127635598182678, -0.41312116384506226, -0.12080629914...
Expected in: dynamic lookup # at (eval 3) line 2 # Compilation failed in require at (eval 3) line 2. # BEGIN failed--compilation aborted at (eval 3) line 2. ``` If you use the Settings for the project, you can mark each setting as either application or user. If they're set as user, they will be stored per-user and when you call the Save method it will be updated in the config for that user. Code project has a really detailed [article](http://www.codeproject.com/KB/dotnet/user_settings.aspx) on saving all types of settings.
[ 0.030199896544218063, -0.044233959168195724, 0.4342716634273529, -0.1308625191450119, 0.20951837301254272, 0.3512594997882843, 0.4503232538700104, -0.1840742826461792, -0.21717438101768494, -0.7080205082893372, -0.23617607355117798, 0.509388267993927, -0.24627825617790222, -0.3161542415618...
What is a good way to remove the code from display pages when developing with PHP. Often the pages I work on need to be editted by an outside person. This person is often confused by lots of blocks of PHP, and also likes to break my code. I've tried moving blocks of code out into functions, so now there are functions spread out all throughout the HTML now. As some pages become more complex it becomes a program again, and processing POSTs are questionable. What can I be doing better in my PHP development? You don't need a "system" to do templating. You
[ 0.49683496356010437, 0.22816862165927887, 0.08946995437145233, 0.24099816381931305, -0.21430104970932007, -0.18949635326862335, 0.27529817819595337, 0.09838251024484634, -0.5117714405059814, -0.5423243641853333, 0.39016661047935486, 0.3342369794845581, -0.30051448941230774, 0.0227916873991...
can do it on your own by keeping presentation & logic separate. This way the designer can screw up the display, but not the logic behind it. Here's a simple example: ``` <?php $people = array('derek','joel','jeff'); $people[0] = 'martin'; // all your logic goes here include 'templates/people.php'; ?> ``` Now here's the people.php file (which you give your designer): ``` <html> <body> <?php foreach($people as $name):?> <b>Person:</b> <?=$name?> <br /> <?php endforeach;?> </body> </html> ```
[ 0.2713891565799713, 0.05279012396931648, 0.30149251222610474, 0.15993596613407135, -0.12389886379241943, 0.1928013563156128, 0.24444031715393066, -0.3955998420715332, -0.4658251404762268, -0.6378417015075684, 0.025457916781306267, 0.44573691487312317, -0.1523694097995758, -0.09746488928794...
I see a lot of talk on here about functional languages and stuff. Why would you use one over a "traditional" language? What do they do better? What are they worse at? What's the ideal functional programming application? Functional languages use a different paradigm than imperative and object-oriented languages. They use side-effect-free functions as a basic building block in the language. This enables lots of things and makes a lot of things more difficult (or in most cases different from what people are used to). One of the biggest advantages with functional programming is that the order of execution of side-effect-free functions
[ 0.11583681404590607, 0.15064716339111328, -0.26638877391815186, 0.2895265519618988, -0.22607724368572235, -0.004164532758295536, 0.4172806143760681, 0.15844698250293732, -0.15306255221366882, -0.5583949685096741, -0.04250564053654671, 0.917244017124176, -0.3196958005428314, -0.272348761558...
is not important. For example, in Erlang this is used to enable concurrency in a very transparent way. And because functions in functional languages behave very similar to mathematical functions it's easy to translate those into functional languages. In some cases, this can make code more readable. Traditionally, one of the big disadvantages of functional programming was also the lack of side effects. It's very difficult to write useful software without IO, but IO is hard to implement without side effects in functions. So most people never got more out of functional programming than calculating a single output from a single
[ 0.15763339400291443, 0.2779761254787445, 0.07745411992073059, 0.36168384552001953, -0.3072993755340576, -0.3322158455848694, 0.18841154873371124, 0.010360402055084705, -0.2940606474876404, -0.4577295184135437, 0.1410476267337799, 0.44360238313674927, -0.3962158262729645, -0.251738280057907...
input. In modern mixed-paradigm languages like F# or Scala this is easier. Lots of modern languages have elements from functional programming languages. C# 3.0 has a lot functional programming features and you can do functional programming in Python too. I think the reasons for the popularity of functional programming is mostly because of two reasons: Concurrency is getting to be a real problem in normal programming because we're getting more and more multiprocessor computers; and the languages are getting more accessible.
[ -0.03259918838739395, 0.1834079772233963, 0.14316563308238983, 0.21639810502529144, -0.4496327340602875, 0.012486588209867477, -0.014879421330988407, 0.24457354843616486, -0.6167714595794678, -0.6554093956947327, -0.15578597784042358, 0.7598254680633545, -0.5663572549819946, -0.27875566482...
I have a page with a Google Maps mashup that has pushpins that are color-coded by day (Monday, Tuesday, etc.) The IFrame containing the map is dynamically sized, so it gets resized when the browser window is resized. I'd like to put a legend in the corner of the map window that tells the user what each color means. The Google Maps API includes a `[GScreenOverlay](http://code.google.com/apis/maps/documentation/reference.html#GScreenOverlay)` class that has the behavior that I want, but it only lets you specify an image to use as an overlay, and I'd prefer to use a DIV with text in it. What's the
[ 0.44587069749832153, -0.14253436028957367, 0.5567686557769775, -0.03601408749818802, -0.08059830218553543, 0.017866965383291245, 0.07061510533094406, -0.02251962572336197, -0.2905014455318451, -0.7952887415885925, 0.294330358505249, 0.31372952461242676, -0.12412573397159576, 0.002906414214...
easiest way to position a DIV over the map window in (for example) the lower left corner that'll automatically stay in the same place relative to the corner when the browser window is resized? You can add your own Custom Control and use it as a legend. This code will add a box 150w x 100h (Gray Border/ with White Background) and the words "Hello World" inside of it. You swap out the text for any HTML you would like in the legend. This will stay Anchored to the Top Right (G\_ANCHOR\_TOP\_RIGHT) 10px down and 50px over of the map. ``` function MyPane() {} MyPane.prototype
[ 0.21850283443927765, -0.1344493329524994, 0.4405602216720581, 0.2582917809486389, 0.211408331990242, 0.10849034041166306, 0.20884473621845245, 0.18589963018894196, -0.17029187083244324, -0.9604203104972839, 0.037748366594314575, 0.020296545699238777, -0.12587635219097137, 0.169870957732200...
= new GControl; MyPane.prototype.initialize = function(map) { var me = this; me.panel = document.createElement("div"); me.panel.style.width = "150px"; me.panel.style.height = "100px"; me.panel.style.border = "1px solid gray"; me.panel.style.background = "white"; me.panel.innerHTML = "Hello World!"; map.getContainer().appendChild(me.panel); return me.panel; }; MyPane.prototype.getDefaultPosition = function() { return new GControlPosition( G_ANCHOR_TOP_RIGHT, new GSize(10, 50)); //Should be _ and not &#95; }; MyPane.prototype.getPanel = function() { return me.panel; } map.addControl(new MyPane()); ```
[ -0.039281439036130905, -0.02111419476568699, 0.7980164289474487, -0.3490252196788788, 0.3267553746700287, 0.44470691680908203, 0.29109397530555725, -0.12034914642572403, -0.30238354206085205, -0.8889641761779785, -0.49567916989326477, 0.47154954075813293, -0.22519086301326752, 0.3083822429...
So I was listening to the latest Stackoverflow podcast ([episode 19](https://blog.stackoverflow.com/2008/08/podcast-19/)), and Jeff and Joel talked a bit about scaling server hardware as a website grows. From what Joel was saying, the first few steps are pretty standard: 1. One server running both the webserver and the database (the current Stackoverflow setup) 2. One webserver and one database server 3. Two load-balanced webservers and one database server They didn't talk much about what comes next though. Do you add more webservers? Another database server? Replicate this three-machine cluster in a different datacenter for redundancy? Where does a web startup go from here in the
[ 0.2012205421924591, 0.10840535163879395, 0.6071757674217224, 0.13026943802833557, 0.08922857791185379, 0.05569284409284592, -0.28465503454208374, -0.2856251001358032, -0.34138455986976624, -0.6186372637748718, 0.43608415126800537, 0.46839603781700134, 0.21003533899784088, -0.03625150769948...