query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
85fd551d-0b5c-1b30-31be-6d5ef404181b
['58949159-68f3-525c-2433-51a30f62ebb3']
. Hi, fellow 'Ham-sters' I hope everyone had a great Christmas, and that <PERSON> was good to you. I arrived back in UK yesterday after spending 7 days with my family in the Gambia, and happily got away before the hold-ups and bad weather here, returning to find all the snow had gone !! Back to reality for another year!! Great news!!! I ordered a UT106 from USA for 98 dollars (56 pounds)two weeks ago, and it got to this address yesterday just after me. I was amazed that I had no duty, VAT or anything to pay on it!! Cheapest here was <PERSON> at £99. The guy is still advertising them on US Ebay, so if you want one - jump in quick !! (RF Components - Walter Zapata). I have not heard much about the UT106 in use, so was anxious to get it fitted (a doddle!), and try it out in the PCR 1000. When installed, it automatically places another 'rack' on the ICOM component receiver which is toggled on the tool bar at the top with an Icon called 'DSP' - which even if you don't use the Icom software, is a good check that you have installed it correctly. I didn't the first time, as there is a little 'ribbon' connector that you have to push in to a connector on the UT106 and the IC1000 main board, I got it in the wrong way round!!! (Well, I am 75 with only one brain cell!). The scanty 'destructions' are not too clear on this - but after opening up the box again (8 screws), I then tried changing all the connections with the case open until finally working. I use 'Talk PCR' as my software interface, and usually MixW as my PSK 31 program, or Ham Radio Deluxe, with MMSTV for Slow Scan TV, and my ears for 'Phone' SSB or CW (Morse)... if the morse goes over about 25wpm, I use CWget. When the UT 106 is installed, with 'TalkPCR' you get a little window on the screen away from the receiver, which allows you to switch it on and off (to use DSP), switch the Noise Limiter on and off, and vary its effect with a little slider, and to switch in and out the notch filter. Hopefully thinking logically, I decided that to test this little Gizmo, I couldn't rely on my ears - nor did I have any of the sophisticated test equipment required to give a practical and subjective appraisal. What better than the 'waterfall' or 'Spectrum' viewer on a PSK31 program? .... I could SEE the results of clicking the DSP (Digital Signal Processing) in and out, as well as the effect on the 'noise' shown on the waterfall when varying the gain on the Noise Limiter. VERY IMPRESSIVE ... on 7.035 MHz at 9pm last night (Wednesday)... PSK 31.. pulled in stations that were hidden in the noise prior to switching in the UT106. Noise level at about half way. I then went to bed, as had been travelling and awake for the last 36 Hours - not good for this old fart !! Also wife not best pleased that I had done nothing else since getting out of the Taxi home, but to install and mess with radio gear, and eventually going off to bed without a 'Good Night' even, let alone the usual little 'peck' !! Thursday Morning - 9am - 11.30am - (Other half still in the land of dreams), have been on 20M (14Mhz) Amateur Band. Listened to some RTTY and PSK, (still impressed, and can HEAR the difference when I switch on and off), but now listening to 'Phone' USB - again... Wow!! the UT106 really sorts out the wheat from the chaff, quiet signals, just 'leap out' - Compared all the time, by the way, to switching to the same frequencies with my Kenwood TS520SE Transciever and my 'Frog' Yaesu FG7700 ...... this Receiver totally out performs them both. Ooooops !! - 'She that must be obeyed' has been activated !! Have you made coffee?, have you fed and walked YOUR dog? Did you go to bed? Are you still playing with your toys? Oh dear ... I had better switch to 'wife pleasing' mode! and I havn't listened/tried CW yet!! Better go folks - or I won't be having a 'Happy New Year' !!! I really hope you all do ... (Have a Healthy, Happy and Sunspottted New Year) Further impressions on the PCR1000 with UT106 will have to wait until next year, I guess. (Its only 2 days!!). 73's <PERSON> ....... or, Old Fart John
fe8c0df9-7800-af26-8400-06ae443ec482
['58949159-68f3-525c-2433-51a30f62ebb3']
. I have just uploaded two (printable - 20-odd pages!!) Files into the Files Section. Firstly a complete list of Amateur Call Sign Prefixes, listing the Associated Country. Secondly, a file (1 Page!) of Amateur Mode Frequencies for each Band for the various Modes, - CW, PSK31, RTTY, SSTV, etc for each Band. I hope these are useful to someone ! Regards, <PERSON>
bb5ee638-d37a-3923-924c-9d56a90cef2a
['5894d137-2412-fb94-a876-65fbb2b66a58']
Hey, I got the whole system to work. Also connected some cpes to genieacs. However, now i would like to directly call the rest api without the gui, but i cant find how the urls are built up and what functions/functionalities there are. I was looking in api-functions.js and nbi.js since that looked like the places to find this. Could you give me any directions as to where i can find this information. If i can find and use the possible calls, im also willing to document these for you. Kind regards, <PERSON>
67476c9b-e7c3-983c-697c-1b3f2dc60bb1
['5894d137-2412-fb94-a876-65fbb2b66a58']
Hey, I am running GenieACS with different CPEs, so far I had no issues connecting them to the ACS. However, now I am trying to connect an Icotera device to the ACS. The ACS does not seem to like that. When the CPE sends an inform request the CPE returns a 500 Internal Server Error. I have attached a packet dump file taken at the ACS. Does anyone know how to solve this? Might this be a bug in the server? Kind regards, <PERSON>
bdba24c9-9a6b-da0e-9038-b9cf2aa818f4
['58a7baf2-5218-51a4-e3e1-f468e6e8be91']
Hi, Can anyone please tell me what does "permutation_test_score"(and the p_value it returns) do in scikit-learn? I am assuming it outputs the statistical significance of the performance of regression models. I am planning on comparing the performance of various regression models if the performance measure they are reporting is statistically significant. To this end, I wanna output the p-value of the prediction first, and if it was smaller than a certain cut-off, I would then report the performance metrics, such as r2 and MSE. Do p-value and score outputs from "permutation-test-score" not provide me with what I want? <PERSON>
522e0a6a-89c6-2ed3-13df-6390560d5354
['58a7baf2-5218-51a4-e3e1-f468e6e8be91']
Hi folks, I have a question regarding how to use permutation_test_Score. Given data X (predictor) and Y (target), I hold aside 20% of my data for testing (Xtest and Ytest) and would then Perform hyperparameter-tuning on the rest (using Xtrain and Ytrain). This way I can get the best parameters via RandomizedSearchCV. I now want to call permutation_test_score to compute the score, as well as the p-value of the model prediction. But the question is what X and Y should I send as input arguments to this function? I could send in X and Y but then my hyperparameter parameters were already tuned to Xtrain and Ytrain, which are a part of X and Y and that would bias the output values. Any help would be greatly appreciated. Thanks, <PERSON>
3945767e-1a75-7d49-5c44-af17f682935e
['58a9c86e-4812-b84d-c515-847bd8c6d502']
Hi folks. I'm just wondering if there is any empirical or analyst data on the adoption and effectiveness of Agile methodologies in various types of organization. It's clear that there are advocates, detractors, critics and zealots on all sides, so I'm thinking empirically here. Specifically, I'd be interested in data or stories of Agile in the development of enterprise products with a large, diverse customer base and/or the development of product suites that must share common user experience attributes. Feel free to contact me offline, to point to references or to engage the group. Thanks much!
88b1d699-f383-8f7b-9312-599318f2c82f
['58a9c86e-4812-b84d-c515-847bd8c6d502']
Hi there folks - Boy, who could resist a goad like that? I'm a usability engineer at a Northwest software company with about 400 people. We have several dev teams that use different methodologies. One of these teams uses "Agile", and I put that in quotes because I can't say that it's "real" Agile. Rather, I have three assertions: 1. "Agile" is like soup stock - it's a set of tools/practices with properties that are beneficial in some cases. As with any ingredient it can be modified, spiced differently and made to be what is useful within an organization. Dogmatists begone! 2. Agile is not a frictionless surface. It is not product development nirvana. Agile can suck mightily, or simply be ill-suited for the project to which it is applied. There's a higher-level debate in business software about the locus of control between IT groups and Business groups. I think it applies to product development methodologies as well, and Agile seems to fall on the side of the Dev/Test side of the house. 3. Sadly, I think Agile can also be used as way to suggest "We can do more - quicker! Better!". This is a "mom and apple pie" argument that upper managements loooove to hear. But there's a cost, I think, in larger projects that end up being done haphazardly. Of course, none of this applies to my specific situation. Thoughts? Thomas
a331ac0a-03a3-72ce-0110-50e77515efbe
['58a9d7bc-403c-c3a3-7d97-f0dd47b75b6b']
Hi, i built my first kernel the Debian way and filed a kernel bug. To Debian for now: https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=798300 I decided not to mention ISO producer programs by "affects". They work fine with the kernel. How long should i wait before trying at upstream ? (Urgency is low. But when it begins to smell ...) Have a nice day :) <PERSON>
77c4242c-9477-f95e-ae60-dd5b4b696f63
['58a9d7bc-403c-c3a3-7d97-f0dd47b75b6b']
Hi, i can reproduce the SIGBUS only when gcc got option -O2. (I removed all -O2 from ./configure, ran make clean and make.) I can silence the error, with -O2 enabled, by replacing a->u = *(union w_list_data *)data; with memcpy(&(a->u), data, sizeof(union w_list_data)); So this seems to be a bug of gcc, unless somebody can point me to C prescriptions which would declare the "a->u =" variant undefined or elsewise ambiguous. I will test the error preventing code change in upstream now and will apply it if no obstacles appear. But we have to expect more such pitfalls if -O2 stays enabled in the Debian compilation of libburn, libisofs, and libisoburn. I really would had thought that -O2 was safe with gcc. Now i have to reconsider. I propose to try whether the strcmp() problem of genisoimage in bug 731806 is related to -O2 too. Currently i am waiting for <PERSON> to provide me with the input directory tree which on the first hand caused this bug report. Just to be sure that no other bugs pester this use case. Have a nice day :) <PERSON>
67752b0a-b3ac-85fc-14ce-cc957f2de62c
['58c8a046-8aa9-ac58-b015-db643d42e501']
Hi There, The discussion about speex triggered the following question: Is there a way to use other encoders, to stream with icecast / oddcast. Is there an format of witch the encoder shoul be standarized to (ACM or something like that), if so would it be possible to use open license encoders and commercail encoders with this. like FHG mp3, mp3pro, speex and so on. regards, <PERSON>
24d10a43-049b-386a-9971-2afd92d4648b
['58c8a046-8aa9-ac58-b015-db643d42e501']
Hi There, Is there a nice helpfile to telle me how to set up: * Multi-Level Fallbacks Multi-level fallbacks allow for specifications of a series of fallback mounts that you could use, for instance, to set up a series of progressively lower bitrate streams that would be cascaded through. For instance, a modem user would connect to the highest bitrate stream, and then could be cascaded down to a progressively lower stream until they reach a value they can handle. When i put up a faalback, do i have to remove other normal settings in my configuration of icecast 2.1.0? Regards, <PERSON>
fb7ba67a-3416-171d-b847-e6c712b8d7af
['58d1b7af-5705-6c3c-43ea-ae3a9d70f2a5']
Hi all, I'm looking for a way to serve random error pages for 504/502 errors. In addition i also want the error pages to be served based on the $request_uri. Something like this: 1) When this URL is requested: http://mydomain.com/search?type=ABC and if the proxy returns a 504 or 502 and when the GET parameter 'type' has 'ABC' as value then serve any random document from "/path/to/custom/error/typeABC/" 2) and the same applies to http://mydomain.com/search?type=DEF just that the directory to be used to serve error document should be "/path/to/custom/error/typeDEF" I'm sure this can be achieved with named locations and probably try_files. But I'm just not able to get anything working. Would appreciate some pointers. Regards Harish
317ba6ed-7167-fb78-8280-65b45b10f9db
['58d1b7af-5705-6c3c-43ea-ae3a9d70f2a5']
Hello All, I have a few questions on backup servers and proxy_next_upstream in general: For the backup server the doc says: "only uses this server if the non-backup servers are all down or busy" 1) So what exactly constitutes down or busy? Is it just response codes - 502,503 and 504 or anything else? 2) Will the same request which received such an error be sent to a backup server or is it only subsequent requests? 3) Can I say 504s happen when proxy_connect_timeout/proxy_read_timeout is reached? Is there a way to find out which among these is happening? 4) Also, In proxy_next_upstream : error — an error has occurred while connecting to the server, sending a request to it, or reading its response; timeout — occurred timeout during the connection with the server, transfer the request or while reading response from the server; I'm not able to clearly distinguish between both of these. What kind of scenarios explain these? 5) In a case when proxy_next_upstream is off and when a backup server is provided, will nginx still try all non-backup servers before sending it to backup servers? Thanks <PERSON>
96863daf-8f3f-5bb5-f079-7841429d11cd
['58f61491-993b-0f58-a7bb-d3d5188fe2ee']
I'm trying to use both these packages and wondering whether they are possible... To make this simple, my ultimate goal is determine long stretches of 1s, but I want to do this within groups (hence using the data.table as I use the "set key" option. However, I'm I'm not having much luck making this possible. For example, for simplistic sake, I have the following data: Dad Mum Child Group AA RR RA A AA RR RR A AA AA AA B AA AA AA B RA AA RR B RR AA RR B AA AA AA B AA AA RA C
dceadccd-a18b-1e97-b701-1637d8524336
['58f61491-993b-0f58-a7bb-d3d5188fe2ee']
Hi All, I'm trying to merge two files together using: combinedfiles <- merge(comb1,comb2,by=c("Place","Stall","Menu")) comb1 is about 2 million + rows (158MB) and comb2 is about 600K+ rows (52MB). When I try to merge using the above syntax I get the error: Error in merge.data.frame(comb1, comb2, by = c("Place","Stall","Menu")) : negative length vectors are not allowed Is there is something that I'm doing wrong? I've merged larger files together in the past without a problem so am curious what might be the problem here... Thanks in advance! ~K
5c9dfe0e-18c3-55a0-0092-4927f2833f19
['58f8f2b2-66a5-f8b6-3ef6-5ff211086bb7']
Hi everybody, it has been a long time since i've been toying around with my wiimote, but i remember that it was'n hard to make it run. Now, i tried to make it functional again (on an Arch Linux system). Wmgui, hcitool sca, everything works fine! BUT, the wiimote external doesn't (i downloaded version 0.3.2 from puredata.info). It compiles fine, but discovery doesn't work (again, no problem with wmgui!). Ok, i can connect with explicit address. But still, only button data is processed, i can't see any acceleration data or IR data ... (i can see it in wmgui). A year ago or so, everything worked quite well ... Any hints ? Regards, <PERSON>
44e6093f-c72c-ae32-e3f9-c90c39425c3c
['58f8f2b2-66a5-f8b6-3ef6-5ff211086bb7']
Hi everybody ... i've been trying to make haskore-realtime (and, in the long run, haskore-supercollider) work in the year 2014 (OS: Arch Linux, 64Bit), but had no success compiling the package so far, even in a sandbox (directory-<IP_ADDRESS> fails to build ...). Would it be hard to upgrade to recent library versions ? Though i have some knowledge of Haskell as a language, i'm pretty new to the development system ... Regards, n
98c7dae1-a654-831c-be40-cc7df6df8219
['59132748-20f1-280b-c3e3-7ab439ae7d9e']
RH 9 Intel I just got an NCD Explora X Terminal from Ebay - a whopping $15 for the unit. I wanted to experiment using my RH 9 installation as a platform for learning this technology. I figured I'd have a cheap X-Terminal solution - I'm new to this. However, to my dismay, I've found that I've got to get the NcdWare software in order to run this unit. The problem is that the software has a whopping $995 list price! Does anybody have any ideas where I can get this stuff cheaper, used, second-hand, etc??? Thanks all.
25ff35fa-7e82-3535-7249-353fdd726f0f
['59132748-20f1-280b-c3e3-7ab439ae7d9e']
RH 9 Intel I've moved my hard drives from the built-in IDE controller to a PCI IDE card. What do I change to indicate to the OS to boot from the new partitions? Previously, the OS booted from /dev/hda and there were partitions on /dev/hdb also. When it boots now, I see it trying to find drives on hda, hdb, hdc, etc. Finally it finds the OS stuff on /dev/hde. Is there a way to avoid this and indicate to the process to go directly to /dev/hd?? to boot. TIA.
587c7e70-476f-9d3b-a174-34a1194d5fd6
['5929378a-662c-3889-eda3-59b80acaa31a']
I agree with others in that the O'Reilly Python books are very good. But if you want to get up to speed quickly or as a good refresher if you've been away from Python for a while, I think there is no better book than <PERSON>'s "Python", part of the Visual Quickstart Guide series from Peachpit Press. Not only is it up to date, having been published this year, but it includes plenty of quick examples and is organized in a manner that makes it very easy to find the topics that you are looking for. -Barron
524519f3-da0e-c269-1c30-b47b86317d42
['5929378a-662c-3889-eda3-59b80acaa31a']
Hello, I'd like to access a database (specifically Oracle) from my Python code and found cx_Oracle through the Vaults of Parnassus: http://www.computronix.com/utilities/ They don't offer a package for OS X (surprise, surprise) but one can download the source: http://www.computronix.com/download/cx_Oracle-2.4.tar.gz How hard would this be to compile for OS X? There is no visual interface for this which I think would make the port much easier. If it's not too hard, would someone tell me the basic steps (or point me to a quick tutorial)? Thanks, <PERSON>
ac4d1d0c-bd2e-8d5a-bb60-7a03b72680ae
['5935edf8-1bc2-992a-bc95-eba55704e664']
Hello, I have difficulty with the search method of the re module, when applied to an object retrieved from urllib.In the following program, f3 is read once and the search is effective. But when I try to read it again, it acts as if the index of a file had to be "rewinded" (e.g. as would do a "seek" on a "true" file). I thought that f3.read was to be considered as a string. Could someone help me understand my mistake and give me a nice solution (means not storing data into a file). Thank you in advance.
8cb9ce0a-6ad0-cbcc-5046-4f52009de85d
['5935edf8-1bc2-992a-bc95-eba55704e664']
Hi <PERSON>, Thank you for your investigation. It helps me understand that my python code is correct. I did not forget any character, while the name of the url comes from regex extract of the page returned by the server itself. But you're right on the fact that the name has something wrong, when used in this way. I tested the web site "manually", that is pasting the entire name of the web site directly into my Netscape. Same result : "Netscape can't locate the server. No host specified" (this message translated from French) Therefore the question is : how does it come that a Javascript function using 'document.location="http://blablabla"' is able to redirect my browser to the said web site, and not 'urllib.urlopen("http://blablabla")' ? More, I do not understand the error message. Does it come from the server which denies the request *or* from the dns which cannot translate the name into an address *or* from Netscape which seems not to receive/understand the host name ? Regards
61c5aaa2-c5d3-9569-c801-4feb19e252ee
['59732b8f-68cb-0d5c-3698-7dd2e150d7d1']
Hi <PERSON>, I've set up some solutions like this for some customers. What we have done is the following <IP_ADDRESS> A web server in DMZ (Apache or IIS, doesn't matter technically, but don't want to restars religious war..) Exchange server in Lan, with IIS. Because IIS as to speak RPC with Exchange, and we don't want to see this protocol trough the firewall, Exchange 2000 permit to put an other web server in Front, and requesting only HTTP (A kind of Proxy) Users have to use a 128 Bit SSL Link from client to Proxy in DMZ. Then HTTP from DMZ to Inside. Concerning Anti-virus : I hardely recommend Sybari's ANTIGEN. A very good and stable AV for Exchange, using 2 different Engine (List is long. See www.sybari.com) So POP 3 was not needed. Concerning Incoming mails, you can add a layer with installing a Mail relay in DMZ. We use sometime an IIS with GFI Mail essential.. Very good and stable. (Win32 shop) Or you can put some *nix mail, with anti-spamming rule, etc... A good option is to mix the OS's : *Nix in DMZ, and Win 32 in Lan, in this case. You can contact me off-list, should you require more detailed info and Schema we already have done... Useless to say that all DMZ and Intern server have to be fully patchen downstripped, etc.. Sometime, a good IDS in DMZ (Snort... works on *NIX and win32) add some more security. Good luck Max
95b99663-7230-9cbb-8a23-7771b07dc87c
['59732b8f-68cb-0d5c-3698-7dd2e150d7d1']
Hi <PERSON> Just my 0.02€.. I tried using Nessus on my laptop, within vmware (version 3 fow Win 2000, demo 30 days). I had no problem with it. Unless with a wireless card, that vmware can't support. In demo version. It has worked like a charm with a Free BSD minimal version, and with some other distro (RH7.2 and other i don't remember). I really think about buying a full version of VMware. So I can stay using Win 2000 for all common stuff, using a win32 client for Nessus, and Nessus runing in a VMWare session. Very fine and worked like a charm. No need for dual-boot, nor to change the OS, etc... And I can use a graphical client, which i feel more comfortable with. Definitely a very good solution, perhaps the best, For me at least. Just my experience. Max
4a6ca489-de14-6c04-8b5d-3fdf18fa8715
['599247cf-6555-e84d-6623-cadc982bb269']
I've got a K2000 that I've created a drum fill program. your basic snare, toms, floor tom and crash cymbal, to use on transitioning from verse to chorus etc. the snare is what I want to edit. at the end of the sample, it's sounds gated, like the end is being sucked down. this I don't like and have played with different alg., env. LFO's...with no luck. anyone know how to rid the sample of this? When I access the snare from EditProg*KEYMAP, the snare does not possess this characteristic. thx. in advance. <PERSON>
aa7a015b-1e8d-177d-f618-9c88be8dec8a
['599247cf-6555-e84d-6623-cadc982bb269']
my original hard drive started intermittently working, then not working. so I bought a new used hard drive off eBay. I was able to format it and save a couple of files to it, when it spun down. the disk page says it's asleep. when I try to access it, to load or save, it says "problem mounting disk". I'm beginning to think my original hard drive is still good. If I power down, wait, and power up, the disk spins up, but I get this clicking sound, kind of a knocking and the disk page reads asleep, even though it is spinning. anyone experienced this one before? thx. <PERSON>
29b6dcf4-7e65-a339-f7a1-0c68065a594f
['59955b2d-6552-26ac-6cf2-a31c02c62eb4']
When we run a test suite, some tests might fail. We get a report in the test-output folder. Now, if we want to re-run the failed tests, we get a different report. Is there a way to re-run failed tests and have the original report updated? Basically, I have a condition like this: I am doing UI Functional Automation using TestNG. We have 200 test cases right now. At times, a few tests will fail. I can re-run the failed test cases, but I then do not have a full report to send out. I can send out the report with the failures, or I can send out the report with the re-ran tests. I would like those combined so, to those looking at the report, it appears there was just one test run. Any way to do this in TestNG? <PERSON>
a5972352-af59-8c33-d050-ef405a2c962d
['59955b2d-6552-26ac-6cf2-a31c02c62eb4']
Hi, I am also new to this community and to TestNG. So far, I think TestNG is much nicer than JUnit, but I am also doing functional testing rather than API testing. I noticed the same thing that <PERSON> mentioned, the full class names are displayed rather than the SimpleName. I also noticed that JUnit includes the time it takes each test to run (in seconds). TestNG is only displaying the total execution time in milliseconds. Having the time for each test I found to be usefull since I am often asked what part(s) are taking the longest amount of time. With TestNG, I can not answer that. Just my two-cents for an improvement to go along with the SimpleName display. Thanks, <PERSON>
c4c1b454-5bf5-6341-3e33-614d8aa7b96b
['59962112-16c1-d942-aeb7-851536308183']
Hi, Ok so i have managed to get my html template working in a file. So the basic examples are like this e.g. main.go ---> template e.g view.html (section) ----->section is read from .txt file e.g. test.txt Now i want the following e.g. main.go--->template e.g. view.html(page section 1) ----- section 1 read from test.txt (page section 2)-----section 2 read from read.txt How would i do that? regards, Will
e075cf9a-2d29-5bbe-694b-3cd0b1f33e41
['59962112-16c1-d942-aeb7-851536308183']
Howdy <PERSON>, Is there anyway of getting web access statistics from a pure golang server (built in a way as described in this article http://golang.org/doc/articles/wiki/)? e.g. number of unique visits, per day statistics, region statistics. How would one go about building this feature in a golang webserver? Note that there is no front-end apache server nor nginx server. regards, <PERSON>
22d07d92-5025-3c0b-e13e-bd677b20ba99
['59ad3b12-1bd0-d6a1-8456-d16fcb1bebff']
I should like to use SQLite from within a VC 2003 NET program. Is this documented somewhere? I do not know which parts to dowload, to start with. I downloaded the .dll but it comes with no .h file and now I am confused. Is there somewhere on the net a sort of tutorial or at least a C++ example? TIA, jb
d468b9e4-8ecc-cb2a-7785-2a74e61a3489
['59ad3b12-1bd0-d6a1-8456-d16fcb1bebff']
For list-related administrative tasks: http://list.cs.brown.edu/mailman/listinfo/plt-scheme I have started playing with plt mzscheme. I should like to extend mzscheme (and later mred) with a C++ file that uses a C++ library. Now when I compile and link the first "hello world" program from "Inside MzScheme" it is compiled but not linked as __gc_personality_v0 is missing. I called the file hw.cpp instead of hw.c. I have had this error before and it is a C++ problem. What can I do? I use Cygwin (and gcc3.2). TIA,
d74f729b-b6c5-26dc-80ce-514b074dc034
['59b034c1-c23a-c2f8-ca69-40767f4633d4']
<PERSON>, I too built up on an open usage platform and yes, when the subscribers logged into their PowerCode portals and viewed usage charts I got plenty of calls. We have not yet implemented metered billing because the pipe is still not capable of delivery, but soon. What I told the concerned callers was pretty much what I explained previously, that a small percentage of subscribers are utilizing the majority of the system’s resources and that it was effecting everyone. I went on to explain how the goal was to charge those that use more services for their usage, and assure resources remain available for low volume users. I also add that based on FCC regulations I can not restrict any specific type of traffic, so this is the only fair way to assure everyone gets what they want. I tell them that our pricing model will not change cost to about 80% of our subscribers, and the other 20% will see increases based on actual usage. Many are fearful because they see the abusive rates charged by cellular carriers for small packages and immediately thing we are going to start hammering them for $150 per month. Like much of what I have read here, I too am looking at about 30-50 GB of transfer as a base with a small per GB cost. The real value to the upgrade for me will be once we demonstrate we can deliver a solid stream that people that are trying to pull multiple streams will have the option to doing so by upgrading to a higher bandwidth package. And that is the point I was making before, that the amount of transfer has little to do with the pipe size, but that size does impact the subscriber’s ability to have concurrent streams. So we are really focusing on three things; first, we are separating the basic and power subscribers, then we are offering those power subscribers the option to get whatever they want, providing they are paying for it. Sure a few will be pissed because they have this entitlement to unlimited service. Tell them you will start the day the power and gas company remove their meters. In the long run, the decisions made will provide maximum benefit to all subscribers. Perhaps we will see a few that refuse to pay and leave, but we will increase significantly as word gets out about our new capabilities. Remember, all those smart televisions need a pipe to connect to these streaming services. And that is the simplest answer, your changes in billing are to accommodate a market that did not exist when you deployed. When you and I put our systems in place Netflix was not streaming. So we absolutely must accommodate these new high demand users, while acknowledging the long time basic users. Just remember that many of them will move to the other side over the next few years and be very glad you were able to accommodate their new requirements. <PERSON>
95feb1c5-ae5f-4088-1f58-628be07f7c7a
['59b034c1-c23a-c2f8-ca69-40767f4633d4']
I believe <PERSON> to be correct. Packages based on speed are not the answer. We call our connection a “pipe”, so let’s use a related analogy; You can have two homes with water service. One is an older home that has a ½ inch water main, the other is new construction and has a 1 inch service main. House number 1 has the original fixtures, so the toilet uses 6 gallons per flush, the shower flow is 7 gallons per minute and the clothes washer uses 40-55 gallons per load. House number two, being built under new codes that promote conservation has a low flow toilet that will use 1.6 – 2 gallons per flush, a low flow shower head that restricts flow to 2.5 gallons per minute and a new clothes washer that uses 20 gallons per load. With a family of 5 in each house, it’s easy to see that , despite the smaller service pipe, that house number 1 will have many times the water usage as house number 2. A smaller pipe did nothing to control the flow because the flow limit of the pipe was not reached. Those two pipes are exactly like a 3 meg and 5 meg Internet connection. Within reason, the size of the pipe will do little to limit heavy bandwidth usage. It only serves to spread it out, creating a longer period of time that it puts a demand on our networks. Like most, we saw our network performance begin to deteriorate as Netflix switched from a physical to a digital delivery system. The others since then have continued to slow our once speedy connections. Now we, as an industry, are faced with a continued rebuild to meet a voracious demand for bandwidth to deliver content that we never intended, or anticipated. Worse yet, we are being positioned to provide these improvements to support the business model of companies that barely acknowledge our existence. And they are getting smarter in their use of our pipes. There was a time when if you didn’t have a good 4.5 meg flow, Netflix would not stream. They have gone to much more advanced encoding that will adjust to feeds of less than 2 megs, rendering a 3 meg rate limit useless in defending against them. The issue of Net Neutrality somehow became synonymous with no caps. It appears we are the only service that is viewed by consumers and governments that should be given away. Services like water, natural gas and electricity are each brought to a home and metered for actual usage, because it is the only fair way for those that use these services to pay their fair share. In most locals, the billing is specifically broken down into two parts. The first addresses the base cost of the connection to the property, and the second reflects the cost of the metered usage. How is Internet different? We are a service that delivers a commodity to be used and never recovered. The bits of data we move for our subscribers are no different than the kilowatt, gallon or therm moved by the others. Could you imagine if consumers demanded there be no metering on these services? We are being restricted by network limits from delivering the full pipe to subscribers. This limitation is a function of cost. Under our current structures we cannot justify the cost of building a large pipe to each subscriber. After all, we are an industry built on contention. This sharing of bandwidth was the impetus of the WISP business for many years, but that concept has outlived it’s usefulness. Our subscribers no longer want to surf the web or check email. Most now do that on their smart phones. No, our pipe has become an unwilling player on the next pervasive shift in the paradigm, as subscription video shifts to a digital delivery medium. Just as VoIP has been disruptive to POTS, and satellite was to cable, we are on the cusp of the next trend in consumer electronics. Televisions today are being built with Ethernet ports and wireless networking. They are coming with built in apps for all the streaming services. And they want all this to work over OUR pipes. So we need to face reality and understand that if we don’t provide these services, we have become useless to our subscribers. Our failure to respond to this trend will throw the doors open for someone to come in to our markets and pluck each of our hard earned subscribers until we are decimated and a faint memory. If you think subscribers are satisfied with basic Internet services today, you are in denial. The answer is we need to build out robust networks that can deliver copious amounts of bandwidth to our subscribers. Our repayment will come by employing the time proven practice of metering for usage. We can divide our subscribers into two groups. The cutting edge-tech savvy type that is creating our issues, and the rest who will be joining them. I am sure that most of us have similar network statistics. If I look at one of my network segments I have the top 4 users consuming 25% of all bandwidth. I hit 50% at the 13th subscriber. This is a change in trend. It used to be just 8 that consumed 50%. And yes, bandwidth consumption has increased accordingly. This change from 8 to 13 subscribers being in the top 50% indicates my high usage subscribers have increased by 120% in roughly the past 6 months. Post holiday season I expect to see at least a 300% increase in my high usage subscribers, which without changes to my network, will bring data flow to a standstill. So build and meter. Don’t ignore the elephant in the room referenced earlier in this discussion. Just look at copper phone lines that peaked at 186 million in 2004 which today number about 84 million. In just 9 years, pureplay VoIP, cable VoIP and cellular technologies caused a 55% shift in a once-thought untouchable market. Joe
9cd3bca3-c9a3-687b-4fef-e7d521bb15e3
['59c7d96a-cbab-d583-854f-465a3da235eb']
[Qemu-devel] [PATCH 0/3 v5] Live migration without shared storage I tested the live migration without shared storage for the patch of the block migration. It succeeds for several times but sometimes it fails. I used the qemu-kvm-0.12.3 and the kvm 2.6.33 module on the ubuntu latest version. In the failure case, i can see the shell on the destination site showed the correct content for the migrated VM, but it just hanged there without any response to the mouse and the keyboard. I can see it prints "migration completes" twice. Is it a bug or did I do anything wrong? Another question is about the command line for the destination site. For memory migration, i will start the vm based on the shared image. For block migration, there is no shared image. What i did is creating an empty image with the same format (e.g. raw or qcow2) as the source migrated VM. Is it correct or it doesn't matter which format it is in the destination site? For the memory migration, it emphasizes that the path for the VM should be the same as the source path. But for the block migration, i change the destination path to a different one and it also works. Does that mean path is not important now? The command line must be "migrate -d -b ..". If it is "-b -d", the qemu-monitor doesn't understand. Does that intend to be designed like this or will it be modified in the future? Thanks! -Jie
bc7c583d-c65e-355e-12bd-2e575f7196a3
['59c7d96a-cbab-d583-854f-465a3da235eb']
Hi, I am confused about the main thread, monitor thread and migration thread. Hope somebody can give me a help. I observe that when a VM (smp=1, which is a file server) boots up, there are two threads for qemu (e.g. i use "top" and "ps" to monitor). I think one thread is for the vcpu. The other thread is for the main thread. Is the main thread the same as monitor thread? When clients connect to the server and start running workload, one more thread is generated. I think this one is I/O thread. Am I right for understanding these 3 threads? When migration (memory and block) starts , 3 more threads are generated. What are those 3 threads? From the code, it seems there is no new thread generated for migration. In the mail list, there is a series of discussion about whether migration-tcp should be threaded. The conclusion is no. Is migration using the same thread as monitor? Is qemu-kvm or linux kernel responsible for thread scheduling? Does KVM has its own algorithm or it only let Linux kernel to handle? If I want to understand the thread creation details, where is the code? I search in the qemu-git tree and only see the VCPU thread is created in qemu-kvm.c. Any other portion that i should look into? Thanks a lot! -Jie
ee4f7708-a134-0ec1-5228-ec2574a47564
['59c7fff9-56e3-dc91-212c-e20a2ee90fb7']
Hi all I tried to install a few extensions to my mozilla on debian testing, especially mycroft and google bar, but did not succeed. It looked as if it were installing something, but then nothing changed. I very much guess that it could have something to do with permissions. Although I don't see a reason why it should not be possible to install an extension just for a single user in his home directory. And since I did not find any hint on the internet I now ask you directly: Is there a howto somewhere for installing mozilla extensions in debian? Thanks a lot in advance, <PERSON>
d0563edd-b562-6bfc-8f01-b38aad5d19ad
['59c7fff9-56e3-dc91-212c-e20a2ee90fb7']
Hi As <PERSON> said before, if you're running unstable, the official version is surely the best. If you're running testing though, you're probably better off using the apt sources listed in http://cran.us.r-project.org/bin/linux/debian/ReadMe since there are some dependency hassles with installing the unstable version on testing (at least I got tired of going through all the dependency stuff). Cheers, <PERSON>
9d9eb052-196d-1007-d2c6-63d7a50f8647
['59cb8ef0-b5cd-2617-4632-9aac5f26f849']
I'm helping start a robot lab and we are considering using aibo + pyro for our first jaunt. One issue is that we will definately need microphone input from the aibo, so I'm wondering if anyone has ever got that working in Python/Pyro before? One goal of the lab is to maintain abstraction from the hardware details as much as possible (e.g. buying off the shelf robots), so I'm hoping it would not be difficult for me to get the microphone working. Any help is greatly appreciated, <PERSON>
a2d66c98-ad26-84db-d063-b2f9e6bb9591
['59cb8ef0-b5cd-2617-4632-9aac5f26f849']
Hello, I want to port Mono to a 2-issue MIPS-like processor and for starters I'm quite happy if only one instruction slot is useful (and the other is a nop) in every 2-instruction-bundle. I need these instructions to be aligned in 64-bit as not all instructions are allowed in each slot. I was going to start with the current MIPS port as a starting point (thanks <PERSON>!) but wanted to find out if the alignment issue is easy to satisfy or if it will be very difficult to enforce the instruction alignment. Thanks in advance, <PERSON>
6d4829c1-cae8-3d23-925c-4561dca963c1
['59e0fd0f-1f29-a997-9e17-24dceb860f32']
Greetings, My LTSP users authenticate to a Mac OS X Open Directory server, and it works great. However, we allow our student users to have relatively simple passwords. The problem is when we try to change passwords from within the fedora core 4/ltsp environment. Fedora requires them to create a strong password. Is there any way I can avoid the strong password, and allow them to choose whatever they want? Thanks, <PERSON>
b91f6785-5436-0f11-9d0c-086ac6b217e9
['59e0fd0f-1f29-a997-9e17-24dceb860f32']
Greetings, Does anyone have any experience using the newer dual-core opterons in their ltsp servers? My most recent servers have all been dual processor opterons, but as opposed to spending $180 for a single core opteron, I can spend roughly $330 for the dual-core versions. I'm thinking that two dual-core opterons would offer incredible performance in our lab of approx 30 clients. Any thoughts or advice? Thanks,
c453c289-e1b0-2d47-71df-3addfa0776b3
['59ed2a33-2ece-6997-89bc-a13c38efd95e']
Hi folks! I have a Newton 2100, good condition for sale.Everything works fine, no scratches on the screen, the only thing is, that the contrast button is pretty much on the left side. I also have a Newton leather case,the one where together with the keyboard, you work like a laptop.... new Newton keyboard, dongle, 2 serial cables, one for mac, one for pc, 56k modem card, 2 wifi wlan cards, lan card, 2 power adaptors, one with european+traveller adapters, a bunch of software on cd,memory flash cards,etc. I also have an apple stylewriter II in very good condition, which is perfect for Newton. If You are interested in all or part, please write, so I can send a detailed list and pictures. I also have an Apple Powerbook 2400c with Newertech G3 processor, Mac OS 9.2 installed, and a second Powerbook 2400c, original, for parts or so. And some RAM to it too. Please reply to: <EMAIL_ADDRESS><PERSON>. If You are interested in all or part, please write, so I can send a detailed list and pictures. I also have an Apple Powerbook 2400c with Newertech G3 processor, Mac OS 9.2 installed, and a second Powerbook 2400c, original, for parts or so. And some RAM to it too. Please reply to: pcMJ+PhCL0DwC8k6@example.com Thanks: <PERSON>
7534df36-8ac3-6558-b75b-f2b46561914d
['59ed2a33-2ece-6997-89bc-a13c38efd95e']
Hi! My newly accquainted eMate is working fine, the only problem is, with different position of the screen and moving backward and forward, the backlight is coming on and off, and sometimes it queeks (makes strange quikee noise and some lines appear on the screen. I am new to eMate, but this is the "hinge problem", right??? I do have a repair guide from the internet already, so I will try. Was just a question to make sure. Thanks: <PERSON>
f148f4d3-3a14-2699-4a4f-b40e3b9dba05
['59f3e545-2f5a-5802-feec-a3c979a7a273']
Greetings all! I seem to be picking up on a couple of different ways to create worlds for croquet. In following through some of the tutorials for squeak and croquet I see worlds as morphic/teapot projects of some sort; though in certain discussions I've been involved in there was a concern for the ability to employ 3DSMAX or similar tools to author the world and export it as an ase-type file. Would anyone care to clear this up for me? I was rather under the impression that world building/artifacting tools would somehow be integral to the croquet environment. Is this a misconception? Thanks in advance!
37311d46-fc75-5bff-26f0-f4161ef5c9b7
['59f3e545-2f5a-5802-feec-a3c979a7a273']
I have three different (squeak) installations right now- one a browser plugin, one that came with the croquet project package, and the last is the free-standing (read no croquet) 3.7 squeak release. Is there any way I can load croquet up independent of underlying squeak? say, for instance, load croquet into my squeak 3.7 context? Best Regards Twitch
3db62ec3-e561-d1be-861c-096826f2e828
['59f5a87d-5491-f202-97ec-6b209d0c8c14']
Is there any work being done to create a TreeTable like component for Qooxdoo? This is a primary component of our interface and if it does not exist or is not in progress already for Qooxdoo, I may have to take a stab at implementing myself. If there is no such component, would it make more sense to build on QxTree or QxListView? Thanks! <PERSON>
33e0bdb8-0c48-7138-be74-70c1971b4284
['59f5a87d-5491-f202-97ec-6b209d0c8c14']
I am curious about some "construction" decisions. Is there reasoning behind making QxListView from "div" components rather than "table"? My main curiosity is we are evaluating QooxDoo for some major internal applications, and want to make sure we understand as much as possible before making a decision. One reason I ask this is we have had to deal with large tables and sorting here, and have had to do some interesting stuff to make them fast - we have at least one app with a requirement for tables containing 5000 rows with sorting capability - with IE support a must. For QooxDoo, if you load QxListView with 5k rows in the demo, it dies.. it has actually crashed my IE, and when it dosnt, it just falls asleep.. even with only a few k rows, sorting is unacceptably slow... Thanks <PERSON>
33bb99b2-f583-3ded-1b7c-c25c4aba868d
['5a02e3bf-2fc2-bf42-02c3-8b11c5800ffa']
We need to change the groupID/artifactID in our POMs but we want to continue to add to the historical data that our project has created so far. Currently when we change it. Sonar thinks is a new project and starts a new history. How can we rename the existing data in Sonar to appear to be from the new groupID/artifactID ? <PERSON>
0cb739a1-f1a5-ad4b-3b6e-3d9ebd2e722c
['5a02e3bf-2fc2-bf42-02c3-8b11c5800ffa']
I was wondering if there was a way to identify newly added classes ? I noticed that some developer checked in about 30 new ones and I don't know where they are. I didn't really see anything in the documentation, though i suppose I can check the Hudson logs for that day :-) <PERSON>
0f1c54c3-0e05-6e3f-67a3-496c02fcf1e1
['5a0cb908-ac51-e8bd-289a-a25d65f3644f']
Dear Fail2ban users, running fail2ban v.0.10.1 on an opensuse box. currently looking into the recidive jail to ban persistent abusers. From what i understand the bans are stored in the persistent database storage so the bans can be added on restart without re-scanning the logs files. If i set a bantime of 1w in recidive jail the jail.conf informs me that i should increase the dbpurgeage to 7.5 days so the bans with 1w can live long enough before getting purged but if i do a permanent bantime -1 what value should I set the dbpurgeage? what's the relation between bantime, persistent storage and dbpurgeage? would be nice if someone could perhaps enlighten me on the topic ;) thanks & greetings <PERSON>
413a7ad5-ba9f-3e8f-b151-ff0fe7857d34
['5a0cb908-ac51-e8bd-289a-a25d65f3644f']
hello again, my current setup is varnish 5 in front of apache web server with 30 different domains / websites in my vlc i can switch between sites and decide to do varnish caching or just passing to the backend with out caching. what's the overhead / delay with varnish when i just pass everything to the backend and back to the client without varnish doing any caching? i mean there must be some overhead as varnish is in between ... ... but how big is it? time delay? does anyone here have any experience with this? thanks & greetings <PERSON>
93357ec4-b35a-a268-5055-d08aa82170c0
['5a45f8f8-1b41-f2ec-6403-7f9c455480d0']
Hello folks, I'm running into a hard time trying to add new OIDs in the certificates issued by OpenCA. We already have a nice structure running, with an online RA and an offline CA, but now we are trying to add more information in the certificates, specially some info needed for the ICP-Brasil (Brazil PKI). The new OIDs should be in the Subject Alternative Name section, and all in otherName format. The OIDs are the following: <IP_ADDRESS>.3.1 - 51 charactes max <IP_ADDRESS>.3.5 - 41 charactes max <IP_ADDRESS><PHONE_NUMBER> charactes max <PHONE_NUMBER> charactes max <PHONE_NUMBER> characters Has anyone made this work? Or even anyone already customized OpenCA to be ICP-Brasil compliant? Thanks,
97abccbe-f446-a63a-f575-6dd919ea6648
['5a45f8f8-1b41-f2ec-6403-7f9c455480d0']
Hello all, Here at my company we use the full set of TrendMicro products, but we're a little disappointed with some facts of the solution, so, we're thinking about changing the product. We've been hearing good things about Kaspersky lately, then we would like to know from you guys if you have any experience with it, went through a migration, what about day-to-day work and, most important of all, if you had any outbreak while using it. Well, Kaspersky is one of the solutions we are looking after, but any other anti-malware product that you recommend for a 10000+ user company would be nice to hear. Best regards,
5f494790-3d68-a18f-3b5d-5b7250b090b2
['5a45ffae-2fa8-0d04-806f-2b15e75e1439']
Hi, The current implementation of netbeans binds many file extensions to mime types in the openide file filesystems/FileUtil.java. Most of these bindings are reasonable. However, The extensions "c", "cc", "c++", and "h" are bound to the mime type "text/plain". This makes it imposible for C and C++ files to be recognised with the correct mime type unless they use nonstandard file extensions. The extensions "C" and "H" cannot be registered as C++ extensions either, because case is changed during the binding. These extensions should not be registered in the openide but should be left for registration by modules adding C and C++ support. An alternate approach would be to bind these extensions to something like "text/x-c" and "text/x-c++". I don't like this approach as well, but would be willing to generate a new patch if thats what the netbeans.org consensus preferred. There is a second mime extension binding problem which could be more controversial. All extensions bound in netbeans have their case changed to lower case. I suppose this could be considered a ``feature'' because windows is very case insensitive. However, this causes problems because one of the less popular file suffixes for C++ is ".C". When I tried to bind ".C" to the C++ mime type it was converted to ".c" and generated an exception saying c was being registered twice (even the error message used the lower case c rather than the upper case C which I was in the code). I've attached 2 possible patches to this email. The full-mime-fix fixes both problems mentioned above. It removes the binding for c, cc, c++, and h. It also removes the forced case conversion and binds the extension in the case it was typeed in. The 2nd patch, partial-mime-fix, could be used if the netbeans comunity objected to leaving the case as its typed. I strongly prefer the full fix but can certainly live with the partial as its not one of the more popular extensions for C++. The netbeans.org/contributing.html page says the way to get a patch adopted is to email it to this list. Please consider this a request for one of these patches (preferably full-mime-fix) to be applied to the code. The end result would be the ability for a C/C++ module to bind the proper extensions to C and C++ mime types. The result with the partial fix would also allow this but would restrict the file extensions able to be used by C++ developers. Thanks, <PERSON>
44168483-c9ff-9dde-6e20-8ce446439b10
['5a45ffae-2fa8-0d04-806f-2b15e75e1439']
Hi, There was a recent change in the module test mode. Previously, loading a module in test mode loaded from the manifest and a jar file wasn't needed. In the new scheme the test module is loaded from a jar. What I haven't been able to figure out is how to reload a new test module without restarting netbeans. Is this still doable? If not then the test module scheme isn't nearly as useful as it was previously. If it IS still possible to reinstall an updated module in test mode then somebody should update the tips document (http://www.netbeans.org/download/apis/org/openide/doc-files/tips.html). If it is no longer possible to reinstall a test module then I'm very curious why this feature was removed. I often work remotely and having to restart a new netbeans can take 1 - 2 minutes! <PERSON>
9bf5e026-f449-04b3-8c4d-f6ed72ac14af
['5a49ba18-3ac7-77f0-ff3f-49645b28cd74']
Hi all, I am having a problem I hope someone can help me with. I am pretty much a complete Tivoli newbie so please forgive me if this is a stupid question. I've searched the manuals, Google, etc. and cannot find a resolution to this. I have a small Tivoli test environment I am using to learn this stuff. I have one machine, that is the TMR server, TEC server, Endpoint Gateway, etc. Basically, all of the "Server stuff" is on one box. Additionally I have one other machine running an endpoint. The server machine is running win2k server, the endpoint is running Windows XP Professional. The TEC version is 3.9, and ITM is 5.1.2, TMF is 4.1.1 if that matters. The problem I am having is this: events generated by the heartbeat monitor, are apparently not being sent to TEC. That is, if I take down the endpoint machine that I'm monitoring, the heartbeat monitor detects that situation (wdmmngcache -l shows the endpoint as "unreachable" ) and a notice is sent to the monitoring notice group. However, nothing appears in the TEC console. Additionally, if I run wtdumprl, I see nothing related to a heartbeat event. However, I know that TEC is (more or less) working OK, as events from the endpoint (for example, a ProcessHoggingCPU event when the virus scan was running) show up in my TEC console. I have run wdconfig -D heartbeat.send_events_to_tec=true and wdconfig -D heartbeat.tec_server=EventServer to configure the heartbeat to send events to TEC. I have also run the task to import the baroc files, and the active rulebase in my EventServer does have the hearbeat events rules loaded. Also, FWIW, I can run wpostemsg and manually send a message, and I am able to see that when I run wtdumprl. Any advice regarding how to resolve this would be greatly appreciated. Thanks,
b51f2472-00f1-b932-ca5c-33b5737c166b
['5a49ba18-3ac7-77f0-ff3f-49645b28cd74']
Hi all, I have a question that I'm hoping someone can help me with. I've recently created a logfile adapter (using TEC 3.9) to monitor a WebSphere Application Server logfile. In order to allow it to co-exist with other LFA's I used the "Identifier Name" field to give my adapter a unique name. Now, I find out that the group who needs to use this adapter is on TEC 3.8 which apparently does not support the "Identifier Name" to distinguish amongst TME adapters. So my question is: given that they are using TEC 3.8, is there any way to run multiple TME adapters? For that matter, is it possible to run multiple non-TME adapters? And if so, is manually installing and configuring the adapter(s) the only way to achieve this? Thanks,
0a457e9c-ac6a-f28f-ab80-ed07168ef12c
['5a90d7aa-2674-ed3d-6769-c7231cd0db94']
BTW: PediaPress has a stand on the Frankfurter Buchmesse (Frankfurt Book Fair), booth E427 in hall 4.2. We'd be really happy to meet people from the community to talk about all kinds of MediaWiki related stuff. So, if some of you are there and can make it... we're looking forward to meet you! <PERSON>
d7636074-0738-e075-bbf2-8f68fd191e5f
['5a90d7aa-2674-ed3d-6769-c7231cd0db94']
We're heavily using the MediaWiki API in our opensource project mwlib (http://code.pediapress.com/ ), so first of all: Thanks to you all for implementing this functionality to MediaWiki! Maybe you're following the discussion initiated by <PERSON> on Foundation-l about appropriate attribution. As there is yet a consesus to be found, we plan to include all authors (minus minor edits, minus bots) after each article in documents (PDFs, ODFs) rendered from article collections. Currently we're using an API query with prop=revisions, requesting rvprop=user|ids|flags. Afterwards we're filtering out minor edits, anonymous/IP edits and bot edits (via regular expression on username and comment) and combine edits by the same author. To retrieve the data for all revisions for heavily edited articles (e.g. [[en:Physics]]), this requires lots of API requests with rvlimit=500. Is there a way (or a plan to implement one) to retrieve the list of unique contributors for a given article (from a given revision down to the first one)? Ideally this would accept parameters for the mentioned filtering. I guess inside of MediaWiki code this can be handled very efficiently (using appropriate database queries) and would eliminate the need to transfer lots of redundant data over the socket. <PERSON>
68249009-6b6f-d030-8340-50ef220b1f6c
['5aaf79e9-aec5-25b1-6d9c-c6365a712fde']
Hi all, I am implementing a power mac protocol and for this I want to send a packet with variable energy i.e. the transmittor node can change its transmission energy during the transmission after some fixed time and that should reflect in the packet being sent.For example if a packet is sent at 0.28W , i would like the same packet to be sent at some lower power level after sometime.Hope I have made myself clear. Any help or even a small hint would help me in giving a direction to proceed.I searched the net too but could not find anything relevant. Plz help... Warm regards, <PERSON>
688ecfda-adce-a1b8-cc57-6af0f8afb4a2
['5aaf79e9-aec5-25b1-6d9c-c6365a712fde']
Hi all, Its a simple question and I hope somebody will answer it.I have just started using ns2 and simulating some wireless scenarios and comparing the results for IEEE 802.11. the topology consists of 31 nodes in a straight line with each node transmitting cbr packets to the next node. Data delivered per joule should be constant with respect to distance between the nodes but I am not getting the exact result. Rest of the parameters are the same(defaults) in ns2. I dont understand why am I not getting the exact results. Somebody plz help... Thanks and Regards, <PERSON>
2aeb2d5c-341c-c598-84a9-a9c45d4150c7
['5ab5a2bc-1bcb-98ba-7abd-a5182fa89ee7']
Hi <PERSON>, If you're going to be using PostgreSQL 7.2.2 (the latest recommended version), then the port number is a setting you can change in your postgresql.conf file. This file is created inside the PostgreSQL "data" directory when you first initialise the database with "initdb". This file is also very well commented and easy to figure out, so you'll be fine to just open it with a text editor and make the required change, then restart PostgreSQL. As a precaution, it's a good idea to make a backup of your postgresql.conf file too, just in case something goes wrong while you're editing (unlikely though). :-) Regards and best wishes, <PERSON>
4a8ca5c8-9a80-6c93-d8bc-de3de716d5a2
['5ab5a2bc-1bcb-98ba-7abd-a5182fa89ee7']
Hi everyone, <PERSON>, a journalist with IDG, is looking to do a brief story on PostgreSQL and has a couple of questions for us. So, being newly motivated to set up the collaborative area for the Advocacy site, it's now ready for general usage (http://advocacy.postgresql.org/documents/) <PERSON>'s questions (from below) have been added into a page in the new collaborative area, so people can make notes/comments/etc directly, and we should be able to get the results to her by about this time tomorrow if it's as effective as I'm hoping... http://advocacy.postgresql.org/documents/NadiaCameron :-) Regards and best wishes, <PERSON>
e4b2f6c7-c927-7700-47e9-6b359480fe92
['5ab6e7cc-39c1-9230-cd79-a73d991b3e09']
I'm a new Squeaker, so I don't know if people even say "source code." I'm making an editor with some new features I like. I'm making it in Squeak. That seems to be going well, but after I'm finished, as well as running standalone, I'd like it to run integrated with such environments as MIcrosoft's Visual Studio. Microsoft's stuff expects a C dll. I'd like to translate my Smalltalk program into C. SPiCE seems to be the ideal thing. It expresses a good part of the Smalltalk environment in C, which is a tough bullet to bite, but once it's bitten, almost all of one's Smalltalk program can be expressed in C without change. The implementation seems to be pretty complete, based on <PERSON> paper. Is this facility available somewhere for me to use? Thanks for any help. Regards, Rick
1d13b7ae-1c85-93e9-d335-1b49934cbb5b
['5ab6e7cc-39c1-9230-cd79-a73d991b3e09']
I asked a few days ago whether the SPiCE program for converting a Smalltalk program to C was available. Apparently, it's not. Thanks, <PERSON>, for the info. The other, less complete but still interesting, solution is Producer, by <PERSON> and <PERSON>. Is that one available somewhere to be adopted and perhaps extended? Regards, <PERSON>
893617db-c266-ac7d-38a6-86fc1f1e44c8
['5aba385e-0947-f1e3-7e61-4f5652bce8e5']
Hello, A sky mote was able to process 10 KB of sensor data using wget-shell example to other Telosb's. It's approximately taking 25 sec for the TX. Currently, wget-shell application in Contiki-OS is using HTTP. Is there any other approach to reduce the end-to-end latency for a large size of file transmission. Will it help if I modify the ContikiMAC (explicitly radio-duty cycle)? Any suggestions. Best regards, <PERSON>
a73e5ef9-f7e0-e0a6-970c-24baad924926
['5aba385e-0947-f1e3-7e61-4f5652bce8e5']
Dear all, I am trying to solve a problem regarding file download from coffee file system. After successful transmission, I was able to write the file into coffee file system for Sky mote. After login to the mote, it shows the size of the file from sky-shell terminal. How can I download the file to my local machine from CFS? There is a (tools/sky/serialdump.c) file can be used to trace the serial port regarding this; but it can't determine the stating and ending of the file. Just print out the hex and binary output to another terminal. Is there a way that I can rebuild the original file using this or directly downloads from CFS. Any ideas. Thanks. <PERSON>
5061aad0-c999-5c56-4ad4-ea884d5a5234
['5abdc934-c8d6-8403-ed9f-f215fe7f5b5b']
Hi there, I'd like to use <asminfo> and <mkiisdir> in my build. So I have a recentish build of NAnt head. And I copied the NAnt.Contrib.Tasks.dll from NAnt 0.8.3 final into my compiled NAnt bin folder. But If I try to use <mkiisdir> I get 'Unknown task <mkiisdir>' Am I being dumb? Thanks, <PERSON>
914baa44-8c53-9f65-fb34-f003a1af8bb3
['5abdc934-c8d6-8403-ed9f-f215fe7f5b5b']
Hi there, I've hit a problem (bug?) using <solution> . If it can't find a reference, then NAnt throws an internal error, rather than this getting reported more nicely. This causes me a problem since Internal Errors don't get sent to the logger, and so I don't see the error appear in a CruiseControl.NET log (in CCNet we use the Logger to capture NAnt output.) I think this would be fixed if <solution> used the <csc> task internally rather calling the csc.exe compiler directly, but that may be too much effort. Cheers, <PERSON>
f38b8328-0b66-9688-8456-889c9a019a3d
['5ac9d506-942a-325a-d787-13ad2c8f6a3a']
I have noticed that the cvs interface in this product, when running the CVSNT version 2.5.02.2088 cvs.exe executable, fails on a sandbox update. However, the same application running with the cvs.exe from the TortoiseCVS folks ( 2.0.62.1817) operates properly. So, if you are trying to resolve this issue, which manifests itself as a 'Checksum failure ' message, use the Tortoise version with EA. A future version of EA will address this issue, apparently. <PERSON>
d641c7b0-2d18-0b53-445e-0edfe33c66bb
['5ac9d506-942a-325a-d787-13ad2c8f6a3a']
There is no limit on the # of repositories that I know of. The only trick is to make sure you enter the right repository name in CVSROOT and that you grant appropriate privileges in the repository project folders. Also, if you are doing special wizardry in your CVSROOT directory like email notification and so on, I think you have to replicate all this to the other repository as well. They are, as far as I know, logically distinct from each other. That means dup'ing all the commit scripts, verify messages, etc. etc. Anybody solve it another way, like with links? Allen Marshall Cambridge Systematics, Inc. 100 CambridgePark Drive, Suite 400 Cambridge, MA 02140 tel 617 354 0167 fax 617 354 1542 e-mail <EMAIL_ADDRESS><PHONE_NUMBER> fax <PHONE_NUMBER> e-mail MVbnGMODyAiXPpbR@example.com www.camsys.com
a291c23a-4ae2-8d4d-cd82-090cc7244528
['5aca5db5-1f3f-bd79-6577-e4b600265e08']
Another dumb question here..... All of the information I've found so far related to parallel rendering using VTK involves the use of multiple individual computers and dates from five or six years ago. Is there such a thing as using one of the new multi-core chip PCs for parallel rendering applications? I notice that MicroCenter is selling a P7 desktop for around $800 now and you'd think the idea would occur to somebody. <PERSON>
23a92084-3b9e-97ff-ba4f-46e7df4c75bb
['5aca5db5-1f3f-bd79-6577-e4b600265e08']
This one finally dawned on me..... the wrapper libs basically don't contain real code, just wrapping code; the real vtkio.dll has to get put in the path for anything new in it to get found, duh..... The problem vanishes when I do that. <PERSON>
d4043cb2-8e3b-ce3d-fb61-bb58c4ef678c
['5accfee4-49e6-f446-79ff-24477c8738a5']
"Is this still around ? I filed a bug several months ago. Which version are you using ?" -- <PERSON> I don't have my laptop handy right now but I'm running the grub that came with Gentoo 1.4. I also preformed an "emerge system" so it should have updated it with something new off the portage tree (this was last week). If this is a bug in grub (sorry about the rhyme), what version was it fixed in? I'll re-emerge grub and try with: grub-0.93.20040125.ebuild http://www.gentoo.org/cgi-bin/viewcvs.cgi/sys-boot/grub/ Sincerely, <PERSON><PHONE_NUMBER>.ebuild http://www.gentoo.org/cgi-bin/viewcvs.cgi/sys-boot/grub/ Sincerely, Rob Elm
db099d09-6ecc-796c-9ad7-d2c1f2621b61
['5accfee4-49e6-f446-79ff-24477c8738a5']
Most modern email clients support multipart MIME which encodes both the text and HTML messages within the body of one email. Besides that, the Bug-Grub listserv does not have a charter which would prohibit the sending of HTML messages (at least as far as I know). If you're truly offended by html formatted email, maybe you should consider using a mail client which is intelligent enough to ignore the HTML encoding. Sincerely, <PERSON>
258d1510-6ae7-3c4e-10b7-e9ab5316ca87
['5af3d5bd-6c32-8417-1b24-1f9b7a120201']
<PERSON>, mam dobrou zkusenost s amd64 na serveru, a ted <PERSON> na IBM x226 s Xeonem s em64t. em64t ma byt kopie amd64, ale <PERSON>, <PERSON> to "obšlehli". Chci na tom valit 64bit debian sarge. Ma nekdo nejakou zkusenost? Google nasel vety typu "xeon slows down a little in general when you enable 64bit mode. AMDs speed up. Given the AMD is usually faster at mosts tasks already in 32 bit mode, the situation is not improved (for intel) in 64bit mode.", coz <PERSON>. Petr
c0afb033-a57c-899d-19be-19c09f09d455
['5af3d5bd-6c32-8417-1b24-1f9b7a120201']
<PERSON>, this is on onboard Geforce 6150 using proprietary nVidia driver. BTW, a workaround for this issue is to try out Alt+Ctrl+F7/8/9/10/.. at step 5) when there is the normal login prompt on screen and user '<PERSON>' wants to get into his/her session. If the user 'B' has really been logged in already then his/her console is running somewhere at F7-Fx. What I am reporting here is that gdm (or who) does not switch to that console automatically. IMHO it should work as follows: whenever any user logs out the gdm should try to locate any other running console. It normally works when second or subsequently logged in user logs out but it doesn't seem to work when the user-that-logged-in-_first_ logs out. Thanks.
d1093ab1-785d-05ae-7b30-717b659159ae
['5b070074-4e93-b4a1-454e-2f8a6cac71f8']
Hi, My name is <PERSON>, I would like to ask you if it is possible to develop an OS based on Debian (similar to SkoleLinux/DebianEdu) specific for psychologist (both applied psychologist and researcher psychologist) built with a set of predetermine tools such as pspp (data analysis), neurodebian, zotero (citation manager), bibus (connection to data base provider) pgp tools for confidential reports about patients,... I am psychologist and professor at University. I don't know about programming, but I would help to search for the appropiate tools for an specific OS for psychologist (asking partners, testing tools in debian packages...). Psychologists as other researchers usually depend on privative software to research and to improve the knowledge about human behavior and mind. We would love a new Free Software OS based on Debian. Currently I am user of Debian Stretch and I think this would be a very useful OS to built an specific OS for psyhologist. I offer myself to contribute of topics related to psychology during the development process and the spread of the new OS if this is released. Please, let me know your opinion about my proposal. Thank you very much for yor attention. Kind Regrads, <PERSON>
5312e527-1942-fa7c-6519-45a28f6a3c1f
['5b070074-4e93-b4a1-454e-2f8a6cac71f8']
Dears, I am <PERSON>. I am a new member of this list. I would like to ask you an important question for me: Using TAILS OS (which forces all Internet connections through Tor Network), I need to get access to data bases related to academic and scientific journals, but connections through Tor Network are refused by those servers, so I must use other OS or other web browser, but I would prefer don't do it. Is it possible to manage Tor browser settings in order to get into those data bases using Tails? Sorry for the annoyances and thank you very much for your attention. Kind regards, <PERSON>
1845bd89-cb81-878f-957f-b2c3b6c06436
['5b263b0b-2537-a33c-ca45-43fb6ebf03e6']
Hi all, I've committed myself to learning Python and have started reading "Learning Python" (<PERSON>) and looking through various online resources. My career so far has involved a little high school teaching and about 10 years of translating Japanese to English, but no programming or coding. I've also joined this list today and this is my first post. My aim is to be able to create Python programs, specifically "activities" that work on the OLPC's XO laptops and SoaS (Sugar on a Stick). My questions are: how long would you estimate it to take and how complicated would it be to create the following as Python programs? (I know it varies depending on the person, but for example, how long would it take _you_?) (i) a simple guitar tuning program involving an image of a guitar and the playing of each of the standard strings of a guitar (E, A, D, G, B, E) upon key input by the user (something similar to this: http://www.gieson.com/Library/projects/utilities/tuner/ (page is 782kb to open)) and (ii) a very basic turtle art program with an intentionally limited set of commands and on-screen display words (say, a total of 30 to 50 specific strings), wherein the entire set of strings is offered to the user (perhaps at first use of the program) in a format that enables easy and full localization of the program so long as each of the strings is translated appropriately and inputted to the program. I know of turtle.py and xturtle.py, but I'm thinking of starting something from scratch. It's the easy localization I'm interested in. Hope these questions are appropriate. I'm grateful to be able to ask them here. <PERSON> (in Australia)
e24df326-1b91-c717-7b41-f2a3fe1bf810
['5b263b0b-2537-a33c-ca45-43fb6ebf03e6']
Hi all, I was just wondering if anyone here can recommend any freeware program that has been written in python and for which the source code is available. Basically I just want to see a program that does something relatively simple and straightforward, but something that is "real world," I mean something that people actually use, hopefully something that comes with it's own installer and GUI. I'd like to try out using the program as an ordinary user and then I want to look at the source code to see how it's been achieved. Any ideas or suggestions? <PERSON>
959330c3-48e7-3c6b-7543-dc5a11713e5e
['5b313f77-759d-e278-92a0-6e11d937b18b']
Hi, <PERSON>, thanks for your reply. I'll test the patch and use memory_cache_shared set to OFF. Sorry, I was wrong. Objects bigger than maximum_object_size_in_memory are not cached on disk. Although objects smaller than maximum_object_size_in_memory but bigger than 32KB were written to disk, I guess they got a HIT because Squid keeps a copy in memory of hot and in-transit objects. That explains why the UFS store was "ignored" when Squid was restarted. Thanks.
92ed2058-1331-f56c-338e-4686ef30797d
['5b313f77-759d-e278-92a0-6e11d937b18b']
Hi! Thanks <PERSON> and <PERSON> for your replies, together they pointed me into the right direction! Based on the the URLs sent by <PERSON>, I edited the "/etc/php5/cli/php.ini" file and tested different values for "max_execution_time" and "max_input_time" but none changed the PHP's script behavior.  Then, I remembered <PERSON> mentioned a 60sec timeout. I saw my cache.log and yes there was an exactly 60sec delay after starting squid and the first Warning. So, I searched the "php.ini" for a similar value and found this directive: "default_socket_timeout". I changed it to 300sec and the Warnings started to show up accordingly. Then I changed it's value to "-1" and the warnings haven't shown up again! Squid doesn't complain anymore about my PHP-scripts, but I don't know if this change has secondary effects or any other consequences.  I'll be monitoring them, but in any case I have the backup Perl-scripts. Thanks again!
53ea4e5c-82dd-a9d1-7bab-3fe0ecf15e41
['5b36bb38-744d-dc81-95ab-063948edbf42']
<PERSON>, How can one configure LyX to use another pdf viewer when clicking the "view pdf (pdf-latex) button? I could not find out. It depends what is set automatically during installation. Actually I would like to disburden Adobe Acrobat Reader from this task because too slow and change to Skim (or PDFView). joachim
5c0bbbc8-d215-c1e4-6629-fac4d7e92a2b
['5b36bb38-744d-dc81-95ab-063948edbf42']
Moin Lyxers, not for the first time I got today an LyX when clicking "show pdf" an error message: Die Datei existiert nicht: (I suppose in english: The file doesn't exist:) /tmp/lyx_tmpdir.L12205/lyx_tmpbuf1/Lyra-Pracher-odt-teil.pdf That was after adding an TeX code for another "footnoteB" (from the - loaded - manyfoot package), because I'm using for that the "handout (tufte)" layout because of the many sidenotes I need. But because - at a certain point, when LaTeX (via LyX) said "too many floats" - I changed to the TeX-box \footnoteB command to get a real footnote. Which normally gives the planned result. If the error message "The file doesn't exist ... .pdf" is a known problem, I ask: ==> how to come out of this situation best? I found out by trial and error, that I deleted the last TeX inset \footnoteB{...}, (but after many other trials, even after clothing and reopening LyX, after deleting the LyX - /tmp ... xyz-files and so on). Regained functionality of LyX in making and showing a pdf result, I reintroduced the TeX-inset \footnoteB and it worked. Fine for the project. But Question: => are there reasons for that behaviour of LyX and/or other ways than just trying out, deleting the last inset(s) and remaking them? ==> What might be the reason of the LaTeX-error about "too many flouts" when inserting "footnotes" = sidenotes in the "handout (tufte)" class, even if on the page where I am working and on the horizontal space on the margin there is place enough for another "footnote"=sidenote? Goutgaun! <PERSON>
dea71aa6-7b2b-0b1f-5e7c-d4b634a916ff
['5b4c2917-71e6-c87b-821d-320d55a85abe']
Hi all, Following a recent conversation with <PERSON> (and contrary to what I initially claimed here) there's value in adding a third storage type to the Quota API: Session storage. Contrary to temporary storage which might not get wiped across UA sessions, Session storage MUST get wiped when the session is closed. Happy to provide patch if there's agreement this is a valuable addition to the spec. Best, <PERSON>
f09153b0-c7d5-9371-179c-14f0ea41d75c
['5b4c2917-71e6-c87b-821d-320d55a85abe']
Hi, I recently stumbled upon a number of use case and requirements docs (such as MediaStream Capture Scenarios[1] or HTML Speech XG[2]) that were published as officially looking W3C documents (for whatever that means, at least, it's not a page on a Wiki). I think that's tremendously useful, especially for authors who can have a much better understanding of the purpose of a specification that way (and therefore use it the right way and for the right purpose). It's also a smart way to get authors involved without corrupting them into thinking like spec writers or implementors. What are the WebApps WG's plans with regards to that (if any)? Thanks,
56d645bc-4778-33dc-3f7f-dd0c0392bccc
['5b4f5fac-1bdb-770f-6cf7-a17b6dff94c7']
Hi, devs By watching recent changes in trunk, I feel the release of Trac 0.8.1 will be done in a few days future. But changes in [1250] breaks the existing wiki pages which has Japanese (also korean, chinese, etc.) only heading string by excetpion. I've reported in #1224. It brings us broken page and we cannot view nor modify the page any more except modifying code of WikiFormatter.py by himself. I strongly wish this issue is fixed before release. --- Regards,
ba7c01f8-032a-6674-bc0a-d25d8ba03981
['5b4f5fac-1bdb-770f-6cf7-a17b6dff94c7']
Hi, I got source code archive: TortoiseCVS-1.6.10-src.zip from SourceForge. http://sourceforge.net/project/showfiles.php?group_id=48103&package_id=41678&release_id=233635 But I cannot expand it with so many claims like this: skipping: TortoiseCVS-1.6.10/xxx/xxxxxx.xx unsupported compression method 12 The bytes of downloaded zip file and displayed bytes on web pages are same. I tried unzip 5.51 and older unzip, but both tools failed. What's happen? Was anybody able to expand it successfully? What tool is used to make zip file? P.S.
dcc96484-1a57-d3d8-baa6-910807510843
['5b56581a-9e13-a4fc-a122-336a088c23e7']
I'm looking to setup a file server using RAID 5 (or the equivilant RAID-Z) and I'm interested in using ZFS. It looks like my primary options are Solaris or FreeBSD and since I'm beginning to really dislike Solaris I'm leaning towards using FreeBSD however I've heard that there are some issues with ZFS on FreeBSD specifically regarding some "kmap_mem" or something like that. Can anybody tell me if ZFS is safe to use for home users? I'm just looking to setup RAID-Z with 3x 300 GB and another pool for 1x80GB (I guess). Thanks for any help!
9f0d093c-1008-4b3c-df93-25601c3c5d9c
['5b56581a-9e13-a4fc-a122-336a088c23e7']
In case nobody saw my last e-mail (from the "Is lua thread safe?"), I have some more questions... (on top of the questions asked in my last e-mail) 1) If I create a lua thread, is the global data shared with the main lua state? The manual seems to contrict itself and says that the data IS shared but then it says it's not. From doing some tests, it seems like the data IS in fact shared. For example, if I load a script (using lua_dofile) and run it on my thread, the script's global functions show up in my main lua states global table. My main goal is this... In my game I will have tons of scripts being run but I need Lua to know the current user (or NPC or object) running the script. I can create a global "CurrentUser" userdata but since it's global it is shared and gets overwritten. I tried creating a thread but that doesn't seem to help me at all... 2) Once a script is run via lua_dofile(), all functions that were declared stay in the global table, right? So I can access them later? 3) How can I load a script and NOT execute it but only run a function contained inside of that script? Also, if anybody could answer any of the questions from my last e-mail (about lua beiing thread safe) I would greatly appriciate it. Thanks!
fc0d5a7a-8172-2ecc-b741-3eaf1dfb3c3c
['5b60aa76-04d8-67db-380c-2eef6c5bf707']
G'day iFolders The download page (http://www.ifolder.com/index.php/Download) for the 3.4 iFolder clients does not list clients for Windows or Mac. And locating the 3.5 client, whilst not rocket science would definitely be easier by having a direct link. For example: http://forgeftp.novell.com/ifolder/client/3.5/current http://forgeftp.novell.com/ifolder/client/3.4/current However, for 3.4 if you backtrack thru the various commits you can eventually find the last Windows & Mac release here (there might be newer, but I cannot find them). http://forgeftp.novell.com/ifolder/client/3.4/20060616-0101 Which raises the question, why have there been no current builds of the iFolder client for Windows and Mac? And why is there no 3.5 clients at all for Windows and Mac? Wanting to be helpful, I tried to edit the iFolder Wiki download page to include the updated download links, but this is a "protected wiki page", so can someone with the correct permissions please attend to this. Thanks for listening!
27fa6814-63a1-0aff-f6fc-34b0507650f1
['5b60aa76-04d8-67db-380c-2eef6c5bf707']
G'day We are actively looking to employ two people in response to the growing demand for FOSS services. Ideally, I'd like to support new local talent, preferrably Uni. or College graduates who are smart/keen and just need a break to get into the IT industry. The work will be a mixture of LAMP development, Linux administration (desktop & server) and FOSS system installs/ maintenance/ customisation. So I expect that you will have some existing skills in LAMP or Linux (if not, why are you on this mailing list?). Having some good graphic design skills would be a real bonus. If you are currently employed but want to change jobs (perhaps working for a Microsoft-monoculture-only-shop?) then feel free to apply as well. So if you are interested or know someone that might be please email the resume to <EMAIL_ADDRESS> I will respond to all applications. Cheers <PERSON>
e9fcad53-48ac-8062-0544-f82ff44a19f6
['5b6423f7-7347-6910-6039-49e864817bc9']
Hi, I just had to do this myself, and I figured some of you may be in the same position, so I'll share it (for Linux / UNIX): We had a remote server vacuum our data, probably seeding a cache and using our box as a free rendering service. Which is fine, but this was very aggressive. So while GeoServer now has the ability to log requests, you may not always want to restart or reconfigure the service, especially if it's transient. A handy tool is ngrep ( http://ngrep.sourceforge.net ), and the command to use is ngrep -qd eth0 'GET' tcp port 8080 If this doesn't reveal the culprit you can try filtering for POST instead, and make sure that 'eth0' is really what you want. If GeoServer is running on a local machine you have plenty of tools to choose from (like Wireshark), but on remote machines where the capture may require more bandwidth than you have) it's trickier. -Arne
6691a87d-8f97-0d05-a8a7-88366f7d0355
['5b6423f7-7347-6910-6039-49e864817bc9']
Hi, I've installed the latest versions of everything, including DIMP 1.1.1, on top of a Courier IMAP server. The problem is that I cannot find a way to sort by message date, only arrival date (which really appears to be message number). This is a problem when I move stuff from my inbox into subfolders, the messages are listed in the order they got moved into the subfolder, not according to when they were received or the timestamp set by the sender. Ideally I'd like to expose both the Message date and the message number, and be able to sort by clicking on the appropriate column. Is it possible to achieve this without too much hacking ? (I realize I can still set this through the defaults in message options, but that's not really something I can tell <PERSON> to do) -Arne
39165f6e-1e94-3eb2-3411-0a013ee0b4af
['5b73c8a5-bd8b-d40a-7ab9-045f59c6fbb1']
I believe that I have made some progress on this, but I'm not confident that my solution is the correct one. There were several factors playing into this that pulled my attention away from the real issue. I originally reported that this problem was CGI specific; it turns out that this is not the case. The way I was reproducing what appeared to be a major leak was by sending a number of parallel requests to CGIs that generate a lot of output. The result of this was that Apache needed a lot of temporary buffers at the same time. After digging around for a while, I discovered that Apache on Windows never calls apr_allocator_max_free_set(), and the default behaviour is to hold on to all memory ever allocated, and allocate memory future requests out of that block. When I spawned the parallel requests, Apache grabbed a whole bunch of memory, and then kept reusing it. After setting a limit here, I was able to move on to the bigger problem. After making the change above, I realized that every request was leaking a little bit, not just the CGI requests. To confirm this, I made a simple configuration file that redirects all requests to a static HTML file. Sure enough, it was leaking. The problem, as best as I can tell, is that mpm_winnt calls apr_bucket_alloc_create() once for each thread, and registers it in the pchild pool. This bucket allocator is then passed through to core_create_conn(), and used for all apr_bucket_XXX() routines during the request. The pchild pool is not cleared until the server shuts down, so the memory used here grows and grows. To solve this problem I changed mpm_winnt so that it creates the bucket allocator using the ptrans pool, which gets cleared after every connection is finished. After making this change, the system behaved much better. Just to check, I then undid my first change related to the maximum memory to hold on to, and the system continued to function corrrectly. So in the end, I only needed the one fix. The following patch shows the changes I made. My question now for the experts is whether this will break anything. Thanks, <PERSON>.
bf7abc6f-bcd7-5c2b-564b-73708e0b03da
['5b73c8a5-bd8b-d40a-7ab9-045f59c6fbb1']
Hi there, I am using Apache 2.0.43 on Windows, and am having some memory issues. As far as I can tell, Apache is slowly growing larger and larger during its lifetime. In stress testing, we've seen it grow as large as 2 gigabytes. I have tried running Apache within Purify, and unfortunately it looks like the memory is properly cleaned up when the program exits, which makes it tricky to try to figure out where things are going wrong. Searching through the Bugzilla database, bug 11427 (http://nagoya.apache.org/bugzilla/show_bug.cgi?id=11427) sounds like it may be similar to what I'm experiencing. I should mention that I do have my own module loaded, as well as some of the standard Apache modules. There is a chance that the leak is in my code, but I don't think this is the case, as my handlers should clean up after themselves each time they are called. Any traditional memory leaks should show up within Purify. The memory growth seems to happen most often if a number requests hit the server simultaneously. Has anybody else seen this happening? Does anybody have any suggestions of techniques I can use to diagnose this? Thanks, <PERSON>.
d00c9397-7d79-f1c4-3ccf-21e66af7ba5d
['5b961b8a-b402-8806-567f-e1026d3e3c95']
Hi, I m encountering problems to compile the IPAQ sleeve drivers. Drivers sources exist in the linux kernel but there is no possibility to compile it. During make menuconfig, it is not possible to activate the option IPAQ_SLEEVE and IPAQ_PCMCIA_SLEEVE, they do not appear in the menus. If I make a search during menuconfig, I can see that these options exist but: IPAQ_SLEEVE is set to IPAQ_SLEEVE (??) IPAQ_PCMCIA_SLEEVE is set to n If I modify directly the .config then the options are not taken into account (.config seems to be overwritten when we do make). If I modify the h3900_defconfig then the option are not taken into account either when doing a make h3900_defconfig Someone here could help me compile these drivers ? Thx in advance, <PERSON>
13e9841e-15af-6046-8cdd-8e9d52394262
['5b961b8a-b402-8806-567f-e1026d3e3c95']
Hi, I m trying to boot angstrom on a ipaq h3900. Since there are no images available for this device, i build the angstrom environment using the following guide http://www.angstrom-distribution.org/building-angstrom Once finished, i got jffs2 images. I tried both minimalist and console images but i got the following error during the boot : "Kernel panic -not syncing: Unable to mount root fs on unknown-block(2,0)" Anyone knows where it could come from ? Regards,
f8393953-5166-497e-c448-5bfdd1fb8fe6
['5b97d700-3902-c5a8-b42f-cc5d4a9052aa']
This is one of the places where IE happens to be a better dev env than FF. FF has a known bug with its caching. If the main page hasn't changed, then it assumes that dependencies haven't either (JS, CSS, XAP, etc). Thus, it never even checks whether they're stale and always uses the cached version. So, if you're doing any web app development in FF, you have to remember to either Ctrl+F5 after each change, or, when that isn't working, close the browser without saving tabs re-open it, and nav back to your page (saving tabs saves the cache info, so you come back to a stale page). I end up setting a lot of shortcut buttons, so that I can single-click to whatever page I'm working on. Also many FF instances, with one page per browser, because using tabs means losing more tabs when I have to shut down the whole browser. Tabbed browsing? Not in my code windows. I do mostly straight JS + CSS dev, not Silverlight dev, but run into this bug there all the time. It's a browser thing, and affects all web app development on FF, regardless of the rest of the technology stack. Sucks, but eventually you just get used to "why is that test failing only on FF? Oh yeah, I have to go hard-refresh all the pages again." <PERSON>
d18cba51-d45b-835d-4748-cf7f340f7b7a
['5b97d700-3902-c5a8-b42f-cc5d4a9052aa']
My apologies, but I announced the wrong date for <PERSON> <PERSON> talk in January. This talk is on Jan 14th, the second Wed in Jan. It is not at our usual third Wed (Jan 21), as I previously announced. I have updated wiki (http://www.xpdx.org) & calagator, so they shouldn't continue to sow confusion. I apologize for any confusion. I hope you all can still make it; this should be an excellent talk. Arlo
0fef63dd-92e9-0e75-0f46-d17693b6681b
['5b9ae334-22b5-5539-fe2d-7b8eadf7361b']
Hello everybody, I had to migrate from debian to centOS and I would like to build rpm packages of opensips on a CentOS dev server. I have retrieved the opensips.spec-4.1 file from the packaging folder in opensips source code and I'm trying to adapt it to opensips-4.4. I would be very grateful if someone that have already done that could help me... Best Regards,
e304dfcd-a214-c07f-b9f2-9603569cd6b0
['5b9ae334-22b5-5539-fe2d-7b8eadf7361b']
Hello everybody, I would be interested, and I'm sure not to be the only one, to have a kind of daily opensips log analyser report each day of: - blacklisted IPs by the pike module - total number of calls/transaction (opensipsctl fifo get_statistics tm: ,......) - flapping or failed gateways of the dispatcher module - ... I used to have a logwatch running on my servers, and before trying to add an opensips file into /usr/share/logwatch/scripts/services I would like to know if you have a better way to do it? Thanks
94699cc8-b948-684b-b7fd-d450632d9785
['5b9b8c4f-30c3-d874-4649-6748c81d6097']
Hi out there! I got a question concerning xterm and rxvt with cygwin. When I run xterm under LinuX, all function keys are passed to the application running (e.g. Midnight Commander). When I try doing the same with cygwin, only F1 to F4 are passed through, F5 and higher result in some key messages on the command line. With rxvt none of the function keys work. When I log into a linux box and start xterm on the remote session (still display is the cygwin box) erverything works (all function keys and even mouse support). What did I miss with the cygwin installation of xterm? Did I make a serious configuration mistake? Regards, <PERSON>
f63c12cb-bf73-99cd-97f9-10126197eb7d
['5b9b8c4f-30c3-d874-4649-6748c81d6097']
Hello! I am sorry to (maybe) destroy your enthusiasm but from my point of view there are some more reasons for not running either gnome or kde on cygwin: The main thing is: they are very large packages and even on my linux box they are that slow, that I prefer just using icewm as window manager and only have kde and gnome installed in case some application needs the libraries. If I start to compare cygwin running under win98 I really can not advise you to run kde or gnome - maybe on NT kernel based machines that is different. Cygwin is (and here the developers have made a real good job - Thanks!) a very good tool if you want to run applications like nedit (which is my favourite editor) and other basic unix routines I really miss under windows - but as any Operating system running on top of another one - it has it`s speed limitations. Another great usage for cygwin is to use it for running xdm sessions or if you have to do admin work on a unix box over the network (it has ssh, sftp and all the other nice tools which windooze does not have). So for me (personal opinion) it is a great (and long expected) extension to that "operating systems" from Redmond which I am forced to use some times... By the way: I would not even use kde or gnome with linux - but that is MY personal opinion about it. Please do not kill me for it! So again: Thanks to all the developers! Regards, Uwe
b92ab33f-2dee-b475-a9ab-0a6483b16e01
['5ba658e8-184d-99c3-f4ac-c4bd3ceeec17']
As it came out, as usual, I unpacked the new revision of the debian linux kernel sources, copied over the .config file from the previous one, made a make oldconfig and built a kernel image using make-kpkg. Everything went smoothly, as far as I can tell. I then installed the new image it on my laptop, and rebooted into it, to find out that _any_ 32 bit executable would segfault! Even the ld-linux linker itself! I reinstalled the old package, rebooted, and I had 32 bit compatibility back. Does anybody know what the heck I did wrong? Or what changed between the -6 and the -7 version of linux-source-2.6.32 that might have caused this? Thanks in advance <PERSON>
6833737e-1831-7172-f6f9-d05b00fa81da
['5ba658e8-184d-99c3-f4ac-c4bd3ceeec17']
Hello. This may be trivial or a known problem, if so please feel free to just point me to where the answer is. For some time now, in debian sid it has become a problem to use the nonfree fglrx driver with gnome 3, and with the latest xorg. So I finally bit the bullet and switched to the free driver on my laptop (sporting a mobility radeon hd 3400 series). The good: xorg appears to work fine, even if quite a bit slower, but since I am not after high-end 3D graphics, it's quite ok for me. The bad: as soon as kernel mode setting is enabled at boot time, the console disappears completely and is not visible anymore. If I disable kernel mode setting, then the console stays visible, but gnome 3 will not work. Is there a known configuration, setting, workaround with which I can have a working text console *and* working xorg with kernel mode setting enabled (to run gnome 3 efficiently)? Since shit occasionally happens, especially following sid, I am used to sometimes drop to a plain text mode console to fix things, and having a laptop which is completely unusable until the graphic login manager fires up is, to say the least, disturbing... Thanks in advance, bye Giacomo
adfe2d3e-713f-a54e-61d7-a495c7622d5a
['5bc86e17-50ac-a947-a831-0d52434cfedd']
<PERSON>, I agree. IMHO default tuning for specific benchmarks doesn't often benefit the average user. Perhaps the default should be the size of the receive buffer modulo blocksize and an alternate default (pre-set) available to revert to the current behaviour, (even then I'd prefer to see the existing limit tunable with a defined safe range, while that preset default was selected in /etc/system). <PERSON>
0623e155-56cd-a822-744d-2326b5cd2c09
['5bc86e17-50ac-a947-a831-0d52434cfedd']
Hi <PERSON>.. you don't in the end say whether you are binding to port 0 or not, but the fact that you are seeing all your pkts in 'sctpOutOfBlue' indicates you are probably not binding at all. Here's how it plays. The incoming pkt is checked to see if it belongs to a specific connect via the Solaris sctp stack. If it's not found in the relevant hash, it is assumed to either be for a raw IP socket or an "out of the blue" packet. It is checked against a hash of raw binds, if there is a specific bind it goes there, if there is a wildcard bind (port 0) all unclaimed packets go there. If there is no bind in the hash then the packet is deemed to be "out of the blue". Either do a specific bind to a port and get only the packets for that port or bind to port 0 and get all the rest. The other web page you refer to may have been written prior to wildcard bind being integrated in the stack. HTH -George
89c139ea-cc46-8061-b479-23c32285ae3f
['5bd79d7b-30d3-ab19-6d79-0d7054755d24']
Hi folks! I'm sure there's an easy way to do this, but what I need to do is calculate weekly returns from daily OHLC data, but I want my weekly return equal log(Friday Close) - log(Monday Open). I'm using timeSeries from RMetrics and I don't see a quick easy way to do this. Any ideas? Best, <PERSON>
cf89d7ab-b321-09b6-5b99-dd7e03fe7256
['5bd79d7b-30d3-ab19-6d79-0d7054755d24']
Hi all! Does anyone know if a vector autoregression package is avaialable that allows binary variables as part of the endogenous system? I'm looking for something along the lines of what is implemented in "Dynamic Forecasts of Qualitative Variables: A Qual VAR Model of US Recessions" by <PERSON>, 2003, Fed Reserve Bank of St. Louis. Another possibility is the autoregressive conditional hazard model by <PERSON>. Thx! Best, <PERSON>
3934f51b-49c9-95e5-7263-cad6fdd1e229
['5c1750e0-1ab7-88f5-1eae-b61ca6bd0aff']
Hi, I don't seem to be able to find any way to trap/handle OS signals in erlang (TERM, INT etc). Am I missing something or is this just "not the erlang way", and if the latter then is there some alternative? I'd like to be able to trap a TERM signal and shutdown cleanly The best I could come up with was to start my erlang process from a shell script which itself trapped signals and then sent the erlang node a rpc message to stop cleanly. Is there a better way? Thanks, <PERSON>
1d6473dd-a377-3a97-3b1a-b58a0043977f
['5c1750e0-1ab7-88f5-1eae-b61ca6bd0aff']
Hi, I was trying to run a php script I obtained as a CGI using a php4 binary with miniserv.pl but kept running into the "No input file specified" error message. After a bit of Googling I found that a common problem withr running php CGIs in this way is that the SCRIPT_FILENAME enviroment variable may not have been defined - however I see that this is defined in miniserv.pl after some discussion of this I found in the archives [1]. Anyway - that wasn't my problem, just a bit of backgroud. I traced the C code of the php4 binary when running my php CGI until I found that it was failing when doing a fstat() on a filename it had opened which turned out to be a directory and not a file as expected. It turns out it is basing this on the PATH_TRANSLATED environment variable which miniserv.pl had defined as the same as my document root. I had a look at the CGI spec [2] which stated the following for PATH_TRANSLATED : PATH_TRANSLATED is derived by taking any path-info component of the request URI (see section 6.1.6), decoding it (see section 3.1), parsing it as a URI in its own right, and performing any virtual-to-physical translation appropriate to map it onto the server's document repository structure. If the request URI includes no path-info component, the PATH_TRANSLATED metavariable SHOULD NOT be defined. I found the last bit interesting so I tried editing miniserv.pl to not define PATH_TRANSLATED and suddenly everything started working. So - is it possible that PATH_TRANSLATED is not being defined correctly here, or iare there cases where it should not be defined as seems to be being suggested here? Thanks, <PERSON>
b45f0528-d8b7-48a4-1da6-48c4fb867ef8
['5c251618-3fd5-1e09-274d-106473a90f73']
Thank you very much for your information! Couple more questions. What do you mean by ideal/quiet environment? I do have other traffics currently in the same channel. But the maximum, so far, I have got is only 12.8 Mbits/sec for a signal strength taking from -34 to -43 dBm and 7.26 dBm for -67 to -77 dBm. What do you think will be the ideal parameters so that I can tune/setup my mesh network to get the maximum throughput? Why is that 802.11s doesn't support 11n rates? Is there any thing like we can configure to achieve this? Thank you, <PERSON>
cbbbe3f2-be42-fefd-a919-4b22c3b9bfd8
['5c251618-3fd5-1e09-274d-106473a90f73']
Hello Friends, I am studying the mesh networking (using ath9k driver, openwrt backfire 10.03 on a TPLINK1043ND device). Currently am running with six mesh nodes and doing basic iperf experiment to find the throughput between mesh nodes. It will be really helpful if you people can tell me what is the throughput expected out of a mesh network. For a single hop (direct link) throughput I get is 11.0 Mbits/sec on avg and for a two hop link am seeing 6.5 Mbits/sec. Is this throughput normal? Thank you, <PERSON>
3d123bbb-f7e0-44b6-d094-4b950f1d4512
['5c2782ba-5fae-0834-c326-1b4b17e20205']
Several weeks ago I set up a server with Koha 3 beta2 in preparation for migrating to Koha once koha 3.0 final was released. I have now set up a brand new server running Debian and Koha 3.00.00 and I am attempting to set up the same configuration that I had on the original server. As part of this I need to import two large marc record files. When I attempt to manage the staged records the progress quickly hangs, and makes no progress for several hours. After attempting to import file1, I could look at the results and see that biblios 1-149 showed as imported, while 150 onwards remained as staged. Fighting with this for awhile, I got file1 to import fine using bulkmarcimport.pl, and attempted again to use the web interface for file2. This again hung, although biblios 1-549 had successfully imported this time, after that again shows as staged. I would rather not use bulkmarcimport.pl again, as there are duplicate biblios that I would like to merge, and from my understanding this is only possible with the web interface. Both of these exact files imported just fine on Koha 3 beta2. Other than upgrading to koha 3.0 final the configuration should be the same. Is there anything that could have broken this functionality in koha 3 final? Watching top after the import hangs, there is a process "background-job-" which appears to respawn every few seconds with a new pid. Nothing interesting seems to show up in any of the log files. I am using zebra, and the koha-zebra and koha-zebraqueue daemons are both running. Thanks, <PERSON>
7543f423-7caf-18a3-1a7b-09a280515a80
['5c2782ba-5fae-0834-c326-1b4b17e20205']
Hi, My library is migrating to koha from Winnebago. I have used bulkmarcimport.pl to import our entire collection into koha, which looks to have been successfull. I am running into trouble scanning in the barcodes on our books however, because winnebago stored only the 5 digit material number in the 852p barcode field, but the barcodes on the books contain much more information, in what I understand to be the codabar format. After the import, Koha now has for instance that a book's barcode is 12345; But the barcode on the book, which we would like to keep is 3055500012345x, where 3 means it is a book, 0555 is our location code and x is a check digit. In the old software typing in either the short version, 12345, or the long version 3055500012345x, would get the same results. Koha works fine with the short version, but reports an invalid barcode if you use the barcode scanner to scan in the book, since it is in the long format. Is there any way to get koha to interpret these barcodes correctly? Thanks,
9f0581ff-517b-856d-fd61-cfdc57e8f82d
['5c31f905-b85f-2f0f-ffd7-dcc0077e219f']
Dear mailing list, I try to get a final solution to the following problem: I have two color microarray data (common reference, dye-swap, multi-factor, multi-group) of which I want to extract differentially expressed genes (DEGs) using LIMMA. Those DEGs I want to analyse for the over-representation of GO-Terms using GOstats. I got as far as a list with GO-Terms and their respective P-values and even could construct a nice direct-acyclic graph out of this data. *However, the output I get does not account for the following: Are single genes/transcripts up- or down-regulated? What about the up- and down-regulation of whole over-represented GO-Terms? * *Could I get there using clustering approaches? * I'm sure somebody on the list has already tried such an approach, *finally yielding over-represented GO-Terms, which are either up- or down-regulated.* Thanks in advance, <PERSON>
8c782d47-6a32-5e1f-ec7f-9d4ba7dcada4
['5c31f905-b85f-2f0f-ffd7-dcc0077e219f']
Hi bioconductor mailing list, I have severall datasets and all of which share 2 reference samples. I want to do a normalization that takes this into account, so that each dataset is normalized against its own reference. In the end I would like to have all datasets normalized in this way, so that when thrown into a PCA the reference samples cluster together. Your help is greatly appreciated! If possible, please guide me to BioC-packages or functions as I am pretty new to BioC. Thanks in advance, <PERSON>
d7b5811c-16ab-e5a2-b9ce-4125c6da3181
['5c41a08f-5950-45e0-9b86-9f107a088e64']
Hi! I've just pushed five changes from Ubuntu xorg to alioth xorg.git.. Hopefully they all are useful, and if not, easily reverted :) There are a bunch of other changes that could prove useful. I'll try to either put them in a git branch or otherwise separately available for a review. I've done that for an older version back in March, and they are still available here: http://users.tkk.fi/~tjaalton/dpkg/xorg/patches I'm not sure how radically the postinst is going to change in the future, so maybe some of the quirks could be dropped now. comments? t
8dcfd402-925c-ba7c-3406-c91665880f32
['5c41a08f-5950-45e0-9b86-9f107a088e64']
Hi I've now even tested the patch against 1.5.3, and it works like a charm. Ok to commit? The rationale is pretty clear I think. "Some" distributions might have used 'vmmouse' instead of 'mouse' by default, and those who upgrade to a newer release could get upset when they can't single-click :) t
fa968a95-cdd0-04b9-96c8-250588f86c82
['5c445b60-086f-b700-2397-7c67790e6186']
FWIW, after the first day of WWDC this year, I decided to ship MoneyWell 2.0 as a Lion-only app. There has been some negative flashback from our customers, but not too much. We'll see if that increases after we (finally) ship it. We will continue to offer our 1.7 release on our website for those people still running Leopard and Snow Leopard, but the development advantages of moving to Lion were too great to ignore. We were able to ditch a huge amount of custom code and add most of the Lion-specific features. Additionally, there is the marketing angle. Apple does tend to promote apps that take advantage of their latest OS more than older apps. Peace, <PERSON>
3094b853-41cc-c027-2692-0a03a9f77fff
['5c445b60-086f-b700-2397-7c67790e6186']
I'm not sure what the issues are with the Lion UX, but I love them. I think that some people will always struggle with change, but if you give Lion a chance you'll see how important the changes in it are. For the average user, not having to deal with opening documents or restoring apps is huge. Snow Leopard was a mess of the old Aqua and modern variations of controls. Scroll bars have needed to die for a long time and gestures are a huge time saver for us trackpad users. :) Lion adoption will be much faster than Leopard and Snow Leopard. I think you'll be doing your customers an injustice by clinging to the past and infecting them with negative comments. Like it or not, Lion is the future of OS X and you should embrace it. Peace, <PERSON>
54bbc8b0-64cf-bfee-9d71-d32e5c1f95ac
['5c44f1bf-f561-6803-b571-f165fc86a639']
Without going into a sales pitch (you can get the detail off our website) we do offer such things... Home Phone Wireless is a fixed-line service running over our mobile network (so you get a landline number). Home Zone tells your phone you're in the house when you arrive home and converts it from a mobile to your landline number (which I think is what you're talking about in the US). And just to clarify a point about mobile termination rates - MTR and Fixed Termination Rates (FTR) are exactly the same in NZ. We don't differentiate. Currently they're both 15c/minute (due to decline by another cent in April next year). Cheers <PERSON> <PERSON>
52fd35ac-1750-4026-b4a5-47e5425b1df5
['5c44f1bf-f561-6803-b571-f165fc86a639']
Hi all. We can't make any public statement about this interference until after the court hearing, early next week, but after that I'm happy to get one of the network guys to come in and chat about it with you if a: that would help, b: that's of interest and c: they're not already here under their own steam. Cheers <PERSON>
a1324183-cc8c-bcb2-2b4d-8dbe94038d99
['5c491191-c639-91a3-02de-53525c3da746']
With TortoiseSVN, labels are decorated and svn operations are available on any folder/file in a subversion repository regardless of its depth in the filesystem tree. I think that this is a great feature. I'm pretty sure that with subclipse, you need to have the root of the project checked out from a svn repo. I was wondering if it is possible to have subclipse recognize any subtree not rooted at the project root as being under subversion control? Thanks, <PERSON>
47af42bd-5b7c-8dc1-ea30-76c924130c15
['5c491191-c639-91a3-02de-53525c3da746']
That comment line in Tix.py seems to be specifying information for Emacs python-mode. Is it? If so, should PyDev even be looking at this at all? I would venture to say that the warning should not be suppressed. If the Emacs specific settings are ignored entirely then the error won't happen in the first place. Is this a worthwhile approach? Regards, <PERSON>
441c9800-c90f-5ef3-7518-9a31f22be44a
['5c54a7f0-779c-e2f0-8f05-beb84a00725e']
Dear all During the last weeks I have learned quite a lot about the QoS mechanisms available on the Cat6500 platform. Now I would like to confirm somehow whether I understand one thing correctly: As soon as I have a connection with less bandwidth than the interface line rate connected to the Catalyst there is no way to achieve decent QoS functionality involving class based and/or priority queueing according to the low bandwidth available behind a (Fast)Ethernet-interface, right? One can use the PFC to globally police the output rate of the port. But although it is possible to use different classes with different policies here, one cannot do any flexible shaping/priority- or bandwidth-queueing at this level. The actual cos-based queueing will be done on port level and its mechanism always works in relation to the actual port line rate (using wrr bandwidth RATIOS). Thus there is no way to achieve something like a shaping down to 5MBit/s AND considering the dscp/cos-classes at the same time. And it does not matter which queueing capabilities the linecard actually provides. I tried this with 6348, 6548 and 6748 together with Sup720-3BXL. Did I get this right or have I overlooked some (more or less obvious) solution? Regards, <PERSON>
b9c17713-7178-158c-d41a-fb62847caec4
['5c54a7f0-779c-e2f0-8f05-beb84a00725e']
Hi all Today we were facing a ddos attack with approx. 1.5 GBit/s and 3 Mpackets/s. The traffic came in on a port of a 6704 and went out on another port of the same linecard. We have two annoying phenomena to deal with: 1. The 6509 is equipped with a Sup720-3BXL and DCEF on all linecards. During the attack it dropped all of its LDP neighbourships (actually they were flapping) although the CPU usage remained below 40% for most of the time and only went up over 50% at a very short peak not related to the LDP drops. What might have caused the dropping of the LDP neighbourships and what would one do to protect them? 2. On the outgoing interface we recorded a huge amount of output drops up to 60kpackets/s. The receiving router here is another Sup720 with 6704 but without DCEF. From my understanding 1.5GBit/s and 3Mpackets/s should be no reason to drop packets on a 10G link. So what might have caused these packets to get dropped? Thanks for any advice or hint, <PERSON>
4c20e305-9db7-dc0c-5378-5e00953e0fb7
['5c5b047b-a330-bfc1-e3c9-4f499b564592']
Hi all, I'm using Apache 2.2 on LE for several years now and i use intensively the virtual host mechanism (300 virtual hosts per http instance). As soon as you start sharing resources between applications, you want to protect your platform and prevent an application from taking all the threads and cause a resource starvation for the others. To do so i used the third party mod_qos module to limit the number of simultaneous connexions per virtual hosts. Now that i intend to migrate to Apache 2.4, my first tests reveal that this module is not compatible with this version of Apache, that's also the warning that developper raised for his module. It appears that i don't find any real alternative to substitute the mod_qos module with something else. Here are my questions : - Can you confirm that Apache does not provide any mecanism allowing to limit the number of connection per virtual host, just to be sure that i'm not missing something ? - Why Apache doesn't provide such a functionality ? :) From my very humble user perspective, i'm surprised that this is not a native functionality as it seems to me that my need is probably shared by many users. Moreover Apache provides many other complex functionalities, the one i'm describing would be probably something simple to implement compared to other functionalities already available. Thanks in advance for your feedbacks. <PERSON>
d93af485-669c-09d2-bd33-7f98f4e98a28
['5c5b047b-a330-bfc1-e3c9-4f499b564592']
Hi everyone, I'm actually using Tomcat on my environment platform (Tomcat 5.5 / Tomcat 6 and soon Tomcat 7). I have a frontend Apache http Server using the jk connector to communicate with Tomcat instance. I'd like to change this connector and use the mod_proxy one for several reasons. The main difficulty to handle is relative to the remote-user information. Indeed the jk connector automatically transmits the information so that the application can retrieve it using a request.getRemoteUser() method call. If i'm not using the ajp connector anymore, i need to handle something on the tomcat side to set the remote user in the request object. I thought i could use a valve to do this. And that's where the road ends, i have watched the ajp conenctor code in order to see how the remote user is set in the request but i can't find it. Could you please tell me how i can do this ? Best regards. <PERSON>
381cf7d1-0322-4479-4ded-628b115646ec
['5c662a1a-b2d8-6a9d-71f5-9e6f498d3deb']
Hi, This is probably off topic, but out of curiosity I wouldn't mind asking. Does anybody experience system hang ups after upgrading to 8.0? IBM T43 8.0 generic (thought that something went wrong during upgrade from 7.2 to 8.0. I did fresh install, but the problem persists). Which would be the correct mailing-list to post the message to? Regards, <PERSON>
0d0d8b83-cbfe-8403-3258-e3f9d342f4da
['5c662a1a-b2d8-6a9d-71f5-9e6f498d3deb']
Hi, Does anyone experience sluggish kmail when scrolling a folder with emails containing attachments? Nothing crushes or hangs up, but it takes +- 30 seconds to open an email. Same goes when sending an email with an attachment (ie. with jpegs). It takes a good couple of minutes to send it (ie. progress bar stops and 99%). Regards, <PERSON>
fa1bd3b5-ea63-590e-b3b0-9d36bad2d9c9
['5c668d29-4e78-b5c8-ce42-31c77f2b6f95']
Hi, I have successfully compiled XMMS 1.2.5 for OpenWindows on Solaris 2.6.But... The problem is that when I play MP3 files they are broken up due to what would seem to be the MP3 decoder not decoding the data stream fast enough. The CPU is only ever about 12-15% utilized but it is almost like something is preventing XMMS from using more CPU in order to cleanly/smoothly decode MP3 files. I have tried turning on "Use Realtime priority when available" but this does not seem to have any effect. Also tried to re-compile after changing the -O2 flags for the gcc compiler to be just -O and still the same result. I am using GCC 2.95.3 If I use a command line MP3 player (mpeg3play) all the MP3s I play decode and play perfectly. Only problem is that being command line it is not very robust. ie. No play lists or user interface etc. Any help or suggestion would be appreciated. Cheers, <PERSON>
caad85db-905b-6044-dfb3-bf9738b9b8f2
['5c668d29-4e78-b5c8-ce42-31c77f2b6f95']
I have what would appear to be a similiar if not the same problem on SUN Solaris 2.6. mpg123 on it's own compiles and runs fine ( takes 8-10% CPU when playing MP3s). XMMS plays the MP3s but the audio is broken up, almost as if it does not get enough CPU time or something. <PERSON>.
f147b2f0-2cc0-a0a7-7dcc-d264ff6652b1
['5c6c99a8-cfbb-76d0-5494-72863bbe5761']
Thank You : ) your post gave me some comfort : ) Yes, the problem is related to paging, and not necessarily activity/places. as for your opinion of master/detail, I think it is a shame to not take advantage of the fact that we can display master/detail side by side, without a full page reload, something that was not possible with traditional webframework, and is nearly possible now with AJAX/GWT, only if there comes a solution to implement the paging correctly. regarding your suggestion on using "continuous scrolling", I am going to investigate that now, and I believe it will work, however my concern is this, lets say we have a list of employees with continuous scrolling, and next to it, we show detail of employee, when one employee is selected from the list. if we bookmark place employee/777 when navigating to place employee/777 for the first time from bookmark, how can the master activity which contains a continuous scrolling table/list, highlight employee with id 777 in the list, when in fact the continuous scrolling list presents only an X number of items initially, and load the rest on demand ? do you think this is possible ? Thank You
ef536cf2-6daf-5808-29c8-c4d43ed6e0b7
['5c6c99a8-cfbb-76d0-5494-72863bbe5761']
Thank you very much for sharing your experience. on this comment, " I also pay close attention to widget lifecycles which is critical to successful implementations of widgets that get attached, detached and reattached and so on and so on. " why widget lifecycle is important ? why is it important when they are attached/detached ? and how can you monitor it ? (logging your java code or viewing resulting html markup gwt creates ?) how can you tell in what state a widget is: attached/detached ? what are the states ? only attached and detached ? Thank You
39001400-8a8f-5e78-a790-0ff65060bcd6
['5c6d97f1-5fa2-fe0a-5ce2-148c8fb43c9b']
Hello, I guess this is the wrong place about asking about qmailrocks and bitdefender.. It looks like it can be done following qmailrocks's install guide, just replace clamav with bitdefender. Howto install bitdefender, ask/read on bitdefender website. QmailScanner will check for virusscanner and set apropriate settings - on compileing/installing, else follow the guide. http://qmail-scanner.sourceforge.net/ But I would suggest you to try clamav if there are no reasons not to do so. better knowledge about it in the forums ( on qmailrocks ;) B/R Ole J
3ee95420-19ae-5c98-17c8-c6d87828b528
['5c6d97f1-5fa2-fe0a-5ce2-148c8fb43c9b']
Obvious the table are missing, For what I have experienced, newer mysql db's does not support . ( dots ) in table name, like vpopmail.mydomain.com Anyway I am not a MySQL wizard, but this is something you can give some thoughts until some of the specialists here on the mailinglist wakes up Try to migrate the data from old table to a new table without dots. Navicat software is a good thing for handling MySQL - transfer, backup and more. B/R Ole J
276b635b-eaab-2268-cc72-87eca90c21e1
['5c701c64-8e8a-ce3d-1d3e-6f4c0b0ff75e']
Hi I am having two problems with my fedora 10 system. 1 Once I tried modifying /usr/share/PolicyKit/policy/org.freedesktop.hal.storage.policy to make it mount removable pendrives automatically always without any authentication. Now I dont know what went wrong, I am not able to mount my removable drives with deskbar applet or from nautilus. I have to manually mount it using mount command with root priviliges. I tried replacing the package that provides this file(brand new file) still no succes. 2 I dont get a shutdown in my system menu. for that matter no where in my GUI, I am able to use shutdown or restart (including GDM). I always hv to shutdown or reboot from command line these days. Please help me if somebody already faced this problem and worked around it. thanks in advance <PERSON>
ebe11f45-ef03-cdc8-517c-68d96ab172b5
['5c701c64-8e8a-ce3d-1d3e-6f4c0b0ff75e']
Hi I am writing a graphicsitem that displays a lot of thumbnails on screen in a big window and then depending on navigation keys browse through them and some action on selecting any of them. I have done most of it using pixmap items and grid layout. But now i am confused whether usign QTableview will give me any advantage other than scalability or it will be overkill for such a small task. Thanks
6e785f67-f01f-7b91-49f8-d7fceef59ba6
['5c83777e-f957-b42f-5c08-b7be0f479299']
Some time ago, i was getting an error while trying to use RTTI functions in a class descending directly of TObject Later i discovered that i needed to descend from TPersistent due to $M+ directive (http://www.freepascal.org/docs-html/prog/progsu44.html), so i did. Now, i decided to go back and descend directly from TObject. To allow RTTI usage i would put M+ around my own class I tested again and using RTTI functions (SetOrdProp, GetPropInfo) worked even without declaring {$M+} So, is it really necessary to put $M+ to use RTTI? Luiz
ca612a5c-82ee-fc12-3089-236896021859
['5c83777e-f957-b42f-5c08-b7be0f479299']
Hi, I've managed to set a COM interface property (tkInterface) through RTTI using SetInterfaceProp. But i could not find a way to set a CORBA interface property (tkInterfaceRaw) since SetInterfaceProp expects IUnknown. In the attached example, when i try to use SetInterfaceProp i get the message "Incompatible type for arg no. 3: Got "IMyIntf", expected "IUnknown"". Is there a way to set a CORBA interface property through RTTI? <PERSON>
9293628e-a431-09fd-3adc-01f71c3ee922
['5c94ccd4-b89f-6cde-92b8-c980abe11c42']
Mifos 2.1.5 is now available. This is a hotfix release that includes the following fixes: * changes to mandatory savings amounts are now accurately reflected in the UI and appear correctly in the collection sheet report * issues/withdrawals amount is now displaying correctly in the collection sheet entry pipeline * Improvements to Question Group query to improve loading speeds of center, group, and client pages * Other improvements to the batch job the generates meetings for savings * A removed fee is no longer showing on the 'Review installments' and 'Preview Loan account information' pages * We fixed the collection sheet error: "The amount specified for <Loan product> of customer <customer name> is invalid. Please specify a valid amount." For the download and documentation, please see http://mifos.org/product/download-mifos. List of issues fixed can be found here: http://tinyurl.com/3et7u55 and in the release notes: http://tinyurl.com/43sga7p. Please note the upgrade instructions as well when upgrading to this release. Cheers, <PERSON>
67170ccc-bb6b-7936-0c58-d880d0130278
['5c94ccd4-b89f-6cde-92b8-c980abe11c42']
Mifos 2.1.6 is now available. This is a hotfix release that includes the following fixes: · Stack trace appears is no longer appearing after submitting Question Groups with response which has more than 200 characters · Client details page are no longer loading too slowly for some MFIs For the download and documentation, please see http://mifos.org/product/download-mifos List of issues fixed can be found here: http://tinyurl.com/427kfya and in the release notes: http://tinyurl.com/43sga7p Please note the upgrade instructions as well when upgrading to this release. Cheers, <PERSON>
3b56dde5-2183-bad8-e626-5b0be52ba20f
['5c96b32f-a765-a2fc-74e3-f23b4c773c64']
Hi <PERSON>, Occasionally I receive strange error report like the one below. You see that the AND part is missing the right side of the "iss_id<>" expression. This is not the only SQL code I get this strange behavior. Seems to me the original caller may had the value passed in the call chain but somehow the value got lost. Reviewing the code this should not happen, but then I see the result like this below. I experienced at another case that, for instance, de-associating a custom field from a project (which has project code less than 10) will delete the custom field values from ALL projects accidently - due to the same malformed SQL as like above. However if I de-associate from project where the project code is equal or greater than 10 the value removal will work fine and removes only custom field values belonging to that particular project. Do you have idea whats going on here? What kind of Eventum/etc logging I could enable to track this down? I suspect the step when a value is taken through the web interface. Thanks <PERSON>
076dc399-af4d-dd6f-6397-e8a357e7b81a
['5c96b32f-a765-a2fc-74e3-f23b4c773c64']
All, We experienced sever data loss recently: while de-associating a custom field from a project all these custom field values have been deleted *across all* the projects. I would expect that only those custom field values would be deleted which belong to the project de-associated from. Say have a custom field "notes". This is associated with several projects (A, B & C) and the projects' issues actually store values in this field. When we de-associate this "notes" field from project A we will see all "notes" values disappear across projects A and B as well. However I would expect "notes" values for project A & B would remain in place. My small investigation shows that there is a table issue_custom_field which stores all the actual custom field values for all projects. During the de-association step I see that all the records belonging to the custom field number (icf_fld_id) are deleted, not taking into consideration the project specific issues (icf_iss_id). I guess on the code level this hits the removeIssueAssociation() at class.custom_field.php. This function makes an attempt to collect together the list of project specific issues which would be used a few lines down for the actual delete statement as part of an "IN" clause. In my case this list gets long (1000+ entries) so I guess this hits the limits of an "IN" list and somehow executes ignoring the list and resulting in the data loss. All this is guess - I need to go ahead and re-work the IN statement to a JOIN to confirm this. There is no error msg logged during the operation. Anyone else had this problem before? If the above right then better off not to de-associate any custom field until this issue gets fixed... <PERSON>
25c9562f-a844-bed0-68ed-e04730b7d094
['5c9ba563-e2ac-6118-316e-694edff9c5a7']
Hi, One of my local users needs to be able to receive specific emails from one particular user, however, the sender is sending attached files which are zipped .exe files. At the moment, I' having to unquaratine these mails daily, but ideally I would like to set MD so that the extensions check is skipped for this sender. If I put in a filter_sender section, where it matches the sender, that returns a ACCEPT_AND_NO_MORE_FILTERING, will this be executed before the extension check, or will I need to do something different here ? Thanks, <PERSON>
6d8c0443-6068-6ca0-aa3b-17d4133aba1d
['5c9ba563-e2ac-6118-316e-694edff9c5a7']
Hi, Could some please let me know what files and the best way to copy an existing AWL and Bayes DB on a system running MD 2.43, SA 2.63, to a totally new machine running MD 2.51 & SA 3.0.3. All the path names, configs etc are in the same place. The procedures that are in the notes only show upgrading the same system, not physically a new machine. Thanks in advance <PERSON>
03d5f8c2-b341-6ef9-31e5-64c4cdbe25f4
['5cba6505-a934-79d5-e952-370699eb49cd']
I'm not that experienced with Maven, or Eclipse.. So forgive me if this is a stupid question. I have a class that relies on a properties file. I place that in ./conf for Maven to pick up and add to the War. However, if I run my test files in Eclipse, those file haven't "moved" yet... So the path for the app is different. Can anyone give me some suggestions for how to deal with that? <PERSON>
1a6a6478-00c6-1456-ed9d-62dc345d4d5c
['5cba6505-a934-79d5-e952-370699eb49cd']
Using maven 1.0.2 deploying to Tomcat 5.5.x and creating struts app with maven genapp struts I can't seem to figure out where to put the context.xml so it get's placed in the correct spot during deploy. I'd normally create webapp/META-INF and put context.xml there... But it seems like xdoclet would want to handle this rather than just letting the file structure of the app do it. Any thoughts? Greg Akins STP Systems <PHONE_NUMBER> tD+8vlTcEam3Gh/D@example.com
434c0e0e-43f2-5daf-842a-88a17b012044
['5cc16554-a78d-6bd3-a7f1-08b38f34b4e7']
According to http://developer.mozilla.org/en/docs/Code_snippets:Post_data_to_window#POSTing_data_to_a_new_window I've tried to post form data to a new window. Providing aReferrer with null, the new window opens up but does not show anything (not even the URL I've provided). If I also set the aPostData to null, at least the URL (and hence the page without processed parameters) is displayed. So, obviously, posting data according to the provided example does not work. Can anyone confirm this or give more hints on what could have gone wrong? Thanks for feedback! <PERSON>.
15370618-5d2b-0d38-881f-3b7d6ad07466
['5cc16554-a78d-6bd3-a7f1-08b38f34b4e7']
Hello all, I've written a Firefox extension and it's doing some initialization stuff when the browser window is opened. Now I'm looking for the easiest/best way to assure that the initialization routine is only run once, ie. if the user opens a new browser window, the extension should run again. Any ideas on how to achieve this? Thanks in advance, <PERSON>.
abe07e37-314e-c288-c05b-8bbe179ef197
['5cc1c17f-6668-72c8-3f7e-2704826b69b6']
Please enter the details of your bug report here The ECB methods buffer is not displaying class members in my C++ source files. However, it does display the non-class methods that exist in the same file. It also is not jumping to the non-class method when pressing return or clicking the middle mouse button in the methods buffer. These things work with the older version of semantic. Thanks for looking into this, <PERSON>
d69a5530-7276-6906-3689-7cf483c68e0b
['5cc1c17f-6668-72c8-3f7e-2704826b69b6']
Hi, If I have an edit window with source code in it, ecb populates a methods buffer. When I start VM (M-x vm) this opens a new frame. I read my mail and then quit VM, which closes the frame, At this point the ecb methods buffer is cleared, forcing me to rebuild the methods buffer. This seems to be new behavior in the 2.20 release. It also seems to be related to VM in some way. I can open new frames manually, or use other packages, like the woman manual reader which opens and closes new frames, and not see this behavior. I tried using the (defadvice fix as described in the conflicts section of the manual and it did not fix the problem. Thanks, <PERSON>
7024aa90-0e74-63c2-2d83-e165db7603a9
['5cc335a4-a5d5-d9ac-f8bd-83a4104458b1']
Hi, Can i have variable think times for different page. Some of my pages people will be spending maybe 30-40 seconds without initiating the next request. some they might do it in 5 seconds, some maybe 10-15 seconds. can i have such variable think times. Does a random timer do something simlar. Or it randomly assigns think time to any request and doesnt take care of these special pagesa that user spends more idle time on. regards, <PERSON>
15299c80-aaf7-3e02-9631-2670fae4449e
['5cc335a4-a5d5-d9ac-f8bd-83a4104458b1']
Hi, Can I assign a different IP address with a users sessions ? I got a load balancer thats got sticky sessions associated to source IP. Most request go from 1 ip as they are being generated from couple of machines only, eventhough i got huge number of users. AS all requests come form 1 ip all sessions get routed to just one app server. I got 4 app servers with this load balances, remining 3 dont get any requests at all. Is there any way to mimic varying ips? regards, <PERSON>
bad7cb20-7880-c086-1640-2f710dcc580f
['5cc53c6a-d543-5488-8c80-95910976965a']
Hi <PERSON>, Here's the promised update with the big multi-BSSID changes, which are related to HE (802.11ax) for which we also have some updates. Note that I took the liberty of including <PERSON>'s rhashtable API removal, I merged net-next for that and verified that the API he removed is also not used in anything new, but if you apply/applied any patches after I pulled you should probably double check that again. Please pull and let me know if there's any problem. Thanks, <PERSON>
383f59f0-1d18-a2ee-e74a-7dd8b404f776
['5cc53c6a-d543-5488-8c80-95910976965a']
I collected everything on http://johannes.sipsolutions.net/Projects/mem1394 It's tested now and works for reading, writing isn't implemented yet. There were some bugs in the previous patch for the dynamic device registration (wrong checking of return value of cdev_add) and in the node_interface registration (misnamed function) that have both been corrected now. If you want, I'll post the patches for a wider audience to lkml. johannes
0e52ee4f-f8e6-02fb-251b-3493a65cf057
['5cc67edc-89b0-5c6f-16d3-dac981c4c3d0']
Not a complaint about the Tapestry list (or PMC members) but I though I voice my opinion about the policy to see if there are others like me. I think I understand the reasoning behind the Apache mailing list policy but I think it's outdated. To me it has two problems. 1. I cannot participate in the list from my work email and I cannot read my personal email (easily) while I'm at work. I suspect there are others like me. 2. There is no real built in archive system which makes web searches a problem. Nabble is considered a hosting site and hosting sites are blocked by my work firewall. These two things make it difficult for me to participate in the email list and make the archives useless to searches because when I click on a result 9 times out of 10 I get a firewall violation. It's things like this that make open source adoption problematic in the corporate world. I realize you are bound by the Apache rules but if no one complains about them they will never change. Thanks for your hard work <PERSON>
def04487-890c-70a8-98a4-c297bed5fb57
['5cc67edc-89b0-5c6f-16d3-dac981c4c3d0']
I migrated 2 sites from 3.x to 4.0.x. I just put in the new libraries and fixed everything that did not compile. For the most part the 3.x html and page files will work. Both took a couple of days to get a site that mostly worked. The biggest problems were things like PDF pages, all the plain HTML stuff was pretty easy. But as you say, you end up the something odd after this. On one project I just put up with the fact that it's an in transition project. Anything new I do the 4.x way. When I have to mess with old code I update things. On the other project I took the opportunity to clean things up and started over. I needed to make some schema changes, I wrote a bunch of hivemind services and pretty much rewrote the entire site. That took about 6 weeks but the new site is better, has test cases and is much easier to work on now. I would try and covert the site, get some experience with 4 and when the time is right rework the site to take advantage of the 4.x features <PERSON>
14633768-e909-92c6-9b47-966e27aa0acf
['5ccc1537-f5f8-6b99-d5a6-777cdb9cde64']
Hi, I am using the YAHOO.widget.simpledialog to show a dialog to confirm if they want to go ahead and submit .But i found that the dialog would show and would not wait for the answer .the code would proceed further and submit .I know that the answer to the dialog can be handled by functions .but since i am going top down on my code to check and submit what do you think would be best possible way to achieve this. <PERSON>
f8fe10d2-b31e-30b9-ae16-35c0f2fee080
['5ccc1537-f5f8-6b99-d5a6-777cdb9cde64']
Hi Everybody, I tried changing the property of the calendar after it is rendered.I think it has to done using the method queueProperty(I am not sure if I should be using setProperty) of config object.After which I am firing the queue using fireQueue of the config object.As per the documentation this sets the property and fires the event.what I found was it was setting the property properly .but would not fire the event . Here is a link to what I mean http://chinnakaruppan.googlepages.com/fireQueue.htm .Try hitting on the next of the calendar .it would throw an alert.But when you click on the button it supposed to throw on alert which it is not doing... thanks for you time.... THnks Chinna
600625f8-74aa-4f1a-715b-a978b22295cf
['5cd4a2f1-5c8f-57c3-d37c-64f8ba119dca']
I downloaded the V9.5 back two months ago, we have NOT applied any of the V9.5 yet, so all my developers are still running V9.0. Do I need to get a fresh download of the base V9.5 or should I continue to use the one I downloaded two months ago? Thanks,
c2549f2d-c736-90a4-a385-dd945b751757
['5cd4a2f1-5c8f-57c3-d37c-64f8ba119dca']
I do not understand what all the hoopla has been, to be totally honest. I just updated my RDi V9.1.1.1 to RDi <IP_ADDRESS> by just simply changing my repository from 91 to 95 in the IBM Installation Manager. To IBM's credit, the Install Manager A) updated my IBM IM from 1.8.1 to 1.8.3, then updated my Version of RDi, and took care of loading up the latest Fixpack. This all happened with just running the Install. Maybe it was my misinterpretation of the Blog over the last several months but this was really nice. I thought I had to download the new version, then download the fixpack, etc. but after all was said and done I did not have to do any of that. AM I just losing it or did I just get great service from IBM? Jerry Roedersheimer Rumpke Consolidated Companies, Inc. Systems & Programming Manager <PHONE_NUMBER>
dd57f6b1-6a44-966c-f07f-55faaa6627af
['5cd8e141-6ea9-c034-1719-18b038bf2f0f']
Hi - I'm using Plone 1.0.1 + CMFCTRAddons 0.1 on two servers: one FreeBSD and one WinNT. I'm uploading content via WebDAV using Windows 2000 Web Folders. - I can upload News Items to both servers - I can upload Plone Events to the Windows NT server - When I try to upload Plone Events to the FreeBSD box, I get an error that says "An error occurred copying some or all of the selected files" - If I uninstall CMFCTRAddons from the FreeBSD server, I can upload Plone Events (Of course, they get turned into Documents in the process) Does anyone know what's going on here? Thanks. <PERSON>
fa6ed431-91a7-ac6b-5fc6-759e957b8583
['5cd8e141-6ea9-c034-1719-18b038bf2f0f']
Hi- I recently posted two problems with Plone Events/Calendar 1. Events on March 31 didn't show up on the slot calendar 2. Could not upload Events to Plone via WebDAV (with CMFCTRAddons) The really odd thing was that I observed these problems with Plone 1.0.1 on FreeBSD, but not on Win32! The answer: Apparently, the Win32 Plone 1.01 installer was built with a patched version of CMFCalendar 1.3 I've installed CMFCalendar 1.3.1 on my FreeBSD site now, and the problems have gone away. Thanks to <PERSON> and others for getting me started with Zope debugging, so I could track this down. <PERSON>
a6a2ea6c-a2a6-6c51-e441-e38b0c08cd5e
['5cf49ef7-b50a-2bac-5863-36053e18670b']
Hi I am having a problem executing following query: cursor.execute("select * from projects where projectid between '%s' and '%s'"%(id1,id2)) But say if i have projectid's from 1 to 20 in database & if i give the project range 1 to 2 then it gives me the result 1,2,11,12,13,14..... When I execute this query in mysql i.e select * from projects where projectid between 1 and 2,I get the correct result. I also tried doing cursor.execute("select * from projects where projectid between '%s' and '%s'"%(int(id1),int(id2)) but no success What should I do?? Regards, <PERSON>
cfb87f8f-ef80-b1a3-4e6b-bd711200356f
['5cf49ef7-b50a-2bac-5863-36053e18670b']
Hi I am drawing a vertical barchart for project vs start date The vertical bars are for projects. Now, at the top of each bar I also want to display the number of projects. Also I want to show the color convention i.e Say I have different bars for H.R projects,etc... Now,if HR project bars are of blue colors then at the side of graph I want to show a small rectangle of blue color with "H.R project" beside it. How should I do this? Regards, <PERSON>
4cd6279e-3d75-e509-450f-d108bd9043a9
['5cf565f7-7d5c-0504-c76c-483aa3f133dd']
I added a node last week which appeared to work fine. But, after a while, it loses its default route, and so cannot access the Internet. It can talk to other nodes fine though, and uses what should also be its default route in order to get to all of the other nodes on a ping. The ETX is around 300. This is a two hop node. I use to have this problem with my one hop nodes, but that was fixed as of rev:3790. Part of the problem with this new node is that it should be a one hop node, but it is preferring to communicate with an antenna I didn't intend it to. Regardless, it is connected to the mesh, so it should be able to find out about the Internet access at the Internet-connected node. And so far it only knows about that for a short period of time right after boot. In HSLS, how often does a node get updated routing metrics about a node two hops away? So far I haven't seen the default route come and go. I see it only immediately after boot. What sort of logical messages would cause a node to lose its default route, and why are they being sent to this node? -John
0b77f02c-3151-7cbd-e3db-209f415b60f4
['5cf565f7-7d5c-0504-c76c-483aa3f133dd']
Is there a known problem with the web server in 0.5.5? I haven't been able to pull up an interface through a web browser using that release. Reverting to 0.5.4 gives me the interface back. Has the way IP addresses become assigned changed as well. I get a different IP on the wired NIC everytime I boot with 0.5.5. Regards, <PERSON>
fc3326f7-109a-e493-11ac-c138b54c32b8
['5cf581ab-a8bd-bfe4-54e0-a4c288e62b54']
Hi Guys, Sorry to bother. But we are really anxious to resolve this issue. We're having a very weird problem with the i915 driver on Opensuse 13.1 and wanted to see if one of you Intel experts can point us in the right direction. We're experiencing this issue on several hundred of our machines, and we're dying for some troubleshooting help. Here is the issue: Randomly, our screen display will change to a solid color. It looks like it takes what we think is the upper left pixel and display that color over the entire screen. We can start a virtual x session into the screen and still see the entire screen just fine, but our customer standing in front of the screen can only see one color. If we reboot the PC everything comes back as it should. However, if we just restart the X server it does not fix the problem. This occurs randomly, but we're seeing it 2x per day with 67 machines running the same setup below. Interestingly, our machines running on a later model Lenovo M71e do not experience this behavior, and I believe that model has the Intel integrated HD Graphics 2000. After some experimenting, it also has those behavior 1. When we keep power to the display but power cycle the PC it DOES fix the problem. 2. When we keep power to the display and just restart X it DOES NOT fix the issue. Do you guys have this similar issue? Thank you <PERSON>
d81e9bff-01dc-5389-466a-461e87ac07c6
['5cf581ab-a8bd-bfe4-54e0-a4c288e62b54']
Hello, We have an Eaton 3S 550 UPS using NUT tool to send power loss event. By default the power voltage threshold is set to 85 volts. If the voltage is less than 85 volts, it will trigger ONBATT event from NUT. Is there a way in NUT we can change the threshold from 85 volts to a lower voltage like 40 volts? We were trying to change ups.conf to modify the output of upsc ..., but looks like it never pickup the new settings. Thank you so much for your help Regards <PERSON>
426e59b6-d34c-5902-3dd3-2c4226d7002d
['5d05c270-917d-2799-8ff1-5d80d4f588b3']
Hello, RubyCorner has been slowly (but continually) growing as bloggers writing about Ruby and Rails join it. Rigth now we have 286 blogs, and 379 users registered at RubyCorner. We must confess, that a big work load took us apart from the RubyCorner development for couple of months, but we are back. We are fixing some issues reported by our users, and filling some gaps (adding some missing feeds). If you have Rails related pictures at Flickr, or Rails related videos at YouTube, just tag them with the 'rubyonrails' tag and they will be displayed at RubyCorner. Best regards from Venezuela, <PERSON>
7932fee2-0fba-7ccb-73cc-bb00c02bc44c
['5d05c270-917d-2799-8ff1-5d80d4f588b3']
Hello, I want to get the production environment configuration from the database.yml in the /shared folder of a deployed rails application. I could use scp to get the file, write it to temporary file, read it and open it with YAML, but is there a better way to do this? Best regards,
61d41e8c-401d-b0af-0ad9-fa38ad464b65
['5d10ad99-8189-8672-303e-49321e112834']
Hello, In Djnago 1.7 Manager class extends BaseManager (actually, it extends its proxy, created using from_queryset). BaseManager is not mentioned in official documentation, so it is not part of official ("stable") api, right? "All the public APIs (everything in this documentation) " (from here https://docs.djangoproject.com/en/dev/misc/api-stability/) But it does not start with underscore and its module (django.db.models.manager) is public module aswell. So, it should be considered as public class according to python codestyle. I believe this fact confuses user. May one use BaseManager or minor api change may brake such usage? The other problem is there is no official Manager reference: there is a good manual, but no reference where all methods are listed (like we have for QuerySet, for example). So, what is the official, stable Manager API? Most developers know answer to this question intuitively, but there is no formal way to answer it, right? I believe that in Ideal Project there should be something like this: 1) Move all "non public" classes to "_non_public" packages (or make classes _NonPublic) 2) For each public class list all members in doc or directly in python documentation. What do you think? Ilya.
4ef10115-b553-1bcf-8f6a-02e638597147
['5d10ad99-8189-8672-303e-49321e112834']
Hello, I have server CentOS 5.7. It has mysql-community installed. I need to run perl-based tool (mysqlreport) and it requires perl-DBD-MySQL (not a surprise) When I am trying to install perl-DBD-MySQL via yum -- it says it requires libmysqlclient.*15 *and it conflicts with libmysqlclient.*16 *which was installed as a part of mysql-community. I tried to make symbolic link and force DBD to work with 16. But it failed with error. Looks like libmysqlclient provides it version or has different API (which is strange for me) The only one idea I have is to fetch libmysqlclient.15 from some packet and put it to my LD folder directly. But what has happened? Where is DBD for new version (16)? What is the best ("official") way to solve it? Thanks.
8be394ab-76fc-40ca-8d4e-e84744f89aa5
['5d1ea9eb-3d97-dced-7615-61dd3a0c49f2']
I am considering starting a project that would design and implement an open-source DRM solution for Ogg Vorbis media. Ideally this DRM implementation could be extended to other formats of media, but I believe Ogg is a good starting point because of its openness. To avoid duplicating someone else's efforts, does anyone know of any current DRM projects for Ogg Vorbis? If not, I would like to start creating a high-level architecture document to get this project underway. I realize that the idea of DRM contradicts with the philosophies of many Ogg Vorbis developers and users; however I believe it would be in the best interests of Vorbis to have an open, independent DRM solution. <PERSON>
4c9eaa11-543b-19be-7710-41fcdc6f09e0
['5d1ea9eb-3d97-dced-7615-61dd3a0c49f2']
I am wondering if there are any active projects to produce Java wrappers for the MusicBrainz libraries? If not, I would be interested in creating these wrappers. From looking at the MusicBrainz code I believe creating JNI wrappers would be relatively straightforward, but am wondering if other developers are aware of any issues I should be aware of. I would like to create MusicBrainz wrappers so that application developers can easily embed MusicBrainz functionality in existing Java applications. Of course all work I perform will be released under the LGPL. Regards, RS
d98f51ca-dbb6-db3e-dcf3-b5a2a84e9f42
['5d1f1f51-fff9-655e-4cf0-82a28869b7ba']
Hello, As of the 0.9x Alsa releases, my SB16 ISA PnP won't work (when loading modules, they cannot find the card). It works perfectly in the 0.5x releases. In addition, the kernel modules work fine. If have read the website and my modules.conf is fine. ISAPNP is working fine on my system ( allthough I have ./configured the drivers with and without --with-isapnp=yes and the same inability to find the sb device is given). Suse's 8.0 rpms as well as their updates rpms have the exact same problem as when I compile my own. Any suggestions? Thanks!
48f6a148-6eb2-cd64-e184-ace1134c90da
['5d1f1f51-fff9-655e-4cf0-82a28869b7ba']
ALSA Project Team, i386 system with a SoundBlaster 16 PnP in the only isa card slot. All other slots are PCI. Linux kernel 2.4.18 with sound support as kernel module only (SuSE 8.0). ALSA 0.5x works great when using the --with-isapnp=yes configure option. In addition, SuSE's ALSA 0.5x RPMs work great also. alsaconf modifies /etc/modules.conf correctly; and no irq, dma, etc parameters are required. Problem: ALSA0.9x wont work at all. I've read all documentation and I am aware of the syntax changes (snd-card-sbawe vs. snd-sbawe for instance). I have not altered my kernel that works with ALSA 0.5x (stock SuSE 7.3 and 8.0; I've tried both!). SuSE's included ALSA rpms dont work. Compiling and installing ALSA 0.9rc3 goes well, but modules wont load; device not found errors are given for snd-sbawe and snd-sb16. I even removed the isapnp.h file before configuring. The kernel sb.o module still works, so it is not a bad card; it has worked for 2 years previously. Lets solve this problem together. My machine is at your disposal. If you can give advice, I will try them and report results back. BTW, I do not have a /etc/isapnp.conf file as my only isa card (this card) was set up satisfactorily by ALSA 0.5x (compiled using --with-isapnp=yes did the trick). Thank you! <PERSON>
6253aaef-3ef4-6ee9-4b64-194a8aaab1d7
['5d203361-f49a-cd7b-7c4d-ac25a35a0b58']
Hi I'm interested in the group's feedback on the most accepted way to deliver a final PT report to a client. Best practices indicate that reports are only sent to a select group of people in each of the Red/White/blue teams, and docs are sent via encrypted email and/or the document itself encrypted with public/private keys exchanged at the start of the engagement. I've even heard that sending electronic copies of the report is a no-no and only a hardcopy should be couried. Could someone weight in on caveats and/or industry standards for report delivery? Also how would report delivery best practices from an internal pesting team differ (if at all) from that of a third party consulting outfit. Many thanks.
0873556f-4ff9-87f6-0504-8dc5f30c0be4
['5d203361-f49a-cd7b-7c4d-ac25a35a0b58']
ok we all know that in addition to good network, host and application security skills, programming in C is a pre-requisite for a decent pen tester or at least one who wants to write their own security tools or simply audit the open source code they use. My question is, despite their similarities should a pen tester be concentrating on C or C++ ? That's it! Thanks.
bf934c22-7dd6-4c69-e16c-5012669014aa
['5d2323a8-15fa-3910-f2e1-767c7f07dbb5']
Hi folks, I have installed the ovirt OS on 3 dl380's. It has been deployed with gluster and all is fine up to a point. When I try to install the engine it is trying to get out to the internet but unfortunately I am behind a proxy with strict rules, is there a way to point to a local copy of the engine using the GUI install or can I scp the iso over to the first node for local install, is there a particular partition where the iso should be stored. Any help appreciated, BR/<PERSON>
f2cf3f82-c4b3-15b6-77da-c8856ca016fe
['5d2323a8-15fa-3910-f2e1-767c7f07dbb5']
Hi There, Our 3 node Data center is up and running and I am populating it with the required vm's. It's a test lab so I want to provide pre fabricated environments for different users. I have already set up a CentOS box for nested virtualisation which works quite well but I would really like to group multiple machines together in to one template so that when a user deploys or chooses the template they get all the vm's together as one. Is there a way to do this in oVirt without nested virtualisation? I particularly want to provide a ceph set up with 3 nodes 1 mgmt server and a couple of clients, which I had planned to do through nested virt but believe that multiple vm's would be cleaner. Any suggestions as to how to achieve this? Any help or hints would be appreciated, BR/David
1229d780-01f4-054b-191b-e63a2f452466
['5d31e210-1c45-f37d-5ca7-b91c4ccabb93']
Hi everyone, I encounter a strange problem. My computer is 64bit CPU, and found the visual studio 2008 expression does not contain the 64 bit compiler, so i use the SDK v6.1 to compile the 64bit itk. I firstly open this sdk console, and run the cmake 2.8.6, in its confuguration UI, I choose to output the makefile of itk. Finally, i build all itk examples sucessfully by using "nmake" command. However, when i run the deformableRegistration15 example (the fixed image and moving images are 50 MB around) and try to output the deformation field, there is a memory allocation problem in the computation of deformation field. The program terminates. I found the size of elements it trys to allocate is 28537730, it is not large. I also remove the code of applying the transformation parameters on the moving image to obtain the registration result (namely, use the resampling function). Then, the progam is still in crash. its very strage. I think the program is not running in 64 bits. Can anyone tell me why? thank you in advance! best regards,
c9d1e3b3-c57d-82b3-eb76-63c1f6c42b1a
['5d31e210-1c45-f37d-5ca7-b91c4ccabb93']
Hi everyone, Now, I try to use the ITK3.2 to do registration work. when I use ImageRegistraion2 example to do simple translation registraion test, and found that the registration process is very very slowly. For a 2D CT slice image with size 350*270, it costs me half an hour, can anyone tell me whats the problem? The time consumption of the ImageRegistration4 is simiIar. I did not modify any code in the example. But when i use DeformableRegistraion15 to do 3D registration, the time consumption is normal. Any suggestion is appreciate. thx
1ce158fd-8066-b220-1fb8-e424da0444a4
['5d36c8b1-efa0-d8db-3f9b-e8290d5f2f42']
Dear all, I would like to crop a particular contiguous area of country polygons to a rectangular bounding box, and return new polygons, from which I can work out the centroids. I would like to do this so that I can use the centroids of the cropped country polygons to work out nice label placements for a map. I am having trouble cropping a SpatialPolygonsDataFrame object by another polygon using gIntersection (I suspect this is not for this purpose?). Can anyone help edit the following to make it work? The gIntersection command fails. I think I need to find an alternative, but I am not sure what to do yet. If anyone can suggest a command I'd be most grateful. Many thanks in advance, <PERSON>
b114566e-8f4f-22fc-64d6-77d6181e1119
['5d36c8b1-efa0-d8db-3f9b-e8290d5f2f42']
Hi <PERSON>, I think you can try spatstat::nncross. You will need to turn your sp objects into spatstat objects. So coerce your locations into a point pattern of class 'ppp', which I believe is done easily enough with maptools::as.ppp and your polygons into a spatial lines pattern of class 'psp' using maptools::as.psp. Then it should be as simple as: dists <-nncross(locations.ppp,sea.psp) HTH. <PERSON><IP_ADDRESS>nncross. You will need to turn your sp objects into spatstat objects. So coerce your locations into a point pattern of class 'ppp', which I believe is done easily enough with maptools<IP_ADDRESS>as.ppp and your polygons into a spatial lines pattern of class 'psp' using maptools<IP_ADDRESS>as.psp. Then it should be as simple as: dists <-nncross(locations.ppp,sea.psp) HTH. Simon
dfa7ec13-ffb7-1845-f6ea-087371f61cbd
['5d37cd6d-b4e0-2cd8-df4e-71596b3439f7']
Hi folks, I had some trouble getting the stackless mailing list repaired, after one message was repeated thousands of times. This was a message from October, but interestingly, the thousands of messages appeared in both the October and the November archives. Looking forther at one instance of the thousands of copies, I recognized that the message had no "Date:" entry, just a Resent-Date, for some unknown reason. After clearing the whole pipermail archive and recreating it with the bin/arch command, the message again appeared in the November archive. I took one of the fields with valid date enties and added it as "Date:" field by hand. Then the archiving worked as expected. My question: When a message happens to have no date, messages appear to be archived into the wrong archive file. There are a couple of other date fields available, like the usual "from" - line on top of the message, fields like Resent-Date or dates from the message passing through other servers. Wouldn't it make sense to do a more exhaustive search for a proper date, instead of accepting a missing date? cheers - <PERSON>
4d1845a4-7b1b-2ad3-3827-2f1ccdf05e85
['5d37cd6d-b4e0-2cd8-df4e-71596b3439f7']
Hi all, as promised, I've uploaded the slides from my Stackless talk to the stackless site. It can be downloaded here: http://www.stackless.com/#EuroSlides The page has also been updated to reflect the recent evalution. Some code concerning thread pickling will be added, soon. EuroPython was just -- great! I'm sure this will be repeated next year or even earlier. cheers - <PERSON>
8926bd59-45b4-d482-b597-532413707a15
['5d455d3a-a927-c7b7-0aba-135bd19bcd5a']
I realize that this might be problematic/devastating if you had an AI PK and did this, however in my case there is no problem as there is no related data in the database yet lol. My question is, how can I reset AI? For example I deleted several rows and now my AI starts with the AI number after the last row that was deleted. Ie. 1,2,3,4,9,10,11 when I want to start numbering at 5 not 9! Even if I insert/update with non-nulls then the sequence is still pushed ahead. Is there a way to reset this? Thanks!
51e05ab6-de90-b9d3-3616-5ff27c0e1a8a
['5d455d3a-a927-c7b7-0aba-135bd19bcd5a']
Unless I am misinformed, subscription to the DIGEST format of this list should cause all messages to come in one daily e-mail? Can someone else who is subscribed this way please at least let me know that this is at least the case for them? I continue to get every single post as an individual e-mail. I have tried canceling and re-subscribing with the DIGEST option selected, to no avail. I have also not received any response from admin. Thanks!
b4994fcd-b83d-35d8-6013-dd23867d9824
['5d5525a5-cb14-9d08-8b41-ddc571001a40']
Hi, <PERSON>, Thanks. I've updated the Turmvilla - Nova as 64bit and now I could start the Noux subsystem. i'm still facing hang issue in starting up the Linux subsystem. This time I could capture the Logs from the /shared/log file. A few questions : 1. For the wifi configuration, I've tried to configure it from within Noux under the /config/wlan_configuration. But it seems to points to WPA_SUPPLICANT file. i have configured both but not sure if it is correct as I checked the /reports for wireless state and it seems to be not working. How can I made the configuration permanent as the files seem to be overwritten everytime NOUX startup. Is there a manual on how to configure NOUX ? 2 . For the Linux subsystem, from the /shared/log file, it seems to indicate some issues with the virtualbox configuration. Writing to non-existent port. Any advise on how to move on. Regards.
9b927e9d-49fc-5545-e169-2ee5b7abde66
['5d5525a5-cb14-9d08-8b41-ddc571001a40']
Hi, With the previous Turmvilla scenario, I wish to setup two virtualbox sessions, one running LInux process and the other running Windows, concurrently. I created a new Win7 subsystem, emulating the LInux subsystem, with new vm_win7.vdi. However, I could startup only one of each sessions from NOUX, either Windows or Linux, but not both sessions at the same time. From CLI monitor, I could see that RAM is allocated separately for both sessions. I suspect some concurrency issue but could not find the issue. I've attached the log from captured from NOUX. Another issue I encountered is that the virtual box mouse pointer in the WIndows session and the Nitpicker mouse pointer can't seem to merge together unlike the Linux session where both pointers are merged. Any advise. Thanks in advance.
5caf8f81-71fa-03e6-a2ea-7ddf8e744da3
['5d6f7a9e-1173-cbbf-40a4-5e8816dff860']
Hello All A thought on printers.... We generally have two types of print jobs 1. light jobs :- are usually needed "right now" typically 1-2 pages economy is not the criteria 2. Heavy jobs :- Its OK if we get the jobs next day, done from a Print Shop in Andheri or so.. SHOULD BE ECONOMICAL and preferably managable via e-mails.. This seems to be a good business proposal and very good for us who need heavy printing jobs If anybody knows such printer peopl pl let us know ONE MORE India could become a PRINT AND POST HUB FOR THE WORLD !!!!!!!!! Any Takers ;) MAHESH DESHPANDE Trainer for Linux Installations,
77633daf-5281-b76c-d5a7-ebd600f58ed0
['5d6f7a9e-1173-cbbf-40a4-5e8816dff860']
CISSP A cissp study book is available at computer Book Shop DN Road Fort Mumbai 400 001 For CISA visit their site and enroll yourself You will start getting their regilar mails and snail-mails Also at Matunga the CISA conducts regular courses for their June exams Bookings usually starts at Nov-Dec and closes by Feb <PERSON>