query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
d67c124a-722f-00ec-381c-a9d665eded29
['17ade0de-c199-bcb7-1569-0b674218071e']
Could you all take a little more trouble with your Subject: lines? People have been throwing my original crude jest, "Mine's bigger" back in my face for the last several days, and each time I have to reconsider my ill-considered soi-disant wit, it looks more and more the way yesterday's Chinese take-out looks lurking in your refrigerator the following morning. Distinctly unappetising. We're way off the original topic, which was memory bloat. So try and choose appropriate subject lines. -Olin
4d8f53bb-4e5a-5a9c-226c-41662813c2da
['17ade0de-c199-bcb7-1569-0b674218071e']
According to <PERSON>, on Solaris 2.5, fileno() is a function, not a macro. This requires you to hack about two lines of scsh source to interface. I was under the impression that scsh built fine on Solaris, no mods required. Is Solaris 2.5 different from other releases? Could someone explain this to me? This sort of thing is why I hate - C - OS's without source code ...and, thanks, <PERSON>. -Olin
eef0b8df-3f2f-e7e8-8317-372c4db5b917
['17ceb9fc-b34d-188a-b730-6f9261e70d3c']
Hi again. As some of you may have gathered, I have a really crappy Internet connection. I was wondering if there are any tools that a relative newbie can use that produces some output that I can shake in the face of my cable representative and tell him to shape up? Changing ISPs isn't possible as there is only one cable Internet provider in Jakarta that is supposed to maintain servers for the various ISPs that provide Internet access through their system. I am getting all sorts of time-outs, loss of packets, etc. Sometimes the connection is really slow or simply kicks off, but for some reason it usually doesn't prompt a cable modem reboot. Any ideas on tools (that run in the background) that will help me to document this would be helpfull. I am using Warty 4.10. Cheers, <PERSON>
26129e63-4980-3c1a-774f-3a127dea5d9e
['17ceb9fc-b34d-188a-b730-6f9261e70d3c']
I would like to install stardict-2.4.4-1.i386.rpm and some noarch.rpm dictionary files into my Warty 4.10, but I haven't been able to find any documentation that I would understand as a newbie. I get the sudo alien -i stardict*, but how do I check if an install of stardict-2.4.4-1.i386.rpm will even work before I do a real install that would also include the noarch.rpm files. Am I just nervous and making this out to be a bigger problem than it is? I just don't want to break anything in the system. Cheers, <PERSON>
1cbf6297-ccc0-c4d5-aee1-882f5d73115b
['17e8044a-a560-a7a5-9f9d-e73e876ded1b']
Hello, I just read this blog https://www.yoctoproject.org/blogs/belenbarrospena/2013/saving-your-hob-customised-builds-image-recipes , which explains how to save your hob customized images . However when it finished to build my image I can't save it because there's not Save Image recipe button. How I can save it? Why I can't see this button? <PERSON>. de Control Industrial HORUS HARDWARE S.A.
9a5bef12-b764-3ad9-8430-8be6ae5c95ff
['17e8044a-a560-a7a5-9f9d-e73e876ded1b']
Hello everybody, I'm trying to test FlexCAN on Sabrelite imx6 board without success. Those are the steps followed : 1) Comprobed that linux-imx has CAN drivers enable 2) Added on the image the packages:canutils,libsocketcan and iproute2 3) Tested the Flex can with canconfig and cantest. Everything seems ok, even I can see transmitted frames incremented. However there's not singnal on the oscilloscope. I don't now why is not working. Any ideas or suggestions? Thank you!
cc1d78c4-d4f9-c71b-ce59-a910e7231e0f
['1803df65-efc6-cfa2-4a22-d89ba504657e']
Hello, Currently we have very few global principles for the Wikimeian projects, namely the GFDL principle and maybe the NPOV principle. We have many recommendations listed on the Meta, which are not taken seriously in many projects. Wikimedia projects have grown tremendously, and in my opinion, it has become crucial that the list of principles governing all projects be a little more detailed. I suggest a few basic recommendations become obligatory. For example, [[meta:Polls are evil]] should become mandatory as a guideline to all projects. This article is very important as it put into practice the NPOV principle as well as the desicion-by-consensus and the differentiation between facts and views principles. I suppose there are other recommendation of this kind that should become mandatory guidelines. I am not suggesting a constitution or a full rule book. I do suggest to carefully single out several basic principles, since the projects' autonomy is a bit too wide, especially as we want to promote cross-contributions between projects. <PERSON>
c6d42eb2-d088-226f-d5c2-4b0a750b49ec
['1803df65-efc6-cfa2-4a22-d89ba504657e']
Hi, As I said, we are not talking about a set of bureaucratic rules nor about a "constitution". I also agree that guidelines should be different for different set of projects, i.e. the Wikipedia guidelines won't be the same as the Wikinews guidelines. We also have several global rules that are already mandatory, most notably the GFDL rule, which is also the easiest to enforce. I have an experience in writing in several Wikimedia projects in different languages. People often confuse the Wikimedia principles with the English Wikipedia principles (I saw this confusion here too), and due to a (strange) resentment towards the English Wikipedia, they reject any recommendation published on the Meta claiming it is just another idea of an American from the English Wikipedia. We should stress that there are certain principles that must be shared by all Wikipedias (or by all Wikinews projects etc.), first beacuase these principles makes Wikipedia an open and reliable source, and second because we wish to encourage cross-contributions, i.e. that contributors will be able to easily write in several projects, and that articles could be easily translated and adopted in other projects. Here are some allegedly global rules, which are unclear: 1. Original research - what exactly does it include? Can a certain Wikipedia redefine the notion of "state" and state: X is a state according to its own criteria? 2. NPOV - Can the Dutch Wikipedia (for example) decide it defines a state according to the EU recognition, and state X is a state only according to EU diplomatic recognition, or is it a violation of NPOV? 3. Polling - can a Wikipedian community vote about excluding a certain person from writing? Can it vote about whether to except a certain source? Can the Spanish Wikipedia decide a source from Japan will never be regarded reliable? These are only a few questions that are not properly addressed, and the system doesn't work well anymore. This kind of problems rise more often as more people write in Wikipedia and as the communities grow, and we fail to supply answers and solutions. <PERSON>
a8c918ed-dfc2-dd7e-0c87-e37b5943d590
['18076470-fb24-cfaf-7627-fe94fc08f743']
Hi ... I have a tiff image file of 16353 rows and 40320 which are uniformly gridded. I want to count the pixels of each type in a gaussian rectangle with pixels which lie on the border be counted as the fraction that is inside the Gaussian grid. Is there a builtin function to do this in pil. Is the histogram overloaded to count fraction of grid bos inside a given rectangle and so on.... Please help <PERSON>
16b01bfe-d5fd-7ee7-eba8-499fa35b1509
['18076470-fb24-cfaf-7627-fe94fc08f743']
Hi The problem that i am facing now is the followng.. I have an image that has pixel values that range from 0 to 23 But when I take a peice( subset ) of this image interior=im.crop((2890,79851,473,3202)) and give interior.getextrema() it shows that the interior has values from (0,255). interior is supposed to be a subset of the original image how does it have more pixel values than 0 to 23. Can you please help Manik
6f00e9ca-e6b7-cf95-9a75-76ad0e18e828
['180a1ee3-bd4c-f646-0e12-b4e1a9a90dd5']
Dear reader, I want to use liquidsoap to stream multiple stream to te web by means of one server running one soap process. When streaming with fallback. is liquidsoap re-encoding these streams or are they just past through. if they are re-encoded is it possible to enforce only a passthrough to the output mount? so no encoding of the audio streams on the server itself wil be done. With Kind Regards, <PERSON>
cac93351-3e9f-a928-8ba5-d41401c1defc
['180a1ee3-bd4c-f646-0e12-b4e1a9a90dd5']
Dear readers, I am currently building a replacement platform to broadcast radio over the web in the Netherlands. We are currently building the proof of concept version of liquidsoap within a cluster based environment. There will run multiple liquidsoap processes to stream the audio after it has been encoded by the encoders to share the load. To controle the cluster we are building a NodeJS application to share configuration bij means of http calls and returning json to liquidsoap. Now we are tying to use the node server through ipv6, however that failed. After looking at the source we saw that there was only IPV4 support. Now is there a plan to include IPV6 support into liquidsoap on the http.post and http.get methodes? And are listeners able to listen on IPV6? Kind regards, <PERSON>
e7d0ee43-44d5-1068-1722-058e4a902615
['181bb91a-4d98-1a6b-393c-3841a6545ee9']
Hi All. Greetings from South Africa. I have a Samba LDAP server (v 3.022) running on Ubuntu 6.10 Its serving about 200 users, with profiles, and domain logons. I want to start server MS Access Databases on it, with the best speed performance as possible. At the moment, the back ends for these databases, are about 200+ MB, and will grow over the next few years. Basically, the share should serve about 4 users, with read/write access. I am using the XFS file system, with ACL support. Has anyone setup such shares in smb.conf? I would really like to see an example. Lastly, I do not think I want to use oplocks. Any help, will be greatly appreciated. God bless.
68c90d5f-9530-ded7-7a8f-50309c7a6885
['181bb91a-4d98-1a6b-393c-3841a6545ee9']
Hi All. Greetings from South Africa. I have a Samba LDAP server (v 3.022) running on Ubuntu 6.10 Its serving about 200 users, with profiles, and domain logons. I want to start server MS Access Databases on it, with the best speed performance as possible. At the moment, the back ends for these databases, are about 200+ MB, and will grow over the next few years. Basically, the share should serve about 4 users, with read/write access. I am using the XFS file system, with ACL support. Has anyone setup such shares in smb.conf? I would really like to see an example. Lastly, I do not think I want to use oplocks. Any help, will be greatly appreciated. God bless.
9b4efd9f-0ea9-cd00-169e-35b480ec47f7
['181c8ca5-35cd-6532-0a22-65fc08832415']
Hi, everyone, I'd like to set it up so that every process that a particular user sets up is niced to 5. I can do this with renice, but the process has to be repeated every time the user logs in. The only solution that searching the Internet has yielded is sticking that command in the user's .bash_login but most of the results are from many years ago. Is there another way of doing this? I'll do this if there's no better way but it seems a bit kludgy to me. Regards,
bf25fa26-84e0-00d2-488f-6f1e297b686f
['181c8ca5-35cd-6532-0a22-65fc08832415']
Hello, everyone, I'm looking to buy an IR receiver that works easily on Linux. I know that it's possible to build one, but I'm really short on time here (I'm leaving for elsewhere soon and I want to make a HTPC for my parents/brother). Also, how does this sort of thing work? I have a Radioshack Universal Remote. Does the receiver need to be built for a specific remote or will it pick up the stuff and do the mapping from button-press-received to keypress in software? In the latter case, I can probably have the Radioshack remote configured correctly rather easily. The less complicated, the better. Any help is appreciated. What I've found so far is that lots of TV tuner cards come with an IR remote and a receiver. However, no one seems to post on the Internet about whether the remote and receiver work with the card or just the TV-tuner part. I could probably pick up an Intex TV Tuner card and try to get it to work, but I haven't used a TV Tuner on Linux in years (last was a couple of years ago with bttv and a Pinnacle PCTV 50i Pro, I think) and don't really know how these things work. By the way, I'm thinking of running XBMC because it looks way better. The output will be an ancient CRT TV for now, but my brother can probably work things out once I have everything set up so that if they buy a newer LCD TV, things will be fine. Any opinions on that are welcome as well. I have little experience with this, but hopefully things will go smoothly. Regards,
867733a5-e50e-d679-1fb4-f37ba695179e
['181e005f-2ba3-7ef8-f15c-304127719923']
Dear Folks, We have been informed of the very distinct possibility of having to close 2 of our 3 branches. I know that this is not uncommon these days, so I'm trying to be as proactive(?) as possible. If you have had to deal with this, I have some questions. 1. How did you deal with the materials in the building? Did you treat it as off site storage? Did you cherry-pick the materials and sell the rest? OR if you kept the materials 2. How did you deal with the records in the catalog for those items if you decided to keep them off-site? Do you go once a week and pull for holds? Did you run a script to have them "Disappear" from the OPAC? 3. Since there is now a distinct possibility I will be RIFed at the end of June, I have discovered that I am irrelevant as a cataloger. The few cataloging jobs that I saw required metadata experience, which I don't have, as well as a whole slew of other things. I have tried, in the last couple of years, to prepare for this inevitability, but where can I get a class on metadata, xml, etc. People just keep telling me that I already know, and I just don't realize it, but that is not helping me. Is there a "For dummies" metadata book out there? Is there a class? Am I going to have to go back to library school for a semester? 4. Are there really people out there that catalog out of their homes, or is that an urban myth? Thanks,
29f785c3-e396-e906-6d29-57a61786460a
['181e005f-2ba3-7ef8-f15c-304127719923']
Dear Folks, Coming from a small system, I am pretty lucky. I made a new years resolution that I would not work more than 200 hrs a month. Sometimes, I make it sometimes I don't. I average somerwhere between 45 and 55 hrs a week, and I try to draw the line at 50, but you need to know that my director is here even more hours than I am. All of us "librarians" are exempt, so we look at the overtime as "points in heaven." Now with that said, when I need to run an errand or if I have family visiting, my director will say, "You have your 40 in, don't worry about the leave sheet, just go." We tend to look after each other here, and we cover for each other at holidays and vacations. My new, New Year's resolution, is to leave before dark, especially in the summer. Cheers, <PERSON>
b0a673a6-0eb4-75b6-ede4-9b498586774e
['18249bf1-d004-bcff-fbb9-122b0f6c1285']
I have a situation where JobA triggers JobB which triggers JobC - upstream job builds something which the downstream jobs use. If JobB is in the process of building, then the code base for JobA changes and triggers another build of JobA, I would want to automatically abandon the currenly building JobB and remove JobC from the queue. Does anyone know how this can be done? Thanks
d016e284-2b1d-4ab0-5ad2-fad0b771fad9
['18249bf1-d004-bcff-fbb9-122b0f6c1285']
Has anyone got a puppetmaster to populate a remote oracle server? The only thing I have found is http://www.djelibeybi.net/2010/08/using-oracle-database-for-puppet-stored-configuration/ I have installed the prerequisties, yet I when I run: puppet master --no-daemonize --verbose --debug I get no indication of any database activity and I get no tables created in my database when a client asks for it's catalogue Any ideas Thanks
ff722258-4f26-cd87-59cb-c75941fbdf48
['1830b2f9-53ab-8eb2-72c3-f58da5cd311c']
I'm so sorry for my stupid request but i haven't found anything with google Could you tell me how to mont my NTFS partition && EXTENDEDDOS where on my fdisk wd0 07 /HPFS/QNX/AUX wd0a A6 / when ui'm tryng to mount -t ntfs /dev/wd0 /mnt it says exec /usr/sbin/mount_ntfs for /mnt no such file or directory PLS helpme thanks in advanced
351cca74-79a5-198c-c742-89080111ef96
['1830b2f9-53ab-8eb2-72c3-f58da5cd311c']
HI <PERSON> Does anybody of you got the pcmcia dfe-380txd working? What kind of kernel do you got whic driver do i need,how can i compile the driver if it says problem with version.h how can i escape it, How can i resolve my problem with 3ccfe575BT i got the source but i'm not able to compile it cause it says version.h problem is there any body who can help me in a step by step my kernel is 2.2.20-ide pci Thanks in advanced
811e8265-525e-7158-b3cf-296204e0b44c
['18328089-a009-1c22-c5f3-dd24e8d1063f']
Hello! Solr (i.e. Tika) throws a "zip bomb" exception with certain documents we have in our Sharepoint system. I have used the tika-app.jar directly to extract the document in question and it does _not_ throw an exception and extract the contents just fine. So it would seem Solr is doing something different than a Tika standalone installation. After some Googling, I found out that Solr uses its custom HtmlMapper (MostlyPassthroughHtmlMapper) which passes through all elements in the HTML document to Tika. As Tika limits nested elements to 100, this causes Tika to throw an exception: Suspected zip bomb: 100 levels of XML element nesting. This is metioned in TIKA-2091 (https://issues.apache.org/jira/browse/TIKA-2091?focusedCommentId=15514131&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-15514131). The "solution" is to use Tika's default parsing/mapping mechanism but no details have been provided on how to configure this at Solr. I'm hoping some folks here have the knowledge on how to configure Solr to effectively by-pass its built in MostlyPassthroughHtmlMapper and use Tika's implementation. Thank you! <PERSON>
7738201a-2da3-5b59-c61b-1d08b1be1ae1
['18328089-a009-1c22-c5f3-dd24e8d1063f']
Hello! I'm sending a HTML document to Solr and Tika is throwing the "Zip bomb detected!" exception back. Looks like Tika has an arbitrary limit of 100 level of XML element nesting (https://github.com/apache/tika/blob/9130bbc1fa6d69419b2ad294917260d6b1cced08/tika-core/src/main/java/org/apache/tika/sax/SecureContentHandler.java#L72-L75). Luckily, the variable (maxDepth) does have a public setter function but I am not sure if it's possible to set this at Solr. Is it possible? If so, how would I set the value of maxDepth to a higher number? Thanks!
2181c23b-aec3-6f72-44b6-b3e05064deb0
['1835445e-3356-3a9d-5f5f-0a505e065753']
Pointlist to Bezier Hi there I wonder if any have code to take a point-list and turn it into a smooth-lined bezier. There are explanations around the net for C-code but it is very hard to understand. (one here said to be good but misses the graphics:) http://www.benknowscode.com/2012/09/path-interpolation-using-cubic-bezier_9742.html What could this code be used for. — Tracing an image — Smoothing the lines drawn by a user plus a lot more Im sure Thanks <PERSON>
7d2cf5fd-473f-035a-f2f9-ae0ee339a7fd
['1835445e-3356-3a9d-5f5f-0a505e065753']
Hi there Icons can be quite space-saving and also pleasing to the eye. But somtimes the hazzle of making the icon in a paint app and import it into Livecode and keep it in a hidden place is to much. On a Mac there is a palette calle "Keyboard". It can show the thousands of chars/symbols in a Unicode font. Im sure you can find a symbol there that suits you. The trick is to add the char to the label of the button, not the name. Scale the char/icon by ajusting the textsize and color it as you please. <PERSON>
c15b4f8e-7147-3fda-f210-ebf40709c35c
['183c2787-c399-3320-9abf-3f0835197d9c']
I just wrote a method with an random generator to issue the secuirty context token. After that I stored the token in a database and then I compare the received sctoken from the client with the token in the database via a callbackhandler. I think that does not make sense or it is wrong to use a policy with a supporting token "secureconversation" without a tokenstore. Am I right? In my opinion, I have two options. 1) Apply a STS, which uses all functionalities of the apache cxf to use a policy with WS-SecureConversation 2) Remove the policy of WS-SecureConversation. I do not prefer option two. Regards, <PERSON>
53f1b67d-d845-0c97-a028-15890af5bdd8
['183c2787-c399-3320-9abf-3f0835197d9c']
A collegue want to consume my services with a .NET client. He cannot use a STSClient. I do not know if it is possible with .NET. Honestly, I am not a expert with .NET. He receive always the error from my first post and therfore I want to prove that my implementation is also working without a STSClient on the client side. In my opinion I think my implemetation also work without a STSClient because the tests with SOAPUI show it. One step further, I planned to implement a java client to show him that it is possible working without an STSClient and my implementation. But in the end I think it is not necessary to implement a client. I hope, that my assumptions are correct. Regards, <PERSON>
52534cbd-b905-379e-fcdb-c75b18ea6288
['184975bc-10e1-ddc9-4b3e-fd868a5022d0']
1. I've just registered BBEdit. No problem with the registration, the program loads OK without the nag screen. But if I hit the blue "i" button, it tells me my demo has expired, and won't take the rego code. Why is this so? It doesn't seem to affect normal operation, just the "i" button Please don't tell me it's a bug. 2. Is it possible to get the Line numbers to stay selected? I can turn them on, but each time I open BBEdit, the line numbers are turned off again. This is really irritating. How do I fix it? DN
15e33a88-7da0-3967-35c7-568337765981
['184975bc-10e1-ddc9-4b3e-fd868a5022d0']
I want to edit some application template files that are embedded down two "package" levels - the packages and the files in question are owned by root. And I want to be able to write safe copies of the originals to the same folder. Is there a simple way to do this in BBEdit? ie a "sudo" equivalent? I'm still learning how this stuff works. DN
c18b4d23-17b7-61af-3e3c-344562b79522
['184ee02a-9d11-870d-81e7-8887f0e0f2a5']
here's some (very low quality) pictures of Opie running on a Cassiopeia E-105 there are some problems, these are probally the biggest problems, - Power saving, its seems to be there and doing something, but i cant disable the backlight/screen or goto hibernate/suspend - Audio, i dunno if the driver in the kernel are even for the casio (the source says its for a few other pdas so i dont think that will work) im using the kernel from http://linux-vr.org its a bit old (2.4.0), there's also a kernel at http://linux-mips.sourceforge.net it has support for casio e15 and casio e55, ive tried this kernel but couldnt get the ide driver and framebuffer working.. i'll try to make some docs available at http://opie.handhelds.org/wiki/ and maybe put some binaries online too, if anyone is intrested.. (kernel + rootfs with opie, should fit on a 32mb cf-card) <PERSON>
51fe4477-6af9-f4f1-49b7-7c6eee7a8ac1
['184ee02a-9d11-870d-81e7-8887f0e0f2a5']
hello all, i seem to have a little problem with software suspend 2 (2.1 with kernel 2.6.9), after resume i lose the video MTRR causing programs like mplayer to use twice as much cpu for playing back video, thanks to <PERSON>_ on #swsusp i was able to figure out that the mtrr entry was gone after a resume, and causing the higher cpu usage, simple echo'ing the entry into /proc/mtrr fixes it (or restarting X also solves it), but i wonder why its gone after resume (possibly some bug in the suspend/resume code?) i put some logs online at http://www.xs4all.nl/~dopez/logs/losing-mtrrs.tar.bz2 (about 26kb tarred) if anyone needs more info please let me know please cc me any replies as im not (yet) subscribed to the list Greetings, <PERSON>
10e83f7b-9460-b6e3-c045-2b685af9d384
['18538664-d6e9-328a-cd4f-8a2711b2b892']
Hello, i use A iSCSI IBM DS3300 bay to store virtual machines. I want to have a 'ready to go' backup on another iSCSI target (NAS QNAP). I would now is there s a way to copy image from one LVM to another one, i tried to use rsync to make it running, but rsync can't replicate special device ! Using dd is not a good idea as it will copy all datas Is there already a way to do this ? Regards Serge
ad3254b1-16fb-4b99-902c-268a2dca5fc4
['18538664-d6e9-328a-cd4f-8a2711b2b892']
Hello, this patch allow to copy a disk from a storage to another (without moving it), it's goal is to replace backup procedures when someone has dual bay store. When we use backup provided from preliminary Proxmox, you have to save (with compression), and to restore to decompress file generated, wich can take some time. In our configuration, we have a disk bay (DS3300) and a NAS. By copying directly disk from the bay to the NAS, we have just to change target drive in case of bay failure. I hope this patch will be present in futures release of Proxmox. Please <PERSON>, answser me on this particular point. Regards Serge NOEL
5e764e15-d157-84f6-0961-3baf97c1cae8
['185caed9-5636-13b0-8ebd-4b57720cb002']
I have been looking at this and I can't figure it out. I am using qmail,tmda, and tmda-cgi. The problem lies in the fact that I have an outgoing program(qfilter-tmda) on qmail as done by someone in the mailing list. It seems that the only way I can get the outgoing program to work is by specifying crypt_key_file or the datadir. Either way seems to work. However, when I specify either, I can't have TMDA-CGI create new users! When I comment out either of those options, I am able to auto install just fine. I need the outgoing program so the mail is tagged properly however I need tmda-cgi for the users. The tmda error will then give Permission denied:/root/.tmda/crypt_key . Creating that file doesn't do anything. For some reason it needs to be specified in the tmdarc. Thanks
54e4d782-2528-c605-7fab-c57fe78b88d5
['185caed9-5636-13b0-8ebd-4b57720cb002']
This is actually more of a "qmail" question than tmda. The way qmail is usually set up by default is to NOT allow relaying from anybody. That means that if the address you are sending mail to isn't mail that it accepts, it will give you the error. For example, qmail is setup to receive mail for rocks.com. Rocks.com would be placed into rcpthosts. Qmail will ONLY accept messages that are addressed to <EMAIL_ADDRESS>. I think you are probably trying to either use your company qmail server or an ISP qmail server for outgoing mail. There are numerous things you can do. Some of them you would have to look up documentation. Either way, you will have to have control of the qmail server. If it is a company server, you can edit a file called tcp.smtp. (mine is in /etc/tcp.smtp) and add something like 192.168.0:allow,RELAYCLIENT="", The relayclient is important. However, don't put RELAYCLIENT on your main address otherwise everyone from the internet will be able to use it. You can also find a SMTP AUTH program that will allow users to login and relay messages. "Paul"
49226531-2503-6575-3fa1-159ebcb25a72
['18682ee4-b2aa-6b6a-1a61-4c6eb390ce6d']
Hi, everyone. I would like to know if it's possible to get rid of the "Enter PGP passphrase:" dialog when sending signed/encrypted emails. The thing is that use a GPG key _without_ a passphrase; nevertheless, mutt asks for it every time, and I have to press the Enter key in order to send the email. By the way, I source the file provided in the "contrib" directory (gpg.rc). Thank you.
ca3287f9-9ca0-dc6f-49ef-f97a497c5c68
['18682ee4-b2aa-6b6a-1a61-4c6eb390ce6d']
Hi, I've been an i3 user for a few months now, and I just began using i3lock instead of slimlock[1]. There's just one thing that kind of bothers me, though: with slimlock it's possible to use an option (called 'tty_lock') which disables switching to another virtual terminal while the screen is locked (using VT_LOCKSWITCH). I think this is a nice "security feature", if you wish. So, I was wondering: first if this would be desirable in i3lock, and in that case if it could be possible to add a similar option. It doesn't seem to be _that_ difficult to implement, but unfortunately I don't know C, so submitting a patch is out of the question. Thank you.
6253b379-01a7-97ae-6f4b-cc5b68122ef8
['186c3ead-4d9a-e1be-b597-7062ba6efb4f']
Hello <PERSON>, What do you mean when you say that you are putting in a bogus login? It seems to me like the process is working correctly. You put in your login information at the login screen, then EZproxy checks that info against your domain, then it goes to the III database with the same info. You aren't prompted for a second login attempt after it searches your domain to which you then put in your III login info. What you need to do is put in a valid ID and password for your III system and see if you get in. If you are putting in valid III patron login info and you are still seeing the itype.htm file then I'm not sure what is going wrong. -Matthew
7ea2cc4a-ba93-ddf6-c7a6-53775f410bcf
['186c3ead-4d9a-e1be-b597-7062ba6efb4f']
I'm having a problem with some of our links to 'PROLA - Physical Review Online Archive' from some 'APS - American Physical Society' journal pages. For example, if you are trying to get to issues of Reviews of Modern Physics from this URL: http://rmp.aps.org/ and you use the drop down menu under 'Available Volumes' and select 'Earlier Volumes (PROLA)' and click on the 'Go' button the proxy is dropped out of the picture. So it seems related to the 'Go' button on the APS pages. If using 'AIP - American Institute of Physics' to link to PROLA for the same journal there is no problem at all. So it seems to be a problem with the APS pages. Has anybody seen this problem before? I saw some talk on the list about APS / AIP and scitation.aip.org and just upgraded to EZproxy 3.0b, but that did not solve the problem. We use serial solutions and EZproxy is running on a Debian 3.0 Linux server. Ultimately I can reroute remote patrons to the resources they need but that still leaves a fair number of links from our electronic journal pages that don't work remotely. Any help would be greatly appreciated. -Matthew Gailey
68ce16ff-724b-6159-411b-f30e30855af6
['1878637a-4956-3ca4-18c2-091b3beee69b']
Does IWAM_MACHINENAME have access to the Oracle client libraries? I've seen recently where some versions of Oracle client will install on some OSs with very limited permissions. In addition, bear in mind that your security principal may write client trace files on various directories of the machine under various circumstances (look for files that end in ".ora") and may actually require write permissions in this case. An easy test of this is to change the anonymous web user to run as you (***only*** as a test and not on a live machine) to see if this fixes the problem. Hope this helps, <PERSON>
a28624aa-1062-ecdd-f473-f6dd074c4b82
['1878637a-4956-3ca4-18c2-091b3beee69b']
Please post this to the appropriate new list (Webforms) in future...thanks. That's three different things. 1. Call a stored procedure from a web app. 2. Insert a row through a stored procedure. 3. Retrieve a refcursor from a stored procedure. Which data provider are you using with Oracle? And, if its the OleDb data provider, which OLE DB provider? I have an example of using refcursors with the OleDb data provider (using MSADAORA or OraOLEDB.Oracle OLE DB providers) posted on my website http://staff.develop.com/bobb. I have a series of Oracle examples that need to be cleaned up for various other data providers for Oracle, which one are you using? The "fill a datagrid" part works the same way for Oracle resultsets as SQL Server resultsets. You can set the DataSource to the DataReader desired. Or set DataSource to the DataSet, DataTable or DataView. If you set the DataSource to a DataSet, you can choose the correct table by using the DataMember property. Hope this helps, <PERSON>
1a1f4d08-d248-6231-1efb-42aacfb5b227
['18960c7a-c02b-f370-c1ec-bf5d17e5b497']
I've finished working on this as part of a project, and I would like to give my feedback for others to discuss. I saw this brand new module as potentially ground-breaking - a feature that would turn general opinion on WordPress away from 'blogging platform' to 'full-blown CMS'. It gives you scope to do virtually anything with this software, from a mini-portfolio to a major corporate website (I know, I've done both). You aren't limited with the opportunities within this either, whether your portal concerns real estate, shopping, music or television. Perhaps with a little SSL integration and a strong server configuration, you can create your own Amazon. However, this defining invention in the WordPress package isn't good enough. Firstly, it's buried deep in the code where only one of two types of people can freely access it. If you're not an advanced PHP developer, then you need to be equally as good at copy-pasting from the few decent tutorials around on this subject or being adept at searching for plugins. If you're one of those that needs to learn by doing, these last two options aren't really worth trying. In addition, the information the Codex gave... didn't seem to even work. I imagined this whole process to be a lot like <PERSON> plugin - an option tucked away in the Settings panel that creates new top-level nav items for each post type / taxonomy. Free cataloguing of these items, but with the additional intuitiveness (that nobody has done yet) added by widgets that easily display the results on any page or sidebar you want. Would it be unreasonable to ask for something like this as an addition to the next gold release of WP? The problem was explaining how to use this feature to idiots like myself. This was not easy to use at all, and it should be. What are everyone elses' opinions? <PERSON>
e0992d82-1973-0365-c0bd-d190b05e4fba
['18960c7a-c02b-f370-c1ec-bf5d17e5b497']
I already use http://gist.github.com for code snippets. Just wondering how this would add value and improve what already exists (i.e.: codex). Here is my suggestions run a system on the site that will scrape for code snippets and suggest to moderator to place them in appropriate places in Codex. If I am stuck with something wp related, I'd search codex and that's where I want those optimal code snippets as an example. Whatever system you use should be integrated with this mailing list or wp.org account, I am not keen on creating another account for yet another site. Kind Regards, <PERSON>
a74da38b-a991-6065-5865-575600417b9e
['189fe7f3-6f96-aa8e-18af-800d46cc2c77']
Hi folks, I'm excerpting this from an ad: Thought Id share something with ya'll. Great fun and it actually works. Met my girl on here www… and so far it's been the best year of my life! Wish you all the same happiness. Thought, met, wish have their subject pronoun removed. I wonder when it's possible for the English subjects to be left out. Is it in colloquial speech? When is it acceptable? Un saludo, <PERSON>
588334ae-553f-27e5-2db5-49641d5ac40f
['189fe7f3-6f96-aa8e-18af-800d46cc2c77']
Hi folks, Dios juega con los hombres. 1) When referring to the mankind, we tend to use in Spanish los hombres regardless of being men or women. This is not politically correct, besides being extremely disregarding for the women. The question I want to pose is: According to the sense mentioned above (hombres as género humano, say, the sum of men and women) which of the following options do you prefer best? God plays with the men? God plays with the mankind? God plays with the humanity? <PERSON> plays with the human gender? 2) In English is there one of the previous options too male chauvinistic? Un saludo, Raúl Martín Díaz
561cc6d3-1cc8-9514-7d13-760a752b1169
['18aa329c-a186-767c-2b03-1d1805b480da']
sorry <PERSON> can't quote when I'm at the library...:/ You are right that wavelets might be useful in different places. My goal is to get a handle on some classes that transform Images and/or rasters (RGB triplets?) using wavelets and let people see if they might be useful. :) sorry about that 'window manager' crack...:) I'll write a more detailed note about what functions I think wavelets help with in the next few days. Arrow <PERSON>
2c668cea-b796-f8e8-ca23-119a8ffee5ec
['18aa329c-a186-767c-2b03-1d1805b480da']
<PERSON> suggested that I throw these ideas out here on -Drawing. I'm currently working on some classes to do image manipulation and filtering based on wavelets. Aside from just compression, wavelets can do may useful operations such as rescale, smoothing, and stretching. What I see is a bunch of classes and methods that can give Berlin fast and advanced rendering of pixels when needed. Some uses would be to display a very high resoultion image on a low res screen or viceversa. Some of these functiions are the same as you would find in ImageMagic and others would be geared to quickly handling operations generally associated with a window manager. Oh...and the kit would add native support for JPEG2000 and a wavelet version of PNG when we get around to that. I'd like to hear peoples ideas on this. thanks
6bef87f5-5427-3800-6023-86020c8f4e2d
['18b062c1-f517-fcc8-204b-3ea767c3dde8']
When I specify the SSL_VERIFY_FAIL_IF_NO_PEER_CERT flag to SSL_CTX_set_verify, it has the intended effect if I set it on the server side; a client not presenting a cert is rejected. Setting this on the client side does not appear to have the same effect; a server that does not present a cert is still allowed to connect, so long as ADH ciphersuites are enabled. Looking through the code, s3_srvr.c has code that does this checking, whereas s3_clnt.c lacks it. Should a client side SSL_CTX understand/implement the FAIL_IF_NO_PEER_CERT flag? <PERSON>
9e958448-d1f0-49d2-0e47-b51182e71cef
['18b062c1-f517-fcc8-204b-3ea767c3dde8']
In both 0.9.6c and 0.9.7-SNAP, the function "ssl_verify_alarm_type" in ssl/s3_both.c seems to be missing the newest x509 verification error codes, like X509_V_ERR_INVALID_PURPOSE, in the switch statment. If such a verification error is encountered, the switch will fall through and an "unknown ca" alert (SSL_AD_CERTIFICATE_UNKNOWN) will be returned, instead of SSL_AD_UNSUPPORTED_CERTIFICATE in this case. I can submit a patch, if anyone is interested. <PERSON>
aabe4490-c557-2574-bf97-f0c9d1290d01
['18c1e9a8-f331-ab2f-9d2c-4d6b9094b150']
Hi, I have the following scenario: We currently have 1 incoming line, that 2 POT phones plug into, and when we have an incoming call, both phones ring. Is it possible to have Asterisks in parallel, so that when the 2 POT phones ring, I can have a Voip phone, which "is" plugged in and configured through Asterisk also ring (so that all three of them are ringing), and then someone can then choose which phone they want to answer? <PERSON>
87ff4a97-ae37-2cc1-a825-bddd6d802c58
['18c1e9a8-f331-ab2f-9d2c-4d6b9094b150']
Hi, I have the following scenario. I have an Asterisk server running on an internal IP address behind a firewall, and I have a remote user trying to connect to my Asterisk box behind his firewall, but he can't seem to get a connection. I have opened up the port (5060) so that he can connect through my firewall, but it still doesn't appear to want to connect. I am pretty sure that the firewall rules are correct, because I have also open up port 21, and he can successfully ssh into my Asterisk box. Any ideas/pointers? Thanks in advance <PERSON>
c9e9ef31-d433-53f1-f406-1465e41bd88a
['18c62d8e-7367-95f0-866f-05aa1aa53d72']
Hi <PERSON>, I'm evaluating GWt as an HTML5 game development tool. I've looked over the JSNI stuff and that covers oen of my questions-- using or writing javascript libraries. My second question is how to create libraires in GWT for other GWT projects to use? In the JavaVM I would generate a JAR containing the library classe.s What is the equivalent in GWT/ Is there soem thing in the docs about this I haven't stumbled on yet? Thanks <PERSON>
57d4d798-1c0c-2f0f-c850-f784fb54775a
['18c62d8e-7367-95f0-866f-05aa1aa53d72']
Okay, so here's the ela. This has happend to me a few times. In all case its been when MonoDevelop barfs and dies. I seem to be reducing the number of these occurrences by being VERY mindful of the synchronization messages when i return to monodevelop after being in Xcode and don't even breath on it until its settled down. When such a MD death occurs, the only way to get MD and Code properly back in sync is to close them both, re open Mono Develop, and reopen the GUI editor from MonoDevelop. This loses the last set of changes in XCode but gets the two back in sync and talking to each other. I hope that helps you iron out the issues. <PERSON>
ca8d58a4-e2fd-f88e-6e58-547f59147b8c
['18cc0ce0-29bf-04b5-8915-9912e910d6b3']
Concerning my previous question, neither a non-store based dgrid nor a store-based one maintain the selection when I apply a sort. A non-store grid however doesn't have a problem with selecting duplicate valued rows. I think you focus on explaining why it behaves like this whereas I am focusing on the effect and I think that the effect is erroneous simply because I am instructing my dgrid to always single select and it doesn't. Therefore this leads me to the conclusion that there might be some kind of design inadequacy. Maybe the selection information is highly coupled to the model, i.e. the store. Maybe I am wrong but please spend some time on thinking if I have the slightest point on this matter. Thank you.
826aac2c-a10f-fa63-5fc2-4f40b2d587a5
['18cc0ce0-29bf-04b5-8915-9912e910d6b3']
Hello, is there a roadmap or any future plans for dgrid support of grouping rows into collapsible sections? I imagine it as having an area above the header where you can drag & drop a header cell to. Then the grid will automatically group the data based on that column. Apart from the apparent reasons I am asking, my need originates from the fact that I am replacing some proprietary grid controls with dgrid and these proprietary controls, despite the fact that they are old and suck, unfortunately do support this feature. Thank you.
1237a258-fce5-1cce-f1f8-5f8f52a511b5
['18d9b57b-c5e2-427e-83f6-1b564718cebd']
Hi everyone, Today I propose you 4 small patches which I needed for a client. I generated them with git which explains the numbers in the patches names. These four patches do the following : - Add of a french translation for a preference setting (default addressbook) - Add of a user's preference to be able to display a contact's lastname before his firstname in autocompletion when composing an email. - Small bug fix where a contact with an empty email was displayed at autocompletion - Add of a user's preference to be able to sort contacts on their lastname in emails autocompletion Have a good day. <PERSON>
0bfd1f1f-b145-3d76-3f69-a5c67f7f0ab9
['18d9b57b-c5e2-427e-83f6-1b564718cebd']
Hello everybody, I'm <PERSON>, and I'm a member of the OBM project core-team. Since we started a few weeks ago using roundcube as a webmail, we're going to do some contributions in the near future to match our customer's needs. I'm currently working on the ticket http://trac.roundcube.net/ticket/1486782 which is about displaying the priority info in the list of messages. I didn't quite finished to write the patch, that's why I didn't join it to the Trac ticket yet. I've already added the priority column on my local installation which retrieves the information correctly (from the headers) but now I'm facing a little problem : I would like to be able to sort my messages on the priority column, but IMAP doesn't seem to be able to do that since I get this error message : "SORT: Invalid Sort criterion priority". Now what I think you can help me with is, what is the best way to "manually" sort the messages list, given the way roundcube is organized ? Thanks in advance,
20b4806d-9a29-70fb-a640-c429bc9a5b0b
['18f1dff0-9072-bbf7-8929-aaf45d824d10']
Okay, I know this keeps coming up, but I've searched the email archives and help files for an answer and I can't solve it. I have a copy of version 5.5.6 for Windows on a network drive and I'd like to use it in a couple of conversion scripts without having to install it on each user's computer. However, I can't seem to find the magick ;-) combination of environment variables that would allow me to do so. Each time I send the command convert test.bmp test.bmp I get back this error: "convert: No decode delegate for this image format" I've been looking at the registry entries for an installed copy of it, then trying to set them as environment variables, as well as the usual MAGICK_HOME variable with no luck. Can anyone shed some light on how this might be accomplished? TIA, Arden
c8921001-7688-9f7e-b0ef-39d33cb08ace
['18f1dff0-9072-bbf7-8929-aaf45d824d10']
Can anyone explain the -clone option in convert? I'm trying to follow the IM v6 examples where the command is given as: convert \( logo_light.png \( +clone -threshold -1 +matte \) \ +matte +swap -compose CopyOpacity \) \ \( logo_dark.png \( +clone -threshold 100% +matte \) \
05db85e3-0174-994a-07ec-2460f4f6e853
['18f3e057-2a1a-0cca-6457-986b3024da66']
Hi I am running cygwin on a laptop, which runs XP. I need to compile Sox with MP3 support. I have tested the c compiler, and I can "configure; make; make install" sox and this works perfectly. However I do not have enough knowledge about C to know how to include the libraries required for MP3 support. I understand I will need to include libmad, libmp3 and possibly a lame library. But 1. Where do I get the libraries from 2. How do I add them to the make file Many thanks Rak
f181c2d5-7bdd-7a20-dbdb-2823c173d623
['18f3e057-2a1a-0cca-6457-986b3024da66']
Thanks <PERSON> & <PERSON> for your response. I have been really dumb,... I just realised I have a c compiler under cygwin, which I have installed on my laptop. I am not a C programmer, but I have some basic knowledge of C, and I am very comfortable with UNIX/UNIX programming. So my next question is how to compile sox with MP3 support (which I will ask in a new thread). Many thanks <PERSON>
f9e6106e-417d-77a5-d686-4819f6dcff67
['190cd04c-d95d-c0be-c576-3bbf0f8697b6']
Thank you very much for the prompt answer. When I delete all the banks, here is what happens: I can now record without a program change message being sent. Fine. But when I playback the track I have just recorded, Rosegarden starts by sendin a program change message that resets all my chained sound modules to GM Piano. Ouch. So what I need is, just as you said, a way of avoiding Rosegarden to send any program change message. Any suggestions you may have would save me from returning to the Windows-based Cubase and Cakewalk (with which I never had this problem). Many thanks <PERSON>
429040bf-ce3e-665f-515d-ba2069a3a7e2
['190cd04c-d95d-c0be-c576-3bbf0f8697b6']
Hi Here is my problem: I don't know how to get Rosegarden to not force me into a particular instrument. I use Rosegarden with several sound modules that are chained together (THRU midi cables). Each sound module uses a different sound. So it doesn't work for me if Rosegarden forces a particular instrument (eg Grand Piano) on all my synths each time I play a MIDI file. But how do I deactivate this function (which I never had to worry about using Cubase or Cakewalk)? I don't know how to record without getting tied to a particular instrument. It's plagued me. Tried every possible button and dialog box. Still end up stuck in Grand Piano, or a particular instrument. How to deactivate this? Many thanks in advance.
8565db07-6574-8d48-7cfc-379445441099
['1911a38e-52c5-dc70-c9af-0dfd53b12181']
Recently, Google Places (aka Yelp Lite :-P) got linked to G+ profiles. This linkage has created a potentially serious privacy vulnerability. To my knowledge it has not previously been disclosed; I know it thanks to a tip from a concerned Google maps user. So, first off, the integration isn't fully obvious; it's not listed on the G+ about page. It is explicitly disclosed when you opt in to reviews that it will be linked to your profile, just not always obvious afterwards. Consider for instance +103351126638314796068. His About page doesn't list anything, which would seem to imply that he doesn't want his reviews linked with his G+ profile (which has what is presumably his legal name). However, if you go to https://plus.google.com/local/*/s/by%3A103351126638314796068 (same ID number) you'll find that he has reviewed +100712323888821655907. (Although that personal reviews link _doesn't_ link to his G+ profile directly, the restaurant's Page _does_ do so, and of course it's intrinsic to the ID number in the URL.) If you do a google search for the review text, you can see that at least one third party site has already scraped it. Now, this wouldn't be too bad by itself. It's a couple UI flaws, and to my knowledge you can't get from here to what I'm about to talk about, only the other way 'round. However, suppose that instead you had started by looking at this map of the West Coast Electric Highway: https://maps.google.com/maps/ms?hl=en&gl=us&ie=UTF8&oe=UTF8&msa=0&msid=214874436355124459198.0004c15567ce4ce290f50 You can see that it was created by someone with the username _jimad_. Click that, and you go to an anonymous Google Maps profile page, which lists another two maps made by <PERSON>… and what seems to be an anonymous review of +100712323888821655907. However, if you google the review text — or just click through the restaurant's name — you can then search through the reviews, and see that the writer of that review was in fact +103351126638314796068. So to review, the improper disclosure — which is _not_ anywhere consented to or explained to my knowledge — is that the Google Maps profile _jimad_ belongs to _+103351126638314796068_. (TTBOMK you can't get the reverse linkage; please let me know if not.) In this case, that disclosure is relatively innocuous; knowing who has mapped the West Coast Electric Highway isn't that big a deal. Consider other cases, though, where the creator of a map may have a significant privacy interest in their identity not being disclosed, like this map of porn stores and churches on I-70, by Google Maps user "<PERSON>" http://goo.gl/maps/7avuJ; or this map of Mumbai attacks by user "<PERSON>" http://goo.gl/maps/dKbcA. Both are currently safe — the only thing disclosed is a separate name, and it's not linked to their G+ profiles or legal names. If either of them were to, say, review a restaurant, they would be told and have the impression that the only link they are creating is between their profile and the review. However, what they would also be creating is a public link between their _maps_ and their profile, and this isn't something they would've consented to. This can be mitigated pretty easily: just patch the Google Maps profile page to remove the reviews section, and/or make explicit the linkage in the opt-in consent for Google Places. However, it's already public, and the data's probably already been scraped significantly, so at this point it can't be fully fixed. I hope that the Google Maps, Places, & Plus teams take immediate action to correct this before it results in a leak that hurts someone — and thanks again to my anonymous informant for the tip. <PERSON>
ea7ad7f5-4c5b-19db-e36d-e0641c6d7bb6
['1911a38e-52c5-dc70-c9af-0dfd53b12181']
A month ago, I notified EasyJet's network administrator, Lance Wantenaar <<EMAIL_ADDRESS><PERSON> <EtBUWIEdOs2YMGy5@example.com>, about a serious flaw in EasyJet's password storage policy. Although I explained the problem and its consequences to him clearly, and explained that I would be acting in accordance with the standards of responsible full disclosure, EasyJet has not corrected this issue despite <PERSON>'s assurances that they would investigate it. I have since attempted to follow up with <PERSON> multiple times, but he has not responded. Since they have both had the standard one month and failed to even superficially patch this problem, and their official contact has chosen to not stay in contact, I am making this issue public in the hope that any other security problems with their websites are also made public, and that public shaming will prompt them to protect their users' security when private disclosure did not. EasyJet is currently storing users' passwords in the clear (or using reversible encryption, which is equivalent). You can verify this for yourself by creating an account at http://www.easyjet.com/asp/en/members/ and then activating the 'I have forgotten my password' link. It emails the password back to you in plain text, something that is completely impossible in a securely designed system that only stores salted hashes. Although I have not tested EasyJet's website for SQL injection vulnerabilities, and have no plan to do so, I would say that in my professional experience, people who make such a glaring security error as storing passwords in the clear tend to have other errors as well. As a result of EasyJet's incompetence, if any such vulnerability is found, an attacker will also be able to harvest all of its users' passwords. For a recent example of why this is a problem, please see http://www.techcrunch.com/2009/12/14/rockyou-hack-security-myspace-facebook-passwords/ - and note the followup litigation at http://gigaom.com/2009/12/30/rockyou-sued-over-user-data-breach/ . If you have any questions about this, or you know of any other relevant security issues that may be of interest to me, please contact me. My contact info is at http://saizai.livejournal.com/info . This has been posted publicly to my blog at http://saizai.livejournal.com/960498.html ; I would appreciate a link from any news story or related blogging. Sincerely, <PERSON>
a929911a-e060-b287-5135-76b988351b1b
['19161a1d-853d-cc53-1e89-dace627abc0b']
Good day list, I hope there is someone out there that can point me in the right direction to find a solution for this. I have two quite large hashes each are several hundreds of MB's in size, now I want to with some logic merge these into a single hash. I this works of course but as one might imagine this takes quite a lot of memory. And can depending on the machine simply run out of memory. Until recently these hashes where quite small about 10MB each which was quite aceptable to do in memory certainly be cause that was the easiest and fastest way to get this done. (call me lazy) So my question is how can I reduce the memory footprint of this application while keeping as much speed as posible. Any pointer in the rigth direction would be very much appreciated. Regards, <PERSON>
85d2963c-a143-9855-ede7-e0b0bd19b81a
['19161a1d-853d-cc53-1e89-dace627abc0b']
Hi list, I'm new to this list so please forgive me if I am asking an obvious question that every member of this list has asked at some point in time... When I look at the FAQ for log4perl I see that there is a mention of threading and having every thread log a thread specific message like a users IP address. This leads me to believe that log4perl is thread safe yet I do not see anywhere in the documentation or in the FAQ a definite Yes log4perl is thread safe, and looking at the source code I do not see any obvious file locking which makes me somewhat doubt that log4perl is thread safe. So my question is; is log4perl thread safe when using the standard file appender module with it? Thank you, <PERSON>.
e96e9f0d-f934-bbf2-768e-7f31bf36c5aa
['19182637-c5f3-1968-f4e1-cbbac96c7896']
Hi, I want to deploy the openVPN client exported through the client export utility using a gpo in MS AD. Wondering if someone has done this for the client exported by pfsense. I have tried this for the openVPN client from openvpn.net following the below link: http://docs.openvpn.net/how-to-tutorialsguides/administration/active-directory-deploying-the-access-server-connect-client-via-gpos/ It works but i need the management UI from pfsense Client export as my users are not given any administrative privileges. Hoping for some help :) Thanks <PERSON>
b5d55128-a201-e368-ac95-1e945b84ffa7
['19182637-c5f3-1968-f4e1-cbbac96c7896']
Hi Everyone, I have installed a pfsense cluster with 2 nodes, during a reboot, the system gets stuck at bringing the Mod_Security+*Apache*+Proxy service up, it says it cannot bind with WAN IP. I have now configured the daemon to bind with WAN virtual IP now, still the same issues. Can anyone suggest a possible fix to this problem. I have tried uninstalling the package, It boots fine then, reinstalling the package recreates the problem. I need the Mod_Security+*Apache*+Proxy package for SSL hosting of my webserver. Regards <PERSON>
2d882597-f208-7357-9edf-e036f3b2cee7
['19287d15-1c05-94d2-94e3-b7ff7e9bb53b']
Hi, I'm installing IPCop on an old Dell P5-120. While the system date is correct, the BIOS doesn't seem to want to go beyond the year 2001. I can manually set the date to 2005 and that seems to work OK (dunno if it rolls over to 2006 or not), but if I reboot or power down the machine, it goes back to 2001 (but with the correct date/time). Is there a way I can set the date to the current year on startup? Thanks, <PERSON>
c9c4a5a6-16c7-438e-1f02-a5a0c087e6c8
['19287d15-1c05-94d2-94e3-b7ff7e9bb53b']
I run an exit node with an ISP who initially indicated they would not have a problem with Tor as long as I was transparent about what I was doing, and ran a sufficiently reduced exit policy. They have now sent me evidence of malicious traffic coming from the exit. I don't think they've had any 3rd party complaints about this traffic, but they have expressed various misgivings about Tor in general. They now also want me to consider running Snort IDS on the outgoing traffic. I don't intend to monitor my traffic. But it occurs to me I don't know whether my ISP needs to be worried about it or not. The last one wasn't, so why them? I've asked the EFF about the legal situation in the UK, who passed me to the Open Rights Group. They've not replied to my enquiry as of three weeks ago. So does anyone know of any reliable source of information on running Tor exits in the UK? What would happen if my ISP pressed me to monitor my traffic, and I refused on legal grounds? I'm not suggesting I actually do that, or that there are even any legal grounds to refuse. In fact right now I'm resigned to closing down the node if my ISP turns up the heat. They probably have me by the balls. But I'm at least curious, and can't immediately find any information about things like public carrier status, or traffic monitoring conducted by people like me when it's done in the context of onion routing. Thanks in advance for any help.
a932e9ee-a064-6e9e-0ade-9aa009c9cfe5
['193cccff-4f9f-efa2-b22d-6ef8945e6c6e']
<PERSON>, Happy to see a fellow New Englander on the list! Thanks for the tip. I was considering doing this, but wasn't sure how it managed to "learn", I guess I should say. Will it eventually work out the O365 blocking over time if I threw it in testmode? Is there anything else I should consider doing before that, considering it's going to tag messages in the subject with "possiblespam", etc? Thanks, <PERSON>
e0e88057-2881-55ae-7925-f8142bf7f877
['193cccff-4f9f-efa2-b22d-6ef8945e6c6e']
Hi guys, I've just recently installed Avantfax 3.3.4 in combination with Hylafax, and am having a few issues that I can't seem to get past: 1) It appears that when I try to add a user that does not need admin or super user rights, upon creation, they're just not able to login. If I give them admin or super user rights, they are then able to login. This seems to be present for me in 3.3.3 as well. 2) It doesn't appear to me that there's a clear way to limit users from seeing other users on the system. I'm wanting to disable some users from seeing some things (beyond just their own modems assigned to them) - for example, users that are not admins or super users should not be able to see the whole user list 3) I'm assuming that users can only see their own contacts when sending faxes, and other users won't have that information exposed to them - but can't really find a clear answer on that Hoping some folks can share their thoughts on these issues that I'm stumbling with. I know this is the Hylafax mailing list, but noticed Avantfax being mentioned here on and off, figured it was worth a shot. Thanks, <PERSON>
750f0c83-99b7-f174-412e-646d5a8e757f
['194d935a-2b20-10a5-0b7d-983ab0cce2d8']
Dear all, Apologies if this has been covered elsewhere. Please also note that I am new to Linux and GRASS (less than 4 days!!) - so it is safe to assume that I know nothing!. I have managed to get GRASS6 up and running on fedora core 5 (after some "fun" installing tcl/tk libraries). When I enter Nviz and try to display the elevation.dem from the Spearfish60 sample dataset the visualisation appears for about 10 seconds and then disappears - no error messages are reported to help sort the problem. The control window is empty. I may have confused the software with my tcl/tk installation antics - I now have both tcl/tk 8.4 and 8.5 libaries in my /usr/lib directory but only the 8.4 library is reported when I test the installation in the command window. I do not know how to uninstall 8.5 (I could not get GRASS to work with the 8.5 libraries). Is this the source of my problem or am I completely missing the point? Any help with what to try next would be greatly appreciated. Many thanks <PERSON>
36f890a2-dc80-acf5-0979-10f651fe04e0
['194d935a-2b20-10a5-0b7d-983ab0cce2d8']
Dear all, Am I missing something simple here? I am trying to import some rectified map scans (I do not wish to go through the whole rectification process again if I can avoid it). They were rectified in ArcMap. This produces associated aux, rrd and tfw files. Using gdal I was able to import the tifs without any coordinate system but running the GRASS image rectify routines produced an extraordinary small output window which I could not read with sufficient accuracy (perhaps my eyes are too old?) - as I zoom in I lose the sight of the map grid lines! Any help or suggestions would be appreciated.It is possible I have missed some obvious online help so feel free to be critical! Best wishes <PERSON>
35226fab-90ea-9907-eb33-45ecaa55d689
['196b8a71-6154-c987-99ad-50457aea5e04']
Hi, The note about application precedence states that: "Note: Data of different TLS Record layer content types MAY be interleaved. Application data is generally of lower precedence for transmission than other content types and therefore handshake records may be held if application data is pending. However, records MUST be delivered to the network in the same order as they are protected by the record layer. Recipients MUST receive and process interleaved application layer traffic during handshakes subsequent to the first one on a connection." This was in draft 8, but in draft 9 onwards it's higher precedence for application data. But section 7.4.1.1 has "Since handshake messages are intended to have transmission precedence over application data, it is expected that the negotiation will begin before no more than a few records are received from the client. If the server sends a hello request but does not receive a client hello in response, it may close the connection with a fatal alert." What exactly is the precedence of application data? and why was it changed from lower to higher precedence from draft 9 onwards? On searching the archives i found a post from <PERSON> dated 15 Dec 2004 about the same thing, but he restated the note in his mail as "application data has lower precedence.." and called the other a thinko. A clarification would be greatly appreciated. Thanks in advance, <PERSON><IP_ADDRESS> has "Since handshake messages are intended to have transmission precedence over application data, it is expected that the negotiation will begin before no more than a few records are received from the client. If the server sends a hello request but does not receive a client hello in response, it may close the connection with a fatal alert." What exactly is the precedence of application data? and why was it changed from lower to higher precedence from draft 9 onwards? On searching the archives i found a post from Eric dated 15 Dec 2004 about the same thing, but he restated the note in his mail as "application data has lower precedence.." and called the other a thinko. A clarification would be greatly appreciated. Thanks in advance, jimmy
b64a2762-a329-11c6-949c-d6735107a81a
['196b8a71-6154-c987-99ad-50457aea5e04']
Hi, In SSH2 during re-keying with compression on, how is the exchange hash calculated? The rfc (4253) mentions in section 8 that I_C and I_S as the *payload* of the SSH_MSG_KEXINIT. Now, is this the compressed or the uncompressed contents of those messages? I understand from section 6 (binary packet protocol) that payload refers to the compressed contents. But i can also see that openssh uses the uncompressed contents. Can someone clarify? Thanks, <PERSON>
5eee5080-62c6-329f-d2dd-dc1be7cec009
['1993dbc6-6d21-a1e0-d2e0-86c050177304']
Dear All Been using IPCop for years, Since I couldn't get any sense from the Nice Person who developed Smoothwall 0.99. Went straight to IPCop 1.1 on a P90 with 64Mb and a 540Mb HD faultlessly, Each iteration of IPCop has had a new Box ( Still got all the old ones , all still ready to boot) Current box is a Pentium Pro 200 with 128Mb ECC ram and 2.1Gb HD servicing 3 machines on 56K dialup. I hope to migrate to broadband ADSL 256/64 in the near future. I can't find much hard info on transfer rates and the suitability of my current setup which runs like a clock. I have a Hewlett Packard Vectra VLi8 p3/500 with 256Mb ram ,4.3Gb HD 3 OEM 3Com NICs and no CD(or place to put one,unless it is a half height drive). Problems arise when I attempt to boot the VLi8 without mouse or keyboard, it halts on error, can't find BIOS tweak to prevent "halt on error " like all my other boxes (except a Compaq) Any suggestions, footroom is at a premium so I would prefer to run completely "headless"and dispense with keyboard and mouse. VLi8 case is very small and tucks neatly under my Cray hub Regards rwindsor
8a053741-c639-5399-dd7e-09817279b6b3
['1993dbc6-6d21-a1e0-d2e0-86c050177304']
, "my slowest IPCop machine (90MHz Pentium with 16MB of RAM) can handle everything it sees on a 4MBit/s ADSL connection. Few of my IPCop machines have more than a 4GB hard disc." Dear <PERSON> I've been searching for years to find info that might enable a graph to be drawn, comparing memory and processor to performance. Lots of speculation but your statement is the first I've ever seen that makes a reasonably concrete assertion. My first IPCop box was a P90 with 128 MB ram on a 4 GB HD ( I still have it with the original install ) Connected to the web via 28.6 K dialup Next came a P 200 with 128 MB (I think it's still in the shed ), this one through a 56K dial-up. Then came a Pentium Pro 200 with 128K ECC ram connecting to the web with a 56K dial-up.( I Still have it too, I booted it up a couple of weeks back). Next was a HP Vectra with P3 /500 slot 1 and 512 SD Ram, This box was a bugger and didn't like the memory in it so was replaced temporarily with an Acer P2/400 with 512 SD Ram running , I am ashamed to say, Smoothwall 3 (I didn't set it up, just borrowed it) The Vectra started on 56K dial-up, moved to ADSL1 when it was replaced with the Acer at a move to ADSL2+. I've only had two "failures" since 2003, the Vectra with its memory chip incompatibility and the Acer needed a reboot after some voltage instability. I suspect that from your statement, I've achieved an order of magnitude of overkill in that I'm only servicing a home network with no more than 3 machines active at any one time. The current plan is to use the most energy efficient box I can rescue from the garbage and replace the Acer which will not reboot automatically after a power outage. I have at present, perhaps the worlds slowest ADSL2+ connection, from 10K bytes to 140 K bytes/sec. Bloody awful but still the best I can manage under the circumstances. System limitation is, I guess, the 10baseT hub through which everything flows (Currently a Cray for a bit of pose).. I've neve been able to make an IPCop choke as far as I know but am still interested as to how to determine a performance graph. The only determinant of the machine I run is the physical dimension of the case, it needs to slip into a narrow slot under my desk. I thought I might have a good box in a Compaq Presario 3200, a tiny box, P3/1000 processor 2 x 512 MB ram but it only has room for 1 Nic. It was missing its hard drive and I've yet to learn how to make it boot properly without the Compaq proprietary software . I should try and install another Nic, the MB has 3 slots but the case has 2 blanks, also the power supply is only 95 watts. It owes me nothing so if I blow it up, nothing lost. A Dremel might fix the problem. As all my machines are furnished by the garbage service, a low profile, multiport Nic is not yet on the horizon (I live in hope) I can actually build any machine from 486DX100 with 64 mb of ram to P4/ 2600 and AMD 2200 with up to a GB of ram so if I had any idea at all of how to go about it I could actually test each machine and create a graph (within the limitations of my lousy ADSL2+ service. Any suggestions? Regards <PERSON>
5fe5b71d-4950-0ec7-07ae-16b406e934c2
['19975db9-dcb8-75c4-e1bf-ae5e194fd3b7']
Hi If you are looking for a web mailer to integrate into the Kroupware "world" then the horde project http://www.horde.org/ might we worth considering. In addition to the IMP web mail system they have some groupware applications including Kronolith (calendar), Mnemo (notes), and Nag (tasks). I haven't looked under the hood, but it might be quicker to integrate these into the Kroupware "world" with hopefully minimal changes to the users interface, then to start from scratch. Using the Horde projects would also give you a larger ready-made user community (and hopefully more developers as well). All the best <PERSON> School of Crystallography, Birkbeck College, Malet Street, London WC1E 7HX Tel: <PHONE_NUMBER>
7f99829e-7800-b80f-37ae-aa1f66217cdc
['19975db9-dcb8-75c4-e1bf-ae5e194fd3b7']
<PERSON> I don't know if this suggestion has any merit but while you are expanding the Bib Brother protocol. How about a modified version of your "config" which only returned hash or stat information for the config file? The hashes could be cached on the server to reduce load. Clients could then check if a file had been changed and only fetch the files when necessary. Apologies if anyone has already thought of this. We currently do the job with a mixture of make and rdist on the server. All the best <PERSON>
ffbeaa2b-1c54-f41a-d0c6-0be235d214d9
['199dcb7d-a2f6-6810-a148-27ea47920155']
Here are the notes I used for a presentation to the local Smalltalk user group. Feel free to comment, or to point out errors or omissions, in case there's a next time. The notes were opened in a file workspace, and the font set to 36pt ComicSansMS (because it was the narrowest font). I had a few projects set up with some demos, using the 3.7 beta image. The turnout was much lower than I expected. I guess die-hard Smalltalkers are not necessarily keen Squeakers. There was a non-Smalltalker who showed up - so some hope for the future. <PERSON>
c7889a0a-7bbd-2c8e-6a65-1e34dbe309f9
['199dcb7d-a2f6-6810-a148-27ea47920155']
In many places the OldSocket code references the new Socket class, probably because the class was just renamed. Mostly, it doesn't matter, but for the #accept method it means that a listening "OldSocket" will return a new "Socket". Since OldSocket should go away, and code should be ported to use the new Socket, this is really just a warning for anyone trying to get old code running, before porting it. --yanni
20ab690c-c03e-5711-0e00-095ee1dfe59e
['19b711b5-c4e1-2435-9e50-92bec82beaff']
Hello, My question is : is there a way to have Icecast buffer the input / increase the input buffering ? Even though the latency would increase ? Why : On some radio streams, the listeners to a Icecast stream experience very frequent silences / disconnections. I strongly suspect that the source has a slow/bursty connection to Icecast so that the data sent to the Icecast server arrives in too discontinuous chunks, even though the average connection speed is ok. What's more, they tell me that the Icecast output is in advance vs the actual radio broadcast (through a link to their FM emitter). So I guess that their emitter has some sort of buffering. Thanks -- <PERSON>
9f7ece65-1f95-43ad-aac0-fb6b3bfecf47
['19b711b5-c4e1-2435-9e50-92bec82beaff']
Hello, I'm not sure whether this message belongs to the "broadcast" side of Icecast, but.. I'm broadcasting a few radio stations but I always have level problems. Either the sound is too low, either it saturates. There's a very fine line between the two and the broadcasters are not always paying attention to the output level. Some radios are on Oddcast / XP, some on DarkIce / Linux. Is there some way I can process the sound (compress it ?) before I send it to Icecast ? Or, if the pb is between Icecast and the listener, compress it before sending the stream out ? Thanks
eee6716f-0b4a-e24c-09cc-98b5399f2dd9
['19b83946-5729-ed71-e885-f99431c16999']
This list seems to have fallen out of date: http://boost.org/libs/python/doc/v2/platforms.html I am working on a GUI widget which uses embedded python. Under windows, the GUI system supports MSVC 6.0, 7.0, and 7.1. If I use Boost.Python will this restrict my widget to being MSVC 7.1 only? I have heard that the current Boost.Python only really works under 7.1, but I don't see anything about it in the documentation. Thanks. <PERSON>
1b0ff8c6-1671-ec8c-00af-f22a42c5047d
['19b83946-5729-ed71-e885-f99431c16999']
I downloaded the boost source from the sourceforge page and followed these instructions: http://www.boost.org/libs/python/doc/building.html#VisualStudio I'm using Python2.3, Visual Studio 7.1. I opened up the solution (and converted to 7.1) the solution in libs\python\build\VisualStudio. I added the python2.3 include/lib directories and told it to link against python23.lib. Everything built flawlessly. However, when I try to compile my pyste generated wrapper I get an unresolved external symbol to "get_override". The really strange thing though, is that this isn't the only use of get_override. In several of the generated cpp files, get_override is called without any problems, but in some of them it results in these errors: http://www.idleengineer.net/misc/errors.txt Here is one of the files that generates errors: http://www.idleengineer.net/misc/Camera.cpp This is one of the files that does not generate errors, which I can't see anything different: http://www.idleengineer.net/misc/TextureManager.cpp Any ideas what I can do to fix this? Thanks, <PERSON>
6b58733d-1a54-1819-2849-e29ae57c3126
['19d50c38-fc58-1703-9749-958b77edb4bb']
"Strange. "Many years ago" I coded for EPROM as well, and weirdly, it was the corss-compiler vendors that were slow to offer C++, while I was constantly bugging them about it. This is kind of interesting, since you apparently knew embedded systems programmers that didn't want C++. Perhaps they were really hardware engineers who soft of evolved into doing software, and even the concept of using C and not assembler was very experimental in their minds." Ha Ha - spot on! I learned a lot from working with them though. Very bizarre to meet people who can do mental arithmetic in hex as well. Racing car engineers are a strange bunch at the best of times and all a bit obsessed - you should see the things they did to their casings - drilling away all material they considered excess. Great place to work though. "Even though stuff like dynamic allocation was irrelevant at the time, I sure could have used the syntactical, organizational help that C++ offers, when doing those 128KB - 512KB ROMs. I think that paradoxically, it was because I was so size and speed conscious, that I was in tune with what the compilers were generating, that I realized that using C++ would have been a great design-time convenience, whithout imposing run-time cost. " And this was indeed my argument too :-) "No, I disagree with you here. C++ is not "marginally slower than C;". But that statement is of course quite vague " I was simply making a concesion to the fact that C++ occasionally has the odd extra pointer de-reference which in some peoples minds that slows it down. You are indeed correct though - by the time its been optimised etc. it makes no difference at all. "But for most practical purposes, C++ is the same as C as far as run-time performance is concerned. At least, from a theoretical point of view, this _must_ be true - both C and C++ are compiled, statically typed, languages. Maybe some computer science graduate could verify or refute this conjecture." Not sure how you would go about proving it but my (many) years of experience bears this out. cheers <PERSON>
643a6969-ec99-b239-5d23-76f6d5334904
['19d50c38-fc58-1703-9749-958b77edb4bb']
Hi, hope this isn't too dumb a question: I am writing a small app using a Custom View. I can use cocoa graphics classes and methods obviously but reading about Quartz there are some nice features that would be cool to use. I can't get a context using currentContext().graphicsPort() because this doesn't work in Python 2.4.1. Is there any other way to use quartz in my view or is that not possible yet? cheers <PERSON>
751eb101-930d-3597-e61c-a6c1e471b5bf
['19e00201-d9ee-b342-1f2d-aced70778597']
Hi all, I would like to notify about my status change. I'm leaving Zend, so I'm also leaving Zend ZF team and continue my work on ZF as independent contributor. I plan to spent most of my time on other things (would like to return to my previous area of interests: relational databases, db2 udb related programming, db2 management), but also plan to continue contributing to Zend_Pdf and Zend_Search_Lucene as well as to other ZF components. My non-Zend email is <EMAIL_ADDRESS>. Use it instead of <EMAIL_ADDRESS> for future contacts (for ZF related things as well as for other purposes, e.g. if you interested in DB2 related development or consulting :)). Best regards, <PERSON>.
ea60923e-c88a-4023-8fb0-4e9bfc06ee93
['19e00201-d9ee-b342-1f2d-aced70778597']
Some people have mentioned having problems where idle PHP processes hang around without freeing their memory (in a FastCGI or Apache module modes). As I have found, it's not really a memory fragmentation problem. The core of the problem is the fact, that free() doesn't return memory to the OS immediately. So, if we have memory hungry PHP scripts and use PHP in a FastCGI or Apache module environment, then large amount of memory could still remain allocated for a process while it's waiting for the next HTTP request. Attached is a patch which resolves this problem by forcing the memory allocator to return free pages to the operating system at the end of the request. Especially in instances where some of your PHP scripts use a lot of memory, this maybe be quite useful in reducing the machine's overall memory consumption at a given time. I think it's probably best for this explicit memory collector mode not to be used in the multithreaded environment (for performance reasons), but the patch works properly if ZTS is enabled. Two new configuration variables could be set in a php.ini: memory_trim_limit (default is 1M) - is a minimum memory grow per request limit until malloc_trim mechanism will be invoked. "0" value disables it. memory_grow_limit (default is 128M) - is a memory grow limit. If process reaches it, then force_exit SAPI global is turned on and could be used in SAPI modules to force process exit. This is for the case when we have persistent objects on the top of memory or memory leaks in external libraries. "0" value disables it. Patch also includes processing of force_exit SAPI global for CGI SAPI module (for FastCGI mode), Apache SAPI module and Apache 2.0 Handler SAPI module (for prefork Apache2 mode). Please review it, try it and send me feedback. With best regards, <PERSON>.
5c4d4ce5-7e28-6165-c3f9-1912089c977a
['19fc9625-b7e8-fa40-cff3-914c7300eccb']
Hi, RT3.0.0, RH Linux8.0, Apache2/ModPerl2 I noticed that with RT3, the display, although esthetically more pleasing than that of RT2, under IE6, the right edge of the display is cut off by a few points. It's minor, but annoying. This is not the case with Netscape6 or 7, which leaves a nice border on the right side. Is there some simple way to fix this? This is not the case with RT2, which, under IE6 looks more like the display of Netscape for RT3. Any ideas? <PERSON>
fe354825-45c9-b070-9b7c-9d56caf0c15d
['19fc9625-b7e8-fa40-cff3-914c7300eccb']
I'm running RT 3.0.7_01 We have a number of custom fields defined for a queue, where one is a drop-down single select list. I have been unable to add new values to the list after the 45th. This doesn't make much sense, and there is no error message or log that I can find that would explain why this happens. Is there a limit to the number of CustomFieldValues we can have for a CustomField, and if so, how can I change it? I have had a look at the code, but couldn't find any indication of a limit. Thanks, <PERSON>
54b07037-aed2-15ea-3af9-d683e98220e7
['1a08053e-e1b2-1100-79b1-12b103c8268e']
Hi, I just discovered xmpppy (which is great, thanks!), and I am a Python newbie. When I try using xmpppy with the example script, it often (but not always) hangs when I call the script. strace tells me, that it's waiting for data from /dev/random. This happens during the 'import xmpp' call. Now, this probably doesn't happen on busy systems, where /dev/random always has numbers, but on my quiet home server, it's definitely annoying. Since I'm a Python newbie, I don't know how things are done there, but from my experience just doing 'import <library>' shouldn't do complex things that can go wrong or take a long time, such as getting a random number from /dev/random. Wouldn't be the xmpp.Client constructor (or however that is called in Python land) be a better place for that? Thanks, <PERSON>.
47bf2dee-714a-c209-2fd0-431a3ca7b8df
['1a08053e-e1b2-1100-79b1-12b103c8268e']
Hi there, I'm using the GPG plugin fore quite some time, and I always sign my mails before I send them. Now that I updated my SquirrelMail install to 1.4.10a (from 1.4.9a, where the plugin worked correctly), I can't sign my mails anymore. When the checkbox "sign on send" is checked and I send a mail, the password entry popup appears as it should. But when I enter my passphrase, I just see my mail again, just as before I clicked the send button, except that my signature has been appended again and is then present twice. I switched on debug messages for the GPG plugin, but didn't see anything useful in the output. My syslog contains no PHP error messages, and also apache seems to be happy. I have no clue where to look/debug :( I tried re-configuring SM and the plugin, with no effect. Encrypting and decrypting still work as they should. Can anybody confirm this? Is there a workaround? Patch? <PERSON>
3ad4d524-cb34-0d22-4d67-7e225338dcea
['1a21075d-2b44-c6bd-6424-01efc3505043']
Hi All, While working on a driver for an in-house built usb board, I found that even when the board failed to register the id strings were printed out. This ended up confusing me for a short while as I expected the strings to only show up if the device registered successfully. What about moving the getstring/showstring sequences to after the usb_set_configuration if bracket? This would ensure that the device id strings print out only if the usb_new_device registered the client successfully. What does everyone think? <PERSON>
fe4ecc7d-1dcf-1c31-95ab-3f4616531910
['1a21075d-2b44-c6bd-6424-01efc3505043']
Hi all I've noticed that during FC4 install if I decide to manually chose packages to install that deselecting printer support seems to be ignored. I still get those packages installed. Not only those packages but drivers and printer fonts from other packages as well. Has anyone else had problems such as these? I certainly would like printer support to refrain from installing itself when I deselect those packages but haven't found any magic incantation to accomplish this. Ideas folks? <PERSON>
746b5228-3ee4-477a-518a-a396b92b1963
['1a30d3a9-9891-bb9a-5edf-0852e5ef3d3e']
My reply to what <PERSON> said is: What <PERSON> said in the first place: #I hope some day we can see computers that come preinstalled with an E #desktop someday. For now though... My tablet PC powered by Enlightenment! :D # #http://www.youtube.com/watch?v=vzgE6gkb2EE Are you using Bhodi Linux? And what did you do to get the touch screen working? Can you tell us what hardware you are using? Thanks. <PERSON>
3c701b79-32b7-2ba3-fabc-9ead5bd34b71
['1a30d3a9-9891-bb9a-5edf-0852e5ef3d3e']
hey everyone, I use documents to go on my treo 680 and am trying to get a spreadsheet sync'd over. i've tried gpilot-install-file, pilot-xfer, and jpilot but none of them seem to be able to get the file across. for example, with pilot-xfer i use the command: pilot-xfer -i publication_spreadsheet.xls -D /Palm and get this error: Unable to open 'publication_spreadsheet.xls'! pilot-xfer: invalid option -- D I generally use gpilot-install-file but can't seem to get any results. anyone have an idea how i can manage this? thanks all! <PERSON>
618cbd49-4b53-8fca-ea75-a9b02248be49
['1a3ca2f6-9e0a-832d-1ce7-b3fdd115e7a7']
Hi, In a wireless BSS, when the link quality is not good enough to support 11Mbps mode, the STA switch to 5.5Mbps mode, then to 2Mpbs... Ns2 have not implement this switching facility. Actually, when packets are being dropped in 11Mb (reason: STA far away, bad propagation...), setting a 5.5Mb bandwidth doesnt provide a better performance. Has someone makes changes in the ns2 code, in order to solve this problem?. Thanks, <PERSON>
a1c1d5e1-713e-db0a-56b3-9b7035f1b34f
['1a3ca2f6-9e0a-832d-1ce7-b3fdd115e7a7']
Hi, I have ns2 installed, and use it to run QoS simulations in wireless (DCF, EDCF). Has someone used ns2 as a tool for wireless capacity and coverage planning?. If so, what kind of simulations, parameters should be tested?. Although I know there are other specific network planning tools, I would like to keep on using ns2 now that I am getting a Know-How on it. Thanks, Best Regards, <PERSON>
92b332a4-30e2-f71f-3b75-e1ea4eb42504
['1a49be70-02ac-0217-5163-2900d7d8558b']
Hi, all I'm new to SDN and RYU. How can I modify the packets' header fields with RYU? For example, I want to change a packet's source IP address from "<IP_ADDRESS>" to "<IP_ADDRESS>". How can I do this? I've scanned the RYU book. But there is no specific codes about this. Can you provided detailed codes? Thanks in advance! Best, Bin
074606c9-3710-50a9-3032-5b39d2b5be58
['1a49be70-02ac-0217-5163-2900d7d8558b']
Hi all, Can I set the counters of the to-be-installed rule when adding a new rule to the flow table? To be more specific, ( ovs-ofctl add-flow s1 "in_port=2 actions=output:3" ) can add a new rule in switch s1. With dump-flows command, we can see something like: "... n_packets=0, n_bytes=0,..., in_port=2, actions=output:3...". So, my question is can I set the field of "n_packets"/"n_bytes" to an arbitrary number as desired? Is is possible? If yes, how? Also, if it is not possible in OpenVswitch, do any commercial physical SDN switches support this? Thanks a lot! Best regards, Bin
bcca56e7-e188-ba1e-8ab9-e6f97478bf3e
['1a5fe83d-6de9-3aee-3f2d-f052676d3904']
Hi, On 2.6.10, the detected hw partition number seems to be insane. All is ok on 2.6.9. I added few printk's to the cfi_intelext_partition_fixup() in 2.6.10 hoping that this will provide some more useful info (probably not). Below I give the code fragment sprinkled with my printk's and both unsuccessful 2.6.10 and successful 2.6.9 boot dumps. <PERSON>
f7c5b163-81e7-573e-7f55-bf30aac48da1
['1a5fe83d-6de9-3aee-3f2d-f052676d3904']
Hi <PERSON>, When enabling CONFIG_DEBUG_SPINLOCK on a board with LH7A400 SoC I noticed that the port->lock is manipulated before it is initialized. The reason is that lh7a40xuart_set_termios() is called before lh7a40xuart_init(). At first glance, I was not able to come up with a proper fix. The boot log snippet is below. I haven't tested this for >2.6.10 kernels. <PERSON>
8294f109-f773-3a12-882e-e5a76748d7d8
['1a6e67cd-c262-c8a6-7732-4cab2582d0a6']
Hey guys, Firstly, great work on the Sony laptop page and mailing lists. It's extremely helpful. I recently bought a Sony Vaio GR370, and installed Debian Woody. I have custom built Linux 2.4.18 on it and it's working great. I just wanted to confirm a few things. For XFree86 4.1.x, I used the Vesa driver for the Radeon video card. It works good with this. Is there another driver I should be using? If so, could someone show me their XF86Config file? Sound. I understand the chipset is an Intel i810, and that ALSA works great for this. Is it possible to see the modules.conf file to load these drivers? I have not had any luck. Is there a page up on the GR370 with Linux? Thank you very much. Afra
f50795cf-89e9-9ac6-f5e6-18e142c38ecf
['1a6e67cd-c262-c8a6-7732-4cab2582d0a6']
Hi, We have run into some problems here which stops us from properly backing up the databases using pg_dump. We do have a tarred version of the data directory (/usr/local/pgsql/data) which contains databases that we need to restore. I found out that simply replacing the data/ directory does no good. Is there another method in which I can restore databases with the data/ directory? Thank you Afra
722b5d5b-ffe0-d4d8-bae3-509000399600
['1a733505-9bce-aa42-fa06-e15503f65b53']
Hello everyone, I want to reproduce global buckling of a 3D column (simply supported) using shell elements independently of its length. I know that each section of the column should behave like a rigid body but I can't do it, so by doing the buckling analysis I get other buckling modes (like local buckling). My initial idea was to use rigid body (tie) constraint in some sections (uniformly spaced) but what I got was some local buckling in the extremities of the column. Can someone help me? I think the problem is in the boundary conditions. Best regards, <PERSON>
a43651d9-c48f-d612-ca20-89493b99fd2e
['1a733505-9bce-aa42-fa06-e15503f65b53']
I use shell elements because I want also to reproduce local buckling of my column. Therefore, for the same model I want to be able to have a certain local buckling independently of the column's length and on the other hand I want to have just global buckling. So in one model I have already one that gives me only local buckling and I want to have another model with just global modes. In order to do that I thought of using Rigid Body constraint or Kinematic constraint but I'm getting wrong results. I tried to put some sections of the column using Kinematic constraint, being the master node, the node in the middle of the section. I defined that the displacement in the section-plane and rotation of the axis should be coupled to the master node. About the boundary conditions of the extremities, I defined each node of the section to be fixed in the plane. And the middle node in the middle of the column can't translate in the axis direction. I don't know what I'm doing wrong but something is. Best regards, <PERSON>
0efc84c6-c029-8871-95a2-08efe7e40734
['1a9274a7-7b77-b1a2-135f-9e5b90f9b102']
Today <PERSON> committed some xml files into the toolchain repo. In Linux we will be using these xml files for: - MMR debug interface in Linux kernel (patches <PERSON> has been sending around) - objdump's '-M reg-names=ARCH' option - which should allow us to print the MMR name, rather that just the address. - Eclipse MMR debug Window (which I will check into svn tomorrow am) The first two options will use a post processed output, and the last will use the xml files directly. Both <PERSON> and <PERSON> have volunteered to update the existing xml files, so they are a little easier to handle for the job we want to do, as well as make them consistant across processors. If you have any comments about these files or where registers appear in the above lists - please email <PERSON> or <PERSON>.
fe2e749a-c53d-3a22-eede-6ddad6514d7e
['1a9274a7-7b77-b1a2-135f-9e5b90f9b102']
If we can define the exact problem - and then come up with a generic solution, it might be easier... The exact problem is that when someone opens a standard memory window in Eclipse, Eclipse starts reading all kinds of things that it really should not. :( For example, to try to make scrolling for the user faster, Eclipse reads the target memory buffer + some other undisplayed memory at the top and bottom of the target/displayed buffer. If the buffer points to the beginning of L1 Data memory - this causes undesired behaviour, since you are reading memory locations that do not exist. It is possible to have the buffer point to MMR space - where Eclipse will just start reading MMRs that do not exist. (or 16/32 issues). It is possible to have the buffer point to SDRAM that does not exist. It is possible to have the buffer point to L1 instruction. It is possible to have the buffer point to the memory that cache occupies. It is pretty clear, that _something_ needs to understand the memory layout of the chip/system. (for now, let's just say chip). This doesn't seem to be a problem with kgdb (but I have not tried in enough detail to know). but it could be that kgdb is making sure things are valid before it returns things. this isn't a problem with command line gdb, since most people don't poke random places in memory, and expect it to work - but it is possible to do, and users are not protected from themselves. Since the JTAG can always read the CPUID, It should be able to self discover the chip memory map (external memory is something else). If we add the chip IDs & internal memory info to a file somewhere, is all we need to do, just tell gdb proxy how much external memory there is on the platform? -mem=64M (or something easy like that?) -Robin
4561af2c-3ebf-9cd6-0297-71c133568a9d
['1ab09cb7-d4fb-e62b-f9a6-962f3e3d7815']
Hey, Quick question if anybody knows. Is there a special way to attach EBS volumes to the new FC images that have XEN aware kernels? We can only attach devices as /dev/sd* but FC uses /dev/xvd*. I am having an issue where an EBS volume is in the attaching state for the last 12 hours. Thanks! ttyl Dima
56f92d4f-57b0-9205-189b-f11277781770
['1ab09cb7-d4fb-e62b-f9a6-962f3e3d7815']
Hi, First, want to say a big thank you for making FC 14 available on EC2, very much appreciated. Second, when I booted the new FC 14 image it came up with a 2GB root partition. Is there any way to increase this parition to say 10gig, and if not, what approach should one take with respect to installing packages, etc, since I don't think that 2GB will be sufficient. Thanks! <PERSON>
62b4f422-176b-991b-e7ec-ca3c4a2270a4
['1ab0a54b-535d-0dc0-4838-21a09a1efe23']
Hi, Before I waste any time on this, is there any reasonable hope that I will be able to transfer information to and from a Samsung Omnia using SynCE? I am working from an i686 Debian Lenny box. The Omnia has Windoze mobile 6.1 -- at least until I despair and try to install linux on it! TIA richard
5da5762d-c748-63ae-4485-fc4c163ae860
['1ab0a54b-535d-0dc0-4838-21a09a1efe23']
I've set up two sites using DIYPlone, and borrowing heavily from its css. They seem to work, but I have two issues I would like help on. I'll address them separately. If you look at the source of the page received by an anonymous user, although lots of stuff is not rendered on the html (eg. I have omitted the author and last change info, the personal tools and other things), all the omitted items are actually there in plain text in the source, just not rendered in the page. This seems a waste of bandwidth, on one hand, and makes a nonsense of omitting those items, on the other. I used the methods in the two tutorials by <PERSON> and <PERSON>. Did I miss something basic, or is this normal? TIA
fc9b7742-0392-06f0-c1a8-5b79428b79f0
['1ab75d6f-03bd-1563-ec3b-350bcfa75059']
I am a sophomore undergraduate at IIT Bombay. I have a very strong mathematical background. I have spent the last one year exploring various areas of theoretical computer science - from graph theory to communication complexity to streaming algorithms through various R&D Projects. In the meanwhile I have collaborated with a friend on a paper- http://arxiv.org/abs/1102.1636. Previous summer I spent Interning at Affordable Solution Labs, IIT Bombay in building various tools for a Govt of India Project to advertise e-learning. The summer of 2008, I attended a nurture camp where topics in Group, Ring and Field Theory were introduced. And previous to that my summers were spent in attending the International Mathematical Olympiad Training Camp. Up to date I have written around 50000 lines of code mostly in C++. In Drizzle I liked “SET and TUPLE data type” topic. I have learnt a lot of theory over the years and look at GSoC as an option to dive in-depth into C++. My strong mathematical background will always make look for simple yet efficient code. While I might not know adequate theory at this point, I believe that there is enough time in summer to continue both learning and coding and make a significant contribution to Drizzle. Please comment on my candidature (considering I am a sophomore) and give me some more insight into this topic. Thanks <PERSON>
fcafce94-ac6f-b4ad-824c-9545022846fa
['1ab75d6f-03bd-1563-ec3b-350bcfa75059']
I am a sophomore undergraduate at IIT Bombay. I have a very strong mathematical background. I have spent the last one year exploring various areas of theoretical computer science - from graph theory to communication complexity to streaming algorithms through various R&D Projects. In the meanwhile I have collaborated with a friend on a paper- http://arxiv.org/abs/1102.1636. Previous summer I spent Interning at Affordable Solution Labs, IIT Bombay in building various tools for a Govt of India Project to advertise e-learning. The summer of 2008, I attended a nurture camp where topics in Group, Ring and Field Theory, Linear Algebra were introduced. And previous to that my summers were spent in attending the International Mathematical Olympiad Training Camp. Up to date I have written around 50000 lines of code mostly in C++. Are there any projects which I might be able to do (further considering my inclination for theoretical computer science and algorithms and also the fact that I am comfortable with C++ and to some extent JAVA and C), taking into account the constraint that I am a sophomore? Thanks <PERSON>
04403853-fb82-24f9-c40b-d0f2e0eb42c8
['1acfa75a-484b-d65c-8f63-1d57e7e79d89']
Because the two editors calling for deletion may be biased vs. woman pianist's ethnicity, I don't know if this really is an article fit for deletion, especially for those of us who are inclusionists. http://en.wikipedia.org/wiki/Wikipedia:Articles_for_deletion/Zvart_Sarkissian I wasn't sure if should where I should post this here or Wikiproject Feminism. This is one of those times a "Wikiproject Gender gap" might be useful in addition to feminism. The older I get, the more I get confused by wikipedia policies etc. Don't tell all those 22 year old guys. Or maybe I'm still not over this darned head cold. That's the ticket! :-) <PERSON> in dc
6c4edc3d-7deb-d016-5cca-215f6b7ad4ca
['1acfa75a-484b-d65c-8f63-1d57e7e79d89']
Number one rule, practice self control: don't reply to list unless it's important for everyone to know,not just that you want them to know it ;-) Otherwise, reply to individual poster or not at all. Of course this applies more to the more assertive posters. However, if one is generally reluctant to post, thinking one's contributions are not important enough, one should not be discouraged by this. And the fact there is less gratuitous posting may mean more reluctant posters feel more free to post. Also self-control cuts down on eventual debate on what should or should not have been moderated. <PERSON> in dc
5c8138e4-5102-8ecf-fa10-10862c9295e8
['1acfb1ab-307e-781f-943f-80f8e1f4cccb']
Hi all! Version 1.1.6 of the mom macros is now available at http://www.ncf.ca/~df191/mom.html I'll be sending patches to <PERSON>, so you have the choice of getting mom from the website or waiting till the patches get applied to mom in the groff repository. The major change to mom in this release is the addition of table of contents generation. Quite a few folks wrote me privately begging for this, so I've obliged. There's also a full suite of "control" macros to help users design the look of their tocs. There are quite a few changes, some visible, some not. Amongst the visible ones are the addition of more macros to control the appearance of endnotes pages. Chief amongst them is that it is now possible to set a document in columns and have the endnotes non-columnized. There's also a new DOC_TITLE macro which tells mom about documents collated together where each separate doc has its own title, but the overall doc has another. By default, the overall document title appears header left, and the individual documents' titles appear header center. To complement this change, there's now a HEADER_CENTER_PAD macro that allows users to put specified amounts of space left or right of the header center string, which helps visually center the string when either the left or right part of the header is unusually long. In RECTO_VERSO printing, mom figures out which side of the header center string to put the padding on odd and even pages. On the "invisble" side, there are significant changes to the START, COLLATE, HEAD, SUBHEAD and PARAHEAD macros, as well as to ENDNOTE and ENDNOTES. Furthermore, I've futzed around with the handling of document leading adjustment (i.e. adjusting the leading so that pages are properly filled with type). Hopefully, this won't affect documents created with previous versions of mom, but if anyone encounters anomolies, please let me know. The big change is that formerly, if one set the leading with LS prior to START, one had to invoke DOC_LEAD_ADJUST in order to get adjusted leading. As of 1.1.6, the reverse is true: if one sets the leading with LS prior to START, one now has to invoke DOC_LEAD_ADJUST OFF, otherwise, mom automatically applies leading adjustment. This seems to be what users who've written me want, and I can see the logic of it. Almost no one uses non-adjusted leading. See the NEWS and ChangeLog files for more information, as well as mom's html documentation. Enjoy!
abf33c68-1a2f-63b4-0021-4cf138355277
['1acfb1ab-307e-781f-943f-80f8e1f4cccb']
Hi all. A number of mom users have written me wondering whether mom supports text colour directly, so I've (finally) added it. The basic procedure is to define a new colour with .NEWCOLOR (whose use is almost identical to .defcolor) or to "initialize" one of the pre-defined colours from rgb.txt with .XCOLOR. Both define a string of <colorname> that can be called inline with \*[<colorname>] or with a macro, .COLOR <colorname>. The (very slight) advantage this has over groff's native way of handling colour inline (\m[<colorname>]) is that it gives newcomers to groff who use mom one less escape sequence to have to research and memorize from the official groff documentation. (BTW --has anyone noticed how good groff's info docs are? A huge improvement over the situation a few years ago.) More importantly, for users of mom's document processing macros, the "control macros" that allow one to set the style of various document elements (main heads, page headers/footers, quotes, page numbering, etc) now have _COLOR variants (in addition to _FAMILY, _FONT, _SIZE and so on), making it possible to, for example, have main heads uniformly colourized red simply by invoking ".HEAD_COLOR red". In addition to adding colour support, I also fixed a number of bugs that had to do with page headers and footers, and the handling of tabs when they're used during document processing. (It appears some users are subjecting mom to creative abuse I didn't foresee.) Werner has applied the patches for mom v. 1.1.8 to the repository, and a gzipped tarball of the whole mom package is available, as usual, at http://www.ncf.ca/~df191/mom.html Because Werner is hoping to release 1.19.1 soon, and it would be nice if the version of mom included in that release were good and stable, bug reports and suggestions for mom are especially welcome at this time. Cheers.
20003450-32c9-2d8b-c8c8-7547f06ae9a9
['1ae03917-2a1c-eed5-1ac0-87907343f9d5']
Hello all, I have not been able to set up the http subversion server in Linux. Can anyone please give me the steps to do it. I followed the steps in the svn book but to no avail. The apache is installed in the server and also the svn. All I have to do is configure the svn to get it working and create the repository. Can anyone help me? Thanks in advance, Regards, Bikash.
5da23c8b-1f61-0e8e-0fb6-8312f59006ad
['1ae03917-2a1c-eed5-1ac0-87907343f9d5']
I have to setup svn on a linux server running Apache 2.0. I have to do it on a vistual host. I have created a repository, made changes to the httpd.include file, created the user list using htpasswd, and also set the authz file. In short I have done everthing mentioned in the svn book but to no avail. Still I can’t get http access to the svn repository. I am connected remotely to the linux server. Can anyone guide me. Thanks in advance. Regards, Bikash.
12bcbcec-fafb-ee77-6694-d8ac9709c464
['1ae8835a-ccd4-12f9-2648-7dc88b82985d']
<PERSON> notes:[[[other surface tricks are high heat, air bubbles, sucking away the turbulent layer periodically, and slime, usually polyethylene oxide which must be renewed continuously. ]]] Re. "air bubbles": There was an article in the now defunct "Small Boat Journal" about a "Captive-Air-Film Hull". that I found intriguing. Air was " sucked" thru ports above-deck, and "fed" (via the reverse-air-scoop termination of a v-shaped bow section) into a slightely-recessed bottom (1.5 inch?) between 2 power-catamaran hulls... creating a sheet of bubble/water mix upon which the boat ran. The claimed increase in performance (speed/mpg) were so impressive that I called and had a very interesting conversation with the designer. He was a down-to-earth, believable sort of guy, I thought. It's quite a story... sorry I don't have time to check my collection for the issue # now. Maybe someone else will tell us, or check library holdings for old issues. <PERSON>
889fce43-66b3-3409-896d-bcb7830a64b4
['1ae8835a-ccd4-12f9-2648-7dc88b82985d']
Hi Everyone, While there are at least a few of us present today, I'd like to ask for some opinion. I'm just purchasing a computer system (Mac) for photo image/design work. I'd like to get 3D modeling software to play with my boat-design ideas (to 30ft.). I will confess up-front to relative computer-illiteracy and naivete... so please bear with my less-than-intelligently-phrased questions. I know there is a considerable learning-curve requirement for the use of such a program, but consider myself "trainable", and I have the time, and the interest. I will have mucho RAM and drive space, etc...so would like responses about the 3D SOFTWARE, specifically. My intended use is simply (simply?) the 3D display of my 2D boat design work. I want to be able to rotate on any of the three axis so as to judge proportion, esthetics, etc. (to "build" models on screen/paper, so I don't have to build them physically). I want the capabilty to put in offsets, and have the program fair the lines (using nurbs?,other means?). "Rendering", which I understand to mean the depiction of surfaces/materials would be nice, but is SECONDARY to the basic shape (wireframe?). (I'm using terms here, the meanings of which are largely mysteries to me... be gentle!) In my beginning research I've been impresseed with "Rhino 3D" , but it is not Mac-compatible, except by using "Virtual PC"... resulting in slowing-down the program. Anyone care to comment on or recommend a Mac-compatible software for this use? (My price limit would be in the neighborhood of the Rhino program..$500-$700.) Please mention also any other software required to work WITH the 3D program. I know that most of these programs have capabilities that I can not now even envision using... but my real interest is in the use I've described above. Much more could easily be wasted on me. Thanks a lot for any help! <PERSON>
f8d15a08-8e71-916b-52a3-e318b5601e78
['1ae92954-5690-b21d-0e37-234072c3e42f']
Hello, <PERSON>, I understood from <PERSON> that you want me to suggest adapted donation sums for the BG version of the landing page. I would suggest 100, 50, 20, 10, 5 (if there is possibility for fifth option) or (if not) only 100, 50, 20, 10. The word "Other" translates as "Друга сума" or - if it's needed to be short - "Друго". Thank you in advance! <PERSON>
e13f5f62-66fe-5e88-104d-0512a4a6ae30
['1ae92954-5690-b21d-0e37-234072c3e42f']
Hello everybody! I'm translating the survey to Bulgarian and I reached question Q3, which I believe needs clarification, especially considering the preceding questions. *Q3. What is your user access level? (Please choose all that apply) * Having asked the editor to first list ALL applicable versions of Wikipedia where they contribute, and then the ONE where they contribute primarily, its obscure then which of both aspects does Q3 exactly refer to. For instance, I may contribute to both Bulgarian and English Wikipedia (Bulgarian being the primary one) and I may have a basic account on the English WP, while being a sysop on Bulgarian WP. Then, how am I supposed to answer the question? :) Hence, the question should either stipulate that the answer is given for the *primarily edited *version of Wikipedia, OR it shall allow the user to provide their * maximal *user access level ever reached. I think both answers would be helpful for the survey, or even one of them, as long as all users understand the question uniformly. I've posted this question on the talk page http://meta.wikimedia.org/wiki/Talk:Wikipedia_Editors_Survey_2011/Translation#Q3_needs_clarification but I decided that it may be a good idea to post it here too, considering some translations have already been completed, while this issue may trigger rewording of the original... Best wishes! <PERSON>
fb0c5290-b284-2ea5-37c3-edceb08ad773
['1af0d91c-99b7-5922-47a2-664eba7149db']
Guys, By now you might have heard of *Cumulogic*, *PASS*, *based on Java*. We've been working with them to integrate http://www.ilikeplaces.com . The approach we took so far on OpenEJB is the drop in war on tomcat. While our Rackspace instances are fine, getting OpenEJB on Cumulogic was a challenge. *However, we got the app to a working point and we had to use Tomcat - 7.0.12 and 7.0.23 with o**penejb-tomcat-webapp-4.0.0-beta-1.war**. * Now, unlike us, Cumulogic will be committing to OpenEJB + Tomcat 7.x when going live, so I would like to know if there are any concerns related to OpenEJB working with Tomcat 7.x. *We used openejb-tomcat-webapp-4.0.0-beta-1.war . Any stable version known to work with Tomcat 7.x ?* Thanks a lot! <PERSON>...
2a4138fb-ea8d-aff4-90c5-5a705da2bf73
['1af0d91c-99b7-5922-47a2-664eba7149db']
I am deploying my same application on a windows 32 bit machine and our server with CentOS 64 bit and JRE 64 bit. Tomcat, upon deployment, utilized a lot of memory, probably five times more than my local memory. I have been checking on if it is a JRE issue of 64 bit. Please help me if you have any experience with this regard. If not, please do not bother with this query as I am all over the net finding and trying out solutions provided. I post this here in case OpenEJB has something to do with the issue and anybody else has had the same experience. Thank you. <PERSON>
07382fc2-89a5-1499-00ca-4c76d4be64e1
['1b42e436-4542-40ec-e767-d5586ae7d279']
Indeed, my question seems to be totally independant of the hdf5 file but related to the way the signal is computed before storage into hdf5. I'm not sure about the exact specification of the signal encoding except this is 32-bit floating points. I cannot choose between floating points or integers as I cannot tune the device producing the signal (the signal is microVolt and floating points by default). I'd rather ask the manufacturer about the possible signal transformation before storage. Thanks for your answer. <PERSON>
c332b0ae-8c3a-0905-a684-0c3920f56b2a
['1b42e436-4542-40ec-e767-d5586ae7d279']
Ok. It makes the point really clear, no data transformation related to hdf5. I'didn't find any information about data transformation values in the rest of the files, just min/max possible values and filters applied to the channels. I will ask the manufacturer to go further in this question. Many thanks for your answer. <PERSON>
1533dea6-4837-bdb3-039f-20568130724c
['1b5565c3-7871-f32f-713b-2b9e1195e3ac']
Hi, Thank you for your answer. Maybe I did gave you a good explanation to my problem. 1st check my blat parameters. blat -body "App.log report is attached." -subject "My App Log Report" -f <EMAIL_ADDRESS> -to <EMAIL_ADDRESS> -server mail.mycompany.gr -port 587 -u loguser -pw logpass123! -attacht app.log This above script goes to EVERY client of ours. As you can see by yourself, to send an email from <EMAIL_ADDRESS> to <EMAIL_ADDRESS> I HAVE to entry the username as well as the password of the send mail which in this case is the <EMAIL_ADDRESS>. This is a problem for my. I dont want my client to KNOW the username and the password of the specific email account. So, what is my options here ? Regards,
10b0da8e-c29f-0ac2-c486-882c8b33de66
['1b5565c3-7871-f32f-713b-2b9e1195e3ac']
Thank you, all of you. I find a custom solution for my problem. I will share it with you. I changed the original source code of the BLAT and everytime it founds a specific smtp mail server ie mail.mycompany.com then it changes the sender, receiver, port, username and password directly from the code no matter what command line parameters are set. Of course the strings are encoded somehow to avoid a lightweight debugging. Best Regards,
599a17c9-355f-3c9c-85a7-62034d6e516d
['1b6d6e56-e827-e9c7-187a-7216981b8dca']
The changes I made to the commit-walker fetch code turn out not to work with USE_CURL_MULTI, and I'm not at all clear on how to deal with it. I think the current code for USE_CURL_MULTI also can't work for having multiple things in the same executable using http.h, which will be a problem for having both ls-remote and http-fetch as part of builtin-fetch, as well as eventually calling http-push without an exec from builtin-push. Is there anyone around who understands it? <PERSON>
c8c03064-6360-94d7-600f-df313be5a8b7
['1b6d6e56-e827-e9c7-187a-7216981b8dca']
I was having problems with WEP (in 2.6.19, but the relevant code doesn't seem to have changed), which I eventually tracked down to this: ieee80211-crypt-wep tries to use "ecb(arc4)", which works fine if ecb, arc4, and cryptomgr are already loaded, but doesn't cause them to be loaded if only arc4 is loaded. I haven't tried with arc4 not loaded, but I did try arc4,cryptomgr and arc4,ecb; neither of these caused the third one to be loaded. (I also had crypto_algapi, michael_mic, and aes loaded, if that matters) I can test stuff without too much trouble, but I'm out of ideas for things to test. <PERSON>
f905be0d-5b36-8dc5-db59-ec7ff7d6f81b
['1b6ea32f-0b59-41dc-3019-2c76365b5d9e']
Hello I'm trying to make my first steps on qtopia on a Freerunner. Still I haven't found any user guide. The tutorial for users on the qtopia homepage is but a joke. Is there any tutorial or howto to configure qtopia on a Freerunner? So far, I could not even manage to get DNS over usb0. Thanks <PERSON>
4e73f0ed-ff32-98ef-f41e-12d6b41dc50e
['1b6ea32f-0b59-41dc-3019-2c76365b5d9e']
Hello I've just succeeded to install Debian on the Freerunner together with xfce4 following the instructions in <http://wiki.debian.org/DebianOnFreeRunner> and <http://wiki.openmoko.org/wiki/Debian>. Most things run fine and just as expected. There are, however, sound problems. The Wiki doesn't give any special instruction on sound, like installing a new alsa-package or building a kernel module. Can I have overlooked anything? 1. There seems to be no sound at all. 'aplay file.wav' just pretends to play but neither gives any output nor returns the cursor. 2. xfmedia doesn't start at all. An attempt to start the program from a shell, too, doesn't give any output or error message. 3. The ring tone is just a lousy noise. 4. During a phone call, voices are poorly transmitted (noisy), for me as well as for the recipient. Help would be very welcome. Everything is working alright under the default system. So, a hardware defect can be excluded. Thanks <PERSON>
cc61846f-0d2b-0e73-2b3d-65b93a2f8c0c
['1b77b99d-54fb-f1ee-b03e-5271d07b0f58']
Hi, you can directly project the crystal structures onto the eigenvectors from the PCA with gmx anaeig. It may be necessary to create an index file for each PDB file with an index group that identifies the atoms that correspond to the vector components. Make sure that the order of atoms is the same in the PDB as in the trajectories/structures used in the PCA. For multimeric proteins, make sure that the order of subunits is also the same. Happy computing, <PERSON>.
ef43f28b-74d7-7a7b-646e-4dddbef8e474
['1b77b99d-54fb-f1ee-b03e-5271d07b0f58']
Hi <PERSON>, The current implementation an older GROMACS version (3.3.1) can in principle be used, but is slow for systems with many titratable sites na d has some other restrictions. You find it here: http://www.mpibpc.mpg.de/grubmueller/constpH The new constant pH framework will offer higher flexibility and performance but still need some time. The project website can be found here: http://www.mpibpc.mpg.de/grubmueller/sppexa Best, Thomas.
673de1e6-b8f4-1aba-30e6-5204e30dd2c9
['1b79da29-9b3d-ef55-d34a-1b72f778bafd']
Hi <PERSON>, I wanted to know if you were able to see something else about the problem we had in the BBB when using LWIP and enabling cache ( the program freezes). I can tell you that here we were using BBB rev. A5C and had this problem, but now we could test this with a BBB Rev C, and it successfully works with cache enabled (using the same sdcard in both boards, one works and the other doesn't). Greetings
41940723-7791-76ee-94e5-7b040dcbcfc4
['1b79da29-9b3d-ef55-d34a-1b72f778bafd']
Hi, i was using some c++11 concurrency classes on ARM (with gcc 4.9.2) that have an underlying use of libgcc atomic functions, for example __gnu_cxx<IP_ADDRESS>__atomic_add_dispatch. In its implementation, (in the file gcc-sources/libstdc++-v3/include/ext/atomicity.h) I can see that the macro _GLIBCXX_ATOMIC_BUILTINS is not defined, and therefore the builtin implementation of __atomic_fetch_add is not used. Instead, the mutex implementation is used which makes it significantly slower. As far as I could see, if I explicitly call one of these atomic built-in functions they work ok and the correct atomic assembly is emitted (using ARM's ldrex and strex instructions). So, the question is why this _GLIBCXX_ATOMIC_BUILTINS macro is disabled. is there a reason why we aren't using the atomic builtins, or I need to enable it. How can we do the enabling? Is the implementation incomplete? Thanks
4cc9cbdf-f5ba-8bae-9695-1609d408ebba
['1b7fa413-d4b3-8bb9-ea09-5a3bfc8c2228']
I know some have been waiting for further loggings from me, from my cliff top DXing on the Oregon coast. I just haven't been able to keep up with all the activity found on the daily 35-45 minutes of Perseus WAVs recorded from Sept. 9-12th! As I've mentioned, I'm on a family vacation so other activities come first. I'm about 75% through the DX for the 10th. When that's done I'll move on to the files for the other mornings. I've been kept busy just digging out signals that didn't appear on the 9th's recordings. There's lots of interesting stuff on these Perseus WAVs made at the cliff, with the compact, broadband FSL antenna! BTW, I measured this cliff's height as 220 feet above sea level, using a GPS. <PERSON> Puyallup, WA DXing at Cape Perpetua (near Yachts, Oregon)
55e73fd2-47aa-79d3-df30-f37795455605
['1b7fa413-d4b3-8bb9-ea09-5a3bfc8c2228']
Using the ability to "rewind the tape" on the entire MW band with the Perseus SDR, it took me 16 hours over two days to thoroughly review the files for the <PHONE_NUMBER> UTC period of September 19. Whew! Having the luxury of going over and over even the weakest signals with possibilities meant that I was able to log many more stations than during "real time" DXing. My logs from the 19th alone are below; full details plus a few audio clips and tips for using Perseus are in the newest post on my blog. It was very much an Asian evening, with a handful of South Pacific stations included. I didn't as much as hear a whisper from DU stations, despite using JB's SW Conti Super Loop up until 1030 UTC. 73, Guy Atkins Puyallup, WA DXing from Orcas Island, WA www.perseus-sdr.blogspot.com
a1bbe0ec-7e50-59c2-3ae2-25c45a9f6a40
['1b9a51f2-19d9-34f9-30a0-29dd18e11186']
Maybe I've missed something in the updates, but after doing a recent svn to my django-trunk, I no longer get a return from the following command: python manage.py syncdb. It just sits there and sits there. I have this hosted on Dreamhost and the other Django sites on different servers that I haven't upgraded the src with yet seem to work just fine. I am using "mysql_old" and wonder if that is causing any problems.I am using revision 8068 Thanks in advance! <PERSON>
d6223ecd-949b-51c9-2531-40f456fbb24b
['1b9a51f2-19d9-34f9-30a0-29dd18e11186']
Thank You All for the help. I ended up using <PERSON> suggestion and installed my own version of python 2.5.1 and a newer version of MySQL- python-1.2.2. After that all I had to do was change a few lines of code, and Voila, working like new. Thanks again for the push in the right direction. <PERSON>
61ca756c-15a9-4b06-94a0-f39e5a6edc7c
['1b9d12f2-605f-a3ce-752a-7a6a0fc706d8']
Dear intltool KINGLONG GROUP WUZHOU CANGWU XINGYE ANIMAL & PLANT WINE FACTORY ( http://wine88.nease.net ) is The scale of Southeast Asia bigger health care wine produces factory. We manufacture "DINGTAPAI" brand series health care wine with high quality and carefully selected ingredient by advanced technics and equipments. Proving by the science, there are more than 20 kinds of amino acid, which are necessary to human, the wine has a strong aroma. It has the quality of good taste, mild alcoholicity and rich nutrition. Drinking it can rise your energy, release fatigue, be good for blood circulation, skin, liver and etc. It is really a top product in wine. The products have been selling well to Southeast Asia, Japan, Europe regions etc. We provide ANIMAL & PLANT WINE with good quality and the best price. looking forward to further cooperation Best regards Mr. <PERSON>
d1956bdf-9bef-5cf7-c87d-73cda1329595
['1b9d12f2-605f-a3ce-752a-7a6a0fc706d8']
Dear pkg-vim-maintainers KINGLONG GROUP WUZHOU CANGWU XINGYE ANIMAL & PLANT WINE FACTORY ( http://wine88.nease.net ) is The scale of Southeast Asia bigger health care wine produces factory. We manufacture "DINGTAPAI" brand series health care wine with high quality and carefully selected ingredient by advanced technics and equipments. Proving by the science, there are more than 20 kinds of amino acid, which are necessary to human, the wine has a strong aroma. It has the quality of good taste, mild alcoholicity and rich nutrition. Drinking it can rise your energy, release fatigue, be good for blood circulation, skin, liver and etc. It is really a top product in wine. The products have been selling well to Southeast Asia, Japan, Europe regions etc. We provide ANIMAL & PLANT WINE with good quality and the best price. looking forward to further cooperation Best regards Mr. <PERSON>
d6f2a80c-944f-5beb-2e7a-1445f3335ce2
['1ba8a22b-ce6e-8d14-b25c-0dc629d8e4a5']
<PERSON>, - Time Division Multiplexing (TDM) imply partitioning the bandwidth of the channel connecting two nodes into finite set of time slots. - Time Division multiple Access (TDMA) imply partitioning the bandwidth of a channel shared by many nodes, typically an infrastructure node and several mobile nodes, where each node gets to access its dedicated time slot. Hope this helps <PERSON>
fe27829a-ac1d-0523-6fae-f3227214700d
['1ba8a22b-ce6e-8d14-b25c-0dc629d8e4a5']
Kostas, PLCP is added to any 802.11 MAC frame transmitted, yes including RTS, CTS and ACK control frames. And CTS and ACK frames transmission starts right after SIFS, even when it is middle of a slot. Best Regards
cd478442-f606-14be-ddcd-6c8e0119eacf
['1bae64e1-3339-2850-aa00-a7eced20bdca']
Thanks for your answers. A big Problem is, that ReactOs doesnt support alpha-icons. With this the icons would look much smoother. Also the antialiasing would profit. I think it will take a VERY long time, till ReactOs supports alpha.icons. It´s also not really important for the functionality of an OS. The folder-icon is still in development. But i swear: in photoshop it looks better. Many details get lost, when converting to an .ico file. Thats the reason , why the "my computer" is more like "my monitor". Today, who has his TFT-Monitor standing on his computer. Exactly - nobody. The Microsoft "My Computer" in W2K represent those old computers. SO, i don´t like the idea of having the "computer" in the icon as well. Also in the Microsoft icon, the most you see is the CRT-Monitor. And that Monitoor represents the WHOLE computer, doesnt it. so, till next time! rené
4c8574ff-8863-f5a9-b6cd-bef8647f8ee7
['1bae64e1-3339-2850-aa00-a7eced20bdca']
Dear ReactOs-Team, first of all i have to tell you that you make a great job. Go on with it. Unfortunately i have no skills in programming, but in computergraphics. Of course the programming is now more important than the design. So i think that you dont pay the design much attention now. But in this point i can help you - if you want. I started to make a new icon set for ReactOs (click on the link). I tried to make it different from Windows, but not too different. Just enough to see a difference, but also to understand immediately what the icons stand for, if you know Windows. They look very serious and have not that "colorful-mainstream-children-windows-xp" design... :-) Notice: The icon for the unknown file types are changed right now by me. As you can see in the forum on the website, there were advices for a better icon. I´ll follow them. http://www.rene.boyke.de/newicons.jpg If you want, take a look. There are just three on the website. If you dont like them, i will stop designing more. If you do like them, i will go on with the work. Of course they are all for free. Here is my eMail, if you are interested: <EMAIL_ADDRESS> greetings
83b38f9b-1e2c-4caa-2288-f35711c72737
['1bc42737-f140-10a8-9e8e-cf5420f90247']
I'm interested in supporting redlining and annotation functionality for my document types and I'm looking for some guidance on how to best approach it. Thus a few questions: Is there anything that XMLMind supports out of the box for any of the supported document types? I've loaded an XHTML document with del and ins elements and it displays nicely, but I don't see how this would be maintained during an editing session. How would I best approach a test of concept implementation? Is replacing some macros sufficient? Do I need to look at the Java API? Which parts? Are there other redlining examples that I could be looking at? Is there anything on the roadmap for future XMLMind versions? Thanks in advance for all your answers. Regards, <PERSON>
08d97e55-abb7-3ea5-79ca-384cc02183d5
['1bc42737-f140-10a8-9e8e-cf5420f90247']
I would like to use variants ala DITA (e.g., audience="general manager"). At the same time, I would like to provide style sheets that show the document with all variants as well as just a specific one. My current thinking is that it would be nice to just provide one style sheet for all specific variants by prompting the user for the variant value(s), let's say "manager", and use that value in the style sheet to filter the content to be displayed. Is this even possible within the architecture of XMLMind? If yes, is there an example to look at? Or, are there other and better ways to accomplish the same? Again, thanks in advance for all the answers. Regards, Andreas
0a413d2d-4efd-325d-50a4-799ed095340b
['1bcaeb5c-f504-2455-b354-333ea1e164cb']
List, I'm looking at the xfree server in cvs and there isn't any patches just a couple of c and h files. Does this mean that I need to replace the ones in the Xfree86 source distro with these and recompile? I'm working on some stuff and would like to use the newest xfree server if possible. thanks, <PERSON>
f8bfd9e1-bef1-f880-1dfd-4500b804a608
['1bcaeb5c-f504-2455-b354-333ea1e164cb']
Hello <PERSON>, I'm trying to understand how initrd works on Xebian. I know that root is mounted on the fs image on the E drive but I don't understand how that could be when the E drive itself is not mounted. I see that E drive is mounted when initrd is ran, but once the rootfs is replaced it isn't mounted anymore. Is it possible to unmount (or swap) a filesystem when another filesystem is using an image (via loopback)? I'm missing something here, anyone have a sec to explain it? Thanks, schu
a7ce245c-aee6-0945-3154-ed8feb1c5fe9
['1bdbc46e-4900-d335-f236-7ef827ae101a']
Hi all, I'm looking for a person with the following experience: • Experience implementing MOSS as a solution for document management, workflow and image capture. • Experience with the recommended Imaging software to integrate with Sharepoint. • Experience using ASP.Net to interface with MOSS documents. In addition, if the person has experience working with public sector organizations to define and develop implementation strategies for MOSS that is a plus. Thanks, <PERSON>
b43a1fdb-e00c-ec52-babc-51f805b635ec
['1bdbc46e-4900-d335-f236-7ef827ae101a']
Looking for two people for different projects: 1.) One person part time architect reviewer 2.) One person lead developer Looking for a person with the following: 1.)Four year degree in CS or equivalent 2.)Experience in developing public facing websites using Sharepoint 3.)Extensive experience in modeling workflows 4.)Experience in planning and presenting architecture details prior to development. 5.)Excellent writing and communications skills Please contact via email <EMAIL_ADDRESS> Please include resume and rate Thanks
cce9deec-b9f1-70bb-55ab-c9ea3d6d92cc
['1beaa395-5f83-4b46-06f7-8cbf692650e3']
Hi list, not sure if this was mentioned here before (I hardly find the time to read the LA* lists regularly these days), but someone from IRC let me know that the VST/VSTi plugin suite by <PERSON> has been released by its author as source, see here: http://sourceforge.net/project/showfiles.php?group_id=230958 I don't have a Windows-based audio system so I cannot judge about the quality of his plugins, but I am pretty confident (from what I know and read about his work) that these could be pretty fine, perhaps an interesting alternative or addendum to, say, the swh-plugins. Should be a cornucopia for interested LADSPA/LV2 porters, anyway :-). As it seems, there is already a binary of a ported package at http://linux-vst.com/download/mda_linux.tar.gz No idea who managed to convince <PERSON> to do this, but big thanks to both of you for arranging this! This should be a quite valuable resource, for DSP coders and finally also for users. Regards, <PERSON>
b860aa57-e2be-81ee-8502-0c270cb2db70
['1beaa395-5f83-4b46-06f7-8cbf692650e3']
Dear all, this should have come one month earlier, but such is life..anyway: This mail is to announce the calls for papers/music/etc for the 4th International Linux Audio Conference (LAC2006). See http://lac.zkm.de/2006 for more information. LAC2006 will take place 27-30 April 2006, again at the ZKM | Institute for Music and Acoustics in Karlsruhe, Germany. We have tried to simplify things a little bit since LAC2005. There are calls for papers, demos, workshops, and music. The former category BOFS has been merged with the workshops. There is no call for project notes anymore; instead we have the call for demos now. The call for posters has been discarded. We hope everybody agrees that this is an improvement and we are looking forward to many interesting submissions for LAC2006! Please feel free to forward this email to anybody who is interested. Thank you for reading! <PERSON> and <PERSON>
04c134e2-846d-aee6-9735-01838c41b58a
['1beee979-df9e-4f9f-ed0e-cb09f2c005d9']
In the past few years, a couple of algorithms have been released which do a better job for tablature. Since Lilypond uses one of the most naïve ones (not much different from the one that TuxGuitar uses), it might be time for an update. Is anyone working on this? I haven't been able to find any information on current projects. <PERSON>
8d612cad-32a4-3d1e-9d96-c4e6adacd321
['1beee979-df9e-4f9f-ed0e-cb09f2c005d9']
This is probably what a lot of you are waiting for ... an actual question about MiKTeX ... Whenever I use pdfTeX, the page length of the .pdf file comes out longer than 11 inches (probably 14 inches). pdfTeX on my system at work doesn't do this, so it has to be something related to MiKTeX. I've checked some of the documentation but haven't found anything useful. So how do I change page length? (A minimal example would be: Hello\vfill\eject\end ) Thanks,
c2c10ad7-9894-62aa-8ef1-e76248d7f75e
['1c15d9ac-4e52-dbb3-f97b-db678baca6ab']
Chaps, I've created a few bug reports (with patches) related to thread safety issues we have found while using Equinox in SpringSource dm Server. I felt that it wasn't appropriate to create massive patches so I focused instead on point fixes, but I wanted to briefly outline what I think is a fuller solution to the highlighted issues. So far we have identified three main issues: * BaseStorage.getNextBundleId() performs unsafe increment (https://bugs.eclipse.org/bugs/show_bug.cgi?id=247520) * StateImpl access bundleDescriptions in an unsafe manner (https://bugs.eclipse.org/bugs/show_bug.cgi?id=247522) * Framework access BundleRepository in an unsafe manner (https://bugs.eclipse.org/bugs/show_bug.cgi?id=247521) For the first issue, I think the patch I supplied is fine, the solution there is pretty straightforward. For the StateImpl issue, I attacked only the issue around accesses to the bundleDescriptions field, but I think that accesses to all fields need to be protected. Moreover, I think that all the state of StateImpl needs to be protected by the same lock to guarantee atomic updates of the internal data structures. In Framework, all calls to BundleRepository.add were not guarded by the bundles lock. I made sure that all accesses to bundles are correctly (I think) guarded. However, I feel a better approach is to make BundleRepository itself threadsafe. The JavaDoc of BR currently states that it must be synchronized before access but obviously it is easy to forget, as is the case currently. Also, I notice that Framework.bundles is used as a monitor for a lot of operations that don't directly manipulate its state and it can also escape control of Equinox via Framework.getBundles(). This is deadlock prone. Ideally, the BundleRepository would remain an Equinox internal data structure, maintain its own thread safetry and not be used as a monitor for other operations. Replacing the use of Framework.bundles for synchronization with something like Framework.bundlesMonitor will help here. I haven't had to chance to explore the code further, but I'll raise any other issues as I find them. If you want to make these larger thread-safety changes, I am happy to submit patches that address them. Regards, Rob
5c0245ef-5830-5906-9131-fbec4e5a9367
['1c15d9ac-4e52-dbb3-f97b-db678baca6ab']
In our tests for dm Server we are firing up an Equinox instance many times in the same VM and after the first invocation (which loads Equinox types with the app ClassLoader), our tests fail because Equinox types are loaded with an EquinoxFWClassLoader. This behaviour cannot easily be stopped because the FrameworkProperties.inUse() method always returns true. The issue for us in using EquinoxFWClassLoader is that we get test failures due to class incompatibilities. In our case, we load StateObjectFactory in the first test run, and for that run the instance of State that the Equinox types see is the same as those seen by StateObjectFactory. For subsequent runs, Equinox types are loaded from the EFWCL and are not compatible with the types that dm Server has previously loaded. The simplest solution for us is to be able to clear the FrameworkProperties. A better solution for us is to prevent the use of EFWCL. <PERSON>
b4aabae0-7c95-6721-0100-f184aadde6ef
['1c173bcd-2d82-8838-502d-cf9ffe00d135']
I've been tagging peaks (natural=peak) with the key prominence=<topographic prominence in meters> Prominence is a natural feature with a use similar to elevation. When I see ele=*, I know how high the top of the peak is, but not how tall the peak is compared to the surrounding land. For example, a hill in my valley may have ele=2000m, but it isn't a mountain: it's a 300m hill that rises out of surrounding land at 1700m. Prominence is calculated by subtracting the elevation of the lowest saddle (or "col") from the elevation of the peak: https://en.wikipedia.org/wiki/Topographic_prominence "The prominence of a peak is the minimum height necessary to descend to get from the summit to any higher terrain" or "the height of the peak’s summit above the lowest contour line encircling it but containing no higher summit within it." Both of these definitions are the same for all peaks except for the highest peak on a landmass, eg Mount Everest in Eurasia: in this case use the second definition, which means that the tallest peak on a (super-)continent or island is the same as it's elevation. This started when I became interested in "peak bagging", where hikers and climbers record the peaks they have summited. There are separate categories based on the prominence of a peak. Gunungbagging.com in Indonesia lists elevation, prominence and names for many peaks here in Indonesia, and the site authors gave permission for the data to be added to Openstreetmap. There are other lists of prominent peaks for the rest of the world, but please check if you can use the data based on the license, before adding it to OSM. Elevation and prominence can both be calculated from SRTM data, eg by using Opentopomap tiles and finding the highest contour lines around a peak, and the lowest near a saddle. Prominence and elevation can be calculated by computer with good data, but for my part of the world the SRTM data is not high enough quality to get good results without cross-checking against aerial imagery. Also the calculations are not simple, and are not precise for sharply pointed peaks or deeply carved saddles, therefore I believe it will be useful to include this data directly in tags. I also find that calculation the prominence of peaks has encouraged me to add more ridge lines and saddle points (with elevations), which should make the database more useful in mountainous areas. Do you think I should write up a formal proposal for this tag? <PERSON>
4210515e-0ce5-afa8-fda1-9bfbbd21aeaf
['1c173bcd-2d82-8838-502d-cf9ffe00d135']
I've just updated the wiki page for the in-use tag "place=archipelago". The main changes are: - Described an archipelago as a named chain, cluster or group of closely related islands - warning against creating giant multipolygons by mapping the whole Philippines as a single multipolygon relation - warning that "all the islands in this sea" is not a useful archipelago; eg Macaronesia (=Canary Islands, Azores, Cape Verde Islands...), "all the islands in the Aegean" - Added info about benefits of using multipolygon relations vs cluster relations (the later is not supported by most map users) - mentioned that some are mapped as nodes (although this is not preferred) The page already suggested that archipelagos should be mapped as a relation that includes all of the coastlines of the islands as members of the relation. I believe with 740 uses this tag is clearly "in use", but not "de facto". Any comments or suggestions for improvements? <PERSON>
10ef26d1-5b9a-f37d-9a70-9a6d74b42031
['1c206871-174a-a781-0e04-8a5cb643331c']
I am really not comfortable doing all this stuff in finalize. Why use finalize at all? If someone forgot a close then he has to find and fix this in his code. Darn. Cannot find the reference I am thinking of why using "finalize" usually is really a bad idea. Was it from Bloch? Can't remember. cheers, <PERSON>
d39161c2-c962-a032-38eb-b0aa71bb53e3
['1c206871-174a-a781-0e04-8a5cb643331c']
Hey folks, I am looking for some more information on the index management of the searchable plugin. I have multiple machines and preferably I would like only one to generate the index (as usual) but then distribute it to the other machines and switch over to the new index. Anyone ever done something like that? Other opinions or pointers? cheers
3d309916-6f3f-bfb9-f654-391e3b4a95bb
['1c24539b-24fb-30c8-3195-770956374c97']
Are there any experts out there on knoda and hk-classes? I am using knoda as a front-end to mysql. I am still learning about it. (I have worked with databases for many years so I mean that I am learning about knoda and hk-classes). I am testing a setup with a form using two linked tables. The tables are 'sales' and 'products' I have set up the relationship all right as far as I can see. For each field on the form I have to select from two datasources. This was the first surprise. I expected to have just one datasource available showing all the columns from both tables. The choice is between A. sales B. products [sales] Using A I get only the columns in sales. Using B I get only the columns in products. Second surprise. I would have expected to have all the columns from both tables available to me from at least one of the datasources, which is how it worked when I was using Paradox. I can't seem to find any way round this. Any pointers or suggestions from anyone please? <PERSON>
e0e1682a-7d88-a3f4-8b9d-27f0cbb59b84
['1c24539b-24fb-30c8-3195-770956374c97']
I already have set up one web site, mainly as a learning medium. Although I have used computers for years (and years and years) I did not have any experience of running a web site. Now that I have a better idea I am considering setting up another one where I can really get to grips with it all. I must keep the original one since friends and family know the URL and are now used to it. Mustn't confuse them. So, any suggestions anyone? Preferably one in the area. I know that that is not vital, just like to keep things local where I can. And one that will not be scared off by the word Linux. Thanks <PERSON>
3451263f-06c9-0bbc-634d-03cc8b26db96
['1c4a09ae-4ef6-cb8f-a5b3-6ee1c920f4b9']
Some of our customers are getting crashes in what appears to be ActiveState or Perl code. They are running both W2k and W2k3 with ActiveState 5.8.0.806. The crashes are happening very rarely. So far we have been unable to reproduce them. Below is a crash dump analysis generated using WinDbg. Looking at the analysis it appears that ActiveState or Perl has started a new thread to access the Win32 api to read version information (uname). Why could this crash? Anyone experienced similar? What can I do to fix the issue? We are also seeing occasional crashes in other win32_xxx functions called by Perl but uname seems to be the most popular. Regards, <PERSON>
3af32433-6f5c-6a51-dabc-e0d51bd791b7
['1c4a09ae-4ef6-cb8f-a5b3-6ee1c920f4b9']
I'm seeing a crash in ActiveState perl during perl58 unload. This appears to be caused by ActiveState calling the SetConsoleCtrlHandler Win32 api with an address which during unload can be invalid. Windows attempts to call the control handler and the process crashes. Anyone know why ActiveState is calling SetConsoleCtrlHandler? Does it always call SetConsoleCtrlHandler or is it calling this Win32 api as a result of some perl code? I cannot find any direct reference to SetConsoleCtrlHandler in the perl code I'm working with so what perl code could make ActiveState call this Win32 api? Regards, <PERSON>
f83653ac-b829-1e75-fe49-cce4479617bb
['1c4e3a48-2e86-c757-8b99-393e9541fc8c']
Dear colleagues Field 264 was added to MARC format to for RDA, but I do not see any statement that it is restricted to RDA records or fields The documentation at: http://www.loc.gov/marc/bibliographic/bd264.html has only examples which fit RDA In the hybrid environment, it would appear that 264 could be used in AACR2 records also If using 264 in AACR2 records would it be correct to continue using AACR2 standard abbreviations such as [s.l.] and [s.n.]? <PERSON> <PERSON>
f17ad83f-7881-6a46-b3d3-3e724b710fd8
['1c4e3a48-2e86-c757-8b99-393e9541fc8c']
Dear Colleagues I have followed the recent discussion of field 264 with much interest. Here in Israel we have decided to wait until after April to decided on when to implement RDA nationally and what national decisions to take. As a teacher of cataloging I find it very difficult to even know what to teach when there are still so many different views of what to do in RDA+MARC, as in the 264 discussion. Worldcat is also a problematic source of RDA examples because everyone seems to be doing something else. Perhaps sometime in the next four months LC, PCC etc. could get their act together and put out clear guidelines or best practices on some of the major common issues (and what could be more common than recording imprints?) <PERSON>
52c63dfd-3542-e34f-f54e-f9e4051fc98c
['1c57ecc9-a4d7-25ae-8ae6-28a91adc7849']
<PERSON>, I believe that I'm running the version that comes with SuSE 11 and updated to SP3. v 1.36.0-12.3.1. I'm using it, because I thought that the easy installation of many of the necessary libraries/tools might make it convenient and easier to build gnu-radio. All of the others worked fine, after I exclude the test for qa_constellation_receiver.     100% tests passed, 0 tests failed out of 175 After the test failure, I've not done anything further, because I didn't think I could. <PERSON>
b88c36ef-7873-ae00-1cb9-43d2594a8874
['1c57ecc9-a4d7-25ae-8ae6-28a91adc7849']
<PERSON>, I followed your instructions for a Windows 7 Ultimate x64 install. This worked the first try. Thank you!! I'd been following the same sort of path you did while I was on vacation, but was trying to make from source for a x64 implementation. I had something close, but couldn't figure out why the WX-GUI part didn't show up. Now I'll go and uninstall the working version and try to build it from scratch again. I've downloaded all of the x64 libraries that are the same as the working x86 version and will try to build it again. Watch for another note on this. <PERSON>
498a6194-181f-db10-3c71-cbde108c0edf
['1c59fc71-477f-a3ae-6309-96a8db4c3f7a']
Hello, I was wondering if there was a log reporting tool that would make it easier for our clients to see the logs. Some of them are not technically proficiant with linux, and therefore might not know how to read the logs. Is there some type of gui or webbased tool that they can easily see what is in the logs? I know about the one in webmin, but didnt see if it generated a report on what sites were visited. Glen
5faad4f0-d192-caf0-eb6e-8b70098f39a5
['1c59fc71-477f-a3ae-6309-96a8db4c3f7a']
Is there a way to make a group that is allowed internet access and everyone else is disallowed? I would like to create a group on our client nt/2000 servers, just for internet access, but block all who are not in that group. I have squid and nt authentification working right now, but it doesnt block anyone who is a domain user. Are there scripts or is it just a simple ACL entry? I checked th FAQ, and also the mailing list I didnt find anything. Thanks, <PERSON>
1f96f9c5-dd09-b8c8-1c07-1445ed38045e
['1c5d5823-9d8a-b407-48ca-55d5b43542cc']
NetBeans 3.3.1 I have myJavaBean (a module form developed) copy and pasted in other form. I delete this myJavaBean from the form. I change myJavaBean. I copy and paste the updated myJavaBean into the form. 1st) it appears as updated. 2nd) some time later, when I edit the form where myJavaBean is, this one appears as the old myJavaBean used to look. Any ideas? Thank you <PERSON>
65e730b1-1dce-255f-49bb-3e565fd88fbb
['1c5d5823-9d8a-b407-48ca-55d5b43542cc']
Once again, suddendly something stoped to work. I have a component that is a subclass of JPanel. So, I have constructed a gui module. Untill today I was able to copy this module from the filesystem and paste it in a form (a JDialog sublass, for example). Now, NetBeans does not let me do that. It gives the error on the subject. Still I have this module in other form and it works! Maybe I have pasted it in the right day. I was hoping to have some coherence. If I did something wrong with my module (JavaBean), the IDE should not allow me to use it anymore, anywhere. Any hints? Thank you <PERSON>
d01fbbeb-6fc7-66d1-bc6c-1ff2a5aadc65
['1c62533d-fce0-17ac-8adb-4bbce64261c4']
Dear <PERSON>, The problem is that sometimes the part of the mesh that you want to remove is the largest. I would be pragmatic and use paraview: 1. First run (your code) the connectivety filter such that it does not remove any regions, but labels them. 2. In paraview, find out which region you want and extract it e.g. via the threshold filter. Cheers <PERSON>
cc7ab2f6-a1c2-0ca6-ca83-514f130560ab
['1c62533d-fce0-17ac-8adb-4bbce64261c4']
Hi I would like to use the vtkPointLocator in the point insertion mode, i.e. I would like to incrementally insert points and search for them. The incremental point insertion mode works with the functions: InsertNextPoint, IsInsertedPoint, FindClosestInsertedPoint There is, however, no method to search for the N closest points. I tried FindClosestNPoints with no success. I tried first calling BuildLocator on the locator but get error messages (No points to subdivide). What do I need to do, in order to use the function FindClosestNPoints in point insertion mode? Another function that is "missing" is point removal, although I can probably live with this. Is there a better code for doing what I need? Thanks for any suggestions & help. <PERSON>
c6585a3e-8f4e-6665-4f40-fee17e93b29e
['1c895177-707b-81c8-7a1c-9e7007a400ee']
I have installed the 5.6.1.1 version of net-snmp on windows 2008 Server r2.  After it is set installed and set up, if I send an SNMP v2c trap, snmptrapd automatically exits every time.  It does not matter if this is running as a service, or as a stand alone application.  There are no extra plug-ins installed, just a clean installation of 5.6.1.1. V1 traps are seen, and printed out to the screen, but v2c just cause the program to crash, with the usual unhelpful windows errors. The system architecture is 64 bit, but the 5.6.1.1 download seems to be only 32 bit.  I do not know if this could be the source of the problem.  Due to legal issues, I'm confined to working with 5.6.1.1 until I get approval to try maybe 5.7.  However, i have not seen any indication that the behavior I am experiencing is normal. I have tested many dfifferent mibs and formats for the v2c traps, and everything so far has failed.  Even the simple examples given in the online net-snmp documentation cause snmptrapd to crash. Any help or suggestions would be appreciated. <PERSON><IP_ADDRESS> version of net-snmp on windows 2008 Server r2.  After it is set installed and set up, if I send an SNMP v2c trap, snmptrapd automatically exits every time.  It does not matter if this is running as a service, or as a stand alone application.  There are no extra plug-ins installed, just a clean installation of <IP_ADDRESS>. V1 traps are seen, and printed out to the screen, but v2c just cause the program to crash, with the usual unhelpful windows errors. The system architecture is 64 bit, but the <IP_ADDRESS> download seems to be only 32 bit.  I do not know if this could be the source of the problem.  Due to legal issues, I'm confined to working with <IP_ADDRESS> until I get approval to try maybe 5.7.  However, i have not seen any indication that the behavior I am experiencing is normal. I have tested many dfifferent mibs and formats for the v2c traps, and everything so far has failed.  Even the simple examples given in the online net-snmp documentation cause snmptrapd to crash. Any help or suggestions would be appreciated. Kyle Grundmann
37d99493-02e4-1248-ca61-7a4f8333ce84
['1c895177-707b-81c8-7a1c-9e7007a400ee']
Right now working on an snmp module which needs to listen to bind and listen to a port for incoming traps. In doing so, I am seeing some very strange behavior, that is a little hard to describe. 1. I did most of my development and testing by starting snmpd via command line (not using /etc/init.d/snmpd).  Why, was new to this and didn't quite know what I was dong.  However, I never had a problem.  2.  During deployment testing, I noticed that the module was unable to connect to the requested port (was using 1162). 3.  After extensive testing, I have found, that when I start snmpd via command line: snmpd -Dportsnmp,TrapInterface,todmain,snmp_sess_add -LS0-6d -Lf /var/log/snmpd.log -p /var/run/snmpd.pid I can connect to any port I want. If I try to start snmpd vi the OS (init script), it fails to bind to any port < 32767, generating a "netsnmp_tdomain_transport failed" error message. Simply stopping the process and restarting by typing the snmpd command manually allows snmpd to connect to the requested port and continue processing the module.  Both methods are using identical snmpd.conf, with identical versions of the module, loaded with dlmod.  Both methods are using the same -D flags. Current OS is Redhat 6.1.   Net-snmp version is 5.5. I could always choose a port > 32767, but that was not what the customer requested, so am trying to meet their requirements.  And, it seems like it should work. Any ideas. Thanks, <PERSON>
dcbb4115-affd-f002-6d10-0ca0bb432c0c
['1c8e3a51-8b39-aab5-89f6-6b84800edee8']
All, I am starting to develope contract first web services. There are many cases where there are objects in my system that closely match those that will be exchanged on the wire. Using a tool that would allow me to generate an xml schema from a given java object would give me a lift in writing these schema files. Is there a way to do this other than using the java2wsdl tool? For my purposes, I really just need a java to schema tool. Any ideas? Thanks, <PERSON>
02479dac-c61c-b038-ab77-b1f0a968efc5
['1c8e3a51-8b39-aab5-89f6-6b84800edee8']
Hello, I am having trouble with serializing/deserializing dates. I am having axis serialiaze a simple java bean that contains a java.util.Date. When the wsdl is generated, it appears as a xsd:dateTime. As you know, when you generate the client, this appears as a java.util.Calendar on the client side. The problem that I am having is that when the client sends a date (using java.util.Calendar), it is not being deserialized into a java.util.Datecorrectly on the serverside. Is there a way to specify java.util.Dates to be mapped to xsd:date instead of xsd:dateTime? I found that if I edit the WSDL file manually and chance the datatype to xsd:date everything works fine. What is the best way to do this? <PERSON>
c0e0c0e0-333e-88bf-a69a-edc47a3cc31c
['1c94901d-0e6c-9348-6af4-b523157cfedf']
Hi everyone. First, thanks for the new release, it works very well over here! And it feels save now to invest in a controller that speaks Mackie control for the sake of wiring it to Ardour. Are there any plans to make the spectrogram in the new (and incredibly handy) post-export dialog also available in the editor? The ticket regarding¹ this is around for a while and contains more information, such as examples and opinions why this feature is considered useful. Best, <PERSON>
27c05dfd-e4ac-f6f8-fdd9-15a0f178b412
['1c94901d-0e6c-9348-6af4-b523157cfedf']
Dear list, currently, I am experimenting with my little and dusty knowledge in C to create a trivial LV2 plug-in. As from what I understand, Ardour calculates the latency when stopping the transport. However, I think I came across a use case where updating the latency would also be useful when the transport is stopped (so that when starting to roll, everything is configured correctly). Would it be reasonable to calculate the latency of a track when a plug-in parameter was changed even when not rolling? The LV2 specification requires plug-ins to update their control outputs when `run` is called with sample_count==0. That appears to be useful for that scenario. (I am aware of the tracker and will create a ticket there if it turns out to be reasonable) Best, Lukas
742059f7-67c1-cfef-ed82-85356c1e522d
['1cb5e48c-14c8-3a69-aee5-f47ed0fca004']
Subject says it all. for just about every version of fetchmail from 5.8.17 - 5.9.5, I've been having random crashes when grabbing mail from my provider via IMAP. See below for logging. After the 'Bad Bogus' line, fetchmail dies, the lockfile still existing, and I've had to restart fetchmail for it to continue grabbing mail. This only happens with IMAP. I don't get this problem with POP3. Any thoughts on this? BL.
119735d1-5d4e-6706-6338-fa7a1c61284c
['1cb5e48c-14c8-3a69-aee5-f47ed0fca004']
This is very offtopic guys, so please excuse me, but this was so one-off, I had to bring it up. I was with a friend of mine, going through his checklist (inspecting) for his plane at McCarran Int'l here in Vegas. While listening to the ATC tower with my scanner, I caught the takeoff clearance for a Learjet, and a Citation, off of runway 1L (facing north, towards the Hard Rock Hotel, for those knowing their way around the place). Normally, I wouldn't give this a second thought, until I saw the planes, both of which, on their tailfin, had Tux (yep, Tux!) for their logo/emblem. These flights came from the SoCal area, as LAS is one of those airports private planes have to stop at for inspection (the security hype for SLC. of from NoCal, they would have stopped at Boise, ID). I was just wondering if anyone here knew who they might have been, or if I were just seeing things from having one too many Mai Tais last night. :) BL.
c9deb5f9-57e4-5525-e46f-12abbc635fae
['1cb9ac94-04a9-7ae8-43d1-02901eb5b63d']
Hi, hope all of you will be fine and doing your best for the development of open source community. I want your suggestions and help regarding my project. I am going to start my master's thesis in the domain of Cloud Computing I want to develop a Signature Based Distributed Intrusion Detection System (DIDS) to detect distributed intrusions in Cloud environment. Yes, I intend to deploy it in CloudStack. I want to modify the correlation module to enhance detection capability already being provided by Snort. Can you please help me in selection of a good technique to improve correlation module? Thanks and Regards, <PERSON>
020ef752-5e88-a844-3b49-51165aa89097
['1cb9ac94-04a9-7ae8-43d1-02901eb5b63d']
Hello everyone! I want to develop a Signature Based Distributed Intrusion Detection System (DIDS) to detect distributed intrusions in Cloud environment. Yes, I intend to deploy it in CloudStack. I want to modify the correlation module to enhance detection capability already being provided by Snort. Can you please help me in selection of a good technique to improve correlation module? Thanks and Regards, <PERSON>
fcb20857-6e6d-4ae4-1e29-28c8f5678227
['1cc86bbc-7915-b237-5455-8bba1143da6c']
Hey all- this is something I've looked for a good solution for for some time, and I'm sure someone else has already worked out. Any ideas appreciated. The scenario: I have entirely too many email addresses, several of which from domains that are mine, but others that are not mine, but am unable to get rid of entirely. My freeBSD system is going to become a mail server among other things, to handle mail for several of my own domains. Not a big deal there, have done that enough times...however: I'd like to also pull email from the mail accounts which are _not_ mine, so I can simply use IMAP to my mail server to access all of my different accounts email. In the past, I've used fetchmail to accomplish this somewhat, but that was on a per user basis via user cron jobs. I'd rather avoid adding user accounts (at the shell/system level) for each email account I have. Does anyone know of an alternative way to do this, that would work well for say, a dozen accounts for multiple domains of my own, and perhaps another dozen accounts from domains that are not my own? Thanks in advance, <PERSON>
70f0e2a2-4e61-3471-d7ff-9ff37ca6ea3c
['1cc86bbc-7915-b237-5455-8bba1143da6c']
Hey all..was wondering if anyone knew of a utlity to copy the contents of a text file into an X clipboard buffer? It's possible via the use of xmessage or any other X editor that allows you to select all text, but something command line only would be useful...I'm sure something exists somewhere, but I'm not having any luck as of yet...anyone? Thanks, <PERSON>
26ae6292-d618-7b95-1bbf-39f8d8d64d9d
['1cd19108-3541-4d75-dab5-3f0f6d58dfa0']
Dear <PERSON> I am undertaking a Principal Response Curve analysis and wish to partial out the effects of additional environmental variables, in addition to the usualTreatment (control and teatment levels) and Time variables. A conventional PRC is a type of RDA set to: rda(response ~ treatment:time + Condition(time)) The first level of the treatment is assumed to be the 'Control' and set to zero in the PRC plot. I wish to include additional environmental variables (in this case continuous ones), but partial out their effects to understand more clearly the treatment effects over time. Assuming that 'envvar' is the new environmental variable, is the best structure for this: rda(response ~ treatment:time + Condition(envvar) + Condition(time)) or should envvar be additionally included in a 3-way interaction term? The R code underlying the prc command in vegan is fairly clear, and so it should in theory be relatively straightforward to make any changes. However, I was also unsure how to partial out any effects that an additional environmental variable would have at the 'Control' treatment level. In the prc R code the Control level is deleted from the model design matrix prior to calling the rda command itself? Many thanks for your help <PERSON>
b0b84779-6324-7d74-fe1e-21666cadf7cb
['1cd19108-3541-4d75-dab5-3f0f6d58dfa0']
Dear list I've had a query from a referee arguing that in PCA and other multivariate methods, one attribute (typically spp) should be omitted when the sum of all attributes in each sample is 100%. This might make sense if you only had two or three spp (not that you'd be doing a PCA with so few), but I'm less clear on why it is needed with larger numbers of spp. Many thanks <PERSON>
0ec18f8f-27e0-f688-3a72-27cf56286762
['1cdd2b35-8f86-fee4-7c18-6376cae06051']
Hi, To my understanding, CMP can be used to put a certificate on hold using crlEntryDetails field of the RevReqContent structure. This is achieved by using a crlEntryDetails that contains a certificateHold CRLReason. Is it, however, possible to remove the hold status from a certificate with CMP? Apparently removeFromCRL CRLReason can't be used for this purpose because it is, according to ISO/IEC 9594-8:2001, for use with delta-CRLs only. To me it seems that using this reason code in RevReqContent to remove certificate's hold status is not what the designers of the CMP meant. <PERSON>
fd338340-2851-e540-e8bc-e8a8d66947ae
['1cdd2b35-8f86-fee4-7c18-6376cae06051']
Hello, CMP RFC (4210) refers to [CMPtrans] which is an expired I-D, latest version being "draft-ietf-pkix-cmp-transport-protocols-05". CMPtrans is supposed to define the transport methods for CMP as they were separated from the CMP RFC. As RFC4210 is a quite fresh RFC, I would assume there is an intent to push also CMPtrans to RFC status. Is anyone working on it? <PERSON>
817cc2df-5cf8-8015-c3fb-b312f3aef328
['1ce119a3-534c-2c0a-afe6-41b7ed5da6d3']
In a well engineered system, all of these factors should become a bottleneck at the same time. In reality, network is often a limitation for me, as I rarely deploy Xen servers with local storage - I prefer to use iSCSI shared storage because the ability to migrate a domain is very useful. This means that storage is attached via 1Gbps or multiple 1Gbps ports. CPU, RAM, and disk are all cheap. 10Gbps ethernet is not - yet. Thank You, <PERSON>
677a45d0-fbf9-6085-a4fa-d014cdfdf1eb
['1ce119a3-534c-2c0a-afe6-41b7ed5da6d3']
I don't normally news-post, cross-post, or even post to devel at all, but I thought this was quite interesting. Apparently, Microsoft submitted 20,000 lines of device driver code for inclusion in the staging kernel. Perhaps this is a step towards paravirtualized Windows guests in Xen? That would be quite compelling. http://www.networkworld.com/news/2009/072009-microsoft-linux-source-code.html Best Regards <PERSON>