query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
65cea7c8-56be-46e1-9c6c-50b9c37ab8ca
['21f126e2-b37c-915b-1a4b-65fef27aa7b0']
Greetings, U-M has posted results from our Assignments performance tests on the WG:Performance Confluence site: http://confluence.sakaiproject.org/confluence/display/PERF/Home As I reported in the Sakai 2.5 release meeting, assignments performance improved from a 1-9 second range to a 1-1.5 second range in the new code. Graphs and details are available on the site. My congratulations to <PERSON>for her tool improvements! Please correct me if I¹m wrong here, but it is my understanding that this code is available in the post-2.4.x branch. /linda
757818d2-2f45-190a-0332-86d23e381a49
['21f126e2-b37c-915b-1a4b-65fef27aa7b0']
Greetings, I must apologize in advance for being unable to participate in the next two Sakai QA meetings. The university sponsored me for an Advanced Leadership Seminar this year and our final presentations are scheduled for the next two Wednesday's in a row. Sorry. I am asking <PERSON>, our new HIGHLY qualified performance engineer, to participate in the meeting on my behalf. [<PERSON>, we'll chat about details at our meeting tomorrow morning.] <PERSON>
87f66c98-89f8-8071-d101-6549c4df5da8
['2205ba62-1626-aa66-fe38-aa1d8f258b83']
All, Can anyone point me in the right direction for the serial Gamepad driver(s) that have been written for Familiar? I picked up one of the Hanaho joypad sleeves. I found some dead links to the Stinger driver, and I think this is what I am looking for. Has anyone else had any luck with this device? --Jon
bbd38316-a62a-5efe-9f8b-a11c53954c92
['2205ba62-1626-aa66-fe38-aa1d8f258b83']
All, I should start by saying "thanks" to the familiar community. I tried this out about a year ago but ended up parting with my Ipaq. Now that I have another, I must say I am impressed with the release quality. Also hats off to <PERSON>! When I last kicked the tires Opie was little more than a duplicate Zaurus feed. Thanks for the great tools and applications! I find myself considering compiling OpenAFS client for ARM. A couple of questions for the community: 1.) Has anyone done this? 2.) Should I mess around with setting up a cross-compile environment, or should I take this right to the Skiffs? Thanks in advance, <PERSON>
1f89513b-7f8d-b5b3-b4cf-fe1d6f335443
['2206504f-307e-6529-fc6a-5d062e3f8486']
Hello. We have a number of Sun V20z Opteron systems which support console redirection over serial. I have been trying to install RHEL4 on one of these servers by putting the RHEL4 boot CD into the machine's CD drive - but unfortunately I don't get very far because (I believe) there is a graphical ``splash screen'' that is displayed early on that totally prangs the serial console. RHEL3 works fine (I tested it) as no such splash screen was present. Delving into the RHEL4 isolinux config a bit, I find that boot.msg references ``splash.lss''. Does anyone have a solution for this problem? I'd rather not have to cut my own version of the boot CD with this screen removed. I don't want to have to plug a graphical head in - this partly defeats the purpose of the serial console, but I don't see many alternatives. Ultimately we'll use kickstart and this problem will go away, but for bootstrapping the first couple of systems this is a pain. Regards, <PERSON>.
37fcc336-a9bd-2e08-c1e4-41cbe28ef549
['2206504f-307e-6529-fc6a-5d062e3f8486']
Hi All. This question is perhaps slightly off-topic for this list, but I'm very interested to hear experiences from others about this. We have spent quite a bit of time here setting up some new infrastructure to provide a single-sign-on (Kerberos/LDAP/PAM) and single home directory (OpenAFS) framework. We had found our old method of managing *NIX accounts individually (on our now 200+ Linux and Solaris machines) was not scaling at all well, so the improvement for us in moving to this framework was substantial. OpenAFS was chosen primarily because it seemed very feature rich, mature and had excellent platform support for both servers and desktops. However, a problem has arisen around the use of Oracle. In addition to our stock of general purpose systems (Web, mail etc), we have many machines that run Oracle databases and applications, ranging from systems with a single Oracle DB instance to full Oracle App/DB (RAC) suites. Now apparently Oracle will only "certify" kernel modules that are provided by the O/S vendors directly (for us, this is RedHat and Sun), plus a handful of special modules for (I'm told) certains SANs and the like. Since the use of AFS relies on loading a kernel module, we are concerned that there will be support problems with Oracle. They regard a kernel with an unsupported kernel module as "tainted". Whilst we will endeavour to work through this with Oracle, I was wondering if other sites are running OpenAFS on systems that also run Oracle? If so, have there been any support issues with Oracle, or indeed any issues with stability on Oracle hosts that could be pinned-down specifically to AFS? Has anyone been able to obtain certification from Oracle to run AFS? (Note that we're only planning on using OpenAFS to provide home directories and some non-critical shared areas, not in any way to interact directly with Oracle). We run Redhat Linux 4+5 (32/64 bit), and Solaris x86 and SPARC. One possible workaround would be to the use the AFS/NFS translator, so that the Oracle hosts only need NFS, but I would really only want to do that as a last resort. I'd be interested in hearing any experiences or ideas that people may have on this subject. Regards, <PERSON>.
27c7cfb2-a067-4f94-ea3b-fc7ba48c437b
['220f98c7-2316-f79e-48db-b3367a78dee2']
Hello, I am new to GPG, specifically GPGME. I am trying to familiarize myself with programming against the GPGME C library. I was wondering if it is possible to explicitly specify the use of AES 256 and choose a block mode when using the OpenPGP protocol ? I am sorry if I have overlooked something obvious. I have looked at the 'GPG Made Easy' Reference Manual, and do not see any mention of these types of options. Many thanks in advance, <PERSON>.
1470196c-67ec-b0c4-5847-b28b86e6aac5
['220f98c7-2316-f79e-48db-b3367a78dee2']
Hello everyone, I have a general philosophical question regarding OpenPGP implementations, and I'm hoping that this is an appropriate place to ask it. When it comes to the most actively maintained implementations, it seems that GPG, and GPGME as an API are the de-facto standards. Correspondingly, libgrcrypt seems to be one of the best choices for using a lower level library to provide quality crypto primitives. Observing the standard "thou shalt not roll thine own crypto" philosophy, I have an ongoing dialog with one of my colleagues regarding the risks around implementing a library that would take the output from something like libgcrypt and format it in compliance with the OpenPGP RFC. I have looked around and seen some efforts at doing this (e.g. http://www.cypherspace.org/openpgp/zerucha/ ). The question I pose is this: Given the inherent risks in rolling your own crypto primitives, is there equal risk in terms of say, attempting to secure private keys that are generated using libgcrypt and storing them in an OpenPGP message format. It seems to me that there is tremendous risk here in terms of implementation details, but I'm unable to put my finger on exactly what it is. If anybody has thoughts on this topic, I'd love to hear them. I apologize if this is not an appropriate forum for these types of questions. <PERSON>.
8dec0570-3750-c8f2-53d2-a7c2ef407924
['2211cbef-cc6f-06f5-ff1f-01022741e602']
175 Notebooks Toshiba 8200 P3 850 128 /20/ DVD These are super nice off lease units , $250 is a take all price , give me a call if you need more info All equipment power on tested. Non DOA warranty. Complete as is No warranty on batteries. Complete with A/C adapters. Retail grade laptops. Good cosmetics. Jimmy Gent HiTech Assets, Inc. 304 N Meridian Ave Suite 9 Oklahoma City, OK 73107 Ph: <PHONE_NUMBER> Fax: <PHONE_NUMBER> Instant Message: jgent12
63a3c556-534d-0baf-efc6-9a7498f2ffe7
['2211cbef-cc6f-06f5-ff1f-01022741e602']
These are great off lease PC's if you need a 40 ft container this is your chance . They are all in my warehouse and ready to go 20 pallets Systems 613 Dell mid size desktop GX110 P3 733/128/10/CD/FD $52 each 178 Dell mid size desktop's GX110 P3 866/128/10/CD/FD $61 each 57 Compaq Towers P4 / 1.7 Ghz/20/128/CD/FD $165 each 20 Dell Precision 340 Tower P4 2.2 Ghz/256/80//CD/FD $210 each 44 6579-RBU Netvista P3 933/128/30/CD/FD $70 each 4 6214-51U IBM Tower P4 2.2ghz/256/40/CD/FD $200 each 7 6214-53U IBM Tower P4 2.2 ghz/256/20/CD/FD $200 each 10 8305-42U IBM P4 2.4 Ghz/256/40/CD/FD $215 each 19 8305-RRU IBM P4 2.50 Ghz/256/40/CD/FD $230 each
74fa52f2-afe4-a1db-9b87-5666907c3d8d
['22138c80-b04a-9b1c-45a3-adab035f4c32']
<PERSON> started a poll, so here is my input: I also agree with <PERSON>, <PERSON> and <PERSON> - RSIP is not a good match IMHO for all the reasons stated in the conclusion of the comparison draft. "The RSIP extension mechanism has the largest impact on the existing protocol and is based upon defining the necessary new parameters. " "In addition, RSIP requires additions/extensions to meet several of the requirements. RSIP would also require several framework elements to be added to the MIDCOM framework as identified in section 1.2.3. " Cheers, L-N
5270f38d-bf6e-f612-a7bf-469d3cb6dea0
['22138c80-b04a-9b1c-45a3-adab035f4c32']
<PERSON>, All, I've been through the comparison draft (a bit late, sorry) and I feel there is an inconsistency with regards to the COPS evaluation in requirement 2.1.1. Currently, it is put as a F. But I believe a P would be much more appropriate. I re-read the definitions provided for F, P, P+, T and according to the definitions it is clear that P is exactly the case for COPS in 2.1.1. If you read the text for COPS in 2.1.1, it is said that only part of the requirement is not met (the directionality) but nothing prevents us from going around this problem by defining local policies at the PEP. Yes, this would be inconsistent with the COPS framework, but that is exactly what the definition of P is. I believe the same comment applies to Megaco. The draft, as it stand, would be inconsistent since other protocols were graded in-line with the F/P/T/P+ definitions, while COPS and megaco (in 2.1.1) are clearly not. Cheers, L-N
f20a00c7-33d4-98b6-13bb-8ddf201f9987
['2221e26e-c66d-0d07-7409-caa97bb9efa4']
Hi everybody, One friend developed an in-house “connection pooling”, the main idea of this component is to open one single database connection. The users request for a statements or prepared statements instead of connections. My friend calls his component “connection pooling” and probably it is not a connection pooling at all. I would like to know the advantages and disadvantages of this component in a multithreading application, for example a web application. I am a DBCP and Pool user and to be honest I like them. Thanks in advance. <PERSON>
8e9ccd1e-2ba1-0f54-137b-ebdb72b741d6
['2221e26e-c66d-0d07-7409-caa97bb9efa4']
Hello, My name is <PERSON> and I am a new OJB user. I would like to know if OJB final release will be ready soon, and could I use it in a web application running in a productive environment where we could reached 500 users concurrently? Thank you for your time <PERSON>
3cbc362f-2320-926b-f24b-3f02d63ab026
['2245d893-514d-d2f2-d78f-fa036735a67e']
Hi again. :) I've gone through the past two years worth of archives to find out how to correctly use the "bounce" command to remove the bouncing emails from the list and place them on my bounces list. I've gotten a decent understanding of what bounce does and what to send it, but I'm still not sure where to initiate the command. I was hoping wrapper would somehow do it automatically when a list gets sent out, but I guess not. Is it in a header that goes out with the list that sends the bounces to a bounce user? Where is a good guide about how this all works? Thanks again for your help. -jason
17d28793-09c3-abdf-3b54-5932f89d2356
['2245d893-514d-d2f2-d78f-fa036735a67e']
Hi y'all. I am working on a mail list app that's pretty simple...don't need to let users post, only admins, and it's very straightforward. I tested it out with mail() on 50 addresses, and it took 78 seconds plus massive server load. Part of that may be the mail scanner we have, but I also know mail() opens a socket each time, and that's gonna make it pretty slow too. Does anyone have any suggestions, or even examples of what you've done for this type of thing? The load is about 3000 emails daily, all at once (the client emails out daily specials to their clients), which at this rate would take over an hour to send. Also, and I might just not be paying attention, but is there any word about this list going phpBB or the like sometime soon? That would be great. Thanks again for your help. -jason
209e1dd6-e108-1701-ca74-33f29c224b9a
['22549b77-7d09-c22a-032f-8d603b460e9a']
'Are there game shows that buyers come to?' Trade only shows include The New York Toy Fair in mid-February www.toy-tia.org and GAMA GTS in Las Vegas in mid-March http://www.gama.org/. Consumer shows that also have buyers attending include Origins in June/July in Columbus OH http://www.originsgames.com/, GenCon in mid-August in Indianapolis http://www.gencon.com/ and Chicago Int'l Toy & Game Fair over Labor Day Weekend www.chitag.com. <PERSON>
4c1b957f-8005-bb7a-8acf-1e57b3ca16bc
['22549b77-7d09-c22a-032f-8d603b460e9a']
Thanks, Tom. Also, at the bottom of DiscoverGames.com is a link to the Toy and Game Inventor Handbook, an excellent resource. Keith Meyers has an Inventor Newsletter, sign up at www.gamedesigncentral.com If you can, attend the Toy and Game Inventors Event Conference, Nov 20-21, www.toyandgameinventors.com. Amazing list of 30 speakers, including Keith Meyers. Mary CouzinFrom: <EMAIL_ADDRESS><PERSON>. Also, at the bottom of DiscoverGames.com is a link to the Toy and Game Inventor Handbook, an excellent resource. <PERSON> has an Inventor Newsletter, sign up at www.gamedesigncentral.com If you can, attend the Toy and Game Inventors Event Conference, Nov 20-21, www.toyandgameinventors.com. Amazing list of 30 speakers, including <PERSON>. <PERSON>: ClmJMsoZmGAKdMH0@example.com
5a8cce80-484e-5912-3d7d-e4d59f331e7a
['22579523-58a2-dbab-afed-7b78167757c3']
Hi All, I have what I believe is a simple task: write multiple protocol buffers to a single file and then read them back sequentially. When reading, I should not have to load the entire file in memory but read each protobuf object one at a time (with some buffering). An example application would to be store a large number of documents in a file, where each document is a single protobuf. This functionality seems to be provided in Java using writeTo (OutputStream output) and parseFrom(InputStream input). However, it seems to be missing from Python: http://groups.google.com/group/protobuf/browse_thread/thread/cfe1955729077132/c8ccf86adecf3b47?lnk=gst http://groups.google.com/group/protobuf/browse_thread/thread/838eb489871a92df/b2863c8b9ebfc433 Have people come up with a solution to this problem for Python? One approach would be to port CodedOutputStream and CodedInputStream to Python. I am wondering if anyone has developed an approach to deal with this. Best, <PERSON>
65ccc544-994b-75e8-2f99-c316bea4abec
['22579523-58a2-dbab-afed-7b78167757c3']
I had the same idea (writing the size of the message first and then the message). Here is a simple reader and writer for Python with this idea. Note that I assume the message size is an (unsigned) integer of 4 bytes. For very long messages, this won't work. However, I am relying on the assumption that messages aren't that large. If they are, then you probably need a different format for storing the data anyway (break up into many protocol buffers.) Comments welcome. Mark
49de4afa-32ed-d8e5-7dba-c7a7083917df
['226d7641-4c42-8ff1-3de6-cb552eaffd72']
My problem is that I installed gspca via apt-get (<PERSON>), configured xorg.conf for v4l ... and no /dev/video* with the camera running in "Live Mode". I tried modprobe gspca and ... the module inserted and ... no /dev/video* device opened, either. No /dev/video* , of course, means none of the viewers work. Details below. What I'd like to know, of course, is what I should try next. thanks
a0599cd0-96a0-3da7-43ca-63c56aeb9608
['226d7641-4c42-8ff1-3de6-cb552eaffd72']
My distro is Debian Etch/Lenny . My kernel version: Linux terrarium 2.6.18-4-k7 #1 SMP Mon Mar 26 17:57:15 UTC 2007 i686 GNU/Linux My motherboard is a Biostar Geforce6100 AM2, with a Athlon 3400 X2 dual core processor. I'm trying to save power on a desktop system. So I want to ***suspend to RAM*** and have it return immediately. I have installed whatever hibernate version from official Debian repository is current. Hibernate works, suspend-to-RAM doesn't. I think there's something obvious I'm not doing or don't have configured, but I don't understand the program well enough to figure it out for myself. thanks
6cbbc270-d8e2-8186-f701-02124a66a631
['226df15d-c9e1-f97d-dcd1-c314bb6a4fb6']
When I create 1000 concurrent subscribers, I find that I get SQL Exceptions saying that there are too many open connections. I assume this is because each subscriber has its own JDBC connection. Is there any way to configure this behavior, or is the only way to modify the source code? Thanks.
b8458e4d-ab8a-72cf-0c90-f3dc40bf6dd8
['226df15d-c9e1-f97d-dcd1-c314bb6a4fb6']
Hi all, Does anyone know of a MySQL persistence adapter for ActiveMQ? I've been using the default JDBC adapter, but I'd like to use Connection/J from MySQL for the pooling capabilities, and I'd rather not write a new persistence adapter (since that could introduce bugs). I think this would solve the problem I previously mailed the list about, which was caused by too many subscribers acknowledging messages at once. Thanks in advance. - <PERSON>
5cf152b2-3ee1-739b-d094-49cc5277cabc
['22780bab-7e8c-73e8-1805-9903e7fd664d']
Hey folks, I have a fix here for bug 18606. I think it's decent, but I could use another few sets of eyes to make sure I didn't miss anything, or even that I'm not way on the wrong track. The problem and solution are described in patch 3. In brief, SHChangeNotify notifies windows when a change to the filesystem occurs that the window asked to be notified about. For example, a window might tell SHChangeNotifyRegister "Inform me about changes to C:\". Bug 18606 occurs because the new directory is passed to SHChangeNotify as '/home/user/.wine/drive_c/New Directory', which doesn't match the IDList created for 'C:\'. This patchset fixes that by converting all paths in SHChangeNotify to UNIX paths, resolving symlinks etc, and comparing the ID lists created from those paths. That way, the paths being compared are standardized and compared fairly. Unfortunately, this is hard to write tests for. You'd have to make an ID list that is significantly different from the one SHChangeNotifyRegister will create, but that actually refers to the same directory. This is easy in Wine, but less obvious in Windows. Suggestions for how to accomplish this are welcome. Logically, though, I think the problem and solution make sense. Please take a look. Any feedback is appreciated. <PERSON>
650cbc38-c60a-221f-388a-87ade454aa4e
['22780bab-7e8c-73e8-1805-9903e7fd664d']
Hi, I'm trying to take a record stream which has been recording for some time, and empty its buffer. I'm especially interested in causing the return value from pa_stream_readable_size() to become 0. I've attached a test program here which uses pa_stream_flush(). The documentation for pa_stream_flush() says, "This discards any audio in the buffer," but it doesn't actually seem to work for record streams. Any hints? <PERSON>
f5ed7ac0-01e6-130a-a5f2-47f35ee27ebc
['22799c5b-1c67-e72f-ca30-446a55e47250']
Hi all, I have a situation in which I am reusing a JSP in multiple portlets in the same war file. Here I need to do some operations on the jsp based on which portlet is using that jsp.. Is there any way to get the name of the portlet, or any property specific to individual portlets from the jsp? Thanks in advance, <PERSON>
be30ff58-90d7-890a-f090-4f006d2190aa
['22799c5b-1c67-e72f-ca30-446a55e47250']
Hi, I have successfully implemented popup in WPS 5.0 and SPF. for the first problem, u need to create a theme which has no header and footer. (essentially, the default.jsp will have only <wps:screenRender/> tag and other bare minimum code in it.) once u do this, make sure of the following things : * the popup content is a portlet by itself. * you are using custom urls to access the popup page containing the single portlet. I assume you might be using javascript to pass the request to the parent window. Most probably this will solve your issues. Thanks & Regards, <PERSON>
dbd89a1a-5de3-879d-cce1-12edcda9db3f
['22812403-5d3b-a797-7fe4-6e0209c4ed04']
Hi All, I've added a "Removed all tracks" to the Tracks menu on my local build. It's also possible to do this by doing a Ctrl-A (to manually select all tracks) and then to use the existing "remove tracks" option. I just find it's easier to have a one-click method to clear out the workspace. It uses the same mechanism as "Remove Track" and the action is undoable. People's thoughts on this? Is this a feature you'd like me to share? The one issue here is the size of the Undo history. If Audacity is used for repetitive capture - export - clear - capture ... cycles I forsee issues with Undo history! Can I put a cap on the Undo buffer size/depth? Regards <PERSON>
9e018814-11ca-0291-4c2b-488a93f20e78
['22812403-5d3b-a797-7fe4-6e0209c4ed04']
Hi All, Introductions first: My name is <PERSON> - I see there is <PERSON> as well. <PERSON> is pronounced <PERSON> (as in "The <PERSON>" - though <PERSON> really has no business assuming my title ;-) ). For my daily bread I write drivers for satellite decoders in C, though I've worked on C++ MFC based apps in the past. I'm living in Durban, South Africa and studied Elec Eng at Wits Univ in Johannesburg. Intentions next: I've got a niche application for Audacity. I'm working on a project to digitise around 4000 audio cassettes. The strategy I'm using is to run 8 auto-reverse tape decks in parallel, feeding into a multi-channel Edirol UA-101. There are high-speed tape digitisers available, but they are too pricy. So in the first instance I'll be using Audacity just for "dumb" capture to wav. If you guys can recommend a better application I'm all ears! I'd prefer Linux, but I didn't have much success getting the UA-101 working, so Windows it will have to be. I want to make some minor mods to the Audacity GUI to stream-line the process. For example, when one hits "record" Audacity creates 8 tracks and off it goes. I'd like to offer some kind of wizard so that track names can be set up before the recording starts. Once the raw wav is captured I want to use Audacity to clean-up the tracks - remove silence, apply noise reduction, etc. I won't personally be feeding 4000 tapes through the system, the intention is to give jobs and training to two or three semi-skilled individuals. The whole thing will therefore need to be reasonably solid, straight-forward and easy to use. Finally, requests: 1. In terms of "dumb capture" - would I better off with something other than Audacity? Remember I need an application which can handle the multi-channel UA-101. 2. Assuming that Audacity is a good starting point, please can I have some pointers on where dialog boxes live and how to create and modify them. 3. I'm trying to keep this legal without spending lots of $$, so I'm running VC++ Express, which doesn't have a "resource editor". What are my options in terms of creating and editing the wxWidget based dialogs and such used by Audacity? And of course, some quid-pro-quo: I work in a commercial environment, and I'm of the opinion that commercial software development could learn a lot from the international coordination of open-source projects like Audacity. As I increase my expertise on your project, I'm keen to give back :-) I've posted once or twice already, and your response has been fantastic. Thanks guys! Kind regards <PERSON>
9176d1c3-0f8d-83f4-7026-b23a32efa72e
['229f72c3-2cfe-c40c-b2c7-fc8781159afb']
I recently installed a Pertronix ignition into my 73 Sonett III with excellent results. I also have dealt with the customer service at Pertronix and although they solved an issue I had I wasn't 100% happy with how I was treated. I installed another of these units into a boat that I have and it fried after about 10 seconds. The reason was a faulty unit out of the box but customer service made it seem like that wasn't possible and that I had done something wrong or wired it incorrectly. They did however honor their warranty and sent out a new unit and I installed that and it worked great. I guess they probably deal with a lot of morons and people that have no idea that red wires go to positive and black to negative so I should have expected them to blame me for the fault. None the less they did fulfill their warranty promise. Very happy with how the Sonett runs with the Pertronix installed. -Matt
f2a56461-a71b-8584-3f0e-3bbf9129a1ac
['229f72c3-2cfe-c40c-b2c7-fc8781159afb']
I had a local decal shop make up an air cleaner decal for the Sonett III that has the 'mann' logo on it. These have vinyl lettering as opposed to the original stenciled/painted letters. The 'mann' logo is slightly larger than original, and the color is a little more red than the original but it looks pretty good. She will make more if anyone wants one. The price would be $35 if she weeds out the decal (meaning she removes the extraneous parts of the decal and leaves only the lettering) or $20 if you want to do it. The only other cost would be shipping which shouldn't be much depending on where you are. I added in the photos section on this group pictures of the decal on my air cleaner I powdercoated this weekend. Couldn't figure out how to attach those pics to this post, sorry. -Matt
20cd37d8-9542-c998-a85e-7f0765ca8084
['22a1c300-1a70-833e-b35d-8787146f032d']
Thanks <PERSON>, <PERSON>, and <PERSON> for your comments and suggestions. I must admit that I'm surprised that it's virtually a consensus that I not attempt to implement my own guaranteed messaging system on top of UDP. I checked into the World of Warcraft using TCP/IP only and that appears to be correct. I've also quickly researched some of the popular MMOGs out today to see what they are using. Asheron's Call - UDP only http://www.fuzeqna.com/asheronscall/consumer/kbdetail.asp?kbid=148 EverQuest - UDP only accept for the Patching server http://www.iss.net/security_center/advice/Exploits/Ports/groups/EverQuest/default.htm EverQuest 2 - UDP only accept for patch servers http://www.eqii.com/phpbb/viewtopic.php?t=20213 Star Wars Galaxies - UDP only accept for patch servers (no surprise since it probably shares code with EverQuest 2 right? http://starwarsgalaxies.station.sony.com/content.jsp?page=Tech%20Support%20Firewall%20Proxy Lineage 2 - TCP only (But this game is mostly played in South Korea and everyone there has super quality broadband connections) http://www.lineage2.wrw.be/faq.php#22 City of Heroes - TCP & UDP (not sure what purposes for each) http://boards.cityofheroes.com/showflat.php?Number=<PHONE_NUMBER> Was looking to find out what Guild Wars uses, but I don't actually think it's a true MMOG anyway. It seems that the "missions" are actually hosted on individual users machines and "cities" are just 3d graphical lobbies. All sewn together, it seems like a real MMOG but it isn't. Well I think I'm going to do what most of you suggested. I'll use TCP/IP at the very least for guaranteed delivery, but I am still debating whether to use UDP for real time unreliable packets or a separate TCP/IP socket. One of the former id software programmers I believe wrote this article http://www.bookofhook.com/Article/GameDevelopment/MultiplayerProgramming.html and in it he mentioned that he thought it was a bad idea to mix UDP and TCP/IP. He recommended users check out http://enet.cubik.org/ which is a reliability layer for UDP. Decisions decisions. -Mike
ab6049e3-31a5-0bc0-0f16-b41b14bc8c9d
['22a1c300-1a70-833e-b35d-8787146f032d']
Hello all! I just found this list. I will be searching through the archives for the rest of the night. So far one question I've not found answered is: "How many UDP sockets should I use in my MMOG?" Right now I'm working on a relatively simple 2d graphical client server game that will serve as a test bed for future projects. The 2d client will resemble the old arcade game "Gauntlet" with the exception of course, that it will support a lot of players (hopefully) running simultaneously on multiple maps & levels. The question I have is what type of UDP socket scheme should I use? I believe RakNet (http://www.rakkarsoft.com) uses one UDP socket per user. I've not used that library, but it seems to be mature and highly functional. On the other hand, one socket per user seems like a lot of overhead and thus losing one of the benefits of the connectionless UDP protocol. Further, with no real fixed range of ports, configuring a firewall could be problematic for users and sys admins. I'm thinking I'll just start off with just one UDP socket per channel (e.g. chat, status updates, state updates) and see how that performs. Anyone care to offer their advice/opinions/suggestions? Thanks. -Mike
fdc091dc-6513-d512-9931-8b249a8b0764
['22a518d5-6aa0-2b59-fe9f-fcd79f321c5e']
I checked out all the open source spell checkers over the weekend and hunspell seems to be the one everyone is using. The names of the file in the project looked promising so I thought that improving it wouldn't be too difficult. As it turns out after looking thtough the code and dictionary files hunspell is actually quite a poor spell checker unless you've missed out a letter or jumbled your letters up in which case it's quite good. Anyhow back to the point, improving the spell checker to a level which would put it on a par or better that google (via pattern matching, re-enforced training based on real world spelling mistakes, quite a bit of stats etc...)isn't actually that hard code wise (I only have one pattern matching algorithm to find and I know exactly what I'm doing), the problem is that it's going to take quite a bit of training to get anywhere near the level that you could call it a 'propper' spell checker because there are no phonetic dictionarys to use as a base data set and even if I could find the data I don't feel like compiling them for all the languages hunspell supports especially when turning a word into it's correct phonetical form isn't that easy. My approach, and this will be a lot better in the long term for the ability of the spell checker to actually suggest the correct spelling and put it at the top of the list, is to write a framework in which the spell checker learns 'spells like' psudo phonetics allowing it to come up with a very high ranking word that should be the correct spelling based on any spelling mistake made in the past that is similar and even spelling mistakes that are similar to the directly referenced spelling errors for the psudo phonetics. The best bit it that I plan to have a centralized server as well as the client app so that all the data from everyone's spelling mistakes (provided they don't turn data collection off, in which case there not going to much better off than running hunspell because they would need to pull a partial snapshot of the spelling database down from the server on first use) and turns them into a huge knowledge base of spelling mistake patters and words not in the dictionary and user profiles that can be pulled down to any machine with the spell checker in it and group dictionarys so that uses can share their words that shouldn't be in the main dictionary with everyone in their office. Anyhow, I think I can make a revolutionary spell checker all I need is volunteers to use the spell checker to train it up a bit. May plan was to create a modified version of the firefox spell checker (currently based on hunspell) as an initial prototype that will work no worse than firefox's spell checker (because I'm partly basing this spell checker on hunspells process) but would train my spell checker the psudo phonetics it need to do the job really well. I expect the finished spell checker (short of tuning some weights and trashholds from their arbitary values) in about 3 weeks time, with the firefox plugin not long after that (I don't know xpi so the firefox plugin may take me a week of two to sort out) Once I've got people up and running with the firefox plugin and ironed out the enevatable bugs in the spell checker I can start integration with monodevelop.. It should also be fairly easy to integrate with open office too as it also uses hunspell at the moment so it's a case of copying over the new library, writing some screens to allow the user to control syncing with the server. And getting openoffice to send spelling corrections back to the spell checker to do the reinforced training. As you can tell from my spelling in this email a really good spell checker it near the top of my Christmas list. <PERSON>.
4912c376-a6ae-75bc-501b-aa55c8662449
['22a518d5-6aa0-2b59-fe9f-fcd79f321c5e']
Is it possible to debug mono-develop in monodevelop, it get's all the way to loading the workbench of me and then hangs. I could do to do this because for some reason code completion isn't working for referenced meta-compiled assemblies but details about those assemblies are still showing up in other parts of the ide. No one else has said their having the problem when I've mentioned it before so it looks like is something I'll have to figure out myself and being able to debug should make it so much easier. Thanks, <PERSON>.
7ab837a3-bb1f-aec3-f337-ae6c9ad3558e
['22a535a5-9c08-93da-4d06-28e59e6f8dcd']
hi there, i installed ejabberd today, got it work with adium/ichat and wanted to ask you all a couple of things: 1. i am working on adding jabber support into my iphone app and i want to configure the server in such a way that only my app can register new users and login into the server. is there anything i can do on the server end to configure this? ssl client authentication or something other client authentication i can build into the xmpp client? 2. since this is an iphone app, the network can drop at any moment. is there a way for a server to quickly discover that client has disconnected so it can begin to store messages for offline delivery? thanks. -- yan
62289f49-9afc-bed9-df7f-e2a647826aac
['22a535a5-9c08-93da-4d06-28e59e6f8dcd']
hi there, we have a server with ~3500 (on average) connections. lately we see more and more of these crashes in our logs and they occur a few times an hour at this point. we have a strong theory that these crashes cause dropped messages for our users. is there anything we can do on our side to debug this further so we can figure out the problem and fix it? thanks, <PERSON>
2a574691-0d9a-1eab-76aa-212353abe687
['22a695f6-ce07-9fb2-2577-588b27e67686']
Hi, I ve got three programs: 1)asp.net client encoding is deafault 2)console server 3)postgre database with database encoding WIN as I use Bulgarian They all run on windows paltform. I only moved the console server to mono on Linux and left everything else on Windows. It worked but the asp.net shows ",,..,." where it is supposed to be Bulgarian. The asp.net and the console server use remoting to communicate. I do not make any conversions. What is included in mono-locale-extras? Please give me some hints to explore. Thanks, <PERSON>
641cee07-50a1-ea25-a20e-53e7e325789d
['22a695f6-ce07-9fb2-2577-588b27e67686']
Hi, Monodevelop has various projects, but where is the ASP.NET project? It should be not that hard to add. -The type of application should be changed when calling msc. -The needed default referrences. -A simple asp.net code should be displayed when the user creates ASP.NET project. -The run button should call xsp every time for the project directory and a browser - let's say firefox. When the browser is closed the xsp should be stopped or restarted cause it will have to load the new dlls if the project is recompiled. -Some options should be added to the GUI. I want to help, but I can not do it without some help? Cheers, <PERSON>
5936bfc0-5c4f-5323-a4d3-e7302c9feb55
['22ad826b-0bb4-48bc-34bd-fe4faa671757']
Hi, I was trying to understand ipsec code and have come accross a doubt. Can anyone explaing how the kernel will call "ipsec_tunnel_start_xmit" of the virtual device [ipsec0-4]. please guide me how this done and let me know where can i find more doucments In the code the comment says "ipsec_tunnel_start_xmit" will be called from dev_queue_xmit() and my understanding is dev_queue_xmit() will always calls hard_start_xmit of physical device. Thanks in Advance. Regards <PERSON>
9cc30e3e-16d3-f9e3-1356-37c126a472c4
['22ad826b-0bb4-48bc-34bd-fe4faa671757']
Hi, we are trying to to integrate our ipsec onto linux 2.6 kernel to do so we have created a virtual device [struct net_device] with dev->type = ARPHRD_VOID and configuring the same MAC address and IP addresses as present on "eth0" interface and when try to send out the data, the packets are seen on the network.But we need all the packets to be processed in the virtual devices "hard_start_xmit" function and this is never getting called. Please let me know how to make the kernel call "hard_start_xmit" function of my virtual device, so i can process the packets in my "hard_start_xmit" and resend them on to "Eth0" interface Do I have to set different IP for virtual device or have to do any other configurations. Thanks in advance. <PERSON>
a7efdda1-69a4-ed70-35c1-e30eee25bb70
['22af487e-3c39-0205-87ed-3433d77edd7b']
Stanford Report, August 31, 2009 Open-source camera could revolutionize digital photography Stanford scientists' open-source camera could change photography by giving programmers anywhere the power to change camera features and create new possibilities. http://news.stanford.edu/news/2009/august31/levoy-opensource-camera-090109.html [excerpt] Stanford photo scientists are out to reinvent digital photography with the introduction of an open-source digital camera, which will give programmers around the world the chance to create software that will teach cameras new tricks. If the technology catches on, camera performance will be no longer be limited by the software that comes pre-installed by the manufacturer. Virtually all the features of the Stanford camera – focus, exposure, shutter speed, flash, etc. – are at the command of software that can be created by inspired programmers anywhere. “The premise of the project is to build a camera that is open source,” said computer science professor <PERSON>.
6f80289f-f68b-43ec-eeab-641e9a866c05
['22af487e-3c39-0205-87ed-3433d77edd7b']
Poor countries' media must tackle climate change ScienceAlert - Australia It was his experience at a workshop on environmental journalism, organised by China's Capital Youth Journalists Association and Internews' Earth Journalism [...] Inexperience and vested interests The reasons why climate change issues are poorly covered or go unreported are many and complex, as I've discovered in heading a programme that provides environmental journalists with training and networking opportunities. Environment, science and climate change are not considered prestigious beats, and are often assigned to young and inexperienced journalists, or to staff who are also asked to cover a huge range of issues from health to agriculture. They don't have the time or budget to research stories in-depth, and often lack scientific expertise. Yet the most formidable obstacles may be their own editors — many of whom aren't interested in or don't understand climate change issues — or vested interests. For example, energy companies are major advertisers and many would prefer such issues were downplayed. Recent studies suggest climate change coverage is improving in developing countries as editors (and governments) are taking more notice of it. But progress is slow. Journalists and media organisations in the developing world who want to cover climate change deserve more support — from research institutes, who could do a much better job at outreach; from national and international climate-related agencies, who should try harder to accommodate the media's need for locally relevant information; and, in particular, from multilateral, bilateral and private aid agencies. Donors who fund climate change work usually treat media and communications support as an afterthought. But the research and action plans they fund will have limited effect if they are not communicated to the public and policymakers. The media has both the reach and credibility to make all stakeholders take climate change issues seriously. Ways to help There are plenty of ways to assist. Journalism schools in developing countries can be helped to cover science and the environment in their curricula. Research institutes can develop better outreach programmes. New media organisations can be established to focus specifically on climate change and its impacts, particularly using digital platforms. Working journalists can be trained, and senior editors persuaded, to improve their coverage. [...] By... <PERSON> is a Thailand-based journalist and the executive director of Internews' Earth Journalism Network.
550b2272-7198-a190-d3e1-10725ec7360f
['22b83772-84b0-0e1b-aa45-6ab023374e94']
Hi Geoserver users and Developers, We recently upgraded our Geoserver from 2.0.1 to 2.1.1 and we are having problem where the Open layers map from OpenLayers doesn't draw anything for some layers. On certain layers namely ran:secchi and ran:biolum no dots get drawn and we note that the initial map is fully zoomed out. Only when you zoom in 1 level (reducing map scale by 1/2) do you get the map (dots) drawn. It took me ages to figure this out as at first I thought there was some problem the getMap request or drawing. To see this issue: 1) Go to our geoserver http://www.metoc.gov.au/geoserver 2) Click Layer Preview 3) Select the Open Layers link for the ran:secchi or ran:biolum layer. 4) Map will come up in Open Layers but no points drawn. Notice how the zoom control is its maximum zoom out. 5) Zoom in a little and you will see the map draw. We don't have this issue with the other map layers i.e ran:beach_temp, ran:xbt and ran:seabed. Something to do data extents of these particular layers or some issue with newer OpenLayers 2.8 ?? Any ideas appreciated... <PERSON>
85a45636-8eee-61da-f53b-7b680b7a7ac1
['22b83772-84b0-0e1b-aa45-6ab023374e94']
Hi Users and Developers, We recently installed Geoserver 2.1.1 from .war file on our Tomcat. We run our Geoserver behind a reverse proxy setting the PROXY_BASE_URL through the "global settings" web form and switching on the 'Reverse Proxy Filter' to true in the WEB-INF/web.xml. Apache 2.2 at the front end does a proxy to our Tomcat/geoserver at the back end. We found that that this setup worked fine with our previous Geoserver 2.0.1 but now we get the following problems with 2.1.1:- 1) Output from the GetCapabilities requests to WMS 1.1.1 and WFS 1.1.0 produces binary garbage. 2) GML and CSV output for all layers produces binary garbage. Other things like WMS getMap drawing and getFeatureInfo seem to be workingfine still as they were with 2.0.1 I note that when I set the Reverse Proxy Filter to 'false' the problems with the GetCap. and GML/CSV XML output go away. Hence my theory that problem is related to the use of the Reverse Proxy filtering. However we can't use that as a workaround as we need the Reverse Proxy filter switched on to correctly filter URL's in within html output coming back to the client outside the front end. Samples of the verbose geoserver.log output attached don't show anything serious i.e [ERROR] and the it appears the filtering and translation is proceeding. The file getCapLog.txt is from a WMS 1.1.1 GetCap. request. The file GML_log.txt is from a GML request to the RAN.SECCHI layer To see the problem with GetCap. output do the following: 1) Go to http://www.metoc.gov.au/geoserver 2) Click on the WMS 1.1.1 link and follow prompts to download the GetCap. file. 3) Open the GetCap. file with a text editor. To see problem with GML output: 1) Go to URL as above. 2) Click Layer Preview 3) Click GML link for any of the layers 4) Look at the output in browser. Looking forward to any replies about this issue..... <PERSON>
6e540aea-179d-485e-ce4d-ec758d9095e8
['22cf9409-b85a-62d1-472e-4af600ee021b']
Hi, Boost Users and/or Developers. Does anybody know why functional objects created by means of boost::bind have no move constructors? It would be rather useful especially when such functional objects consist of some smart pointers (like boost::shared_ptr). For example, move constructor of boost::shared_ptr is very light (comparing to its copy constructor) - it doesn't use any CAS at all (which is the main overhead of boost::shared_ptr). May be someone know the reason of move constructor absence or maybe someone know any plans of implementing it. Regards, <PERSON><IP_ADDRESS>bind have no move constructors? It would be rather useful especially when such functional objects consist of some smart pointers (like boost<IP_ADDRESS>shared_ptr). For example, move constructor of boost<IP_ADDRESS>shared_ptr is very light (comparing to its copy constructor) - it doesn't use any CAS at all (which is the main overhead of boost<IP_ADDRESS>shared_ptr). May be someone know the reason of move constructor absence or maybe someone know any plans of implementing it. Regards, Abrarov Marat.
b786c76a-73d8-04fd-e8d3-cd06c0c981e7
['22cf9409-b85a-62d1-472e-4af600ee021b']
At least at Windows asio is the best - better then ACE proactor because of strands and custom memory allocation support for them. You may look at http://sourceforge.net/projects/asio-samples/ for more complex echo server than the asio examples have. Stack corruption may be because of wrong lifetime of used objects... and it may be not seen and debug because of different time scale. Regards, <PERSON>.
2db5583e-f381-01d1-bc0b-370ff7ca4e47
['22d12f4b-8426-859e-cc7c-ea8ac477675b']
Hello everyone, to allow concurrent buffer access by different engines beyond the multiple readers/single writer model that we currently use in radeon and other drivers we need some kind of synchonization object exposed to userspace. My initial patch set for this used (or rather abused) zero sized GEM buffers as fence handles. This is obviously isn't the best way of doing this (to much overhead, rather ugly etc...), <PERSON> commented on this accordingly. So what should a driver expose instead? Android sync points? Something else? Please discuss and/or advise, <PERSON>.
6c56582c-a6b3-dab0-054b-082096962eab
['22d12f4b-8426-859e-cc7c-ea8ac477675b']
Hi everybody, the following crash happens when booting one of our E350 based test system in EFI mode, booting with in BIOS compatibility mode works fine. Bisecting initially turned out to be commit 8d57470d8f859635deffe3919d7d4867b488b85a, but that turned out to be a false positive and already fixed in 3.9. Complete bootlog is attached, so does anybody got any more idea what to try? Thanks in advance, <PERSON>.
fb79765f-2f3b-b941-a8e9-653f82846ccf
['22dde1fb-4826-6f7b-0cdb-e400464b599e']
[This is from an email about the impact of not having a static Threads lib available anymore. I wanted to post this here to see if anyone else has similar experiences since I have only seen posts about the problem of distributing RTL dlls so far - not computation time impact.] Hello, I was reading some some of the posts about your thread implementation and the removal of the static RTL option. I understand the technical reason for this, but I wanted to submit another compelling reason (for me) for the need for static RTL linkage even if this means loosing some of the features under win32. Some of the software I develop deals with different types of analysis (such as scientific and financial). The runs can take as much as 20 hours to complete. If I do nothing but change to link with dynamic RTL, the run time is increased on average of 35%. This means and additional 7 hours of run time! So dynamic RTL is really undesirable. Do you see any way of making a more limited static Boost.threads available in the next Boost release? [Above I say "limited" because some of the features of the Thread lib must be implemented via dll to work under win32.] <PERSON>
b2671b89-4e91-50b0-da81-b48b1a2d6065
['22dde1fb-4826-6f7b-0cdb-e400464b599e']
I'm not sure other than the reasons regarding optimizations possible with statically linked libs. I know that there is also some overhead involved in calling dll functions as opposed to the static counterpart, but this surely is a small cost. Other than that, I am not sure. I am very glad to hear that you are working on the static problem though. The latest checkout from thread_dev branch revealed the use of thread specific storage in thread.cpp, so it looks like the latest work there may be making it more difficult to make a static subset... <PERSON>
76f366e6-d5b0-ab2a-5a9d-6c5fca373f1e
['22df17e1-c44f-bf9d-878f-eaca658fcf8d']
The first Tails [HackFest](https://tails.boum.org/blueprint/HackFest_2014_Paris/), which took place on July 5-6 in Paris, was a great event: about 120 people attended on Saturday, 70 people on Sunday. We shared some good workshops, some very good food, and homemade beer in a nice and friendly atmosphere. The slides of some talks have been [published](https://tails.boum.org/promote/slides/), but there is more to come. The first Tails stickers were distributed and you can also make your own! Just use [this SVG file](https://tails.boum.org/promote/tails-sticker.svg). We are particularly happy that the HackFest attracted very different people, so we were a quite diverse assembly. Probably thanks to our [anti-harassment policy](https://tails.boum.org/news/../blueprint/HackFest_2014_Paris/policy/) (thanks <PERSON> for the frensh translation)... Some people talked about the HackFest: [jvoisin](http://dustri.org/b/tails-hackfest.html), and... more? (please send links!) Thanks to:
feab96c7-da08-2c88-5c99-77eae9e658d7
['22df17e1-c44f-bf9d-878f-eaca658fcf8d']
Every day, Tails helps thousands of people to safely use their computers online and offline. Tails is available for free because we believe that nobody should have to pay to be protected from surveillance and censorship. Tails is free software because if our source code was closed, there would be no way of verifying that it is secure. In today's Internet, the most rewarded economical strategy is to rely on advertisement and tracking. But one of our reasons to exist is also to protect you against targeted advertisement, and we obviously don't want to rely on these sources of revenue. The number of people using Tails is constantly growing and so is our responsibility to provide the easiest and safest tool possible. In 2019, we want Tails to be smoother to install, upgrade, use and to include even more security protection. Making Tails costs us around 200 000€ each year. This is a really small number given all the work done, and many of us still do a lot of work for free in addition to paid hours. Our money comes from: * Passionate people like you * Private companies like Mozilla, DuckDuckGo, or Lush * Foundations and NGOs like The Handshake Foundation or The ISC Project * Funds from the US government like the Open Technology Fund
19b5d234-f19a-2a0b-92f0-93fcfaddf630
['22e70180-c41e-16ba-c1b3-d8c8528d5241']
Hi, This is caused by a bashism, ${var/old/new}, which is used in a few places in debian/rules. Attached is a patch which replaces it with a call to basename, or an `echo $var | sed s/old/new/`. debdiff shows no differences between the package with and without this patch applied, but it seems like some md5sums change. I can't assess how serious this is. Anyhow, hope this helps. <PERSON>
faad41c5-b26c-af8e-ac87-fba777e49ebf
['22e70180-c41e-16ba-c1b3-d8c8528d5241']
Hi, Please forgive me for emailing you directly, but I couldn't figure out how else to send feedback for this library (imaplib2). I am cc:ing the offlineimap mailing list, since it concerns my efforts to integrate the two projects. I'm trying to get offlineimap to support the IMAP IDLE command. An attempt was made once to get this to work using your imaplib2, but this caused stability issues for some users. It looks like fixes to imaplib2 were made that may address these issues, so I decided to take a little while to try to integrate offlineimap and the newest imaplib2. Everything seems to be working OK, except for when connections break -- and my ISP is such that connections break with some frequency. offlineimap has a setting "socktimeout" which is given to socket.setdefaulttimeout(). In the old, pre-imaplib2 world, this let the user specify the maximum amount of time he was comfortable waiting for a possibly broken connection. However, since imaplib2 uses non-blocking input in a separate thread, socket.setdefaulttimeout() is useless for us, meaning that (on my machine) broken connections stop the program from doing anything for 15-20 minutes. I have been poking around in the code for imaplib2, trying to figure out where would be a good place to put code to "simulate" the effect of this timeout. First I thought it would be good to put a timeout in the Request.get_response() method, since this seemed like it could let the connection languish for as long as it needed to, up until a response was needed, and then it could break. However, adding a timeout there did not seem to fix the problem; connections would still take 15-20 minutes to die sometimes. I have no idea why, but my working hypothesis is that the get_response() call never fired because the writer thread never finished actually sending the commands. Currently I am experimenting with accumulating timeout in the reader thread (something like the attached patch). This "seems to work", but I think I've seen one case where it hung for 20+ minutes trying to stop idling (so possibly the same problem as above). I was wondering if you could comment on this or any other approach you think would be good to implement this functionality, and your willingness to update imaplib2 to include functionality like this (assuming a suitable patch were provided). I'm especially interested in: is self.tagged_commands the best way to check if we are expecting input? What do you think the timeout variable should be called, given that there are already idle_timeout things throughout the class? Thanks for your time! <PERSON>
1958c8a9-2914-9f0c-30d3-6788249c7fa5
['22f7efce-0a71-4f9d-343d-6a7ebf5c684a']
Now is your chance to show your expertise and knowledge to the 4D community! Speaker submissions for the 4D Summit 2016 are now being accepted until December 15th. Potential speakers are asked to submit a topic for approval and if selected, develop slides to be used at the event in Portland. A 4D team member will be available to assist you with your presentations. Presentations should be no more than 60 minutes, including time for Q&A. All topics, from beginner to advanced, will be considered. 4D Summit 2016 will be held in Portland Oregon, at the Portland Marriott Downtown Waterfront, April 5th & 6th, 2016. Submit your proposal: http://www.4d.com/company/events/4d-summit-2016/speakers.html -Brian Young
33a6b373-2a14-2981-7e37-2ee13dd2c667
['22f7efce-0a71-4f9d-343d-6a7ebf5c684a']
<PERSON>, thank you for this writeup. :) I also want to thank <PERSON> and <PERSON> for putting so much effort in to making a active, growing, online usergroup. Plus, thank you for inviting 4D to to participate. (including the 'hey 4D' hot seat) You should watch the entire meeting. It's is worth the time. But, if you would like a direct link to skip the technical difficulties, you can jump to here: https://youtu.be/eEjrHR2BelE?t=36m57s -Brian
5c714b9e-8404-31eb-2074-4016a8b36e24
['22fbb047-271e-7fc1-779f-9cbcd54e3030']
Hi all! I am happy to anounce that I have just submitted my application regarding File System Changes Notification Support. Some of you may have noticed that I have already discussed it here on the mailing list. Please be welcome to post any comments if you are interested on this subject, I will be grateful for any feedback. Best regards,
31650b66-2922-a336-ba30-b9193373d024
['22fbb047-271e-7fc1-779f-9cbcd54e3030']
Hi all! I am a student from Warsaw University interested in taking part in Google Summer of Code 2009 with wxWidgets. To start with I just wanted to introduce myself and say 'hello', so that we can get to know . I'd also be glad if I could discuss my idea with somebody to build a proposal which will be a successful project for wxWidgets and me:-). I am interested in building a File-System Notification API for wxWidgets, an idea listed here: http://wiki.wxwidgets.org/Development:_Student_Projects#File-System_Changes_Notification_Support. It looks like a good project for someone from the outside, not familiar with the wxWidgets codebase and it is a well-defined one - in my opinion this is perfect project for GSoC, because if a good quality code is produced, it will easily fit in the codebase and be fully utilized by wxWidgets users. I have already begun working on a concept and reading documentation on wxWidgets site. Who should I contact to further discuss the idea? Or probably I should ask whether wxWidgets' developers are interested in the idea in the first place:-)? Any comments or suggestions are greatly appreciated. Best regards,
6c2e948e-9294-e33d-c5db-e013729323a4
['23077492-fe05-a194-373a-7904be177a5e']
Hi <PERSON>, Since osdteletext page drawing is rather slow on my C3-600, I investigated the page rendering speed of osdteletext-0.4.2. Average rendering time is 0.74s, dividing down to 0.11s rendering, 0.45s scaling, 0.09s osd drawing and 0.11s screen update. (numbers might be slightly off due to additional debug output) After a rewrite of the scaler, the scaling is down to 0.1s, speeding up screen update from 0.74s to 0.39s - a noticeable improvement. ;) The attached patch adds the new scaling code. Cheers, Udo
a04c388a-f6d7-325d-637c-a3fac5e17c8d
['23077492-fe05-a194-373a-7904be177a5e']
... or why two legs are better than one. Hi list. Due to some unexpected S2 capable hardware, there's a new version of my S2API wrapper patch available for download. The new version adds S2API auto-detection at runtime, so that one binary vdr executable can handle old DVB V3 drivers and DVB V5 (S2API) drivers without any changes. As before, the patch compiles fine against the old V3 header files and the new V5 header files, and even if vdr was compiled against V3 headers, the resulting vdr can still tune to S2 channels when using it with V5 capable drivers. As a fallback, the patched vdr understands the command line switches --dvbapi=3 and --dvbapi=5 to manually select the API to use. The patch applies to VDR from 1.7.13 at least up to 1.7.15. If there's any need for VDR versions 1.7.12 and earlier, contact me. Get it at: http://www.udo-richter.de/vdr/patches.en.html#dvb-api-wrapper Cheers, Udo
d741f6cd-1591-9eeb-47c9-6481760eda79
['2309daf4-0ebe-ced5-88f0-4e49aafa0693']
Hi there, I was just curious if anyone had experience using their ibook (dual-usb model) with a pcmcia wifi card that is not a lucent orinoco/airport. I have a card which fits neatly (without having to do anything ridiculous with a dremel =), but I just can't seem to get it to even detect. I really have no experience with anything PCMCIA on linux, and I cant even get the computer to show that something is there and plugged in....although I'm sure its a card supported; it's just a re-badged prism2. Although the LEDS on the card light up when I plug it in, cardctl returns no socket() or something like that...nothing shows up under the boot cd loading with the dopcmcia option either. Am I missing something just obvious and stupid? any help would be appreciated. Best, alden
8c73414b-4e30-c3c2-7119-c5387e011afc
['2309daf4-0ebe-ced5-88f0-4e49aafa0693']
Theres a little plug on most recent iBooks and powerbooks that has the label of a monitor : |O| except the O is a rectangle ...I wanted to use my PPC linux laptop to give a presentation...and typically, if one were using a powerbook with OSX you can simply plug the projector into that port and it automagically works upon boot. I was just wondering if anyone knew if this held true for an ibook on linux as well...im running it on a iBook4 G3...with the latest vanilla sources. its the white one and the first to have 2 usb 1.1 ports. I imagine it should, seems hard-wired...but I just wanted to be sure before I went to put on the show... Also, is a wireless airport card the only card available? Does anyone know any non-usb alternatives that are supported in linux and also physically fit in an iBook? And if anyone who has contributed to the Gentoo Linux project happens to read this...I wanted to extend a big thank you for putting together the best damned linux distro, or any os for that matter, out there for people to enjoy...you're do an exceptional job. Thanks. Best, alden
feef5fdc-ab3d-d94c-fb11-350d825ba9b5
['2311d762-7640-2ad6-0d76-d98169400cd1']
I've attached a patch for QuantLib-Python to compile on darwin (the BSD core to Mac OS X). Basically darwin distinguishes between shared libraries and loadable modules so the patch detects if darwin is the build environment and if that is the case is uses different compiler switches. That's it. Unfortunately I wasn't able to create this before 0.3.5 but compiling on OSX is the only think affected so the fink quantlib-python package uses the attached patch on the 0.3.5 tgz. <PERSON>.
65472c38-51bf-e287-27a5-6d1087858bdc
['2311d762-7640-2ad6-0d76-d98169400cd1']
I'd like to post in support of this patch. I've spent the past day trying to track-down the cause of a DllNotFoundException in beta3 that turned out to be cause by specifically the same problem quoted below. Had this information been present it would have same me a lot of time since the problem is not with the Dll, the problem is with the linking. Although a better solution might be to throw a DllLinkingErrorException derived from DllNotFoundException but this probably deviates from the standards but it would be a lot of help to many people. Thanks <PERSON> for posting this solution (even if it never goes into mono). <PERSON>.
b9f28dad-5051-94b2-dad2-356a911473ad
['2314b9c8-a646-abc2-4281-b4ee7e5ad9e0']
My Webfaction account uses Apache with workers for MPM. It seems that any memory allocated to a process spawned by an Apache worker actually clings to that worker after the process finishes. That's why there is a maximum_requests_per_worker directive. Could it be that big database migrations cause huge memory allocations across most of the available Apache workers? Thank you, <PERSON>
af9c6725-76f0-a396-2fc7-eabdd68f2510
['2314b9c8-a646-abc2-4281-b4ee7e5ad9e0']
In November I gave a talk to the Atlanta Python Meetup group on role based authentication and authorization. I mentioned several times during the presentation that Web2py features role based access control, and showed some demo code which is a simplified extract from the Auth class. The demo code is available at https://github.com/cliffkachinske/role-auth-and-auth I post the link here because it might serve as a gentle introduction for anyone who wants to delve into some aspects of the Auth class. I plan to do a little more spreading the word in February with a talk on the DAL. This time I won't be doing any derivative code. <PERSON>
b07cdf64-d11f-b18a-deb8-9743a08cddc5
['2317f052-9ca8-3898-3c32-be60af47a1bc']
I personally have no dog in this fight (have no reason to favor one side or the other, for those unfamiliar with this Americanism), but I do have opinions I think are pertinent. For those new to Mandriva, getting an iso with all the updates is a major major benefit. The new backgrounds are a nice touch but of no great importance to most. If everything had worked as intended, existing Mandriva users would have had the change imposed on them, but no changes on their part would have been necessary. All would have happened auto-magically. All did not go as intended or expected. We have all been there, done that, got the scars to show for it. This time the scars were small and will soon disappear, I hope. Arguably, simply renaming the new isos would have been enough (technically), but making the change obvious might have garnered a little favorable publicity for Mandriva's attempts to serve new customers better. When the change went somewhat wrong, this did backfire. The decision by Mandriva/Eugeni perhaps was an error (or perhaps not, honorable men may differ on this point) but you cannot convince me it was wrong-headed. Time to continue on. Cheers! <PERSON>
ff0e6524-2acc-0b45-dda2-38b5eb5e119a
['2317f052-9ca8-3898-3c32-be60af47a1bc']
Attempts to compile and install the fglrx module are failing for the 2.6.34-tmb-desktop-0.rc4.4.1 and 2.6.34-tmb-desktop-0.rc5.1 kernels. X-windows will load and run, but I cannot "leave" them from the menu, and cntl-alt-backspace gets a blank screen but does not kill x-windows, At that point cntl-alt-F2 does not get a new terminal window.. alt-printscreen-[rseiub] to reboot seems to be the only way out. If this is of interest, I can run make and collect error messages for fglrx and maybe play around more with trying to kill x-windows. Cheers! <PERSON>
9f466a55-d0ab-6261-c86b-d028b54869b7
['23266986-3ee1-e867-a576-18f49e1fe55a']
Hello, I use script provision.pl to populate SNMP configuration of equipment into OpenNMS. The SNMP configuration is taken into account by opennms. However I saw that the opennms configuration file snmp-config.xml is not updated with this process. Do you know if it is possible to write configuration on snmp-config.xml on command line like in web user interface "Send Locally" ? Regards.
808215ae-6417-68bf-427d-1a2d0d85ad00
['23266986-3ee1-e867-a576-18f49e1fe55a']
Hello, I want to know all methods to select a varbind on a mask field on event configuration files. For the moment I know the vbnumber selection (the number of the varbind). It depends on the order of the varbind. For example, can we select a varbind with : - his name (defined in translator for example) ? - his OID ? Regards, <PERSON>.
9ce32a50-9562-0c91-a94a-d2b4d70db234
['2327479f-31da-20d0-cfa3-3eef3151b750']
Hi, Sorry for the of topic question, but where else do you get so many telco guys in one place. I have a customer who is moving to Australia and was on ADSL here in the UK. Q) Is ADSL a standard? and will his router/modem work in AU? I have told him a tentative yes but would page the oracles for clarification. Regards <PERSON>
9b22aad9-9c0e-bc1c-b214-5c60af1ebe4f
['2327479f-31da-20d0-cfa3-3eef3151b750']
Hi All. When I run "modprobe zaptel" I get the message that the zaptel.o was compiled for kernel version 2.4.20-4GB while this kernel version is 2.4.20-4GB-athlon. And fails. When I run "modprobe wcfxo" I get the message that the zaptel.o was compiled for kernel version 2.4.20-4GB while this kernel version is 2.4.20-4GB-athlon. And fails. How di I get zaptel & wcfxo to recognize my kernel? I am not a linux guru so layman terms would be appreciated. Thanks in advance <PERSON>
30febd33-224b-61fb-bd66-8b4927d216b8
['232ba3c3-fb20-6b5e-f49f-4e96a10a2c3d']
Hello! I try to mux an mpeg-ts file into an rtp context for a unicast / multicast (is this possible?) video stream. What I did so far is the following: 1) I opened the mpeg-ts file into a common AVFormatContext (ts context) and detected the streams 2) I created an AVFormatContext for rtp 3) I added the same streams to the rtp context which were detected in 1) 4) I wrote the headers into the rtp context (avformat_write_header) 5) I read the packages of the ts context and wrote them directly into the rtp context (I tried av_write_frame and av_interleaved_write_frame) 6) I wrote the trailers into the rtp context (av_write_trailer) When I try to receive the rtp stream with ffmpeg -i rtp://xx.xx.xx.xx.:1234 I get the message "Invalid data found when processing input". When I try to receive the rtp stream with ffmpeg -i udp://xx.xx.xx:1234 I receive just one video stream, no audio. When I try to receive with vlc, I get nothing. Does anybody have an idea how this could work? Or where I make the mistake? Thank you very much for your help! Best regards, <PERSON>
6bbd7226-45bc-2a35-7a6c-42bd60e09599
['232ba3c3-fb20-6b5e-f49f-4e96a10a2c3d']
Hi! I try to include a data stream into mpeg (ts). I wanted to ask if any encoder / decoder is implemented for CODEC_ID_TEXT in ffmpeg 0.7.1. I always get an error when I try to do avcodec_find_encoder(CODEC_ID_TEXT). In one forum I found the statement that this codec is not implemented for version 0.5. Now I'm not sure if my code is not correct or if the codec is still not implemented. Thanks for your help!
b32c3a1e-df40-1f36-6b95-eb1ebc990514
['23313373-b18e-996b-418e-c126be5533c9']
Hi Team, Request to let me know where I can post some suggestions and ideas. I am a new to Selenium. Had been using Mercury products for quite some time and found what Selenium exciting. Few things that came to my mind when going through the web site and documentations are as follows: Just my two cents, not sure it might be useful. 1. Can Selenium merge QALiber to give us a awesome Desktop Testing tool like it does for Web Applications? 2. Can it use features of Fiddler to create a cool tool like Selenium IDE for Firefox in IE? Reason why I am asking? 1. I see Selenium is a powerful tool for Web Browser Automation targeting web applications. It would be good if we have a version of Selenium that can do Desktop applications and Java Thick client applications as well. I might be greedy but I used to use a open source tool called QALiber on Sourceforge.net. Unfortunately they stopped development for quite some time. I read in your documentation and found that the group found two powerful tools like Selenium RC and Web Driver and merged them to give us Selenium 2.0. Can this happen with QALiber to give us a desktop tool or any other open source? Not sure is this is a vision for Selenium? 2. Selenium IDE is a real cool tool embedded in Firefox. It helps a lot. But it is restricted to Firefox. Fiddler is similar kind of tool on IE which can be extended to provide similar IDE on IE. I am not sure on the complexity of the implementation. I know you folks are busy, just wanted to share what I had in mind. Thanks.
6556b17b-d854-92df-e4b2-0108bb322d85
['23313373-b18e-996b-418e-c126be5533c9']
Hi Team, I am new to vagrant. I could find virtualbox images for Windows, LinuX and Mac in Atlas. But I did not find any images for Portable devices like Andriod OS and iOS. Not sure if vagrant can be used for portal device VM (Andriod tablet and phone). VMs are only for PCs? Thanks. <PERSON>
b8469f3a-4f0f-86fb-cee8-956e1fe90d6d
['233ceb15-7822-c6e5-5c06-be33312996fd']
I am currently working on the digitisation project. Some of the scanned files are very big; the file itself can even go up to several hundred pages in hard copy and in order to make it smaller when scanned I saved it in few 'smaller' pdf files. Those smaller files have in between 6 up to even 20 MB. Then I had to add those files to one record only. This means I ended up with a record that can have up to 8 files attached - some of the files up to 20 megabytes. The problems I have with this solutions are; a. The sort for these files can get messed up - for example: 154.v2 shows before 154.v1. b. When the size is bigger than 5 MB, instead of just opening a pdf in the web browser; the window pops up alerting you to either save or open the file. This does not happen for smaller files. Is this a known problem? Do you have the same situation? How can I re-sort? How can I skip the pop-up window unless this is the browser configuration issue. Any ideas to go around this are welcome. Thanks,
fb1ed167-9565-f4ab-0037-27ba8dd52212
['233ceb15-7822-c6e5-5c06-be33312996fd']
Dear dspace-tech list; I am new to the dspace world and need some basic help. This is what I am trying to do: I have exported over a 1000 records from the library management system into a csv format and would like to batch import these records (no attachments here) into the dspace. I found some documents on the wiki but I seem to miss the initial starting point. I have created the collection and the fields that will match the fields in my csv file. Then after creating a collection I go to the Import metadata and select the file to load it - but I got the error message: An unknown error has occurred I just have a feeling that I am missing the crucial step; do I need to have a separate plug-in installed? We are on Dspace 1.6.0. Sorry for asking a stupid question but I thought this would be a very time-consuming investigation for me while being an easy answer for an expert. If anyone has done something similar I would appreciate if you could share your procedures. Many thanks,
1bb05e8f-49e6-c2e3-e6f1-035a1349802a
['233fb9d2-1135-ebe5-54f5-26c53504dcfd']
Hi <PERSON>, what is the status quo regarding your automated wxWidgets bindings generation ? Your last message sounds pretty optimistic, so what's the matter ? Is the wxc (C++ to C) part working, but you still can work around the Multiple Inheritance problem to generated wxD ? Then it seems that /mixin templates + Interfaces/ as well as /alias this/ completely fail to mimic MI. Or are you waiting for 2.9.4 ? well, frankly, I am only guessing. Would be nice if you can tell a bit about show stoppers, unsolvable problems, etc. However in case that the C++ to C doxygen based generator is working... why not sharing ? <PERSON>
4aff9d67-fe42-d84d-c68c-87d5b6420675
['233fb9d2-1135-ebe5-54f5-26c53504dcfd']
Seems that my last reply was jammned or whatever .. so in short x509 certificate support... Why not ? SOAP/WSDL SOAP service consuming is what makes my bread and butter nowadays but to face it : Parsing a WSDL stream sucks Creating a SOAP service in D is .... Ouch.. Consuming a SOAP/WSDL in D is super...... anyway, Smart Devices are not realy interested in SOAP/WSDL... Means Vibe.d should support REST out of the box. I hope that you give ---- Interface to REST ------ (code-gen-wise) a try.. <PERSON>
5fa25ade-c976-59a5-0e6b-ccc3bfad38fa
['234a6a78-fc1f-473a-4cad-14c32351529b']
Hi all, I have my own application which takes my robots raw sensor data and generates an occupancy grid from it. I wanted see how my data faired in other mapping applications, and thus far the most widely supported file format for map view software seem to be player log file - am I right in thinking that's waht it's called? As yet I have implemented the navigation side and being able to use another app would help verify my design/code. Anyway I would like to add to my application the ability for it to output the raw sensor data in such a format - my question is where can I find the specification? URL or something. I also assume I would then be able to use stage to look at my data also? My robot is at www.toeminator.com Thanks in advance <PERSON>
d18b6379-3e52-d5be-4c8e-a312122c15e8
['234a6a78-fc1f-473a-4cad-14c32351529b']
Hi all, I've been trying to find a sucent guid to setting up users and SSL with SVN. I've got it running ok (Windows) with apache 2.0. All seems ok, what I'm uncertian about is the security aspect. I want to be able to permission only specific users to the repo. What I've read it seems as though the passwords are simply put in a basic file, not encrypted - not correct surely? http://better-scm.berlios.de/subversion/Svn-Win32-Inst-Guide.html If there are any tutorials re: this please can you let me know. Thanks Mark
897ad439-cdca-b89f-c26d-953d1868ee21
['234c3908-c39c-aa38-9ed6-bd50c534c05d']
Hi, I would like to know (or estimate) how much memory an opened index consumes inside the JVM (heap) and outside the JVM (fs buffers?). To my understanding the amount of memory inside the JVM depends on performed searches and search results which might be cached by Lucene. However, I am not yet able to see a way to get that information on a per index basis when multiple indices have been opened and queried. I would be grateful if someone could point me into the right direction for this one. Many thanks! Cheers Florian <PERSON>
9d9596bf-9f08-d510-9818-1fc1f62f73c6
['234c3908-c39c-aa38-9ed6-bd50c534c05d']
Hi, I am in the process of updating a large index from Lucene 4.x to 5.x and have two questions related to the sorting order. 1. Is it correct that stored fields can only be sorted on if they become a DocValue field in 5.x? 2. When "updating" stored fields to DocValue fields , is it required to update all documents in the index at the same time? Thank you in advance for your help. Best regards <PERSON>
89d17780-fac5-f410-5617-28ac226fce49
['236a41e9-1899-42bf-5f3b-e9048401858c']
Hi, <PERSON>, I want to make multiple spotlights. So I checked the spotlight example and have a problem as follows. When I use createSpotLightNode to create two spotlights, Always the first spotlight works. What is the problem? Is it also a problem about texture uint? I want to make many spotlights. Are they possible? Thanks for your reply. <PERSON>
bf19274d-0a8c-a931-164f-4e143bf7a811
['236a41e9-1899-42bf-5f3b-e9048401858c']
I want to use osg under MFC envrionment. I test the example OSGMFC, which uses rendersurface and sceneview in MFC MDI. However, I find two problems. 1) After two osg files are loaded, and close one, then the other osg data can not be correctly rendered, some textures are missing. 2) When the window is closed, the memory leak is very huge. Does anybody know how to solve the problem? Is there another way to use osg+MFC? How about to use MFC+osgProducer::View? Thanks for your reply. <PERSON><IP_ADDRESS>View? Thanks for your reply. Yigang
978d3548-1451-de58-b82a-1b4e3e9b10be
['237019bc-55c0-bc65-fbec-1ae46cd2564e']
Hi, I am using a modified 3.3 CD. The kernel has been changed to 2.4.26 with support for serial console enabled. When I start with knoppix console=ttyS0, this works for the boot messages output by the kernel itself, but once the linuxrc script starts running, output is again on the monitor. When starting a real Debian GNU/Linux with console=ttyS0, output stays serial until the getty is started, with no output on the system's monitor at all. So there must something go different between knoppix and Debian. I wasn't able to figure it out. Can anybody help? Greetings <PERSON>
2b6abba4-567d-151a-3e24-7ba4a6debc3e
['237019bc-55c0-bc65-fbec-1ae46cd2564e']
Hi, I have not been on this list for a long time, but it seems like things have changed a lot in the mean time. My main interest is building a small knoppix subset that will fit on a business card CD. I have accomplished that by hacking <PERSON>' build script a year ago, but there seem to be a number of local build scripts available to help this option. I would like to know if there is a web page that compares the different scripts. Which script is the recommended way to go? The script would need to be able to exclude certain directories from inclusion to the CD since I had to ditch the "my knoppix is a fully featured debian" stance to get the distribution sufficiently small. I would appreciate comments. Thanks. Greetings <PERSON>
65996de4-d14b-85cd-db05-583de406affa
['238b0564-9151-9826-5bd1-d2fbdef5457e']
Morning all, I've knocked up a simple iPhone client for the TWFY / Public Whip databases - mainly to show how active & responsive an individual MP is/isn't, but it also scrapes the Register of Interests to show how much they're trousering from directorships etc. The original idea was to have it in the app store ready for the election - I've missed the boat on that one, but the way things are going we may not have to wait another five years for the next election... If anyone's interested in beta-testing the current early version, drop me an email off-list with your phone's UUID number and I'll send over the app. I'm particularly interested in getting some feedback about how best to display this kind of info on a small screen, so if you've got l33t visualisation skillz I'd love your input. Thanks, <PERSON> <PERSON>
93a48699-b1b7-4fc8-fd68-3aef32dc73a8
['238b0564-9151-9826-5bd1-d2fbdef5457e']
Hi all I'm in the process of putting together an iPhone app that uses the TWFY and Public Whip APIs to create a directory of MP data suitable for quick lookups in between shouting at the screen during Question Time. This is very much a case of scratching some personal itches - one is that Googling for that kind of info mid-rant gets a bit distracting; the other is that it's a nicely constrained application with which to level up my Objective-C foo a bit. The source data is constrained by what is served by the TWFY and Public Whip APIs, and how much screen scraping of the new BBC Politics site I can bear to do without slowing things to a crawl The interface is also somewhat limiting, although I've been surprised by how much you can cram onto a 160x320 palette. Where some input from the list would be useful is around what sort of data would be useful to present in a palm-top app; and how that could be presented. My l33t dev skills aren't necessarily up to building particularly flashy interfaces, but after spending hours deep in the bowels of XML parsers it's easy to lose the perspective of what people would find useful in an app like this. Also, if anyone's interested in being a guinea pig, I'll be looking for willing beta testers once things are stable enough to deploy onto somebody else's device. Any ideas and suggestions will be gratefully received! Thanks, <PERSON><PERSON>
593a9a9f-75ea-6e38-a125-6d1e641eb1bb
['238ba853-3ef0-247d-7911-abda91c6a7f1']
In short, apt-cdrom add writes the following line to my /etc/apt/sources.list: deb cdrom:[Debian GNU/Linux 3.1 r0a _Sarge_ - Official i386 Binary-1 (20050607)] / unstable contrib main As you can see, the "dists" component is labeled "unstable". I can't figure why. I have downloaded the following ISO DVD IMAGE: d6eeef5a49ff851839bc514cc56f305a debian-31r0a-i386-binary-1.iso using jigdo. I downloaded the .jigdo file http://cdimage.debian.org/debian-cd/3.1_r0a/i386/jigdo-dvd/debian-31r0a-i386-binary-1.jigdo and the corresponding template to get the ISO image. Has anybody else had a similar problem? Is it a jigdo problem? Thanks everyone for the great work! <PERSON>
49b085a0-a1c0-dac3-8ad0-55ce35050e04
['238ba853-3ef0-247d-7911-abda91c6a7f1']
Hi everybody, Here is my problem. I have two classes, 'Body' and 'Member', and some attributes of 'Body' can be of type 'Member', but some may not. The precise attributes that 'Body' has depend from instance to instance, and they can be added or deleted. I need any instance of 'Body' to keep an up-to-date list of all its attributes that belong to the class 'Member'. How do I do this? Best, <PERSON>
70f30f58-81d0-0846-9467-bac5226e5495
['238cd7e3-a588-4088-2961-856b420411cb']
Sorry, I sent this earlier from the wrong account. Hello to the Hadoop Core development group, I am interested as to weather or not I could use Hadoop in the pseudo-distributed mode to launch multiple client processes on all cores of a multi-core processor and then re-assemble the results. Map-reduce is probably overkill for this kind of thing, but I have very limited engineering resources and can't really start from scratch with pipes or sockets or an MPI implementation. If Hadoop can be configured to do this out of the box, that would be very useful. I am also dumping my results into a Java database (H2), so it would be nice to keep the entire application in Java. I would greatly appreciate any information that could be provided. <PERSON>
ef8debd3-8bcd-46d4-b35a-74544bb75892
['238cd7e3-a588-4088-2961-856b420411cb']
Well once again I tried to update to SM 2.0. This time the executable actually opened, which is quite an improvement over endlessly double clicking on the icon with nothing happening. I tired to import my old profiles with the manual method; create a new profile with the exact name of the old profile, select the folder containing the old profile, close SM, open SM and select the profile from the list. Now when I open the profile, it is just a new profile with none of the data from my old profile. When I go and look at the profile folder, I discover that SM has DELETED ALL OF THE OLD CONTENTS. If I had not backed up the folder before trying this, I would be looking for someone with an ax right now. Whose moronic idea was this, or am I the moron for not knowing how to properly import a profile? I think that at least a warning message would have been in order here. I have literally tens of thousands of emails saved, along with lots of other data, and the main reason I use SM is so that I can keep my profiles in their own folders and back them up properly. This lets me just import them if I have to re-install the OS, or SM. Is there some reason that there is not an "import profile" tool where you just point to a folder, say import, and done? This seems like such an obvious no-brainer, it makes one wonder about opting for the wonky manual method instead? ...still on SM 1.1.17 LMH
fe927bf0-7e58-9c1b-c9bb-bd0d7c186103
['23968d8b-22f1-8b9b-9454-ca4b5037f87a']
Hi, I work for Brightcloud and part of my work required me to write Out of Core Sparse Matrices. I was thinking of submitting these to Scipy as it currently has Sparse Matrices, but not out of core. I was wondering if this code would be a desired addition to SciPy. Also, currently it uses the Python Sqlite3 library. Is it okay to use the Sqlite3 package? Could I also get documentation about writing test cases? I notice that you use NOSE. It there any specification about integrating my test cases into SciPy. I would only need a simple modification to the existing sparse test cases. Thank you,
d29dcee4-f51e-dcfd-5062-1422342ec999
['23968d8b-22f1-8b9b-9454-ca4b5037f87a']
Hi, I would like to profile my code similar to "ocount" in that I measure the amount of events (ie cache misses) that occur, but specify that they happen at or around certain lines of code. Is it possible to count the number of cache misses within a single function? single line of code? a single loop of code? etc? As a side note, when I use ocount on OpenMP compiled code, are the event counts for a single thread, or over all. Thank you,
482bd6e1-62a6-05e2-4828-5d63c1efbe68
['23979741-7849-4b60-328b-325b8d26e4e1']
Hello everybody. I've found that in WS:ca it's possible to create pages within the "Page" namespace(?) as in WS:en or in WS:fr. One example is :ca:Page:ConstitucionsCatalanesVolumIr.jpg, which takes a digitalization from Commons. There is a header, a body and a footer, and the image zooms as the mouse passes over. But it doesn't exist a Page namespace on WS:ca. Is this a bug? Why this behavior is not present on WS:es or WS:it (no page namespace, nor "Page" functionality? Why/Who/when/how it was activated on WS:ca? Should there exist an "official announcement"? And if so, should be meta or oldwikisource or local WS the best place to announce it (this present list is not as accessible as Scriptoriums are)? Where can I find more info on the behavior of "Page" pages? Am I doing too much questions;)? Thanks you all, <PERSON>
3aeacf02-cf50-617e-f0a4-b7bdf2468c5e
['23979741-7849-4b60-328b-325b8d26e4e1']
And what about the LST template combined to section tags? One example: - Page 1 is http://ca.wikisource.org/wiki/P%C3%A0gina:Cansons_de_la_terra_%281866%29.djvu/48 and contains some text with one full reference, also the beginnig of a second ref, and we can also see the transclusion of the following portion of the second ref (from Page 2, thanks to "Lst" template). We see the full second ref, but just proofread it at its "physical" page. - Page 2 is http://ca.wikisource.org/wiki/P%C3%A0gina:Cansons_de_la_terra_%281866%29.djvu/49 and contains 2 sections: some text with one reference, and the following portion of the reference (transcluded at Page 1). So, the final transclusion uses the "Page" template with the "section" parameter when necessary: http://ca.wikisource.org/wiki/Cansons_de_la_terra_-_Volum_I/Introducci%C3%B2 It's a bit uncomfortable, but also in books made of paper!! Why do they break refs?? Regards, <PERSON>
51597dc7-d6f1-4a16-884a-892c2f8894dc
['23991197-1d66-494f-a71e-72a5a0f061d6']
Hi all. We use live-build to make out distro Huayra. I consider this tool is mature and work fine for our requeriments (include our "hand made" UEFI support). I see the inminent change from live-build to live-wrapper. and search for documentacion/help/recommends to start migrate our build system to this new. ...also i'm confused because i want to known how to this tool can replace the actual live-buid if still not in the debian archive? https://packages.debian.org/search?keywords=live-wrapper Sorry, your search gave no results Can anybody confirm the current status of live-wrapper? is usable? it's posible to move to live-wrappaer without start over? =| thanks!
2bd9ea6f-bb16-9611-e3bb-4cf2567ed465
['23991197-1d66-494f-a71e-72a5a0f061d6']
Hi, we working on Huayra GNU/Linux a derivated distro from debian. At this time our 3.1 release is made over jessie. I found this behavior on the repository configuration screen of synaptic. we solved it adding the correct *.info and *.mirrors files into /usr/share/python-apt/templates i know that is a very old bug, but maybe this info can be useful for other people. Saludos!
81cc67ee-6c5e-9e6f-30ee-f4f6ada7c7b4
['23a4c4f8-d044-d8cb-8f7e-e83f9c64ade2']
Hi. I'm a bit confused, because I just pushed changesets to my private repository and TortoiseHg didn't asked for user credentials. Well username is present in the url, but not password. How can this be? if someone knows my repo url, everyone can push/pull changes? Sorry for this dumb question, i'm very new here :) Thanks
8fa681b1-c2ff-a8b7-7629-33ac7bc778b0
['23a4c4f8-d044-d8cb-8f7e-e83f9c64ade2']
Hi guys I'm a new to NHibernate. I just started development of my next project using NHibernate. One of the requirements is that entity classes should be localized. I'm doing localization the following way: - Every entity which should be localized is split into 2 table and 2 entities. For example, I have an entity Service which holds some information, and another entity ServiceLocale which holds localized properties, such as Name, description etc. Same goes for tables. - Service has a protected field which holds a list of ServiceLocales. ServiceLocale has a property Culture which specifies which language the object belongs. - The list of ServiceLocale is loaded eagerly from xml mapping file. So when Service is loaded from database, all ServiceLocales also is loaded for this service object. - Service has public properties, Name and Description. From this properties, I check what's current culture, load appropriate ServiceLocale object and return ServiceLocale's Name and Description. - Service is persisted using a repository. The repository checks saves or updates the service, and also all of it's ServiceLocale objects. So my question is: is there a better way to achieve this kind of transparent localization via NHibernate? Thanks
bb3eb852-a7f9-4f83-8eaa-40adfe8ed201
['23bd8362-5492-7a97-90d7-e3e2902a38c2']
Hello, we use Skolelinux Squeeze with KDE. We have one tjener and two ltsp-server. The computerlabs are connected to the 192. network of the ltsp-server and we switched from Thin-Clients (runlevel 4) to Diskless Worksations (runlevel 3) but the difference in speed is not very big. The harware in the labs varies (oldest system 2GHz singlecore CPU, 1GB Ram). I used to test the hardware with a standalone workstation installation and everything worked fine. The diskless clients are so slow that they are hardly usable. Any idea how to find out, why the systems work so slowly? Greetings
4258a132-08cc-d9c6-76fa-40e9f5dd9d6e
['23bd8362-5492-7a97-90d7-e3e2902a38c2']
Hello, the manual recommends the "roaming workstation" profile for the use with laptops. It says "The users' files and profiles are stored on the local disk." how are those files synchronized with those on the mainserver? We have some notebooks for use in a classroom and the students and teachers using a notebook should have acces to their files stored in the computerlab and vice-versa. Greetings
76d9959c-e291-145b-813a-f8bbaf6e11e2
['23bd980a-aff0-46cf-6d49-c679aa180178']
Hi list, I would like to represent part of a postgis table as a WMS layer in Geoserver. Is this possible? It seems really basic, until now I have never had to do this, and now that I need to, I am stumped! For example I have a postgis table of point objects. I would like to create: A first geoserver layer of all point objects where type = 1 as blue dots another geoserver layer of all point objects where type = 2 as red dots, another where type = 3 as black dots  etc. As far as I can see it seems to be one layer per postgis table. Is there a better way of doing this? I could make the layers WFS instead if that offers a better solution. I am using Geoserver 2.0  Thanks,   <PERSON>
393d164c-5c78-11e6-8fdb-b8f08489f527
['23bd980a-aff0-46cf-6d49-c679aa180178']
Hi <PERSON>, I am having trouble getting started with this problem: I have a table of points and a table of polygons. I would like to calculate the shortest distance between (each) pair(s) of points, avoiding the polygons (obstacles) in the polygon table (as they represent holes in the plane). From what I have seen, using pgRouting is the best way to approach this, however, any documentation I have read refers to a "network" which is used to calculate this distance. Do I need to generate a network of all my points and polygons to do this? Could anyone suggest some basic steps to follow please? Many thanks, <PERSON>
a7c250c3-a9f2-e769-7402-f6d4c0227d2b
['23c5df6c-5c74-51bc-b86e-3fee9310dcf4']
When using OLEDB (MDAC 2.71) and you get the parameter information for DML statement (example statement below). If the parameter is a varchar(20) for example, it return the parameter size as something like 2354453, and the flags for the parameter has the DBPARAMFLAGS_ISLONG bit set. This bit represents a BLOB in OLEDB. Can someone take a look into this? Ex. INSERT INTO CUSTOMER (FIRST_NAME) VALUES (?) Thanks
ad83ebec-7440-bc75-a8c3-3a59a2389c97
['23c5df6c-5c74-51bc-b86e-3fee9310dcf4']
I need a way to determine if a record was updated by another user. In Oracle they have ROWID, in SQLBase they also have ROWID and in SQLServer they have ROWVERSION. SAPDB has a SYSKEY but it can't be used to determine if a record was changed by another user. Does SAPDB (SQLMode = Internal) have some way to determine that a record you are currently dealing with was update by another user? Thanks
3ccd3e02-afae-37f5-ba33-3be76a939726
['23c907ad-fa3d-8a89-e391-4ea441122f89']
Hello <PERSON>, I have almost same setup: Mac PowerBook G4 15"/ M-Audio-FIREWIRE-410/Axon AX100/ FCB 1010/Gibson Blues Hawk/Logic Pro, Live and Guitar Rig. For the first. It is really good equipment for live performances. The same way as you, i use Logic for sequencing and use all menu effects and models, besides use Live like a ReWire. They working together capitally. You can use classical scheme for connection your guitar and Axon:- the magnetic signal into the Axon guitar output (back Axon) - than amp. It is more flexibility for a live performance. I asked before in Group about FCB replace the chain and hold pedals for Axon AX 100, but it is impossible. I have try many times to solve this problem but unsuccessfully yet. However you can try with Axon MKII. It's seems to me there is way to connect them. Needs right assign controllers. I use FCB to control Guitar Rig and other software. But i bought FCB specially for Axon. However...! Ask about it somebody else. Be bless.
3028bb43-4d28-5449-9009-cd4df3f3116f
['23c907ad-fa3d-8a89-e391-4ea441122f89']
Hi <PERSON>, Its Slava again ( if you remember I wrote about problem wiht AX 100 power supply). I still have not solved a problem with my AX 100 SB. On former does not work mix and guitar mode. I have checked Axon with GK-2, result same. If this Axon prompt where possible repair it. On Ukraine of no service centre. And one more question. Can MIDI foot controller substitute pedals EXP and switch? If yes that what better ( I have felt FCB 1010 Behringher) but beside me be not got ajust controllers to pedals EXP and switch. Best wisheses. <PERSON>
e9ff3a6e-fa53-ba02-fd67-bd0f0d60d2fe
['23cbce40-e019-1bfe-cf65-877a074c9557']
Hello everyone, Note that yesterday I moved CI artifact storage from AWS to Dreamobjects. However, I was not able to move the old artifacts themselves (transferring >500 GB of data that will largely be expired in two weeks between cloud providers is simply not economically feasible). Consequently artifacts produced earlier than today will be unavailable. Cheers, - <PERSON>
fe1b0d89-f7c2-c9c4-c00c-19394a7aa131
['23cbce40-e019-1bfe-cf65-877a074c9557']
Hello everyone, For the last few weeks Trac's password change functionality has been disabled due to passwords being leaked through the ghc-tickets mailing list. I have updated the Trac configuration to fix this leakage and reenabled this interface. Moreover, we shouldn't see any further email address verifications messages on ghc-tickets. Thanks to everyone for their patience while this was handled. Cheers, - <PERSON>
800b2248-743b-5af6-43a3-a548f9196d27
['23d38075-d5e6-93c2-ee02-1012f5a4377f']
Hi, and thanks for many great tips from all of you SW users around the world. I have a problem with the <Dial on demand> button in PPP settings, when i enable this function, Outlook and IE 60 doesnt seem to find my modem on the SW box. If I disable the function there is no connection problem, and everything runs smoothly. My web proxy is enabled and transparent. I have just installed the beta1, prior tho this version i have run the 0.9.9SE version with absolutly no problems at all. Hope someone can help me... regards <PERSON>
bb1c1bef-210f-5f6e-94e1-c55e611fd28f
['23d38075-d5e6-93c2-ee02-1012f5a4377f']
Hello everybody.... I am using SW2.0 beta2 everything appears to work almost smoothly. but there are two things i have to ask about. 1. The Dial on Demand function does not work properly on my smoothie, on 0,9,9 there was no problems at all, but now... 2. My ISP now demand the users to dial with a prefix, 4 digits, but mye SW wont connect when i use prefix in the phone no. in PPP settings the SW dials the no but i wont establish connection...? Does anyone have a clue??? thanks Regards Stig V
6eb495bc-6765-0ed1-0018-f145d1ca11c7
['23d85888-59ee-4d8c-9c3f-27ebb3de174b']
Hello everyone, you may remember that a few months ago I was trying to fix KOTOR to work with Mesa to use the Gallium llvmpipe software renderer. Well, it's been a while and I'm happy to see that things are a bit better with Mesa 17.2. The game still crashes, but we're closer to fixing it. Here's what I found using 17.2.1: With frame buffer effects and soft shadows the game crashes at the end of loading; the crash is inside a function that amongst other things, generates mipmaps for a texture used in a pbuffer (function at offset 2FB37D in my exe). The crash happens when gluBuild2DMipmaps is called, however doesn't seem to be a null pointer like it was back in march: it's an access violation alright but no longer a null pointer. So I think it's a different, hopefully simpler, problem. Back in march, <PERSON> suggested that changing the checks for the pixel format could fix the problem, and he was right; without those checks we definitely got a step closer to fixing it. My first thought was to just NOP the entire section that generates mipmaps and a bit of code later that uses it. The game no longer crashes, however it displays nothing, but I can hear it running in background. So this is the last issue! We're almost there! Now, I'm bothering you again because I think that at this point it's just a problem with the texture format used there. The call to gluBuild2DMipmaps uses LuminanceAlpha' as texture format as well as internal format (0x190a). I tried changing it to RGB and RGBA just to try something, but that didn't work because I guess the texture was already generated with another format. What could I do to investigate this further? And where should I look inside Mesa if I wanted to say... force a specific texture format for pbuffers? I feel that we're very close to fixing this. Your help would mean the world to me and the whole KOTOR community. Thank you ;) P.S.
866e368e-3dad-1fc6-5100-6041bf1b9148
['23d85888-59ee-4d8c-9c3f-27ebb3de174b']
In the last week I've been trying to bring an "old" game back to life, Star Wars Knights of the old republic (KOTOR, for short). It's from 2003 and uses OpenGL 1.4. I have used Mesa, libtxc_dxtn and some trickery to decompress the textures to boost performance, and right now I have it up and running smoothly with Gallium on LLVMPipe, compiled on Windows. (I can upload a copy if someone is interested). This took me about 2 days of compiling and figuring out stuff. Here's where the weirdness begins: Turning on framebuffer effects or soft shadows make the game crash right after the menu. Using a disassembler and debugger and what little knowledge I have of reverse engineering, I managed to track down the issue to a function which uses wglGetProcAddress to get the addresses of several OpenGL functions. Some of these calls return a null pointer (even if there is a valid context and it is current), and when the game tries to call them, it crashes. The first one that makes it crash is a pointer to wglBindTexImageARB, but there are a few others. NOPing the offending instructions did not work, and returning a nop function just makes the game display artifacts. This issue is not new: a guy named <PERSON>, here in the Mesa mailing list somehow managed to fix it in Mesa 11, but his patches do not seem to work anymore (https://lists.freedesktop.org/archives/mesa-dev/2016-February/106695.html). I tried to contact him but without luck. The issue affects LLVMPipe, Softpipe and OpenSWR, so I think it's something in Gallium itself. Nvidia's OpenGL driver is the only implementation that seems to work properly with this game at the moment, so if someone here's willing to help me out with this, a lot of people would be very grateful, and I am willing to pay too. Thank you
3953e9ff-0ea1-de8b-c0e7-24d37a96d89f
['23ece270-259a-6d4b-53ad-7fe512247149']
Hi <PERSON>, Is copy of AAD data to destination buffer when dst != src is mandatory requirements for crypto drivers or we can skip this copy. Actually I am bit confused, In following link <PERSON> had mentioned caller will memcpy the AAD to destination buffer but authenc.c also copies the AAD to dest. buffer. http://www.chronox.de/libkcapi/html/ch02s02.html Secondly When AAD data remains unchanged in AEAD encryption/decryption operations. Why we copy the same data to destination buffer? Thanks & Regards <PERSON>
dbeb89ad-9d1f-0de6-fb7a-7dc3d017abf3
['23ece270-259a-6d4b-53ad-7fe512247149']
Hi All, How can we open socket of type "authenc(hmac(sha256),cbc(aes))" from userspace program.I check libkcapi library. It has test programs for GCM/CCM. There are 3 types of approaches to Authenticated Encryption, Which of them is supported in crypto framework. 1) Encrypt-then-MAC (EtM) The plaintext is first encrypted, then a MAC is produced based on the resulting ciphertext. The ciphertext and its MAC are sent together. 2) Encrypt-and-MAC (E&M) A MAC is produced based on the plaintext, and the plaintext is encrypted without the MAC. The plaintext's MAC and the ciphertext are sent together. 3) MAC-then-Encrypt (MtE) A MAC is produced based on the plaintext, then the plaintext and MAC are together encrypted to produce a ciphertext based on both. The ciphertext (containing an encrypted MAC) is sent. Regards <PERSON>
78f5a9b5-06cc-6c9f-4559-4a3fad900506
['23ef897f-8dba-f38c-fa7a-7326a804161f']
Hi Would you be interested in targeting Apache Spark, Datameer, Bigdata and Apache Hive user client base to enrich your marketing and sales efforts?. *Also, we can connect you to the Users of:* AWS, Microsoft Azure, IBM, Salesforce, OpenStack, Oracle and many more. This file includes– Company name, Website, Contact name (First, Middle, Last), Title, Direct email address, Phone, Postal address, Industry, SIC codes, Employee size, Revenue size and technology/application details. Kindly review and let me know if we can connect and discuss further. Await your response! Thanks,
625cfb2f-4212-0230-61b1-fdee1978a3b0
['23ef897f-8dba-f38c-fa7a-7326a804161f']
Hi, Would you be interested in *Targeted Symantec B2B List*? We are happy to provide the database across globe, specifically North America, EMEA, APAC and Latin America. We also have other users like: *McAfee, Trend Micro, Microsoft, Kaspersky Lab, Sophos* and many more. Please review and let me know your thoughts, I will get back to you with counts pricing and more information in my next email. Await your response. Regards
df8c7783-14fc-a249-b0af-03c977449319
['23f898cc-4fe0-62f7-de7a-109e69509394']
tshark -i eth0, eth1, lo are all empty. Works under 2.6.23.0 just fine. A quick scan of the log between 2.6.24-rc3 and current tip (-rc5) doesn't show any obvious fixes, but then again, what do I know. I'll check current tip on the weekend when I'll have the luxury to have my main system down long enough for a test. Right now I'm kinda up against a deadline, but didn't want to leave it unreported. Should be easy for someone else to confirm or deny whether current tip has the problem. <PERSON><IP_ADDRESS> just fine. A quick scan of the log between 2.6.24-rc3 and current tip (-rc5) doesn't show any obvious fixes, but then again, what do I know. I'll check current tip on the weekend when I'll have the luxury to have my main system down long enough for a test. Right now I'm kinda up against a deadline, but didn't want to leave it unreported. Should be easy for someone else to confirm or deny whether current tip has the problem. Ray
5226686d-d670-13a1-3070-1607143074ed
['23f898cc-4fe0-62f7-de7a-109e69509394']
Hey all, I read through (what I believe to be) the current EPlugin docs at http://www.gnome.org/projects/evolution/developer-doc/eplugin/ , and couldn't find mention of this anywhere. The itch I was thinking of scratching was handling the receiving side of a drag'n'drop from Epiphany or whatnot. I follow a few very high-volume email lists from the web, and replying to those messages is a pain (at least if I want to be a nice guy and keep the CC lists and headers for threading). Dragging from an mbox-formatted URL such as: http://lkml.org/lkml/mbox/2005/4/1/160 into a folder would be handy, though would likely appeal to a quite small target audience :-/. Presumably it could do whatever the automatic importer does, and react appropriately on a vcard, or pop up a new message composer on a raw text chunk or a file attachment. If the event isn't there, and may be a bit of a hassle to add, don't sweat it. What I'm doing really isn't that much work, my available time to work on this is spotty right now, and I really should just bite the bullet and subscribe to the dreaded gig-a-year mail list. It's just a thought. (Heck, allowing full URLs in the import would be almost as good.) Thanks for providing one of my life-blood apps for my day-to-day work. <PERSON>
f586d69c-835e-999c-2881-0530d3126418
['23fb0128-40bd-a747-b735-8313c677233d']
I wrote the cgroups stuff, but in fact have very little experience on different systems (and understand very little the ins-and-outs of shared cache, memory banks, NUMA architecture, etc.). I can say that on the J1900, though, where two L2 caches are each shared by two of the CPU's four cores, allowing OS processes to run on a core sharing cache with the core running the RT threads killed our excellent results, and jitter measurements returned to those for conventional configurations without CPU isolation. Your guess about memory infighting seems to make sense in our tests, but if it's not too difficult, maybe it's still worth a try since your system isn't exactly the same? Also, I don't understand your comment about the DDR4 memory; is that a second stick you'd add to the J5005? (Our J1900s have only one SODIMM.) In any case, I'd appreciate hearing about any results on any CPU. OT, we encountered one interesting result entirely by accident: With CPU isolation on the same J1900, a misconfiguration caused POSIX thread selection instead of the intended RT_PREEMPT; however we didn't notice because we still saw the same excellent sub-7uS jitter! <PERSON>
19208add-401a-f9dd-58c4-3b26fc4d62fe
['23fb0128-40bd-a747-b735-8313c677233d']
<PERSON>, Very nice! This is not an easy thing to do. As <PERSON> said, we do something very similar, all inside Docker containers, to speed up armhf builds on the amd64-arch CI systems the project uses to run tests and builds. Building packages is an additional challenge, and requires installing even more package dependencies alongside the build-arch tool chain to get `dpkg-buildpackage` even to *agree to try* building your packages! The basic trick we use to get around incomplete `MultiArch` support is to install conflicting packages in a multistrap root under `/sysroot`, and tell the compiler and linker to grab headers and libraries from there with the `gcc --sysroot=/sysroot` argument. After that, the real fun begins! Your `libboost-python-dev` hack is a good example of the many tricks we use to make this work, and the result is a big, stinky mess. Happily, Docker neatly contains all this ugliness in an image that's easy to reproduce and easy to run, both in the project's CI systems and in our developer environments. You can see what exactly goes into it in the below GitHub repository (PRs are very welcome if you can improve it). Hopefully these kinks in `MultiArch` will be worked out some day and this will all be much simpler. https://github.com/zultron/mk-cross-builder <PERSON>
67147f78-95ec-c94e-1787-a40d79874d3f
['241e162b-256a-c1e7-9b9f-017dfd8d8b10']
Hi, i was trying to figure out if there is way to demodulate traffic by tapping a modem line which is called i think *blind demodulation*. I've looked for solutions like using a sound card or using the lucent winmodem. I'm not interested in if the impedance of the line decreases as a result of interception. I am just looking for theory and maybe some basic application. Thnx.
2deb6353-6031-9c7f-117a-aa1b2f48510c
['241e162b-256a-c1e7-9b9f-017dfd8d8b10']
New additional information from my survey: I've finally understood that my usual sb16 can be used to emulate a modem. 8Khz, 8 bits are sufficient. The most informative project on this topic done by <PERSON> (http://fabrice.bellard.free.fr/linmodem.html). This project is a generic modem driver for linux and includes the source code. It has complete v34 modulator,demodulator and a core v90 source but most of them are not working since it's unfinished. Also, source includes a pipe to sound card which uses linux oss driver to /dev/dsp. Readme states that two computers interconnected via sound card can trasfer v34 data but do not states how to do it. Most of the stuff is ready like dft(discrete fourirer) and fft (fast fourirer). Btw, i couldn't find enough info so, i'm going to pick Proakis Digital Comm. book from library and look for PCModulation explanation. Also, i've got my phone line properties. 62.5DCvolts when offline. 23mA current. When rings, something like 113 DCVolts appear on line but this is not true i think since AC is sent to ring the phone. While connection 5DCvolts is seen. Could anybody having enough knowledge on electronic circuits send me a circuit diagram so that i can connect my phone line to the line-in of my sound card? If i can do this, i would be very glad and continue to try to code the dsp stuff. Thnx.
dbec9dff-9e3d-9747-4675-62a0d672295c
['243abbc5-faae-70f9-7ae5-ad11dc3e0d08']
Hi All? I tried to compile php in a very similar way that it's recommended to do to support mysqli, but to support mysql: ./configure --with-mysql=enabled,/usr/bin/mysequoia_config I'll exucute ldd mysql.so genereted and get that's it's linked dinamically against libmysqlclient.so, Although it didn't work to connect to a VBD in sequoia, it can't find the host. Is there any Howto to compile PHP to support mysql extension? I suppose that i forget to change some link to mysqlclient but which one?... like i see that mysql.so is linked dinamically against libmysqlclient.so, i suppose that it would work. What i'm forggeting/missing? Regards!
00e4d23c-e4bf-4937-41ba-9965942c49d4
['243abbc5-faae-70f9-7ae5-ad11dc3e0d08']
Hi All! I was making test with mysql client and libmysequoia when i detected a strange error... if you try to execute '\s' to get the status information of the server you'll receive a "Segment Violation" error. Really this is not an important bug but i don't know if this error could hide another important error behind. Regards.
e373ee75-e0a6-7a81-a702-8eba644fc2e8
['244e1ec5-8088-8c66-f4c4-440f9ffe5f2c']
Hello, I am trying to start a vpn connection from the command line. My interface and connection are called vpn. The command i use is: tincd -n vpn -D -d2. This is what I always do to test if I setup everything correctly. In this instance I get the error message: : De opgegeven klasse is niet gevonden. ICaa.....mp4H Could not set up a meta connection to ............. I left out some om the key and the hostname, I don't no if the key should be kept secret. The Dutch error message is windows error 10109: Class type not found. I have installed the latest version of tinc (1.0.26). The computer is running Windows 7 pro. I am running as administrator. Thank you,
61003d1d-7295-6947-4f9b-51049c14411a
['244e1ec5-8088-8c66-f4c4-440f9ffe5f2c']
Hello, With regards to my previous message and to help others. Using the source I tracked the problem down to a call to getaddrinfo that failed. After a lot of trying I found out that somehow my host file got notepad++ macintosh file endings. Never even knew it existed. Anyhow changing that to Dos\Windows fixed my problem. Maybe the error message could include the error number. Looking/google-ing for a not English (Dutch) error message is hard. This proofs the added value of having the source code. being able to build it would even be better. So if anybody has some pointers than please. Henk
ec6b6722-4c5e-594a-8757-071437863b51
['244f8d03-3409-b91b-d68f-9a1b54df57f3']
Hi, Thanks to <PERSON>, we now have a port of the latest Linux kernel (2.6.33+): http://github.com/tmatsuya/linux-2.6 It compiles, but I have not tested it on the board. Once there are no regressions compared to the 2.6.23 currently in use, I will delete the old repository (http://github.com/lekernel/milkymist-linux). When the port is mature, we'll look forward to merging it into the official kernel.org releases. <PERSON>
6b572ece-27d5-0f3a-c936-cd1671f766c1
['244f8d03-3409-b91b-d68f-9a1b54df57f3']
Hi, I have a few questions about your USB stack for RTEMS: * is the code available publicly somewhere? * what type of controllers does the driver support? UHCI? OHCI? We will start developing quite shortly a USB host controller core for the Milkymist system-on-chip (which will run RTEMS) so I would like it to interface nicely with your driver. Regards, <PERSON>
1e25fb26-093a-6b45-30fb-d3fb6e6ba552
['2457b771-09c0-49d5-8d4c-65d1ba39dfb7']
The crime scene: Win NT, Python222, <PERSON> book on Win32, and win32comport_demo.py The problem: I connected my machine to another machine running Hyperterminal, using a null-modem cable, then fired up the demo. Works just fine. Now I want to learn in detail how it works. The demo script refers to SetCommMask, SetupComm, GetCommState, etc., but I can't find an explanantion how they work. I want to poke and prod my com ports, including the UART registers. Any pointers? Thanks.... <PERSON>
de3ec86d-e042-f2c7-bfb5-8b162fe8c081
['2457b771-09c0-49d5-8d4c-65d1ba39dfb7']
<snip a lot of good stuff> I, for one, find this thread very interesting. It's nice to learn that other people have different rates of learning. My personal observations/comments are: I too can pick up a programming book, and in a a couple of evenings, do the "simple" stuff...like for loops, procedures, etc. But I always seen to hit a leveling off (or several) in the learning curve.. I can't seem to progress in the sense of expressing a programming task in terms of the current lanquage of interest, until, (and this is a big UNTIL) I see a book which explains concepts in a way that makes the light go on. One example....I'm a fair C programmer. When C++ became popular, I read the usual books "Learn C++ in 10 microseconds", but I never really grasped the concepts of OOP nor classes. So I shunned them. Then along came Python..and the light dawned. It wasn't the first Python book I read, not the 8th..it was the 9th that broke my personal learning dam. Why? Because it explained the process and offered concepts that stuck, and to me, made sense. Another example: I was using Python to do some heavy duty parsing and report generation of log files. I knew I needed to use regular expressions, so I hacked at it. Some things worked, others didn't. I was getting frustrated. Then I found "Mastering Regular Expressions" by <PERSON>. Bingo! I learned more in the first 2-3 chapters than all the other books combined. I was happy to see that others cringe at their early code when they revisited it later. LOL..welcome to the club. Regards, <PERSON>
eb14dbc2-e738-fab7-3500-6b9e8bab0f0a
['245b0910-79ff-8c89-4dbe-59a9fcbbb76a']
While working on the web pages today, I realized that we should have made this ages ago. Fix a comment. Add explicit support for dealing with .htaccess and .htpasswd. This does not change actual behavior, as the default used to take care of these files, and I tested it with a full run on gcc.gnu.org. <PERSON>
6d654af1-801e-321c-a9dd-f5db5ff5c94b
['245b0910-79ff-8c89-4dbe-59a9fcbbb76a']
This is the analogon of what I did with snapshot-README earlier today, though these two files are not completely in sync (for whatever reason); I'll probably address that later. I have also update gccadmin/scripts on gcc.gnu.org. 2003-05-28 Gerald Pfeifer <<EMAIL_ADDRESS><PERSON> <a6sdIYXhYn0GeMAC@example.com> * snapshot-index.html: Refer to .bz2 files instead of .gz files. Simplify the note on not using unannounced snapshots. Gerald
9f6215ed-0e91-3c86-283d-7a6e10d3209a
['245f45ff-2c2e-907d-af8e-40305a7015db']
It would be nice to have hr keymap for ts 5 but.. I need croatian (hr) keymap for rdesktop, as i see in /lib/rdesktop/keymaps there isnt one. But there is one in the tarball downloaded from rdesktop.org. (1.7.1) If i log into my terminal server and manually switch from en to hr it works fine, so i presume that i dont really need ts 5.0 hr keymap, just rdesktop one. I also suppose when building a ts image it only includes keymaps for locale set in build.conftime in /lib/rdesktop/keymaps/ since there is only the us one there... so i tried to trick the bastard and put hr keymap too, now the -k hr switch does not output any error but just does not work. I again have en layout on the server. freerdp on the other hand works fine with en ts locale and hr keymap, but i prefer rdesktop. As i obviously do not understand how this works, any help is appreciated :) Tnx. <PERSON>
50a642be-28ee-ae21-a32c-578e125f60fb
['245f45ff-2c2e-907d-af8e-40305a7015db']
My clients are booting as they should but they wont pull the thinstation.conf.network from tftp server. As i see in the log they dont even request it ( using win/tftpd ) I put it in both tftpd root and /ts5.0 directory, and tried with thinstation.default, thinstation.conf on top of .network I used 5.0 default build.conf Network is working fine when they boot. What am i missing? I think this worked that way few years back when i tried 2.2.2.x.. or maybe im just getting senile tnx
42120aa7-54c5-312d-1156-246dcc5ccc64
['246bb2cf-1ecd-4719-32fb-f24b91975680']
hi, I have just installed debian 4 on a machine. i am running sshd on it. However, when i try to login to the machine from Putty (on another windows computer on the network), it goes ahead and asks for user name. I enter the username, and when i enter the password i get 'Network error: Software caused connection abort'. I believe the connection is lost at a very basic level. When i then try to ssh to the box, i get 'connection refused', as if nothing is listening on the port. Sometimes i am able to login successfully, but after a while the connection is completely lost at random. When lost, i get 'connection refused' when I retry. But I still can ping to the machine all this time. I checked ssh logs as well, but nothing of use there. So i tried using VNC, and same problem occurs (meaning i get connection lost), but still ping works. I am completely lost, and dont know where to look for Any help is appreciated. Thanks <PERSON>
8e49f680-924f-9ccf-4476-26d080702c2c
['246bb2cf-1ecd-4719-32fb-f24b91975680']
hi, I have a small LAN with about 3 windows machiens, 1 debian server (etch), and adsl router with 4 ports. Currently, I am using the router as a gateway (192.168.0.1), and the debian server is 192.168.0.254. The client machiens are configured to use 192.168.0.1 as a gateway. What i need to do is to use the linux server 192.168.0.254 as a gateway for all the client machines, thereby i can monitor traffic and gives me more control. But the problem is the debian machine has only one interface and it is not directly connected to internet. It is also using the router as the gateway to internet (just like the other client machines). So my question is, can this be done? if so any pointers is appreciated. I have checked routing/iptables info on the web, but in all the cases, the linux machine is directly connected to the net and has two interfaces. thanks, <PERSON><IP_ADDRESS>), and the debian server is <IP_ADDRESS>. The client machiens are configured to use <IP_ADDRESS> as a gateway. What i need to do is to use the linux server <IP_ADDRESS><PHONE_NUMBER>. The client machiens are configured to use 192.168.0.1 as a gateway. What i need to do is to use the linux server <PHONE_NUMBER> as a gateway for all the client machines, thereby i can monitor traffic and gives me more control. But the problem is the debian machine has only one interface and it is not directly connected to internet. It is also using the router as the gateway to internet (just like the other client machines). So my question is, can this be done? if so any pointers is appreciated. I have checked routing/iptables info on the web, but in all the cases, the linux machine is directly connected to the net and has two interfaces. thanks, meekaaku
94be5c00-5da6-3629-1ae9-59c157be66ec
['24726abc-9b8a-9e77-4f16-3c0bba34aa79']
Hello, We have Linked Servers set up on our SQL database and can successfully write to the related View Forms. We've had this working for years. Now, when trying to set up a new Linked Server and process, we are getting the following DTC error: ARERR [552] The SQL database operation failed. : The operation could not be performed because OLE DB provider "SQLNCLI10" for linked server "SSASVR1" was unable to begin a distributed transaction. (SQL Server 7391)OLE DB provider "SQLNCLI10" for linked server "SSASVR1" returned message "No transaction i Of course, we can write to it at the database level but not through Remedy. I have gone through all the usual steps to resolve this but have not been successful. I'm thinking that the problem has to be on the external database because Remedy is successfully writing to the historical databases. DTC is definitely enabled and has permissions. The server can ping Remedy. Component Services is configured...but we cannot seem to resolve this issue. Does anyone have any ideas? We're both running on SQL 2008 R2. Thanks. <PERSON>
ea43c2ba-313e-ccfd-7a2d-7bd29f70ba2f
['24726abc-9b8a-9e77-4f16-3c0bba34aa79']
Hi, I need to know how to disable, preferably globally, the 'Set Refresh Interval' option on Tables and on the Results List. I've found that I can do it if I remove the 'Preferences' option altogether, but I'd like to keep the Preferences and just remove that option. Does anyone know how? Thanks! <PERSON>
33c75f76-fbd0-4033-9c6a-2cd96553d47b
['24847e4a-3024-cb2f-7bdd-a040306b74c1']
Concern has been expressed on the debian-devel list about license status of PennMUSH and its legitimacy. PennMUSH was relicensed under the Artistic License as of version 1.7.6p0 in November 2002. Aspects of PennMUSH's code have been drawn from, of course, it's TinyMUD roots as well as its 'sibling' codebase, TinyMUSH (2.0, 2.2, 3.0). I spoke with the PennMUSH source maintainer and one of the developers/upstream authors, <PERSON> (aka <PERSON>), about any information regarding how the relicensing was handled and the concerns expressed about it. This is what he said in response: TinyMUSH 2.0, whose authors relicensed all their code in 1995 to a BSD license, so that's clean. We also contacted the TM 2.0 authors (<PERSON> <PERSON>) and got their agreement anyway. The next bit is TinyMUSH 2.2, which their devteam (<PERSON>, <PERSON>, <PERSON>) all agreed to relicense under Artistic (from 2.2.5, I believe), and I have their email saying so. Then there's the PennMUSH copyright holder (me, <PERSON>, Raevnos), and we all agreed. So Penn's clean. TM 3.0's dev team also switched to Artistic (as did tinymux, I believe) at the same time. I don't know anything about 2.2.4unoff, which we've never incorporated code from to my knowledge. He has also stated that he did track down all authors which followed TinyMUD, which was cleanly licensed under the BSD license, to get their approval, and has emails from them granting permission. I would appreciate any comments regarding whether concerns about PennMUSH's legitimacy under the Artistic License are valid, and legal obstacles for its inclusion as a Debian package. Sincerely,
8b211e02-f9c2-1fb1-5ad8-484fa2b60a1b
['24847e4a-3024-cb2f-7bdd-a040306b74c1']
Hello, I recently found the time to really delve into Debian packages and such and with the support of the upstream author have created a package for the PennMUSH flavor of mud server. I am also beginning to look for orphaned packages which may relate to the same category or which have been useful in the past, Jitterbug is one which has been added to my list of possibilities. In any case, I would appreciate any interest on the part of a possible mentor. The current PennMUSH package files can be found at http://debian.korongil.net/ Any input is appreciated. Also, any Debian Developers in the Michigan area who are willing to meet to verify and cross sign pgp keys, I'd be interested in doing so rather than following the alternate methods of verifying my identity. Sincerely,
6b538d4a-4a25-0d97-0dad-7a5ba130421a
['248c965c-7a61-032e-bebb-67009608f4bf']
Hello and happy new year! Im Spanish, then excuse me my English. I have just read your message so...: It is not a good way create a trigger for this? Your question, I supousse,is about how parametizer the <PERSON>, isnt it true? I Think that is not posible. Good Luck! Un saludo, <PERSON>, Mario Departamento de Sistemas ISOFOTON, S.A.
fda9807b-3076-0672-6c13-d9105571db94
['248c965c-7a61-032e-bebb-67009608f4bf']
Hi! How do you do? I have a problem maybe you can help me... I have to turn into the reports (Oracle) format to excel format and I saw some manuals but noone tell about this. If you have some information about this, I will be very happy!! Un saludo,
7e0e6b8d-fc9f-bdec-8c61-31d7b9a87b12
['24965cbf-7e33-3aa0-75e2-411ee7c91de2']
Hi all, After almost a month of hacking I finally finished (well, not quite) the new GdlDock. It has a completely new architecture, even though is like 95% source compatible with the old API. The changes to anjuta2 are also in place so you don't need to worry about that. One thing that has changed is the format of the XML layout file, so you'll probably get an empty screen when launching anjuta2. Either you remove ~/.anjuta/layout.xml or just open the layout manager and check all the visible checkboxes. There is a file called README.gdl-dock in the gdl module which you can read to learn a bit about the new architecture. It allows things like docking items to floating items, but the changes are mostly not visible. The design is a lot cleaner, which will allow us to extend the widget more easily. For example, layout saving/building is now completely separate from the dock itself, and it's handled in a separate object. There is one thing that still needs work: placeholders. Even though they work for simple things such as hiding and restoring an item's position, I expect them to be a reliable way to dock a new item in a given position, independently of how the current items are docked. The code is simple right now, so it doesn't handle special cases very well. Anyway, it's failsafe, so the worst thing that can happen is that new items float instead of being docked to the main window. That's it for now. Please, test and file bug reports to bugzilla. Have fun, <PERSON>
37113c96-5e5d-10db-5237-60ea3fafd686
['24965cbf-7e33-3aa0-75e2-411ee7c91de2']
Hi all, (<PERSON> and <PERSON> I included you in the CC list since you're the authors of BonoboDock and GimpDock) As <PERSON> requested, I made a comparision of the different dock alternatives currently available. I focused the analysis in the architecture/design, but in all cases I mention notable features. I tried to be as objective as possible, but since I wrote GdlDock I'm probably somewhat biased (because it's the widget I know better). Please correct me if I made a mistake or I missed something important. Regards, <PERSON>
bb3f999d-27f4-67ca-dcb5-8305f6de2658
['24a2ca74-9995-4ba7-d5e4-f175ebd4e0a1']
Hello, I have spent too much time on this problem and am beginning to think that it can't be done in Django. I am pulling a list of pathogen names from a postgres database to a drop down box. The user selects the pathogen, which requires the id to be passed back through django to the database for further retrieval of more data. Please see http://dpaste.com/119358/ for the drop down code. The problem is that the select id = statement is above the point where the form retrieves the pathology id. The django statements must be placed at this point or the drop down doesn't work. Has anyone written something similar and found a solution? Max
3ac42f47-d21a-9076-e9a6-45c7c7c7629a
['24a2ca74-9995-4ba7-d5e4-f175ebd4e0a1']
Hello, I have the tags <em>name</em> around particular words in the titles of my data. When I bring the titles from the PostGres database and display on the html page the title appears with the tags (<em>name</ em>). I am expecting the page to read the em tags and convert the name to italics, not display the tags with the unitalisized name. How do I get django pages to display the results of html tags? Thanks, <PERSON>
e7d6dad1-af8d-77a0-f3a4-c493197cc70a
['24a941a7-f5c4-1ec1-b12a-8a62bc02ad3d']
In my Native Client application, I often have long-running operations which may take 60 - 80 seconds to complete. I see that the web page GUI is unresponsive during this time, and the Chrome dialog warning that the page has stopped responding occurs a few times. I would like to avoid this behavior, and instead return control to the web page GUI while the operation is occurring. Specifically, my application is doing this: 1. Call postMessage from JavaScript on the NaCl module. 2. The postMessage call returns. 3. In the HandleMessage method on the class in C++ I have derived from pp::Instance, perform the long running operation. 4. When the operation completes, call PostMessage from the C++ HandleMessage method and return from HandleMessage. I suspect that I am misunderstanding something. Is it a better practice to create a thread for the long-running operation, and return from HandleMessage in C++ immediately? Please forgive me if this question has been previously answered here or in the Native Client SDK documentation, as I have not been able to find the answer yet. Thanks. <PERSON><IP_ADDRESS>Instance, perform the long running operation. 4. When the operation completes, call PostMessage from the C++ HandleMessage method and return from HandleMessage. I suspect that I am misunderstanding something. Is it a better practice to create a thread for the long-running operation, and return from HandleMessage in C++ immediately? Please forgive me if this question has been previously answered here or in the Native Client SDK documentation, as I have not been able to find the answer yet. Thanks. Josh
fc1f45e4-0259-7eb4-4793-4aad12a30d62
['24a941a7-f5c4-1ec1-b12a-8a62bc02ad3d']
It seems that Boost.FileSystem does not implement some methods accepting wpath arguments when compiled with GCC 3.4.3 on Linux. This occurs because BOOST_FILESYSTEM_NARROW_ONLY is defined, which according to the comment in the code occurs for "badly broken compilers or libraries". Ideally, I would like to use the boost::filesystem::copy_file() method on both Windows and Linux, passing wpath arguments in both cases. Is the lack of wpath support simply a problem with GCC 3.4.3, or is wpath never supported for copy_file() with GCC? I can see from the implementation of boost::filesystem::detail::copy_file_api() that open(), read(), and write() are used to copy the file. Since open() accepts only a const char* this may be the limitation. If this is the case, has there been any discussion of providing a copy_file() method which accepts wpath arguments and converts from wchar_t to char before calling copy_file_api(). I do not have a lot of experience with locales and character encodings, so I may be missing something. If this has already been discussed on this mailing list or elsewhere, please point me to the correct discussion, as I have been unable to find it. Thanks. <PERSON><IP_ADDRESS>filesystem<IP_ADDRESS>copy_file() method on both Windows and Linux, passing wpath arguments in both cases. Is the lack of wpath support simply a problem with GCC 3.4.3, or is wpath never supported for copy_file() with GCC? I can see from the implementation of boost<IP_ADDRESS>filesystem<IP_ADDRESS>detail<IP_ADDRESS>copy_file_api() that open(), read(), and write() are used to copy the file. Since open() accepts only a const char* this may be the limitation. If this is the case, has there been any discussion of providing a copy_file() method which accepts wpath arguments and converts from wchar_t to char before calling copy_file_api(). I do not have a lot of experience with locales and character encodings, so I may be missing something. If this has already been discussed on this mailing list or elsewhere, please point me to the correct discussion, as I have been unable to find it. Thanks. Josh
f375fa6e-45e0-a6dd-4544-c14e88d17ee3
['24b375f8-60be-d77c-eeac-23d818c4abbc']
I've seen this on our network in recent months also. It ususally has to do with virtual machines that default to using 192.168.x.x (VMware) and 10.211.55.x (Parallels) addresses. They either exit their physical machine not properly NATed for your network, or when they interact with some applications, such as MS Exchange, the Exchange server may try to reply to the original 192.168.x.x or 10.211.55.x address. Apparently this original source address must be buried somewhere in the data portion of the packet. Either problem makes the origin very difficult to trace, because you can't route to, or ping, that source address. <PERSON>
a8fdbab2-9af3-6189-8a05-cb5c3fa71e86
['24b375f8-60be-d77c-eeac-23d818c4abbc']
We've been told to install a vender solution for video/audio streaming. The vendor, RAW Communications, feeds their on-site server (MS Win2K) via a satellite download (receiving only, no transmission back to the satellite), and then uses multicast to supply the video stream to the local desktops. The vendor requirement is that all ports be open from the server to the desktop for a single multicast address. Is there any way to do this securely? With minimum exposure? My initial suggestion was to isolate a couple of machines and just allow the service to those desktops. But unless we can come up with some real world examples to show how unsafe this can be, we will likely have to open this up to our entire LAN. Thanks, <PERSON>
ca1b2dba-fca8-9c71-b154-1b3476cb6450
['24be7124-404c-2cdd-7883-78e3f0fdd99b']
can anyone tell me how to enable the options menu? i already read some of the old posts which points me to horde.conf... can someone paste the exact line i should alter... cant seem to find it.. i tried looking at prefs.conf hmmm almost all conf in imp and horde but with no luck... TIA
e26c1739-72e2-eaa5-5168-f4c14cbe5ff0
['24be7124-404c-2cdd-7883-78e3f0fdd99b']
Hi, Im new to apache.. initially the box is running on apache2 ( w/c was installed via rpm together with php ).. now im trying to downgrade my apache2 to apache1.x... i thought everything is working as planned but... unfortunately I got this error.. Directory index forbidden by rule: /usr/share/toaster/htdocs/admin/ any idea how I can fix this... TIA
5013a615-3a93-744d-d7a2-8f742a39fb71
['24d3981c-a564-45a8-615b-11d3db4daddf']
Hi, I am trying to do some computations using Laplace transforms. I used the continued fractions approach to compute laplace transform of a birth-death process as described in [ http://www.columbia.edu/~ww2040/continuedfractions99.pdf Abate 1999 ]. But I cannot find a simple numerical routine to compute the inverse Laplace transform (evaluated at 0 in my case). Does anyone have ideas on how to do this? Much appreciated. <PERSON>
4e3f6863-af55-bd07-6eb0-c8e9fca3169b
['24d3981c-a564-45a8-615b-11d3db4daddf']
I am trying to use http://rss.acs.unt.edu/Rdoc/library/stats/html/constrOptim.html in R to do optimization in R with some given linear constraints but not able to figure out how to set up the problem. For example, I need to maximize $f(x,y) = log(x) + \frac{x^2}{y^2}$ subject to constraints $g_1(x,y) = x+y < 1$, $g_2(x,y) = x > 0$ and $g_3(x,y) = y > 0$. How do I do this in R? This is just a hypothetical example. Do not worry about its structure, instead I am interested to know how to set this up in R. thanks!
3343be0b-1764-ac2e-e46c-b26b9fe2939b
['24df451b-4e46-88a5-3f9b-b4fa3d0bd9e0']
Hi, Here are some commits out of which #3 would be the most important. It fixes the signing of keys for a specific account/service, then verifying the resulted signature. This is a harmattan platform specific functionality for the moment. 1. Added aegis-crypto, aegis-builder to build deps and created aegis manifest files. 2. Increased timeout for the test_blocking test. - Test was failing with the original timeout 3. Fixes: NB#Account::verify and Account::verifyWithTokens() returns TRUE all the time irrespective of whether Account::sign() called or not. Br, <PERSON><IP_ADDRESS>verify and Account<IP_ADDRESS>verifyWithTokens() returns TRUE all the time irrespective of whether Account<IP_ADDRESS>sign() called or not. Br, Aurel
559377df-40fb-e05a-217a-2d21f2df6b24
['24df451b-4e46-88a5-3f9b-b4fa3d0bd9e0']
Hi, Please review a few fixes and improvements: 1. Fixed permissions for the Signon storage directory. This will allow `Restore` operation to work. https://gitorious.org/accounts-sso/signon/commit/955acb690aab318633a64d0c536f5137362e95e6 2. Allow authentication core cache to append BLOB authentication data. In case of authentication cache multiple inserts for the same cache key, any newly inserted BLOB data will be appended to the existing one, while already cached data is not deleted. https://gitorious.org/accounts-sso/signon/commit/e4aa472c350aaef8cbdf817bb749c21e08d89641 3. Removed TODO. Not allowing duplicate parameters in the auth. session data. https://gitorious.org/accounts-sso/signon/commit/adbb62c233f0fe4d2fb7850243d3e981622fb80e 4. Improved authentication data caching If the secrets db is not open, the SignonIdentity caches authentication data upon a successful storing operation. This helps signon clients to perform SSO after storing credentials, without sending username/password through the auth. session data, even when the secure storage is closed. https://gitorious.org/accounts-sso/signon/commit/fe9cc7a93270f735b24c68c9298e60315a4f14b7 Br, Aurel
a201079c-1c0b-7be9-cb5b-21f1bb88438c
['24dfd5bf-fbf0-6d06-1dfc-6e53c412a158']
i agree that it doesnt look very nice that way, with ->property or ->method() , that doesnt negate its usefullness in my mind however. if it was every imlemented, i would suspect that properties and methods wouldnt be syntaxed the same way, which, for that reason alone, makes it unlikely it would ever be implemented i guess. Too bad, it is handy. <PERSON>
96839adf-64e3-91c0-e91f-6f05fb24a995
['24dfd5bf-fbf0-6d06-1dfc-6e53c412a158']
Hi, im having problems disabling the service that sends my pc into standby with redhat 8, i've shutdown the apmd service not only by stopping it (and pc still went to standby) but also by shutting down the service on this runlevel. ps ax shows no results for a ps ax | grep apmd (its actually shut down..:)) ive tried things with xscreensaver, that does nothing. any insight? what have i forgotten. <PERSON>
84e6f574-4efa-0d88-2cd1-534634ca849f
['24ec2204-cd37-bf6e-8a74-331b1fde9230']
Hi, I have a machine with 1G RAM, running Xen 3.0.1, and Linux. I want to create two domU VMs and move ALL the memory, flexibly, back and forth between the two VMs. To do this, I created two domU VMs running Linux, both with kernel boot parameters mem=1G, to tell Linux to prepare memory tables for 1G RAM. I booted the VMs, with xm parameter 'memory=256', i.e. telling Xen to only give them 256Mb RAM during the startup. I wanted to later be able to give 700M RAM to one machine and 300M to the other, by using the balloon driver ('xm mem-set' command) Creating and booting the VMs worked fine. But then I tried to use 'xm mem-set' to give more memory to the VMs, and that did *not* work. It's because Xen has set the 'Xen hard memory limit' (visible in the VMs' /proc/xen/balloon) to 256M. I tried to use 'xm mem-max' to change that up to 1G, but 'mem-max' does not seem to do anything. Is that correct? Can it be fixed? There is a bugzilla entry asking for 'mem-max' to be removed, but it would be nice to have a way to boot up a domain with less memory than it is allowed to have in the future. http://bugzilla.xensource.com/bugzilla/show_bug.cgi?id=212 Is what I'm trying to do perhaps doable in some way, without using mem-max? One way that I thought could be an alternative way is this: I tried to create the VMs one by one, each with 1G of RAM, and ballooning them with mem-set down to 256M before saving them. Then I wanted to restore them, as they then would have the mem-max paramters and Linux memory tables set up correctly. Sadly, they couldn't be restored, because Xen tries to allocate all the memory (= the memory given at the first boot) when it restores the VMs. So only the first VM could be restored, even though they only used 256M each. I mentioned this about a year ago, and got the impression that it was being worked on. Was it ever fixed? "out-of-memory problem when restoring ballooned domains" http://lists.xensource.com/archives/html/xen-devel/2005-03/msg00480.html http://lists.xensource.com/archives/html/xen-devel/2005-03/msg00498.html http://lists.xensource.com/archives/html/xen-devel/2005-03/msg00502.html Kindly
0c7e5f9e-f752-4a38-0923-1304867154d2
['24ec2204-cd37-bf6e-8a74-331b1fde9230']
Hi, I am serving a repository with "git instaweb". In the web browser, the name in the Project column is only ".git", and not the "myrepo.git" which I would like to have, and as it is on http://repo.or.cz/ (which is a site that also uses gitweb). How do I configure it to serve the project directory (or even better, all the repos in a directory), and name them appropriately? (Is it possible to do with instaweb, or can it only be done with a more heavy approach like using apache?) Thanks for any quick ideas, cheers, /Lars
43823646-ee3e-cbb7-66c6-3966eab72191
['24fc6303-525b-f357-489b-81f936306ae2']
Hi, Do you know about some efforts of making a free alternative to wolframalpha.com with maxima as engine? (I suppose tha maxima is the best candidate, at least for simbolic calculations) Sincerely, I prefer the philosophy of having installed the programs in your computer instead of using internet for solving your needs, but sometimes it is useful to use software as a service, for example, when you are away from your computer and you need to use other computer for making calcs, and it is not good to install that computer ~100MB of software temporally. For this cases, I consider that having a website under a free license as the AFERO is a good idea. It could be good relaxing the matched as correct sintax of the input, or suggest what you may mean. Also to give as output the exact command that would generate that answer in your installed maxima. Thanks for your experience. <PERSON>
2c3a707a-66f3-38d6-1803-2a26670a5b04
['24fc6303-525b-f357-489b-81f936306ae2']
Hi <PERSON>, I'd like starting learning the skills for helping to write a firmware replace for the 88WW8388 chip. I was reading on the wiki, specially this entry http://wiki.laptop.org/go/88W8388 However, I don't know which is the current state or if some code is already available. I couldn't find any yet. I found this link on the Libertas wiki page: "Discussions about how to open libertas firmware" http://lists.laptop.org/pipermail/devel/2007-January/003720.html I ask you for information and colaboration. Cheers, <PERSON>
a1adbe3c-cbbe-fd37-773f-eee0520a0e31
['24fd807e-c5ea-8f73-9983-390ef77a305f']
My name is <PERSON> from Buenos Aires Argentina... I want to join you for spanish documentation. I have download from cvs php-gtk-doc tree. Someone can write me the steps that i must follow to do the work. I can do it well and fast..i need some help to start... conventions,software to write, test? thx
02e21d23-e8cf-01f3-4c6f-573a425b8bba
['24fd807e-c5ea-8f73-9983-390ef77a305f']
i have almost translated doc to spanish...only i have to translate some files on references/gtk but i need some help on using cvs...so i could create ./es directory but i dont now how to add all translated files and directories on /es/* please help me...... after translating what about the creation on php-gtk site of links to spanish doc.... thx joseluis
342a1cc5-27c9-b0d0-600e-6f8d99ed2e76
['25233046-c9d4-0e65-aabd-61a1ab99446b']
Dear colleagues, A reminder that the deadline for submission to IAEP 2018 is a week away! Please submit your abstract by February 15th. Our Call for Papers for the 22nd annual meeting of the International Association for Environmental Philosophy, to be held in State College, PA, can be found at https://environmentalphilosophy.org <https://environmentalphilosophy.org/> All best,
85570e9b-01c2-9c77-6b42-8b9d1b67a1f7
['25233046-c9d4-0e65-aabd-61a1ab99446b']
Dear colleagues, We have extended the deadline for the 2018 IAEP meeting until March 4th. Our 22nd annual meeting of the International Association for Environmental Philosophy will be held at Penn State in State College, PA. Additional details and the CFP can be found at https://environmentalphilosophy.org <https://environmentalphilosophy.org/>. I hope you’ll consider sharing your work with us, if you haven’t already. All best,
6ac9f9fb-9ed3-3d0b-1b15-d32596a57eae
['252cbce4-350b-fddf-ac70-9ce0425c1017']
Usage of getauxval() within the float128 libgcc IFUNC resolver functions is causing problems: https://sourceware.org/bugzilla/show_bug.cgi?id=21707 <PERSON> describes why we can't have relocations in IFUNC resolver functions here: https://gcc.gnu.org/PR81193 With the addition of __builtin_cpu_supports (), we no longer need to call getauxval() to query the HWCAP/HWCAP2 masks, so let's use that instead. I have verified with some small test cases, that we do call the correct __{add,sub,...}k3_hw() functions instead of the *_sw versions. I did that by running the test cases in GDB and manually setting the IEEE128 bit in the HWCAP2 mask stored in the TCB before the resolvers were run. This bootstrapps and regtests with no regressions, ok for trunk? I will note that this patch causes issues in some tests in the GLIBC testsiute, which Tulio is working on fixing (it's a GLIBC issue, not a GCC issue), so if this patch is "ok", I plan on holding off on committing this, until the GLIBC fix is committed. <PERSON>
f78dd349-edd8-72bd-f51c-f9ebc2e43dc3
['252cbce4-350b-fddf-ac70-9ce0425c1017']
This patch adds support for __builtin_cpu_init(), __builtin_cpu_is() and __builtin_cpu_supports() builtins for PowerPC. We use the same API as the x86* builtins of the same name. These builtins uses the new GLIBC 2.23 feature where we store the AT_PLATFORM, AT_HWCAP and AT_HWCAP2 values in the Thread Control Block (TCB) which offers very fast access to these values. As part of the agreement with the GLIBC community, we always emit a reference to a special symbol exported by LIBCs that support the AT_PLATFORM/AT_HWCAP* values in the TCB, whenever we expand one of the CPU builtins. We do this so that we will never attempt to access the TCB on old LIBCs. <PERSON> also asked that we conditionalize the enabling of this code with a configure time check for GLIBC's version and that is included here. I'll note that since GLIBC initializes the TCB before the application gets control, we don't actually need __builtin_cpu_init(), but we have implemented it anyway, to keep the same API as x86. It's just our init expands to nothing. This passes bootstrap and regtesting with no errors. Ok for mainline? <PERSON>
efa887bc-75a1-2c53-54a1-ec2a27121ea6
['2535cc40-82ec-9cce-30ae-ad10f1d4ef8e']
Hello and happy new year to everybody, I am looking for peers for a new SKS keyserver installation. I am running SKS version 1.1.4, on keyserver.sincer.us. We are an small IT company, and as part of a new project we are starting to run our own keyserver! The server is physically located in Germany (EU). The machine has IPv4 connectivity. I have loaded a keydump from pgp.jjim.de, dated 04-Jan-2013 at 01:37. I see 3190051 keys loaded. For operational issues, please contact me directly. keyserver.sincer.us 11370 # Petru Ghita Sherar <<EMAIL_ADDRESS><PERSON> Sherar <RJvz20do285Ri8tk@example.com> 0x7CF29D04
421affd5-6bd9-ff91-9a16-2a92b09bafb7
['2535cc40-82ec-9cce-30ae-ad10f1d4ef8e']
Hello everybody, keyserver.sincer.us needs more peers as it seems it's falling out of the pool due to not having enough connectivity and therefore missing keys. So if you'd like to peer please add the following to your membership file and drop me an email so I can add you back to mine. keyserver.sincer.us 11370 #Petru Ghita <<EMAIL_ADDRESS>> 0x7CF29D04
e359d9bb-6e66-b628-ad63-a8f188bd60dd
['2539b83a-f9ed-b2bf-5c32-7aa61cab28b1']
Hi, I waited for it to finish for almost 5 days nothing happened except creation of those files. I had posted a similar mail some time back. Is there no solution for this. Is it something to do with speed of the system? I ran the program on HP dual core system. <PERSON>
a4f49b42-b9ef-1162-c03b-b51f6d0a6404
['2539b83a-f9ed-b2bf-5c32-7aa61cab28b1']
Hello gmx_users, I wish you a very happy and prosperous new year 2007. I have a very fundamental question in trjcat usage. I have two *.trr files of 5ns and 10 ns runs . I would like to concanate the two and make a 15ns trr file. How to give the command so that the starting time of the 10ns file should be 5001ps and not 0ps and end time is 15000ps and not 10000 ps. May this is a trivial question. Kindly help. I have tried using the following command : trjcat -f nchnp3_5npep.trr nchnp3_10pep.trr -settime -o nchnp3_15pep.trr -b 0.00 -e 15000 its taking 0ps to 10000ps only hoping for help regards sharada
1b06cc33-043e-4bd6-b580-57aecffe7c2f
['254327ed-aab2-2e25-5712-aa6a83b3172d']
Hi <PERSON>, Thank you for your help. I would definitely like to port the video modifications to a newer release, if time allows. However, I could resolve the problem on my own. Here are the points that I consider as important: 1. Using laptops, the CPU should not be set to the power save (or power on demand) mode. Automatic CPU frequency adaptation causes artifacts ( "cpufreq-set -g Performance" is the solution). 2. I've replaced the port audio lib with the stable snapshot from the portaudio website. It reduced the application start-up time. Additionally, I've replaced pasound.c (and adapted other files respectively) to avoid the sampling mismatch for OSS. 3. However, using OSS sound driver (dsp devices under linux) causes some minor artifacts on my hardware, so ALSA must be used. Thanks, Blazej
f7026fd6-659b-6ec0-6257-98fc06579075
['254327ed-aab2-2e25-5712-aa6a83b3172d']
Hello, We modified the PJSIP modules/libraries to support audio/video. For the video functionality, we integrated the ffmpeg framework as a third party code. Our aim was to adapt our code to the original pjsip structure. The general concept results as a high-customizable Video Phone. So far, it performs well. However, the client still needs some development time to finish and make it public. We would like someday to share the code and contribute to pjsip by making our code open-source. Cheers,
a664acea-c9b7-335a-c866-3bb37893e286
['2543d5b7-abe1-0f6f-647a-9a980ec0d972']
Having just returned from a tour of central and southern France I have to report that the lack of a "France addresses nationwide europe.obf" is a nuisance of the first order. It makes navigating across the country so much more difficult and when a friend texts me the name of the town in which he is located and to which he is hoping we will go to meet up, but does not give me (or even know) the name of the region he is in, how am I to find it short of resorting to my paper road atlas, a book I hoped I had finished with. This applies to all countries split up into regions. Come on developers, get this sorted or I, for one, am off and no more recommendations from me. As a result of this I am now looking for a different mapping application, has anyone got any good ideas? <PERSON>
1ca7ebf3-8382-c5a5-0f7f-d6856c226b52
['2543d5b7-abe1-0f6f-647a-9a980ec0d972']
The fact that OsmAnd tries to take one back to a missed waypoint is the reason I do not use them at all when navigating. It is a real nuisance and the developers should take note of all the dissatisfaction this is causing. I have taken to having my next waypoint as the destination and resetting for each of the stops. What a bind but the risk of being sent back is too great. Yours <PERSON> (R)
3f2c0eb7-74b6-c8c9-9217-78c0a3aa5b65
['254a2b24-b8ee-7263-6fdd-67edefd981c0']
I'm trying to simulate adding an OSD to a cluster.  I set up an OSDMap::Incremental and apply it, but nothing ever gets mapped to the new OSD.  Apparently, the crush map never gets updated.  Do I have to do that manually?  It seems like apply_incremental should do it automatically.  My test case is below.  It shows that the OSDMap is updated to have 11 OSDs, but the crush map still shows only 10. Thanks, <PERSON><IP_ADDRESS>Incremental and apply it, but nothing ever gets mapped to the new OSD.  Apparently, the crush map never gets updated.  Do I have to do that manually?  It seems like apply_incremental should do it automatically.  My test case is below.  It shows that the OSDMap is updated to have 11 OSDs, but the crush map still shows only 10. Thanks, Adam Crume
b78a2327-ef6a-b7cd-db4d-b7cb6d0fa38c
['254a2b24-b8ee-7263-6fdd-67edefd981c0']
LTTng requires tracepoints to be linked into a program only once. If tracepoints are linked in multiple times, the program crashes at startup with: "LTTng-UST: Error (-17) while registering tracepoint probe. Duplicate registration of tracepoint probes having the same name is not allowed." This is problematic when mixing static and dynamic linking. If the tracepoints are in a static library, that library can end up in an executable multiple times by being linked in directly, as well as being statically linked into a dynamic library. Even if the tracepoints are not linked directly into the executable, they can be statically linked into multiple dynamic libraries that the executable loads. For us, this problem shows up with libcommon, and could show up with others such as libosd. (In general, I think anything added to noinst_LTLIBRARIES is static, and anything added to lib_LTLIBRARIES is dynamic.) There are a few ways of solving the issue: 1. Change every library that has tracepoints, like libcommon, from static to dynamic. This could be a big change, as at the very least we'd have to rename the library to something like libceph_common to avoid conflicts, since now it would be an installed file. This has the advantage of keeping code and its tracepoints in the same library. 2. Keep tracepoints in a separate static library. For example, libcommon and libcommon_tp. Unfortunately, every executable (but not library!) that links in libcommon (directly or indirectly) would have to manually link in libcommon_tp, and I don't think Automake could automate that. 3. Keep tracepoints in a separate dynamic library. In this case, I think libcommon could depend on libcommon_tp, so executables would not have to manually link in libcommon_tp. (I'm not an Automake expert, so let me know if I'm wrong on that.) Again, libcommon_tp would be an installed file, so we'd want to rename it to something like libceph_common_tp. I attached a minimal test case of the problem. Thoughts or suggestions? Thanks, <PERSON>
87b96c54-c955-3ea3-e718-f43aa66e6236
['254bede6-d4b4-8807-10a9-ead65cfaf797']
Hi, this is mostly a resend of the series. The only change is in patch 9/10, where I removed the dev_warn that could be an issue on some laggish operations on Synaptics RMI4 devices (when retrieving the heatmap). It also appeared that these warning were sent quite regularly with the Elantech touchpads, so let's just disable those given that they are mostly harmless. Cheers, <PERSON>
34ef1ba7-826d-beb2-490e-8d467682e5ae
['254bede6-d4b4-8807-10a9-ead65cfaf797']
Hi guys, well, here comes the promised v2 of the ll_transport cleanup. As I said, I removed patches which need some more work, and kept only the trivial ones. I also added <PERSON>'s documentation, which gives us a net difference of +210 lines of code :( Let's say that we still have a net worth of -106 lines of actual code :) Cheers, <PERSON>
76980d3f-c573-1260-928e-9e26905c208d
['254c0c7b-f8da-c6fa-33f1-2b4e8c8fe8be']
Many people seem to be having similar problems, but I've tried everything people have posted, and nothing seems to work. I cannot post the dmesg unless I use my digital camera. I am not aware of a way to dump the dmesg from the installation image boot. If needed, I will post it on a webserver and link it. I am using: OpenBSD 3.4 (I also tested 3.3 --same problems) Tyan Tomcat S2850 with Opteron 140 Newest BIOS -- 1.03 Basically, OpenBSD will not install at all on this board. Fishing through the newsgroups this also seems to be a problem in the S2466 and S2469. I heard lots of different solutions, none of which seem to work for me. I freezes right when/after it probes the npx0 device. Oddly enough when you hit the CTRL key specifically it unlocks and continues. I do not expect this to make sense. CTRL should send the same interrupt AFAIK as any other key. I reproduced this many times because I didn't believe it myself. It then says ... "stray interrupt 7" ... then freezes again later. I proceeded to boot into UKC (User Kernel Config) and disable npx0, and then it doesn't freeze, but freezes later. Ok, so it has something to do with the numeric coprocessor and interrupt 7. It would seem like a blown board, but upon googling for it, I see many people who have almost the same exact problem. Notably, they're often AMD-based Tyan boards -- and they're not even the same one I have. http://www.netsys.com/openbsd-misc/2003/04/msg00437.html He posted a solution as well up top, but it doens't seem to work for me. Someone also said to try disabling the LPT port in the bios. I did that too. First, I'm baffled as to why "stray interrupts" should occur, and I'm baffled as to what it has to do with the npx0 module. And I'm further confused as to why it still crashes even after the npx0 issue is "resolved" by removal -- and disabling the FPU is not something that is reasonable anyway .... Lastly, OpenBSD does not recognize the Opteron as a 686 class CPU. It sees it as a 386. This is probably a bad idea, but I don't think it's causing this mess. If so, can I go in with a boot -c and change my CPU to a 686 class somehow? I couldn't find a way. I am posting this seperately to the list since it is definately an issue. <PERSON>
57611798-bc58-9d61-0520-484f871d72a6
['254c0c7b-f8da-c6fa-33f1-2b4e8c8fe8be']
Sorry to repost, but I should have posted this as a seperate message ... : I believe some things have changed with the php4 port in 3.4. I assumed "flavors" would be easier to use than to pkg_add all the stuff I need from PHP, but ... "make show VARNAME=FLAVORS" returns an error, and I see no documentation. I could piece through the make files .... but I'm not confident I'll know exactly what I'm doing even after that. I'm using the standard chrooted apache on OpenBSD 3.4. How should I be building PHP in 3.4? Maybe I should just use the package system already, but it's bothering me that I can't figure out how to use the "flavor" system. Can someone give me a hint? make; make install works .... but i want to tinker with the "flavor" ... <PERSON>
212035fb-4ee0-16e3-11d1-89bffa6b3784
['2570cfa8-9d57-f2c9-eb21-d9f146371c19']
Would you all be so kind as to recommend some help systems for the 4D help files? And a builder for custom applcations. We'd like something that runs on osx that works like the old Microsoft Help system. Searchable, cross referenced, and fast. We have used the os9 Microsoft Help, but the files are too old. I don't think the 2004 help files will work with it. Ideally, what we'd like is something like the RealBasic Help app. Something that would work with all the 4D product lines and version and <dreaming> have a component for displaying extensions like the command and parameter heavy ALPro.</dreaming> HTML help would be fine, but it needs to be searchable and provide a persistent list of references, indexes, etc. We develop mostly on osx, for cross-platform apps. I know we can use the built in HTML help within 4D, but it is not really searchable. And not fast. I know MacOSX help is searchable, but it is not fast and does not keep a list of matches once you hyperlink around the file for a while. It is not very good for a programmer's needs, am I right? I have seen ASG's Commander a long time ago, but the version currently on the site seems to need a password to extract it. Not ready to buy until I see what it can do. So I haven't looked at recently. Thanks, <PERSON>
158fab5e-6a23-f646-403c-732594ad5587
['2570cfa8-9d57-f2c9-eb21-d9f146371c19']
Is it possible, with native 4D, to drag a file or folder into a 4D form and have it trigger an event so we can read the file? We need this for both Mac and Win. The plug-in Escape's Qdrop looks perfect, but I thought I saw this as a new feature in 2004. Thanks, -Michael Check
c7a862ec-322b-0790-45d1-93b68dac8785
['258be056-ab30-b22f-86fe-db382d684085']
I have recently installed a big disk (1.4 TB sata raid with scsi interface) on an i386 computer running FreeBSD 5.4-RELEASE-p1. I have 3 partitions on the FreeBSD slice. The disk is working OK, but the disklabel seem a little strange. For one of the partitions the "fsize bsize bps/cpg" values are "0 0 0". For the other two partitions these values are "2048 16384 28552" which I beleve is more normal. The filesystem is initialized (with newfs). fsck gives no errors. Partial or full output from the commands "dmesg", "df", "mount", "fdisk da1", "bsdlabel -A da1s1", "tunefs -p /dev/da1s1e", "fsck /dev/da1s1e" and "dumpfs /dev/da1s1e" is attached below. Should I worry about this, or can I ignore it. <PERSON>
639b8841-c53f-cab6-8f71-130657fe9891
['258be056-ab30-b22f-86fe-db382d684085']
I have recently tried to upgrade a horde installation on a FreeBSD 5.3 x86 computer. In the new /usr/local/www/horde/imp/config/servers.php.dist file there is a reference to http://wiki.horde.org/ImpUWIMAPNamespaces. This web page recommends me to apply a patch to the /usr/ports/mail/imap-uw/work/imap-2004g/src/osdep/unix/env_unix.c file in order to change the mailsubdir from '' to 'mail' for the imapd server. I have tried to apply the patch, but afterwords the imapd server still doesn't seem to work as intended. I have probably done something wrong. Here is what I have done: 1. I saved the patch to the /usr/ports/mail/imap-uw/files/ directory and named the file patch-src-osdep-unix-env_unix.c-horde-imp 2. cd /usr/ports/mail/imap-uw make patch (I had to replace some spaces with tabs in the patch file before the patches was applied correctly. Probably because I pasted the patch from the web page). 3. I checked the file /usr/ports/mail/imap-uw/work/imap-2004g/src/osdep/unix/env_unix.c and the patch seemed to have been applied correctly. 4. make -DWITH_SSL_AND_PLAINTEXT -DFORCE_PKG_REGISTER reinstall 5. I checked the date of the /usr/local/libexec/imapd file, and I confirmed that it had bee updated. 6. reboot 7. At last I tested the imap server from a Mozilla Thunderbird client where I changed my "IMAP server directory" setting from ~/mail/ to nothing. (Removed my ~/mail/ setting). Now all my files in my home directory appeared when I opened the Subscibe window. This is not the intended behavior. I know that I can make a /etc/c-client.cf file to fix this, but this is not a recommended method, and it also affects pine and the location of the users mbox files. Has anyone managed to change the behavior of imap-uw imapd server as intended using this patch ? If so, how did you do it ? <PERSON>
ac6d5cb5-deda-c733-b1f0-fecdf7e0c53f
['259757b0-72a2-33c4-f009-139354055641']
Folks, I know we're kind-of behind the curve here, but I'd just like to let you know that the SIM team's 3.4 plans are now available online. The plan itself is at http://sim.netbeans.org/plans/features34.html and a rudimentary user view is at http://sim.netbeans.org/nb34/userview.html I'd appreciate any feedback. Particularly about the user view, as the SIM core is largely infrastructure and so I didn't know quite how to write that. Regards,
70e039be-30d1-1753-21e1-d1efba63fd7f
['259757b0-72a2-33c4-f009-139354055641']
Folks, If anyone has a quick answer to this, I'd be grateful. Otherwise no doubt I'll find out eventually by reading the source... How does autoupdate decide where to put the module jar it gets from an NBM? We have an NBM which seems to install into the system dir, rather than into the user dir as we expected. This is with FFJ4.0 EA, if it matters. As a test, I installed serialversion and that one *did* go into userdir, so it seems it must be something about our NBM. Thanks in advance for any tips, Regards, <PERSON>.
fecbe3f7-b50f-40f3-1eae-6b47ab9f294e
['259830e6-11e1-4d64-ba0d-825b2ef2387c']
Hello, I have very recently deployed my application on Google App Engine. I have also enabeled billing of a large amount. However my application failed at the cliets site today. It was working fine on all the clients before then it suddenly stoped working on some of them. We also observed that it continued working on some clients while on other clients it dint respond to any of the requests. First thing I did after it failed is checked the logs on the Admin console. There was only one message of some High CPU usage for some requests. Then I ckecked my enabled billing usage and it was just 1%. The only red mark I could see was on the usage of some url requests. Please help me fix this issue. Any help is appreciated. I am really sorry if this is not the right place to post this request. But please could you direct me to where I could find help regarding the same. Thanks again.
dd47042a-6f8e-1e58-807e-6e2e42a3e332
['259830e6-11e1-4d64-ba0d-825b2ef2387c']
I often see the following message in my application logs: " Request was aborted after waiting too long to attempt to service your request. This may happen sporadically when the App Engine serving cluster is under unexpectedly high or uneven load. If you see this message frequently, please contact the App Engine team. " Also our application is currently under development. We get this message for a few of our requests: "This request used a high amount of CPU and will soon exceed its quota" What is it suppose to mean? The latency for these requests is very high (about 12000 ms) however we have assigned a large quota for our application(100 $ / day) when my application still generally runs under its free quota. We are working on the optimization of our application. However our algorithms would still require a large amount of CPU. Please let me know how we could resolve these issues, Any help is appreciated. Thanks.
cb16d312-88ce-34b1-950d-c408f2e0531e
['25a0b751-75cb-6e4b-5bf5-4a248cafe358']
Hi, from time to time I get files like xyz.asc which may contain PGP keys. Sometimes I only need a few keys from such a file, but it is a bit annoying to make a "gpg --import xyz.asc" just to see which keys are included in this file. So, is there a quick way to check and print the key id's of keys contained in a file, without having to import the keys ? Something like "gpg --viewkeys xyz.asc" would be nice... <PERSON>
ff03426e-17b3-ffaf-0d0b-63f39027d596
['25a0b751-75cb-6e4b-5bf5-4a248cafe358']
Hi, one of the common capabilities of all AES candidates was to offer different key length, e.g. 128,192,256 bit. So one can edit his key in order to make use of AES256 instead of "standard" 128 bit AES. Why is this not the case with Twofish ? If I look at cipher.h it seems 128 bit Twofish is deprecated and only experimental and the cipher option "Twofish" always refers to 256 bit Twofish, so why is there no Twofish/Twofish192/Twofish256 just as it is the case with Rijndael ? Are there any special implementation issues regarding Twofish one should be aware of ? I can't imagine that 128 bit Twofish is dropped due to security reasons... <PERSON>
55a852b8-4526-e07b-a7df-c0152615ff2c
['25ac5f09-882c-db76-6557-10473c0b629f']
<PERSON>, Thank you for your answer again. I try to check the comment's lines at HTML, but there is a lot there and I don't know with sure which ones I can edit. I save our HTML in a text archive that I'm sending to you. I hope that this can clarify better our problem and that you can help us. Our blog have a lot of visitors each hour and they are complain about this problem. Thank you so much for your help. <PERSON>
fb1d2624-0a76-fc6f-2672-2bab8852ff96
['25ac5f09-882c-db76-6557-10473c0b629f']
<PERSON>, I did your suggestion about the forum but I didn't get any answers there to help us. Meanwhile we work hard to solve the problem of comments and we did several tests, since starting a new blog with everything new or even using the simple models templates. Summarizing we did all possible tests, but after we did using different ways, the problem of comments is still there when start the second page of comments, this is, after 200 comments. We got a conclusion that maybe this is happening by a blogger problem that we can't fix without the own blogger help, that this is probably a developing problem. Can you verify this, if the problem is yours instead of us? I will wait for another answer from you and I thank you in advance for your help again. <PERSON>