query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
6aa4f389-82dd-6b4a-fe4f-cb6a0321a7e0
['72066748-6c97-65a5-5540-3cdcef93d0b2']
Don, Is this the option you are looking for? MQRO_PASS_DISCARD_AND_EXPIRY If this option is set on a message, and a report or reply is generated because of it, the message descriptor of the report inherits: v MQRO_DISCARD_MSG if it was set. v The remaining expiry time of the message (if this is not an expiry report). If this is an expiry report the expiry time is set to 60 seconds. <PERSON>
34ee3d40-0166-4b3e-ee6e-27191d122c33
['72066748-6c97-65a5-5540-3cdcef93d0b2']
Intra-Group Queuing will allow your QMGRs that are part of a QSG to send messages to each other without using the channel initiator. The QMGR you connect to would need still need to know where the local Queue resides (just like when you use channels). This does require shared System queues to be defined. Gregg
3f6141e9-c867-bfdb-9afd-78a55541d965
['7207f195-5d2b-09bd-d516-25259960a714']
Hi all, I took a look on the exception classes generated by wsdl2java and got it to work now: The point seems to be that the exception class needs to be derivated from AxisFault, then the deserialization works. I'm not happy at all to have to derivate my exceptions from AxisFault, as I want to keep my business logic interfaces (and the exceptions which belong to it) as independent from the underlying RPC technology as possible. Does this really mean that it is not possible to use custom exceptions which are not derivated from AxisFault? Regards, <PERSON>
6e885be1-0e6e-6983-f75c-781f0fdec5ae
['7207f195-5d2b-09bd-d516-25259960a714']
Hi, Once again, as no one replied: Did really no one else face this problem up to now? Don't you use exceptions in your service interfaces? I took a look on the JAX-RPC specification 1.0 now: It defines that service specific exceptions must extend java.lang.Exception - and not something like RemoteException or AxisFault. This means that Axis 1.1 doesn't comply with the specification in this point, unfortunately. Should I write a bug report? Regards, <PERSON>
2b3fe6b8-8752-896a-767a-e4446d23071f
['72261789-2813-75d9-5eae-73340fc58092']
I'm noticing a RealVNC viewer issue when running VNC over an SSL VPN connection versus an SSH connection. Under the SSL VPN that desktop connects and is displayed properly. The desktop window manager is Ubuntu Unity-2D. When I left click on the file browser icon, no problem that works fine. When I left click on Ubuntu Dash icon, the entire desktop session is disconnected and the VNC server is terminated. I have a similar SSH VNC session that does not exhibit any problematic behavior. Does anyone have any clues on what may be happening? I suspected something relative to the mouse and keyboard mappings, but I have been unable to identify anything specific that will correct the problem. Kind regards, <PERSON>
3bc5d78a-6695-fcd1-1c7b-e785f2e01cfb
['72261789-2813-75d9-5eae-73340fc58092']
Sorry for inundating everyone with this new Jython questions! I am trying to use the Xerces DOM parsers from Jython. Please notice the object casting below. I have been working through Chapter 10 in the Jython book to figure out how to accomplish this Java object casting in Jython. Can someone help me with how to do this? Kind regards, <PERSON>
f5b2799a-2100-28ef-8404-dfd1657008b6
['72299629-052e-2c9b-0038-70c75a700cef']
I have to implement a single sign on solution, but would like to know when we really use CAS. When we say CAS as a single sign on solution for many appliactions, does this mean that all these applications as internal applications for an organization? I mean, with CAS , we do the authentication for all these application against the SAME db/ldap/any-other-authentication-data-store .Is this idea correct? Or if I need to have a single sign on solution just to have integrations with external applications like Salesforce,Siebel-on-demand etc. does CAS really help? Pleas let me know. Thanks <PERSON>
55f80257-feea-675f-9aba-62287d4ac846
['72299629-052e-2c9b-0038-70c75a700cef']
<PERSON>, This works now for the IDP client , when they tried with another certificate ,which is a 1024bit/SHA1 certificate. :) Mostly issue should have been with the wrong key as you guessed. And it might have worked now, when they tried with a new setup , with another new certificate. But they think,the issue was Our SP implementation(using OpenSAML) , doesn't support 2048/sha256 certificate. But I had tested using my test certificate which is of 2048/sha256, and it worked fine . So just wanted to confirm once again that Opensaml doesn't have any restriction on certificate key size/algorithm as 2048/sha256,? Thanks <PERSON>
20c214ba-9f1b-9b13-263c-3b8ea6dc1b01
['722acd2e-05e1-b81b-7d8d-7fd9a402a5e8']
*This message was transferred with a trial version of CommuniGate(tm) Pro* I am trying to install mysql. I have completed 2 steps 1. running the installer 2. putting a my.cnf file in the /library/mysql/var directory 3. Installation has already made a mysql startup file and as per the instructions below Next we need to start the MySQL server before we can set the root account password. The installer will install a Mac OS X Startup Item into /Library/StartupItems/MySQL. This will make sure that the MySQL server is started at boot. You can execute this startup item manually to ensure that the MySQL daemon is properly started. You can do so by executing the following command as root in the terminal: shell> /Library/StartupItems/MySQL/MySQL & i need to check this when i run this it says you must be root to run ConsoleMessage what do i do. I am aleready the admin user. For your info I am trying to install on Macosx Jaguar 10.2.2 Bhavul
5f8249e8-400a-d6bc-c015-ddf28a441ae6
['722acd2e-05e1-b81b-7d8d-7fd9a402a5e8']
*This message was transferred with a trial version of CommuniGate(tm) Pro* I have recently loaded mysql on macosx 10.2 It has loaded it in the directroy /library/mysql presently I have my user prompt on the terminal screen. I know that I need the sql prompt in order to be able to issue sql commands. How can I get to the sql prompt. Please help <PERSON>
f16a89b2-07af-3b24-4954-c5cd25088dbd
['7235ac7b-8c60-c9b4-3609-5abf552cf8b9']
Today I found myself asking, why it is that I continue to read this list and use this project. And the answer I came up with is.... Beats the hell out of me. Could it be the friendly, helpful people? No that's not it. Because once again I find that instead of an answer to a question, there is a discussion on the "correct" way of asking the question. Once again I find people being accosted for having the nerve to ask a question in a way that someone finds to be improper. And that making sure that they are asking properly, is more important then them getting an answer to their question. Makes you feel all warm and fuzzy, doesn't it? So I've asked myself again, why do I continue? Because it's my opinion, that it's one of the best projects of it's type to be found. BUT, and there is a but, ... <PERSON>
1955aa20-2581-1689-322e-67c0cd7b2706
['7235ac7b-8c60-c9b4-3609-5abf552cf8b9']
For those of you that are using / have used IPCop with a compact flash, What capicity are you using / would you recomend? Looking at moveing from Smoothwall to IPCop and using a smaller case. And does IPCop come with lm_sensors, or something simmlar? Have a micro-atx case, and want to do away with as many of the fans as I can ( ie. the power supply fan, and the case fan. Keeping just the cup fan, and a bit of case modding ( completely diskless. ). So I will meed to monitor temps under load for a while to see if it over heats. Thanks, <PERSON>
b44fc5a7-13ea-0c6d-8bb9-d2b182b592d4
['7243cfcc-86ca-e13b-a41c-bdd796fb395e']
Hello, I've noticed quite a lot of non-album track additions lately, and am a little unsure how to handle them. This is only for tracks that haven't been released in album form at all, right? My default is currently to vote 'no', with ideally one note to the moderator to add albums instead, and to vote 'yes' only if it's clear that it is really a non-album track (by note, or suggestive title). And the occasional 'abstain' when it's unclear. Is that ok? Since I don't really see a way to add non-album tracks via the web interface, it seems these are usually submitted through the tagger. Does it encourage submitting unsorted MP3s as non-album tracks? And if so, could it be made to warn users (more explicitly)? Cheers <PERSON>
507fd8ac-c806-ac04-f8c1-69f921dd2c89
['7243cfcc-86ca-e13b-a41c-bdd796fb395e']
Hello, my flatmate and I just tried using the newest sarge-i386-netinst.iso to install Debian on his Compaq notebook. Some incomplete notes on how it went follow. I realize that this is not released yet, and don't mind it not working. I just wanted to give the new installer a test, and hope some of this feedback is of use. Please CC: me on replies as I'm not subscribed. Here goes: The kernel booted flawlessly. Actually, hardware support didn't seem to be a problem as far as we got. The keyboard worked fine with English layout. Selecting the German keymap didn't actually change the keymap. We tried both the PS/2 and USB keyboard -- I don't really know how it is really connected, though I'd guess PS/2. Not a big problem, though. CD-ROM detection worked well. The installer module installtion seemed to work fine, although there were some error messages I couldn't place. I didn't write them down and don't have access to the machine currently, but could try to get them tomorrow. The network device was apparently probed successfully. However, the static network setup failed because ifconfig didn't work -- busybox printed its list of commands when calling ifconfig from the console. We then tried the harddisk setup, using the all-in-one option. The drive had already been partitioned with Partition Magic, which also created filesystems. However, we didn't get the option to mount these partitions without first choosing what filesystems to create. This also happened on a second installation attempt, where the filesystems had definitely been created correctly. After entering the mountpoints, this module finished successfully. As checked from the console, the partitions were mounted correctly. However, /target/etc/fstab was not created, which caused the next modules to fail. In particular, installing base from CD-ROM failed. Cheers <PERSON>
7e50a9f6-b068-11aa-f0ed-5b7e28139924
['7258e62e-de0e-cdd1-2a82-3111dd8dfab0']
Hi! I've just installed OpenCMS 5.0.0 this week and started to create a web site. Everything runs great, except I would like to use another source code editor than the built-in one. Is there any way to let OpenCMS call a .exe file when I select "Edit sourcecode"? Also, I tried installing "LEdit" just by copying LEDIT32.DLL into the WINNT/SYSTEM32 folder, as the readme file suggests, but it won't show up. Thanks for helping, <PERSON>
0797348e-9fed-a487-0361-f86180e642bf
['7258e62e-de0e-cdd1-2a82-3111dd8dfab0']
Hi all, I am using multiple content areas for selection in the WYSIWYG editor as <PERSON> described in http://mail.opencms.org/pipermail/opencms-dev/2003q2/005734.html Is there any way to check whether one content area is empty (i.e. the content editor did not enter anything in that particular body)? The return type of CmsJspActionElement.include(...) is void, so I can't directly check it. Is there any way to retrieve the content of a content area as, say, a String, where I could further process it in my JSP code? Thanks for any hints, <PERSON>
79f1449d-2306-c414-4933-3bcea8ed52ed
['725fe1da-187a-6936-373b-8a7424bd3af1']
Hi, I'm working in the catalan localization of Sunbird. I see that ca is red in the tindebox http://tinderbox.mozilla.org/showlog.cgi?log=Mozilla-l10n-ca/1220715480.1220715552.11380.gz But now I'm a little lost with all the recent changes. Today I have checked out calendar project, from CVS TRUNK, MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. And I see that dir calendar/locales/en_US/ is missing in MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. Is that correct? When I compare locales with TRUNK I see that everything is ok. So why is tinderbox red ? Thanks, <PERSON><PHONE_NUMBER>.<PHONE_NUMBER>.11380.gz But now I'm a little lost with all the recent changes. Today I have checked out calendar project, from CVS TRUNK, MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. And I see that dir calendar/locales/en_US/ is missing in MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. Is that correct? When I compare locales with TRUNK I see that everything is ok. So why is tinderbox red ? Thanks, Quim Perez
90230f70-5c2f-555a-ba79-de26e666c4f4
['725fe1da-187a-6936-373b-8a7424bd3af1']
Hello, I don't understand why Catalan locale is rejected. https://l10n-stage-sj.mozilla.org/shipping/pushes?locale=ca&av=tb3.1 I see that my dashboard and tinderbox are green: http://tinderbox.mozilla.org/showbuilds.cgi?tree=Mozilla-l10n-ca When I finished and commited my work on Saturday 30th , I saw that https://l10n-stage-sj.mozilla.org/ didn't get the update, as fast as http://l10n.mozilla.org/dashboard/. So I couldn't validate https://l10n-stage-sj.mozilla.org/. After that I've been busy until today. I would like to sign in if it still possible. Sorry for the inconvenience Regards Quim.
5858c363-abaa-be18-7248-383bcf3983b9
['729088ef-3e73-c250-f248-f690b97137b4']
Hello, we should have a mechanism to manage the state of the documentation. The documentation is only helpful if the documents reflect the actual situation of Qubes OS. For example, documents which are written for Qubes v1 or 2 are outdated. Wikipedia displays certain meta information above articles which are outdated and/or need to be reviewed to reflect the latest changes. My idea is to have a similar mechanism. For example, it would be helpful to look after every Qubes OS, Fedora or Debian release through the documentation and mark the documents which need to be updated for the new release. In addition, it would be helpful to create an issue for every outdated document. This way it would be easy for the community to contribute to the documentation. Best regards <PERSON>
ca7e3543-3886-7ca6-8a1e-348ec4497676
['729088ef-3e73-c250-f248-f690b97137b4']
Hello <PERSON>, thank you very fast response. You did not answer my question about Axt-Style vs. Non-Axt-Style. What do you prefer or better what should be used Axt-Style or Non-Axt-Style? I will create a script to remove all {% highlight trac-wiki %} qvm-grow-private {% endhighlight %}, let this be my concern! Best regards <PERSON>
2cb80650-5752-7796-5c00-80a28d8717ed
['72936592-6484-d638-ad5a-89ff677acb1c']
Okay, I patched the doc/Makefile.am and got this. I'm a really dummy with autotools, so I can't say I know what I'm doing. When I tried adding aclocal -I m4 and automake to the compile script, I got the same error as before I patched the doc/Makefile.am. Any suggestions? <PERSON>
a5ec4455-f719-22df-4838-59c61a94f72f
['72936592-6484-d638-ad5a-89ff677acb1c']
An updated 0.7.0 beta for Red Hat/Fedora systems (one that works this time) has been uploaded to: http://www.videolan.org/pub/testing/vlc-0.7.0-test1/rpm/redhat-fc/ If anyone is using Red Hat 9, let me know which dependancy files need to be updated. Since I'm using Fedora now, I tried installing RH9 on my old laptop, but it'll need a new keyboard to work right. <PERSON>
38cd4dce-d078-6d85-1636-3e402927482b
['729adf5c-1b23-0cfa-fd81-7cc7c09dce69']
Hi, First a warning, Io is new territory for me and I'm using VisualStudio 2005. Could someone post an example of the IoBindingsInit() function for the IoVM-2005-10-17-patched.tar.gz so I can try some OpenGL examples? The MSVC solution IoFull-2006-05-26\projects\win32vc2005\Io.sln is out of sync so I try to put the pieces together without using the _build.io. If I'm completely off track some guidance is much appreciated. Regards <PERSON>
8488663c-eb4a-5f6b-fae0-3f0acbdad8c3
['729adf5c-1b23-0cfa-fd81-7cc7c09dce69']
Hi, I have created Visual Studio 2008 solution and project files for the new structure in the git repository. It is not complete but includes Mesa (softpipe), glu and glut. I have also added a demo application (gears) to see if it works. I had to make some small modifications (cast from void pointer) to the mipmap.c file in the glu project. I have also included some c files generated by python when using the scons build system. This is my first post at mesa-users so be gentle with me ;-) Regards, <PERSON>
6b1f86fe-050c-ebdd-398a-9518b28da266
['72b20c9b-ab19-fd41-58a7-2d5389aebc37']
Hey folks, as it turned out to be very hard to revert all patches from v3-2-test due to conflicts, the following will happen: On May 23 (release date for 3.2.0rc1), I will create a patch to remove everything from v3-2-test which is not in v3-2-stable. If you are interested in having a history, please revert your patches yourself before that date. Cheers, <PERSON>
e028897e-b647-da8d-4122-2e18dfb9c704
['72b20c9b-ab19-fd41-58a7-2d5389aebc37']
Hello, I have a question about winbindd design. We noticed, that with "security = domain" winbindd enumerates all members of a group plus all users who have this group as their primary group. "Security = ads" only enumerates the mebers of the group, not the users which have it as their primary group. A Sniff shows, that the MMC works in the same way as rpcclient / "security = domain". There ist a second ldapsearch request for users having the group as their primary group. The question is whether winbindd should do the second ldapsearch request, too. This means to behave different to Linux, but like Windows. With kind regards, <PERSON>
8fd930bb-ef4a-58f3-9d2a-f875ac597791
['72b3f5ef-6781-3c81-2194-b31b6a632908']
hello there in our company we use two different trac. an internal and an external. is it possible to put a prefix in front of the ticket id's so we can differentiate between internal and external tickets. cause till now it's possible to have to tickets with the same id (intern, extern). thanks for your help regards pascal
af6e1351-d71a-35cb-7cc8-e01ea0f4baf5
['72b3f5ef-6781-3c81-2194-b31b6a632908']
Hi all I want to install trac on a real webserver. I made a subdomain trac.mydomain and got a new folder "trac" in the subdomains folder. how can i install trac on my server? what source should i use? i'm confused about the infos on TracInstall site. thanks for your help regards
91b3ead4-20fb-b6ee-3196-b7dbe67514fd
['72c3134a-6243-14df-62e0-d830268f5cf2']
just as the topic describes. I'm currently doing SNAT to force some destination tcp ports to be routed through a specific route rather than the default route. To accomplish this I mark thoes packets with iptables, use 'ip' to specify marked packets via the specified route, and then use iptables to change their source address. one problem i've noticed is that in netstat (and hence /proc/net/tcp) the connections are shown as their original source. this causes problems with some things such as identd, because identd won't see the SNATed to ip with that port open. hence it will reply that it has no connection. anyway i can get around this? <PERSON>
a9ee4a2f-9e0e-1d14-b210-d2b211f6d31e
['72c3134a-6243-14df-62e0-d830268f5cf2']
I am trying to secure a machine that will have potentially untrusted users. I would like to set a max memory usage per user though if this is not possible then I can possibly work with setting a max memory per process and then setting a max process, though that mix and match is not so adjustable. As the topic states, the question is if options that say 'in core memory use' (memoryuse setting) or 'in core memory size' (memorylocked) are per process or per user, such as the process limit is. thanks
da52d5a0-e14b-35c4-19e4-7c989e157aaf
['72c6b4be-0871-3182-df63-8bbb2f5dd874']
I'm trying to compile in support for the 64-bit ifXTable counter support for the IF-MIB for Solaris 10 Intel (and Sparc ultimately). Based on bits I've read in the archives, I've a) run ./configure with enable-mfd-rewrites b) updated the source to 5.4.2.1 distro What I'm running into is that 5.4.2.1 creates a package (my package installation is an outboard creation process that's been updated from many successful installs of the result for 5.4) that installs and starts up the agent successfully if I use the ./configure options *without* enable-mfd-rewrites (which is to say, the options I've used to date). If I reestablish the build environment with -enable-mfd-rewrites specified, I get an agent that starts up printing over and over to stdout insert (-1) error on subcontainer 'interface container' No other loggable output (I tried -Dread_config) gets a chance to print anything else out before the log/stdout gets flooded with this message over and over again in a loop. Now, before I get into too many more specifics, should this work, based on experience anybody has had putting the -enable-mfd-rewrites option in per the instructions of README.solaris? I've read bits where folks suggested just going with the Sun-issued net-snmp binaries, but the reason I'm going to this trouble is that we have customizations (at least one of which has turned into a bugfix update) into our version of the agent. Apologies in advance if the word on this has been brought up on the list or in the archives and I just didn't find it. Regards, <PERSON><IP_ADDRESS> distro What I'm running into is that <IP_ADDRESS> creates a package (my package installation is an outboard creation process that's been updated from many successful installs of the result for 5.4) that installs and starts up the agent successfully if I use the ./configure options *without* enable-mfd-rewrites (which is to say, the options I've used to date). If I reestablish the build environment with -enable-mfd-rewrites specified, I get an agent that starts up printing over and over to stdout insert (-1) error on subcontainer 'interface container' No other loggable output (I tried -Dread_config) gets a chance to print anything else out before the log/stdout gets flooded with this message over and over again in a loop. Now, before I get into too many more specifics, should this work, based on experience anybody has had putting the -enable-mfd-rewrites option in per the instructions of README.solaris? I've read bits where folks suggested just going with the Sun-issued net-snmp binaries, but the reason I'm going to this trouble is that we have customizations (at least one of which has turned into a bugfix update) into our version of the agent. Apologies in advance if the word on this has been brought up on the list or in the archives and I just didn't find it. Regards, Wayne Tackabury
88303fe7-166f-c3a0-3b81-068bdecca38b
['72c6b4be-0871-3182-df63-8bbb2f5dd874']
Hi all/<PERSON>/Wes: For all the pass scripts I've written in the past, I'm surprised I haven't run into this before, but I haven't. I'm writing a pass script which will effectively synthesize a gauge and string object, to effectively detect a condition worth trapping on via a snmpd.conf 'monitor' directive. So the point is, my pass scripts need to GET other object in the management scope to make dynamic and conditional decisions on to drive the value of the OID's that the pass scripts themselves expose. The problem is-I run the pass script no problem when directly invoking it with -g, and it works with the agent if I return static values without doing any snmp gets or get-nexts (in other words, the basic protocol through the pass script and its return is AOK). The problem is-if I actually do a snmpget from the pass script, it times out. Now as it turns out, the get I need is from my application proxy agent, listening at an alternate port, so I simply updated the pass script to explicitly target the get at localhost:<that_port>. and all works fine. So I have a workaround, but I'm trolling for a general understanding of how to fix this the "next" time or when part of what I need is from, say, the ucd-snmp MIB or something whose instrumentation is built into the agent. I'm suspecting the issue here is reentrancy. Now, there's a build option, which has had some reports of not working over the years, called "-enable-reentrant", which appears to affirmatively define (i.e., actually compile in) a fair degree of mutual exclusion primitive API code for the agent based upon what I can see from the #define hierarchy. On the other hand, I can't find in the code where this would actually affect reentry through the top of the GET PDU request processor. OK, that's a coding question. Either way, after a rebuild with "-enable-reentrancy" in my target environment, the timing out on GETs issued from within my own pass scripts still is happening. And along with that, what my real questions are (I'm doing this on Solaris 10, gcc compilation from the 5.4.2.1 distro, etc.): -- should the reentrancy support target the exact kind of thing I describe above, where a pass script servicing a GET request needs to issue another GET request for another item? -- is this some kind of Bad Thing for pass script usage anyways? If so, can you folks suggest an alternate strategy? Thanks in advance! <PERSON><IP_ADDRESS> distro, etc.): -- should the reentrancy support target the exact kind of thing I describe above, where a pass script servicing a GET request needs to issue another GET request for another item? -- is this some kind of Bad Thing for pass script usage anyways? If so, can you folks suggest an alternate strategy? Thanks in advance! Wayne Tackabury
ba2b162b-dc9e-dbe8-205a-390d26d2ea46
['730d51b1-7247-ebd1-93e9-74f3ede44d20']
I have down loaded disk1 of FC2 and cant get the loader to run. the system boot till i get running /sbin/loader. then ti gose to the blue screen and hangs for a while and the screen starts filling up the random char's. I tride it on 3 computers 2 pentiume II and p4. tryed to burn it with 2 diffrent burners each time I checked the md5sum before burning it. the new computer I'm trying it on is a intel main board (D865PERL) with intel p4 processor 256M DDR 333MZ ram 20 gig ide hard drive <PERSON>
db277683-be9f-33cf-19a5-8447c879a4a5
['730d51b1-7247-ebd1-93e9-74f3ede44d20']
Hello I have an IBM e-server x-series 335 with 2 xeon 2.6 GHz and 1 Gig of ram. It uses a built in raid controller I have set up as a Raid 1 with 2 36 GIG hard drives. If I try to install using 64 bit version it tells me that there are know 64 bit CPU and to use other version if I install 32 bit version with the ACPI=off I can get it to install but it locks up all the time. Most time I get a CPU 1 not responding error also when I check the hardware it reports back 2 cpu I would have thought that it should have shown 4 cpus since I have 2 dual core cpus running hyper threading. The severe is an old one and the problem could be in the hardware, I have been using Fedora since core 1 and never ran into so much trouble dose any one have any experience install fedora on this server? <PERSON>
918d946b-b05b-6e4a-99de-3a400458d189
['731a236b-9a23-545b-9987-d944d11d3988']
Hi all. What do I need to change so that Apache creates a world-readabe access_log file? I use a small tool (GeekTool - http://projects.tynsoe.org/en/geektool/) to display the Apache log. Since I'm not normally logged in as administrator, I need to change the read permissions of /var/log/httpd/access_log to all. Only problem is - whenever the logs get rotated, Apache creates a new access_log with the previous strict permissions and I have to change them again ... I thought this might be a FAQ but couldn't find anything at apache.org or via a google search. While this isn't a thing of life and death, changing the permissions every week becomes tedious. So thanks in advance for any help. :) Regards, <PERSON>
26feda41-9b13-f43b-27f0-14d3bcb2b8d4
['731a236b-9a23-545b-9987-d944d11d3988']
Hi all. I think this is of potential interest to all Mac software developers, so please excuse me in case it feels a bit off-topic to you. _Introduction_ MacPAD is short for Macintosh Portable Application Description. MacPAD files are used to get information about Macintosh software, like prices, versions and updates. Read more about MacPAD here: <http://macshareware.net/sdk.html> or here: <http://macpad.sourceforge.net/> _Discussion_ Some people, including <PERSON> and me, felt the need to add some missing features to MacPAD. We started a private discussion and thought that it is time for the community to join; since it is you, the developers, that are supposed to benefit by using the format. We set up a wiki at <http://wiki.furrysoft.de/phpwiki/index.php/MacPAD> to collect ideas and discuss possible implementations. So please join us, add your thoughts and spread the word! If you know other developer lists or forums that you think might be interested, send them a copy of this message. The topics we are discussing currently are: 1. Localization 2. Currency 3. Plugins 4. Private Extensions 5. A New Name Hope to see you soon at the wiki. :) Regard, <PERSON>
f529d2fb-3d20-49e9-db66-014ec8163672
['7325a291-f672-ea55-1b78-6d7446ee523d']
i have sent this to both lists in hopes that someone out of both would have had experience with the framework. I have read the docs in the source code download of wonder but I need a little more guidance. I have a website (http://visionworxpro.com , not done yet) and would like to integrate paypal payments seamlessly into the site. Can someone provide me with a little more guidance than the docs provide. I don't like the deal about the Listener deal. Can I just have a component that uses some sample paypal conneciton code (api's) to connect to them to make the transaction and then handle everything else my own way ?? Thanks, <PERSON>
69aa280f-ad49-d003-4e5e-a51eb96124d8
['7325a291-f672-ea55-1b78-6d7446ee523d']
This is the first part of a two part series on Display Groups. Also learn how to create a custom project template for XCode for code reuse and to make your future projects a breeze. Covers To-Many Relationships, Editing Context issues, API docs, learn to use Jakarta's Commons Lang Utilities and create a fetch spec in code and in the EOModel. This episode offers a lot of tips/tricks. Video is available for $14 dollars and is nearly 2 hours long. <PERSON>
b16f4698-1640-6479-b8d1-5548b0867180
['73290adb-70fa-a2c0-eaed-4b839a9ff8b2']
Hi, I wanted to fresh up the 'Joining Active Directory as a Domain Member Server' HowTo on the wiki for 4.x. So it was the first time, I had to compile samba 4.x as a normal member server (not as DC). And there a (maybe stupid) question came up: On 3.x I run the 'configure' from inside the source3 directory. Is this still the same, if I only need a normal member file server (smbd/winbind) with version 4.x? For testing I configure/make/make install from the source3 directory and the fileserver seems to work like expected on a first test. I only wanted to know if this was the right 'configure' or if the one from the top level of the sources should be used in this case too, like when compiling for an DC. Regards, <PERSON>
19454e71-71dd-2c0b-8a56-f1768063c1a4
['73290adb-70fa-a2c0-eaed-4b839a9ff8b2']
Hello, at work we are currently looking for something that allows to store files on a Samba share. After a defined time, the files/directories should be accessible only read-only on the share. So it should be a kind of WORM feature (of course it wouldn't on Unix side. But that's OK). So my thought was: Maybe this could be done through a VFS module (create time + x days < current time: write access denied) My questions: - Is there already a module that does something similar and can be easily adapted to provide a kind of WORM feature? - Can this be done by some who is familiar with C? Or should it be better done by someone who is Samba code experienced? - Do you guess it's much work? 1, 2, 3 weeks,...? If the total costs wouldn't be to high and we could realise it, the final module should of course find back to the Samba project, as it could be interesting for others, too. Regards, <PERSON>
209740a0-cf72-dec4-a79b-3642db8fddd5
['735f1c2b-58a1-c8ac-a46b-4d423ef2bf0d']
What small tasks can a relatively inexperienced person do to help along the alpha port of redhat? I got KDE3 to work, and I thought that at the very least I could test some packages and report where they are breaking. Is this information useful? I understand that because of the lack of interest (read: money) behind the alpha port it often lags behind. I don't think this is really anyone's fault but I would like to help where I can (and as time permits). <PERSON>
0f4b3a43-d281-a72d-c97b-653a09d494a3
['735f1c2b-58a1-c8ac-a46b-4d423ef2bf0d']
Until the problem with the adaptec driver is tracked down (I understand these things take time), I was considering using a symbios card (I have one lying around) in the place of the adaptec one. Are there any known issues with generic ultra-wide symbios chipsets in 3.2? If not, I"ll proceed to upgrade! Thanks all, <PERSON>
6f9cd0a2-dca6-a30c-50c2-02c98154212b
['7367912f-c737-7aa1-0796-ba409e6229ba']
Dear All   Can all of you who responded to this subject line share their mobile phone numbers and the area they live? If most of you are working in Hi-Tech area, we can plan for Lunch in Hitech Area  and share the lunch bill.   I think working day will be good start with and then plan for next meetings.     Regards USMurthy
bf94178a-93a8-890b-1ed4-8a3eb1720f97
['7367912f-c737-7aa1-0796-ba409e6229ba']
Dear All I would like to have a Linux System dula bootable with Window XP To start with I want the following functions in the Linux - Internet browser - DVD Writer in Linux Can somebody suggest a place where I can get the PC and Linux CDs in Hyd? Thanks Murthy
0d3875c1-2896-dbeb-998a-9978a59b7225
['736ef59b-5a77-e5ee-25db-cf4e5f4a698d']
Hey <PERSON>! I've found it easy to hook up with a designer, even from Christchurch ;-) Actually I've always worked with designers half way round the world, even before that from Munich. However, that was mostly doing icon designs. Based on what design work you need your mileage might vary.. discussion intensive topics like complete UIs could become tricky, but then again there's always Skype and Co. Just search the list archives for people recommending designers for various topics, that's how I've found my guy! cheers, Jan
fc9e2d60-a819-0cee-3a6a-a67569fc1dca
['736ef59b-5a77-e5ee-25db-cf4e5f4a698d']
Hi all - can anybody suggest some good site, directory, online service, etc. where a Mac dev can find an experienced UI designer? There's sites like 'http://builditwith.me/' and tons of places to find icon designers on the web - but these seem to be swamped with people who want to give it (UI design) a try but don't quite seem to have the experience.. at least that's what it looks like to me :-/ I'm currently looking for somebody re-imagining the UI of my new Cocoa app. It's a rewrite of an existing Carbon app with a healthy user base so I'm quite confident when it comes to the set of features to be included in the new app - however exposing that functionality in a way that won't overwhelm rather casual computer users and just making the UI super-awesome is where I'm in desperate need for somebody who has a designer mind set :-) Any input appreciated!   Cheers, <PERSON>
ec11d111-8949-7114-3480-b7c29c4d2d3e
['73719722-206e-cf53-ff16-8b992cac34ed']
Hi all, I wanted to share this proposal we arrived at after discussion with some AlterMundi hackers, so we can discuss it during our future IRC session. I've previously shared it with <PERSON>, who had some observations that he can better explain, but agreed on this initial assumption which triggered the proposal: The scenario we see in our networks is that over a certain link quality which is considered acceptable, we want the clients to choose the gw with better bandwidth. So if for example this quality floor is TQ 100, then if a gw has 6Mbit/s advertised b/w and another has 3Mbit/s, the clients that see this gateways with a TQ above 100 will choose the faster one between them. We observed that in the current implementation, advertized gateway throughput is used to modify the final gw selection by publishing unrealistic bandwidth. The proposal tries to fix this, as well as the "dynamic switching" for selection class 1. Looking at the current code involved we also believe it would allow to make the implementation simpler. This would be the proposed options: gw_sel_class [1,2] 1 will consider gw throughput, 2 will only consider TQ. When using selection class 1, clients will switch gateways if one with better throughput becomes available and reachable with a TQ above gw_tq_floor (see below). Defaults to 2. gw_tq_floor Only relevant for gw_sel_class 1. Above this TQ floor, the gw with the best advertised throughput will be chosen.* Defaults to 100(?) gw_tq_threshold
bb41ebb4-4466-b78a-a65a-6ec1661efe8c
['73719722-206e-cf53-ff16-8b992cac34ed']
I'm curious as to what's the performance of current real world batman-adv networks. Throughput, latency, concurrency behavior... We have been conducting some tests on new hardware we plan to use as replacement for our current recommended node setup and published the results in our blog: http://blog.altermundi.net/article/multiple-hop-mesh-performance-with-multi-band-dual/ We will publish a follow-up when these nodes are put to work in the real networks. Do others have data on their network performance? I'd also like to know if anyone has some explicitly designed set of performance tests to share. cheers, NicoEchániz
26c59d79-f214-5860-b9c6-85eef2090355
['737952ed-b0e4-fdc9-7d8b-4c7d38ffb07d']
Hello I'd like to know if there is a way to get the resulting coordinates (i.e. x/y position in cm on the page) of a musical event in the compiled ps/pdf file; what I want to do is to typeset a lilypond source, and LATER post-process the resulting image. The typical case would be for example highlighting a specific bar. Note that I do not want to do that directly in lilypond, but as a completely separate step. I tried looking at the documentation, but could not find anything. Any hint? Thanks in advance <PERSON>
f821e8ac-f12c-ca88-d438-dd6d7533248e
['737952ed-b0e4-fdc9-7d8b-4c7d38ffb07d']
Hello I'd like to know if there is a way to get the resulting coordinates (i.e. x/y position in cm on the page) of a musical event in the compiled ps/pdf file; what I want to do is to typeset a lilypond source, and LATER post-process the resulting image. The typical case would be for example highlighting a specific bar. Note that I do not want to do that directly in lilypond, but as a completely separate step. I tried looking at the documentation, but could not find anything. Any hint? Thanks in advance <PERSON>
b3e0caad-1568-2b93-52c4-7108f2b4ccf3
['73861ba3-95b1-dd5a-6e1f-6745af5cd13d']
Check on the web interface to the Patton under: Telephony / Call-Router / ISDN Interface IF_PRI_1, then look at the lower section and see if the Accept/Emit button is checked in the Caller Name section. If it is, uncheck it and try the call again. We had this issue - unfortunately, in our case, either caller-id with name will work, or we can use directed call pickup - in our scenario, they were mutually exclusive with the Patton. We opted for number only... Nathaniel Watkins IT Director Garrett County Government 203 South 4th Street, Room 211 Oakland, MD 21550 Telephone: 301-334-5001 Fax: 301-334-5021 E-mail: <EMAIL_ADDRESS><mailto:<EMAIL_ADDRESS><PERSON> IT Director Garrett County Government 203 South 4th Street, Room 211 Oakland, MD 21550 Telephone: <PHONE_NUMBER> Fax: <PHONE_NUMBER> E-mail: h+eJ282PLSu8ObZd@example.com<mailto:h+eJ282PLSu8ObZd@example.com>
079ef70e-8972-d36a-d639-8e61cbe450c4
['73861ba3-95b1-dd5a-6e1f-6745af5cd13d']
Hoping this is an easy question - I am connecting an older PBX with Sipx via a Patton PRI gateway. I'm pretty sure the call from the older PBX will come with a '6' in front of the 4 digit extension that needs to go to sipXecs (or whatver it's called...) What is the best way to get rid of the '6' (does this happen at the Patton level or SipX). <PERSON>
96865cdb-6e96-f9dc-18ad-988835b9d240
['7387c44b-1337-ccf9-b434-1ad77abff6ca']
Hi, I have a feature file with 3 scenarios. The first one is a setup scenario and if it fails I don't want to run the other two. I can set a variable like setupPassed = false if the first scenario fails. I can also check this variable in a Before hook associated with the remaining 2 scenarios. My question is, how could I force these 2 scenarios to fail immediately (skipping all the steps) if setupPassed is seen to be false. Oh, and I'm still on cuke4duke. Thanks. <PERSON>
573665b4-97d9-df38-69a1-58555d878c1b
['7387c44b-1337-ccf9-b434-1ad77abff6ca']
Hi, I'm actually running cuke4duke. Suppose my feature file hierarchy is rooted at C:\cuke4duke\features and that I have 2 subdirectories here A and B which contain A.feature and B.feature and also A.txt and B.txt. I want to have a step in A.feature that reads something like: Then extract information from A.txt The implementation behind this step would then look for A.txt in the same directory as A.feature. How do I find this directory? Any suggestions would be greatly appreciated. Thanks, <PERSON>
48eb0cac-f60b-46f5-80d8-0705209be7f6
['739a9e87-fac7-c259-e689-4276f3ed9143']
I have files I move around between several different environments, and in each environment I have a config file (web.config) that is unique to each environment. I make changes on my machine, push them to a repository, move to a development machine and "pull" them from the repository, then move them to to a test machine and "pull" them from the repository, etc. Through some trial and error I figured out that "most of the time" if I put my web.config into my .gitignore file everything works fine, but not always. Right now I'm spending my afternoon trying to figure out why whenever I checkout the "master" branch it deletes my web.config file. It happens like this. I will have my web.config file in my directory and then checkout a branch into that directory (e.g. git checkout new-search) and everything is fine. If I switch to another branch (e.g. git checkout fix-search) everything is fine as well. But when I check out the master branch (e.g. git checkout master) it deletes my web.config file. Why is it deleting my config file and more importantly, how can I get it to stop doing that so that I can spend my time working on code and not fussing with git? Thanks!
88b5880e-f56c-6e41-6af9-ecaee05cb764
['739a9e87-fac7-c259-e689-4276f3ed9143']
I swear I'm going to figure out how to work with git at some point and stop bothering everyone with these stupid beginner problems! So here's my current problem. I'm trying to "checkout" (not sure if that's the right term or not) files from my repository into an existing folder (a folder created as a NetBeans project...something NetBeans recognizes as a project folder). Clone doesn't work...tells me there is existing content. When I deleted all the content and cloned it again, it put all the contents into a sub-folder. So I tried using fetch. When I fetched, it took a very long time and looked like it was doing work (it was showing me some kind of progress), but when it got all done the folder was empty. I tried this a couple of times and then tried pull, which did the same thing. At one point after a fetch I typed "git status" and got a long message saying all my files were deleted. Finally after searching around for clues decided to clone into a separate directory then copy and paste all the contents of that directory back over to my project directory. I guess my question is, how do I get my files out of the repository and into an existing folder on my computer? And why doesn't fetch or pull actually fetch or pull any files down from the repository? Thanks,
0039e97a-d5ca-08fe-dcee-e3736a954f9b
['73a393f9-639b-ee2c-3742-03757588400d']
Hi <PERSON>, we need to troubleshoot a problem around excessive dynamic sampling and library cache locks at my client. Probably we are facing a bug 14542720. For the moment I'm looking for a way of how to match the sql_id's that have used dynamic sampling during parse to the sql_ids of the dynamic sampling itself (those queries with the SELECT /* OPT_DYN_SAMP */ ...) any idea? <PERSON>
87ee6caf-634e-003a-3409-b6557b6a7a91
['73a393f9-639b-ee2c-3742-03757588400d']
Hi there, i wonder if anybody of you already made some experiences with DBMoto from Hit SW http://www.hitsw.com/products_services/dbmoto.html It promises data replication between a lot of database systems including oracle, db2, sql server. And it should be based on grabbing the logs not on triggers? Can anybody confirm this? Experiences? Actually Goldengate was my first idea for an upcoming project but it does not support DB2 on i / AS400 ... Thank you for any of your comments and thoughts.. <PERSON>
61990ea4-10ab-bc41-21e4-adc953cdc1a7
['73baab60-8818-0f8f-14b5-58a6ccaa6694']
The insert-at-cursor signal in GtkTextView doesn't seem to be emitted. There are a few references to a similar problem with GtkEntry, for which using a signal for the GtkEditable ancestor worked around, but I cannot see any suggestions for GtkTextView. Anybody know what is going on? The key-press-event is working fine, but that leaves you with a lot more work since you have to decide whether a character would be inserted... <PERSON>
2acb4854-683c-4451-cbd3-ee5d1a674e1d
['73baab60-8818-0f8f-14b5-58a6ccaa6694']
I am displaying a structure that seems to be circular, and it shows (x_LIST . #-3#) x_LIST is my own symbol but what is the meaning of the other member of the pair? And how does one track it down this notation in the guile manual? I am sure I came across this before, but cannot locate it in the manual (again?) <PERSON>
08bf5834-943c-63cc-3bc1-2b8f1674d704
['73c34315-a0aa-c795-3d41-376beadbf6b6']
Hi, I'm trying to use OpenSSL RSA routines into my custom GINA for authentication with USB token in Microsoft Windows XP. I've used the RSA_new() method in a normal application and it works very well (Encrypt a random challenge with OpenSSL RSA public key and then Decrypt with PKCS11 hardware method to retrieve the response). But when I use the code in my GINA, the system crashs. I've discovered that the problem is the RSA_new() function. Probably, there is a problem in memory allocation... Is it possible to allocate the memory for a RSA key in other ways?? Thanks
0554026d-ec42-35b5-65c7-61a00884f681
['73c34315-a0aa-c795-3d41-376beadbf6b6']
Hi, I\'d like to export a public key (generated on-board with a USB crypto-token) to verify a sign with OpenSSL functions. (I don\'t use certificate) The data are signed with USB token private key and I must verify the sign on the computer (it\'s a challenge-response authentication). I have the Modulus & Public Exponent on a file (user database on the host) and they\'re in HEX base (byte..no PEM or DER format). Here an example of public key modulus on the token: äkò(jÖ o0¡áÛÑ&#8596;Ü[ÉGþ&#9574;;&#9600;$&#9608;&#9788;7á|Ú¯¿/¦9¾Û¿úÍúÕ ¿ÁÌ&#9658;&#9571;&#9500;&#9474;&#9532;Im*©:ÿcÍý¶wW Can I export the public key from the token in DER/PEM format and use it with OpenSSL code Verify function? Must I use the OpenSC pkcs11 engine to export the key? It seems impossible to use a public key generated with an hardware token... Thanks
2798942b-15b6-0b27-9759-48d113a1fa60
['73d7abb9-26af-b150-75e2-b980c55b4a29']
Hi, I successfully managed to work with ceph jewel. Want to try luminous. I also set experimental bluestore while creating osds. Problem is, I have 20x3TB hdd in two nodes and i would expect 55TB usable (as on jewel) on luminous but i see 200GB. Ceph thinks I have only 200GB space available in total. I see all osds are up and in. 20 osd up; 20 osd in. 0 down. Ceph -s shows HEALTH_OK. I have only one monitor and one mds. (1/1/1) and it is up:active. ceph osd tree gave me all OSDs in nodes are up and results are 1.0000... I checked via df -h but all disks ahows 2.7TB. Basically something is wrong. Same settings and followed schema on jewel is successful except luminous. What might it be? What do you need to know to solve this problem? Why ceph thinks I have 200GB space only? Thanks,
ffea3930-2d20-3eb0-81af-48d91e69f26e
['73d7abb9-26af-b150-75e2-b980c55b4a29']
I already tried 512MB but re-try again now and results are the same. Both without tuning; Stripe 2 replica 2: dd performs 250~ mb/s but shard gives 77mb. I attached two logs (shard and stripe logs) Note: I also noticed that you said “order”. Do you mean when we create via volume set we have to make an order for bricks? I thought gluster handles (and do the math) itself. Gencer
480c18cb-c6b9-9d30-1a95-9b39262c2b8e
['73d7dc3d-6820-9628-f651-35764675d65f']
Since I installed Observium a few weeks ago I have noticed that the overview timestamp for syslog and event log delta is wrong for most (not all) devices, this is prevalent on the main overview and the specific device overview page. By wrong I mean "-1y -1d -xh." etc.. Hovering over the incorrect entry pops up the correct timestamp so I think it's just the logic in calculating the delta between current time and that event that's broken, also viewing the syslogs and entry logs directly with the original timestamp is fine as well. I checked my local timezone and php timezone just to be sure and they are reporting correctly, I would imagine this would break the original timestamp anyway which is not the problem here. Hover over example: ​ I find it particularly weird as not every device is doing this, the ones that are have no real pattern either, mixture of N7K,N5K,ASA etc and some even have a mixture of working and broken entries (individual device overview): [image: Inline image 1] Anyone else seen this, any ideas? Cheers.
5fbf0076-4afd-756d-3c29-e8ac1359543b
['73d7dc3d-6820-9628-f651-35764675d65f']
Hi, I have quite a few alerts stuck in the "Other" state, any idea on what causes this? I had this problem a while ago and then it seemed to disappear after creating a few more alerts and several svn updates so i'm not really sure what cleared it but after adding more devices it's back again. I've tried regenerating the alerts a few times but that makes no difference, is there some debug that can be turned on to work out what is going on here? It's across multiple alert checkers and even something as simple as "device status eq 0" for a device down check has multiple alerts stuck in "other" state. Hovering over the alert shows an empty status: [image: Inline image 2] [image: Inline image 1] Cheers
5bd3ca72-9a34-89e3-0ac1-5aa992a404ac
['73e591da-1aee-8d22-a028-0a64728a3148']
Issue #2669 has been updated by <PERSON>. Regarding the ums0 issues: How was ums0 doing with the LATEST-ISO? It seemed to fail with 3.6.2 but I didn't see any mention of it in your report using LATEST-ISO. If it works better with the latest ISO, it could well be that it is related to the mouse not working with our old USB stack but working better with our new USB stack (which is default for LATEST but not for RELEASE). Regarding the keyboard issues: If it is atkbd(4) (i.e., PS/2), then it could be related to the removal the KB_CONF_FAIL_IF_NO_KBD flag from our kernel configs in master (it's still present in RELEASE). It was originally done to fix issues seen on Lenovo laptops but it might well be happening on other machines. Regarding booting off mfi(4): You said you could install to it, so that's a good thing. :) The message you quoted seems to be from your box itself and not from DragonFly. Are you sure there is nothing else to care about for booting from it? I remember some RAID controllers have some option in their BIOS to enable booting. I'm not sure what's the cause here. <PERSON>
ae7802e0-fb1d-4aec-ebd1-5a2f129a6549
['73e591da-1aee-8d22-a028-0a64728a3148']
Issue #2828 has been updated by <PERSON>. Hmm. Did you measure the actual power consumption? The default (2) is supposed to be a hybrid approach that falls into ACPI. But looking at it, it is quite possible that the code in kern/lwkt_thread.c that is responsible for resetting gd_idle_repeat to 0 is breaking the concept. -Matt
bd73ff8b-6b82-67c9-3819-67e80c83c4cc
['7400d89a-66be-3cdd-bd84-0928efdae52c']
""hey everyone, i need some big time help, and fast! one of my songs has some how become corrupt, and i can't figure out what to do to get it back. when i ever i try and open it up and causes logic to freeze up and die. all my other files seem to be working correct. i even went so far as to reinstall my entire system, and that didn't work! (i didn't do the whole system just because of that, it's been acting very unstable lately, this was the final excuse). so please please please help me. thanks in advance. -mark""
198f5511-df5b-8bd3-e0a8-22c1b91eb118
['7400d89a-66be-3cdd-bd84-0928efdae52c']
I know this is a bit of a dumb question...but how do I buy an unlock key for the logic instruments? My demos expired and I want the exs24 and es2, but emagic doesn't really say how to buy them. I've ssen them at audiomidi.com, but they look like they are vst versions. Anyone know the best, fastest way? I'd prefer to get an unlock code, with the physical materials shipping later. Thanks, <PERSON>
b3c55654-3745-e8a7-01a0-ae9def63d899
['7402cc2a-d41b-1286-06fa-8588c3df9b1e']
Hi I'm trying to set up courier on a FreeBSD based virtual dedicated server (from johncompanies). Installation and configuration has gone well but I can't seem to get esmtpd to run. All the other services (including authmysql, pop3d, pop3d-ssl, imapd and imapd-ssl) are running perfectly on the machine. The tests that I have run on the system worked perfectly. Mail is delivered locally and to virtual users from mysql without a hitch. From looking at my logs I see no reference to esmptd being started (from either my startup script or from the esmptd script itself). When run (certainly with "emstpd start") then nothing appears to run and nothing is returned. I have tried esmtpd-ssl to check if the same thing happens. Something slightly different happens here with "ll_daemon_resetio: Permission denied" being returned when run. I wonder if the fact that the virtual dedicated server has no loopback address has any relation to this. Something else I have noticed, if I use webadmin to configure any of the esmtpd conf then the esmtpd config are made into 0 byte files (all contents deleted). I have since stopped using webadmin ;) Any ideas on how to get esmtpd working? Any help would be greatly appreciated (this one of those. had to be done yesterday kind of projects) <PERSON>
23d3f0a0-03d4-3eda-fda8-6844d31fedcf
['7402cc2a-d41b-1286-06fa-8588c3df9b1e']
Hi I'm trying to set up courier on a FreeBSD based virtual dedicated server (from johncompanies). Installation and configuration has gone well but I can't seem to get esmtpd to run. All the other services (including authmysql, pop3d, pop3d-ssl, imapd and imapd-ssl) are running perfectly on the machine. The tests that I have run on the system worked perfectly. Mail is delivered locally and to virtual users from mysql without a hitch. From looking at my logs I see no reference to esmptd being started (from either my startup script or from the esmptd script itself). When run (certainly with "emstpd start") then nothing appears to run and nothing is returned. I have tried esmtpd-ssl to check if the same thing happens. Something slightly different happens here with "ll_daemon_resetio: Permission denied" being returned when run. I wonder if the fact that the virtual dedicated server has no loopback address has any relation to this. Something else I have noticed, if I use webadmin to configure any of the esmtpd conf then the esmtpd config are made into 0 byte files (all contents delted). I have since stopped using webadmin ;) Any ideas on how to get esmtpd working? Any help would be greatly appreciated (this one of those. had to be done yesterday kind of projects) <PERSON>
477b39f6-d685-f09b-9772-d124912ebff0
['7415b05f-6aa5-4f30-c427-479890841b07']
I had gotten the DDR core to work in simulation, and it synthesized, but did not work in hardware. Has anyone been able to get it working on a real board? I'm using an Avnet Virtex-II evaluation board, with an XCV2P7ff896, speed grade 5. It has two Micron mt46v16m16 chips. I haven't been able to figure out what the problem was (and have since moved on to trying to get some of the controllers from Xilinx working). Does anyone have any suggestions on what it might take to get the core from OpenCores working? Thanks, <PERSON>
574ef423-9461-dd30-384f-7c1cc3ad2eb8
['7415b05f-6aa5-4f30-c427-479890841b07']
Hi, I have a Radio Shack USB to serial adapter, but I can't get the driver to show up for the device in /proc/bus/usb/devices, and I saw a warning message at least once in /var/log/syslog that said something like the driver dosen't know what it is. I did try using insmod specifing the vendor and product id's, but that didn't help as well. Does anyone have any suggestions, or should I return this and buy one that's supported? thanks, <PERSON>
c252a45e-d891-71e6-a2a1-36f1a4cd529a
['7420f44e-fe75-7326-f0ec-4950d161d44b']
Hi, Given a certain data.frame, the lattice xyplot function will plot the data as.is and join the data point in the order of the data frame. It is my (probably flawed) understanding that, using the same data frame, ggplot orders the data by increasing order of the x-axis variable. Can one control this behavior? Thanks <PERSON>
338d223b-a5c9-c68f-eb77-4d88104a3874
['7420f44e-fe75-7326-f0ec-4950d161d44b']
Dear R-users, [Sorry to have to send this again but my former posts were trimmed because of some embedded HTML code] readline is a nice function to get user input when one is working with the terminal. This function does not allow any interaction with the user when a script in batch mode (I guess by design)... Unfortunately, that is exactly what I am trying to do. I would greatly appreciate if any R-user could: - tell me if this is possible - indicate a function I could look at, that would allow this type of interaction during a batch run. Thank you in advance,
d8b956a6-1957-7f4f-9ac5-b52ae284e6ef
['7430a7db-e9e3-dfd8-7081-facefab1632e']
Hello, Is there a way to build omniORB entirely in its own namespace? For example, references to CORBA::string_free would have to be qualified as omniOrb::CORBA::string_free, etc. The problem we are dealing with is that a single process happens to use two CORBA implementations. We are providing a shard library that uses omniORB, but the process we are linked with at run-time uses Visibroker. If we could isolate ourselves from the global CORBA namespace, we would be able to make it work, I think. Thanks for any suggestions, <PERSON><IP_ADDRESS>string_free would have to be qualified as omniOrb<IP_ADDRESS>CORBA<IP_ADDRESS>string_free, etc. The problem we are dealing with is that a single process happens to use two CORBA implementations. We are providing a shard library that uses omniORB, but the process we are linked with at run-time uses Visibroker. If we could isolate ourselves from the global CORBA namespace, we would be able to make it work, I think. Thanks for any suggestions, Slawomir
47978e97-0332-6820-d9aa-a19adab375cb
['7430a7db-e9e3-dfd8-7081-facefab1632e']
Hello, Why does the value of JSESSIONID cookie change per requested portType within a WebService? I used the same Service instance, with the maintainSession property set to true, when obtaining different port stubs via the Locator and still calls on those stubs had a different JSESSIONID sent out. JSESSIONID value was the same when making calls on the same portType only. Any idea? <PERSON>
ff688274-5d30-f0d1-bec8-23b323cc9fa7
['74359334-9bbf-5bfd-3c60-0f1ff7a2bc2c']
Hi, I just want to say thank everyone who contributed to this project for all the hard work that was put into making this powerful system work. I have already figured out and implemented a simple windows XP/2000 PXE-based installation with partially auto-generated unattend.txt, and I look forward to configuring various post-install scripts to install various software and hotfixes. I thought that the declarative structure of config.pl was pretty clever until I learned about how todo.pl works. Then I was really blown away! Another one of the coolest things I like about this Unattended system is how dosemu is used to run Windows Setup inside a Linux OS! Thanks.
5d9560ee-c00d-5a73-bccc-912e5dff0447
['74359334-9bbf-5bfd-3c60-0f1ff7a2bc2c']
I understand that you guys are busy with GSoC as it is, plus you are already volunteering your time to this project, and I would like to ask whoever can to please comment on my Google application for this testing project. I asked a question about the "abstract" text/section and I would like to get feedback on it as soon as possible. That way I can edit it before the deadline after which the abstract becomes read-only. Thanks for your time.
1d02b81b-a0e0-2aa7-7b61-33eb800f7133
['7438cdb3-2b7f-0227-d66d-60ef2dc8405c']
Hi I've installed version 2.2 without problems. Then, I ve changed line Listen 80 to Listen 8080 and restarted the server. Nevertheless, my server is still responding on port 80 !! This is not the first time I install apache on windows or Linux but is the first time on Vista and I did it as administrator. Which is the way to change the configuritaion on vista ? Thanks in advance J
74bbad26-79cd-27d2-911a-66fb65fc6ca2
['7438cdb3-2b7f-0227-d66d-60ef2dc8405c']
Hi After a big fight with proxies I finally got maven working. I've generated a sample project and an eclipse project without problems. When I tried to get an ant build I realized the generated build file ignored my local repository and put a lot of lines trying to download jar and install them in my profile folder. I wonder why maven did it and how could I fix it ? Also, I'm curious about how to customize goals. I've seen pom files from some software I've downloaded to use it as examples but all of them use jellu scripts and are made with maven 1. I've been looking maven 2 docs but I couldn´t found a I guide step by step to customize goals or use ant tasks as they did in maven 1. Could somebody give me a clue in this matter ? Thanks in advance J
ab3ef331-a7ca-56ef-4b21-f4f26caa5e58
['74511451-6989-c377-a926-df9fecba8339']
Thanks for all the advice, I learned a lot today about making presumptions when writing my scripts. He was running my script from the Windows desktop, which was why the window was closing after he finished entering the search query. Once he started the script from the DOS prompt, it worked great. Thanks for pointing that out to me. I won't make that rookie mistake again. :) <PERSON>
c096ef6f-8cb9-3007-fe2e-43ff41a55221
['74511451-6989-c377-a926-df9fecba8339']
I've written a python script for a fantasy hockey league I'm in, and I'm using Python 2.1 on Mandrake Linux 8.2. I have a friend who downloaded Python 2.2 on his Windows 2000 and when he runs the script, it 'crashes' after he enters a search criteria and hits enter. The whole screen disappears. I'm hoping if I include the section of code I have questions about that someone could offer advice. The following script is stored in the same directory as two text files, "fwfha_skaters.txt" and "fwfha_goalies.txt" and the script works well on my computer. If someone could tell my why it isn't working on Windows I'd be grateful. <PERSON>
dda5dc8c-6eca-9808-ce8d-98b26d646749
['74546d01-a587-ccbe-c2e5-817e38529dbc']
I have installed octave on an old Linux system that I cannot upgrade. I tried installing 3.03 but the gnuplot version was not recent enough so I went to 2.1.73. This version compiles but I need to specify the --disable-readline flag in ./configure. Once at the octave command line, a minor inconvenience is that Up-Arrow does not recall the most recent commands. Instead it displays [[^A. I was wondering if this is a readline issue? Is there a way to modify the key bindings in an Octave setup file? Thanks for your help!
77d82e7d-82b6-42cf-ff20-2bc70b504279
['74546d01-a587-ccbe-c2e5-817e38529dbc']
Hi all, I would like to carry octave around with me on a USB stick. I've compiled octave 3.05 (on Suse 11.4) and it runs fine on that machine. The problem I am having is how to copy it over to a USB stick. I copied the entire build directory and launch with ./run-octave. I type ldd octave to see the dependencies and copy all the .so files into a directory on the USB stick "lib" inside the build directory then add this to LD_LIBRARY_PATH. However when I try to ./run-octave from the USB stick on a new machine, it doesn't work. It looks like maybe octave is looking for the .so files using absolute paths for the original machine rather than in my "lib" directory. I apologize if this message is unclear/naive - I know I'm hacking without a clue and really need some "big picture" help how to accomplish what I want. Thanks. <PERSON>
ff194594-e839-f85a-f0ba-62ee2d9c1086
['7461e39c-0f8a-a5d6-9bbc-5b851d8fd12c']
Hey all, I'd like to phase out cwiid [1] (as it's super old and unmaintained). Currently it's an optdepends for supercollider, but its use will be deprecated in the next version [2], so I will remove it there eventually. The only other packages depending on it are kodi (make) and kodi-eventclients. However, is it really required there? For now, I rebuilt cwiid, but I guess it could be dropped midterm, if it's not a hard requirement for kodi et al. Best, <PERSON>
12d91acc-8005-4e4b-1398-dd1c10616b11
['7461e39c-0f8a-a5d6-9bbc-5b851d8fd12c']
Hey all, I was wondering, what the current upstream status is. I'd like to package the latest tuna for Arch Linux, but I'm a little confused about the current state of things: While in the past I used [1], it seems that <PERSON> repository [2] is currently holding the latest updates. Will the latter eventually be merged into the former, or is it time for me to switch? Additionally: What can I use as the current upstream website for tuna? Currently I have: https://rt.wiki.kernel.org/index.php/Tuna Best, <PERSON>
bffb942a-7591-b047-becf-c2f67e3ff77a
['7465d810-50bc-992a-752b-8170e6602650']
Hi, We are new to packetfence and trying to setup packetfence with Meraki access points. Stuck at the point where to add switch on Packetfence config to add the new AP. How do I go about this setup, what is the IP of Meraki cloud controller and do I need more setting to get this working. Anybody tried Meraki with packetfence? Kind Regards,
d30d6f63-5f5c-eb8d-281a-020734eeb6cd
['7465d810-50bc-992a-752b-8170e6602650']
Hello, I am setting up PF Zen 5.5 in inline mode. I have configured two interfaces one management and another inline. Everything is working OK except that I do not get a registration portal on the inline network. The default gateway is the IP for PF inline interface, I can ping it but even if I input the IP on a browser I don't get a Portal/Registration page. Users on the LAN are not prompted to register when they get an DHcp lease from PF. Any ideas will be appreciated. Thanks and Kind Regards, <PERSON> . S
17d1c316-6efd-b6b2-e644-45cf3eb6047b
['74799ebb-7031-8603-322a-d0eb87b30a6b']
I was trying to run the Wizard tutorial and the instructions produce code that doesn't compile. It appears to be due to differences between version 7.1 which was used in the tutorial and 7.3 which I have. Can I install version 7.1 and still have version 7.3? Is version 7.1 available? Thanks
38282cfc-8313-adb7-c9a9-1221456dcfb5
['74799ebb-7031-8603-322a-d0eb87b30a6b']
I frequently see reference to layer.xml files when I lookup tips on how to organize my netbeans platform application. The latest version of Netbeans doesn't have that. Is there a way to accomplish what one used to do with the layer file? Specifically I want to control the startup window sizes of TopComponents. Thanks
284bad9d-247c-4a5f-212d-19678230e00d
['74879d9c-3405-714d-3291-77c861dbbdfb']
It is true that Jews and Christians share only the Hebrew Bible (Old Testament). However, I would suggest that if you look through the BS class, you will find texts of the Bible edited and published by both Jewish and Christians, as well as Jewish and Christian commentaries on these books. I think it would be unfortunate for readers of any persuasion not to have editions and commentaries on the same literature in the same place. It should remind us all of our common heritage. Just to give you an example: the Biblia Hebraica Stuttgartensia is the standard text of the Hebrew Bible (slowly being replaced by the Biblia Hebraica Quinta, a joint project of the German Bible Society and the Hebrew University of Jerusalem). The Stuttgartensia was based on the work of <PERSON>, a German Protestant, who based his work primarily on the St. Petersburg Codex (or Leningrad Codex, if you must). The St. Petersburg and Aleppo codices were both copied by Karaite Jews, and not accepted by orthodox Jews as authoritative until <PERSON> declared them to be the most accurate manuscripts of the Bible. Michael S. Borries
efd3d2cc-9343-322b-9abd-94f8566a65db
['74879d9c-3405-714d-3291-77c861dbbdfb']
For those with the title Shui hu zhuan (translated into English under the titles Outlaws of the marsh, All men are brothers, <PERSON>, and Marshes of Mount Liang), LC has added the two authors to whom this work is sometimes ascribed (<PERSON>, <PERSON> and <PERSON>). The correction should appear in OCLC in the next few days. You may want to check you authority records and your bib records. <PERSON>
df58ac5c-1a78-e55c-7da7-8faad59a7a6e
['7493753d-67dc-9888-38be-919a5bb2d943']
Hi, Have 2.1.5 with 2.6.9 (Debian kernel-source) and hibernate package 0.99-1 from cp.yi.org Hibernate doesn't run : Your kernel does not have any recent Software Suspend 2 support compiled in. Please follow the HOWTO linked from http://swsusp.sourceforge.net/ for instructions on how compile Software Suspend into your kernel. hibernate: Aborting. Have all as module, as in the wiki here: http://swsusp.sourceforge.net/wiki/DebianInitrd What can I do? Thanks <PERSON>
27f6b299-354b-f09c-ba12-101da8dab597
['7493753d-67dc-9888-38be-919a5bb2d943']
Hello everybody, Hi run Tomcat 4.1.18 Standalone on Linux debian(sid)and use J2sdk1.4.1_01. I have a servlet which manipulate exif informations in jpg images. it works well but when I try to run it on tomcat it fails (error 500) I have the following error: "Can't connect to X11 window server using '0:0' as the value of the display" What can I do? <PERSON>
a9b6efb7-9e12-628f-4b81-10301695675e
['7495048c-aa95-0911-35c7-06b35d673593']
This patch adds the possibility to enter the prompt at boot time by typing the sequence of buttons "123" rather than a single button: http://www.freebsd.org/~attilio/Sandvine/STABLE_8/boot123/boot123.diff This is useful in the cases where a serial console is likely going to be used which can carry on spourious character, leding to the prompt erroneously. This option is wrappered into the BOOT_PROMPT_123 option, in order to maintain the current POLA. This patch has been contributed back by Sandvine Incorporated. Please review. Thanks, <PERSON>
5c2f802f-abd5-ad93-110c-2e837e29ec87
['7495048c-aa95-0911-35c7-06b35d673593']
Hi all, first of all, I have to say I'm not too much familiar with DragonFlyBSD kernel (so I have no idea if this is completely new for the project), but I contributed some for FreeBSD kernel and I'm rather experienced with IA32. What I would like to work on are some x86 'improvements' to DFLY kernel. For improvements I mean adding different versions of some critical functions for different versions of the CPU (i.e.: in P4 you could just implement an atomic memory barrier using mov + *fence instructions against using xchg or stubs like these). First of all, a good discussion point is about optimizations activation: do you prefer compile-times stubs or run-time patching? Actually, I'm planning to improve and add a run-time patching concept inherited from Linux to FreeBSD, and maybe you would be interested in its port. Run-time patching is very useful, but on the other side it faces a lot of problems (inlined functions can't fit and it deals with dimensions problems), if we don't want to loose performance matters. Compile time stubs are simpler and possibly quicker but the kernel needs re-compilation in order to get benefits, that is not so nice. So, some feedbacks about this is very appreciated. Once that we have choicen a method in order to apply changes, the first thing I would like to add (BTW, I don't know if it exists alredy) is sysenter/sysexit support replacing interrupt 0x80 (I have an item in the FreeBSD list for volounteers about it, since I think I would like to add it there too) and possibly evaluating the usage of FPU/MMX/XMM registers in order to gather syscalls parameters. Feedbacks about it are very appreciated too. I have other ongoing projects we can discuss later, but they are very architecture dependant so, since maybe they can fit in FreeBSD, they cannot in DragonflyBSD. Thanks for your time, <PERSON>
395667df-0e5a-adba-fd14-d13096909bae
['749c496b-fdd7-be6e-63d5-17edc279c84d']
A week ago, my car refused to start in a parking lot after a routine drive to the store. Now it won't start or charge. I am stumped and could use some advice where to start looking. The drivetrain is an Azure Dynamics DMOC445 with the AC50 motor. It has run perfectly for two years. The wiring and relays are from Electro Automotive and are wired according to the book. Nothing obvious has disconnected. There is pack voltage at the controller (267v - should be 288v, but 267 should be enough). My cells all seem to be OK (I can't reach every one, but the individual banks sum up to the expected amounts). The key switch relay clicks when the key is turned, but the DMOC contactor does not. I tried to connect my laptop to the DMOC, but nothing was received - it seems to not be turning on at all. When I run the charger, it cycles on and off (the fans spin, but I hear a relay clicking every 2-5 seconds and no charge is being applied to the pack). It never did this before. I am suspecting a 12v issue, possibly one of the charging relays, but not sure how to figure it out or where to start. I am learning why non-engineers should not do conversions L Has anyone else seen an issue like this, or can give me some ideas of how to start troubleshooting it? If anyone in the Portland area is willing to help in person, there would be beer and/or cash involved. Thanks! <PERSON>
6fd81dde-2829-19ae-0495-3e8d78f39f4f
['749c496b-fdd7-be6e-63d5-17edc279c84d']
I used a heater from CANEV.com in my car and it was very simple to install. This heater already has an over-temp switch built-in, and comes with a contactor and fan switch. A very simplified wiring diagram can be seen here: http://www.evsource.com/datasheets/heaters/heater_wiring_diagram_v2.pdf CANEV has a more detailed diagram, but you'll need to email them to request it. The fan switch is connected to the 12V+ side of the contactor with a one-way diode (29 cents at Radio Shack) on each of the fan switch posts for Hi-Med-Low. So you have three 12V+ wires from your switch, each passing though a diode, then all three connected to the contactor 12V+. 12V- on switch and contactor go to chassis ground. Otherwise it is like the drawing. The heater comes on whenever the fan is switched on. And you can't have heat without the fan, so fire risk is almost nil. If you don't have an over-temp switch in your heater, you can install one between the three combined leads from the fan switch diodes and the 12V+ on the contactor. Over-temp then shuts off the contactor. Another one-way diode is needed between the fan switch leads and the over-temp switch. Good luck! <PERSON> Electric '73 VW Fastback Portland, OR
c80fb4dd-6a84-6158-284b-7f81e50e66a8
['749c8193-94ed-0a40-3c33-b177aeea72c2']
Hi, I tried posting before but something must've gone wrong as I can't find a trace of it on the list. Recently I started using Leo on a documentation project so I'm mainly using Leo to edit and get rst output. A bit of context. I am using rst mostly because it's easy to write and it provides structured output in the form of XML. I convert this XML to DITA which is a topic-based documentation format. I chose Leo for it's outlining capabilities and cloned nodes. The outline structure in .leo files can be converted to DITA maps (kinda like table of contents) and the node content to DITA topics which are pulled into DITA maps to form a manual. What I am looking for is a way to get the unique ids that Leo gives each node and use them to link topics to maps. This fits in very well with topic re-use as cloned nodes all have the same id. I looked a bit at the rst3 plugin code but I wondered if there is a way to get the node ids into the rst output without patching this plugin code. I am curious if there is an easy way to do this. Thanks <PERSON>
3dfa12a8-92e5-470e-3e6d-80b8bdbe4584
['749c8193-94ed-0a40-3c33-b177aeea72c2']
Hi, Recently I started using Leo for a documentation project at work. I mainly use it to generate rst files (using @rst nodes and rst3). These files are converted to XML and then to DITA which is a topic based documentation format. DITA based documentation consists of small re-usable topics which are linked together via so-called maps (similar to a TOC). Re-use of topics is perfectly captured via cloned nodes. Great. But I somehow need to be able to give each topic a unique id. The id's that Leo generates inside the .leo files are just fine for this purpose. However I need to get these id's (I mean the v element's t attribute) into the generated rst file (I think this is like sentinels?). This way I can generate the DITA maps from the .leo XML file and reference the individual topics based on the same id. I can then write out a big collection of topics as an rst document and use other @rst nodes to capture the hierarchy of topics containing only cloned nodes. It doesn't matter much if these id's are written out as rst fields or comments under the heading as long as it's in the generated file. Is there a way to achieve this in <PERSON> without coding or patching the rst3 plugin? Thanks a lot. Cheers, <PERSON>
0c2a24ad-619d-66bf-bdcc-c7051897ef76
['74a46c09-a3dc-db75-8ba7-a3f4574fbf11']
Hi everyone, As part of the Groovy planning process, I've posted a review for multi-echo laser rangefinders: http://www.ros.org/wiki/sensor_msgs/Reviews/2012-08-01_MultiEchoLaser_API_review_API_Review This review focuses on the ROS API for scanners that return multiple ranges (and intensities) for a single beam. Example lasers with this feature are the SICK LMS151 and LMS511, as well as the Hokuyo UTM-30LX-EW. Please take a look and contribute your thoughts. Thanks, - <PERSON>
a1892a63-0a43-2dcc-9e8a-599a5f4bd878
['74a46c09-a3dc-db75-8ba7-a3f4574fbf11']
Hi ROS Community, I've been working on a driver that connects the sensors in Android devices to the ROS environment. At this time, it only publishes sensor_msgs/NavSatFix messages, but I will soon introduce sensor_msgs/Imu and sensor_msgs/Image to publish data from accelerometers, gyroscopes, magnetometers, and front/rear cameras. To get more information and to install, please see the following: http://www.ros.org/wiki/android_sensors_driver/ https://market.android.com/details?id=org.ros.android.sensors_driver To file bugs, request features, view source, or contribute UI, translation, or other improvements, please see the Google Code project: http://code.google.com/p/android-sensors-driver/ I hope everyone finds this useful and I look forward to hearing your feedback and seeing cool uses for Android devices in robotics. Thanks, - <PERSON>
a5bfe4e4-776e-368e-bb66-eeeef48825ac
['74b20919-1950-f5ae-2902-b2b3abdc126b']
Hello - I've noticed that converting a PICT to TIFF (grayscale 16-bit with alpha channel) in IM 5.4.7 - RH 9.0 is resulting in a change in the computed signature obtained with "identify -verbose". The command I am using is "convert foo.pict foo.tiff" - no options are involved. The reason I am concerned is that this apparently means the underlying image data is being modified, and that the exact same command in a previous version did *not* result in a signature change. Is there a way to prevent this? BTW - if I remove the alpha channel first - "convert foo.pict +matte foo_nomatte.pict" followed by "convert foo_nomatte.pict foo_nomatte.tif" - the signature does not change between the two "nomatte" images. Anyone know how to force preservation of the data or prevent mucking around when the alpha channel is present? Pls forgive me if this topic has been discussed - I didn't find it in the archives. Best regards - Ken
129d3689-0734-0cae-e557-38be57f0cabd
['74b20919-1950-f5ae-2902-b2b3abdc126b']
Hello - I upgraded my version of IM to the latest downloadable on the imagemagick.org website - however converting PICTS to TIFFs with an alpha or opacity channel is still failing. When viewed in display, only one half of the original image remains in the TIFF - the other half is black. Again the problem doesn't occur if I remove the alpha channel first. The conversion goes fine and the image signatures are preserved - which isn't the case when I perform the conversion described above. It appears that the presence of an opacity or alpha channel is corrupting the conversion of PICTs to TIFFs in IM. This did not occur in the pre-IM 5.4.7 version. Has anyone else experienced this problem? Is it a bug and is there a workaround? Pls forgive me if this topic has been discussed - I didn't find it in the archives. Best regards - <PERSON>
7f6ae547-25a8-d314-bd54-5b0a58e927fc
['74b6b66f-e005-71d9-dcd9-6775d46aabe5']
Hi all, (please feel free to copy this mail to your local list) Version 406 (June 2004) of the MozOO.org CD has been released. The MozOO.org CD contains a simple set of productivity tools for Windows. The CD ISO image is small enough (110MB) to fit on a MiniCD. The main components Mozilla and OpenOffice.org, with language files and other stuff for both, and a nice installer that takes care of all necessary things in the proper order. It's very simple, targeted. The primary target audience is home and small office users. Non-geeks. They won't move to Linux yet, but this offers an easy and appropriate first step with the tools everybody uses. It allows people to not worry about upgrading their office suite, having a good browser, a bayesian spamfilter built into their email program, etc... Plus it's an excellent answer to the "open source is losing the desktop battle" nonsense that's appeared in the press ;-) For more info as well as the ISO image, see http://mozoo.org/. Yes it would be excellent to see this CD with a little booklet available at news agents around the country and beyond! Feel free to talk to me about this: <EMAIL_ADDRESS> (you can also ask me for my phone# there, if you don't already have it - if you do, feel free to call to discuss ideas.) Ahyes, the design allows for easy localisation, so that's another thing we'll be looking into now that the glitches in the earlier release have been ironed out. If you can help with that, or anything else, please also contact me. All assistance and ideas will be much appreciated. Thanks. Regards, <PERSON>.
e7e6c19e-0177-08ff-43ba-cce52f542b3f
['74b6b66f-e005-71d9-dcd9-6775d46aabe5']
Hi, We're restarting a MySQL User Group in Sydney. <PERSON> is sorting out a venue, but if you have offices or another suitable place near the CBD, do let us know. We're aiming for a first meeting Tuesday 7 Feb 7pm, since I'm actually in Sydney next week for a MySQL training course. If you're interested, please sign up at http://mysql.meetup.com/142/ so you will receive further announcements about the venue etc. It would also give us a better idea of scale. Hope to see you next week! Thanks. Regards, <PERSON>.
d21add2f-e214-b13c-4433-14dca63c0d01
['74bc1b43-4fb7-52d9-ecd9-ff31a58a7873']
Hi, I am using NB6.1Beta on Windows XP (Japanese edition) with JDK 1.6.0_5. My project is encoding as UTF-8 (project --> properties --> java source --> encoding ) and comments are wrote in Japanese. After updated the project using CVS, I try to resolve conflicts using Versioning --> CVS --> Resolve conflicts, however the comment in the source files, both local and remote source file, are changed to unreadable characters in *compare editor*. Yes, only in the compare editor. n the Java source editor, they are displayed normally . I guess the compare editor doesn't show the source code with the same encoding type as the project setting. Is it a already known bug? I hope it can be resolved when NB 6.1 releases. Regards <PERSON>
549ff6fe-f243-19d2-a1c4-9cc20c965aac
['74bc1b43-4fb7-52d9-ecd9-ff31a58a7873']
Hello everyone, I made a patch to rdoc -- just 3 lines -- which makes the rdoc(generating html) run faster and use less memory. I listed the test result in the below. (Tested on 32 bits Linux, Ruby1.9.2-trunk, 4GB memory, Intel Quad 2.6GHz) Using rdoc to process whole lib directory: original : 15 minutes, 1.3GB memory patched: 7 minutes, 65MB memory I compared the generated results(patched and unpatched) using "diff -r" command, and confirmed that the patch does not break anything. I attached the patch file, please check it out. However, I only profiled rdoc with html output and found the problem. So I am not sure whether or not other output format code has similar problem. BTW, I found the *bug* using a memory profiler I created for Ruby1.9. Right now, the memory profiler can show how many live instances for each class, size and age(survived from how many times of GC) of each instance, alos can show the allocation call trace of each instance. However, I am still working on it. So, if you have interest on the memory profiler, you can contact me. Best regards, -Tetsu
3d692c95-0189-cf64-f55b-131bbfe187e3
['74d5811d-bdbc-fab3-7c95-fd1b533dd699']
if you are referencing Tapose it's metaphor is really based o. The defunct MS Courier project. I do see some similarities to Newton. I will be checking out the Notes Plus app as it seems to give a more Newtonish approach of HWR (instant and delayed) and drawing smoothing. Those were things I used a lot with my Newtons. Sincerely, <PERSON>
001025ca-a103-d5e1-05a3-604d78f3ad56
['74d5811d-bdbc-fab3-7c95-fd1b533dd699']
I agree with your statement <PERSON>. "Regarding Mossberg's review, I think it's wrong to assess who is and who is not a plausible user of the iPad. This product will define a new way of using mobile devices; way beyond what the iPhone did. " With the apps more and more users will flock to the iPad because of the overwhelming useful of the product. At BestBuy today the demographics of people buying the iPad ranged from 20's to 80's, males and females. They all had different uses and were all happy. We had to activate some just so they could play with them on the drive home. (the passengers I hope). My MacBook Air server us perfectly fine and you cannot beat the weight! Sincerely, G.
52cbce79-bc9c-0b6a-ef78-82e734c66e00
['74e1a09e-946d-176c-f3f8-9f910119d600']
Hello everyone, I am <PERSON>, a third year computer science undergrad with an interest in Information Retrieval and Natural Language Processing. I am interested in doing the "Improvements to Debian search" project with Debian. I've been involved with open source development for some time now and have also implemented a couple of state of the art weighting schemes for Xapian , Given my experience with open source development , Xapian and Information retrieval ,I really think I'll be able to justify the project by implementing state of the art weighting algorithms , query expansion using relevance feedback , highly efficient query expansion algorithms, spell checking ,resistance to keyword spamming and many features in Debian search to drastically improve the search results and the search experience for the Debian community. The search on search.debian.org, mailing lists , Debian wiki and the packages site is the gateway for the huge Debian community to gain information about Debian and hence, improving search will definitely aid both Debian developers and users a lot. I'm already in touch with <PERSON> ,the mentor of this project and am working on my proposal which I'll put up on the wiki in a couple of days. -Regards -Aarsh
32bf3e27-4257-a80a-035f-872389d2b70a
['74e1a09e-946d-176c-f3f8-9f910119d600']
Hello admins. I apologize for not being able to do any work in the past week as I was busy packing for debcamp and debconf and with my visa interview and visa procedure which took place in a separate city as the embassy is located there. However, I am way ahead of my schedule already and the admins need not be worried. :) I shall resume work today itself. :) Can't wait for debcamp ! :) -Regards -Aarsh
392774dc-bf42-1d01-a389-699ee5726bb4
['74e84488-3a66-c025-3b45-30e88fa43cff']
This reverts commit c51aa112f53adc87250177002aa3e008305e0777. There are several serious issues with the stated rationale for and intended use of the "Adult" Additional Category that have yet to be resolved. - The menu specification is designed for worldwide application, and thus must be culturally impartial. Different cultures have different criteria for something that is "adult-only". Ambiguous examples provided on the XDG list include "Bible" and "Art Gallery" applications. - Given a set of criteria for what constitutes "Adult" content, the categorisation is binary, but whether an application is "Adult" is not. Ambiguous examples provided on the XDG list include a "Breastfeeding Tutorial" application; the "PornView" application, which does not in fact include or provide access to any pornography; web browsers, which provide access to a wealth of pornography on the Internet; and text adventure games that include the possibility of violent death.
7fa74e9f-647d-c1f4-ba18-2cd14c3bf31b
['74e84488-3a66-c025-3b45-30e88fa43cff']
pathconf(_PC_NAME_MAX) was implemented on top of statfs(). The 32bit version therefore fails EOVERFLOW if the filesystem blockcount is sufficiently large. Most pathconf() queries use statvfs64(), which avoids this issue. This patch modifies pathconf(_PC_NAME_MAX) to do likewise. <PERSON>
e2119d0c-9992-d0ea-e04f-df803514a3a3
['74f7656a-54c2-0a86-53cc-cb8a83410a94']
Hi, As I just subscribed to this list, I would like to introduce me first. I'm an electronic engineer, writing from Brazil. I'm 27 years old, work for the auto parts industry, and I have been a Spectrum user since 1987 (actually, it was the TK90X, a Brazilian clone of the Spectrum +). Back in 87, my dream was to have a Sam. I know that some of them came to Brazil at that time by the hands of some game pirates. We could not import those machines at that time because there was a law that does not allow us to buy imported computers, so most Amigas, Macs, Commodores and even some Ataris wer bought by "not so legal" ways (this includes my Amiga 1200) I've downloaded SimCoupe for my Mac, and it was really nice to see the dream machine emulated on my monitor. Two things then came up in my mind: First, I don't know how to use the Sam. The basic is different from the Speccy, and I could not find any manual.. Second, I still want a Sam !! So, if anybody has a second Sam to sell, or knows were I can get one, I would appreciate very much. Best Regards, Marcelo
0154d87a-6a28-3432-51e4-6bb8a56d747e
['74f7656a-54c2-0a86-53cc-cb8a83410a94']
Hi Folks, Take a look at this. Just amazing... I'm really considering to purchase one http://virtuals.atlant.ru/peters/e-index.htm Unfortunately, their english is even worst than mine :-) Seems to be they used software translators <PERSON>, <PERSON>
902052b5-71c8-a35b-8b64-205a3835b9b2
['74fb815c-f364-813f-50d2-da2fd6cbc3b9']
Hi, I've tried to set up the biblio.frameworkcode at import time by setting up a link between the 099$e subfield and biblio.frameworkcode in koha. But it seems this fields in not updated in either way. Even if i create a new non-default notice and that I export the notice, the 099$e subfield is not created in the export files (marcxml, ...). Do i miss something? I'm using koha-3.00.06. Best Regards, Frédéric
cafc21b8-081b-e4f5-9afa-ed63fa7dfe3e
['74fb815c-f364-813f-50d2-da2fd6cbc3b9']
Hi, I'm trying to switch our old ILS to Koha but I'm stuck with how I'll import Serials in the system. The information I have in the old database is rather sparse but I have a value I don't know how to map. I have a publication date, which is normally linked to one item notice (and not a serial notice). The problem is there is no mapping in the koha standard (sample?) UNIMARC definition. Moreover, I don't know which koha field to map to? I've found "timestamp" and "dateaccessioned" but do not know which one to choose and if i'm in the right direction. I plan to add a 995$w field to map to the koha field. I have to admit i'm computer scientist only ... and that the person who manage the old ILS is not a librarian, so I'm on my own to handle the issues. Hope you can help! Thanks a lot! Regards, Frédéric
54d572fc-5be6-8e9c-6c38-cfdc86e18cf3
['7509fcb2-51ed-91bf-3a2a-538473468a47']
I am very new to DDD. I have just started reading the book Domain-Driven Design: Tackling Complexity in the Heart of Software(by <PERSON>). Following the Cargo example in the book, I am trying to figure out how one can keep track of which Vessel (Truck or Ship etc) carried cargo from one location to another. I am not sure whether a Vessel should be an attribute of a Voyage or Carrier movement(Registration No etc ), or model the vessel as an entity on its own. I am assuming that in the process of transporting cargo many vessels operated by different carriers will participate in the journey. Pleas help. Thanks in advance <PERSON>
d23fd4b1-4c17-fbee-c013-a2820131a0ed
['7509fcb2-51ed-91bf-3a2a-538473468a47']
I am very new to DDD. I have just started reading the the domain driven design bible. someone please help? I need to keep track of which Vessel (Truck or Ship etc) carried cargo from one location to another. I am not sure whether a Vessel Registration No should be an attribute of a Voyage or Carrier movement. I am assuming that in the process of transporting cargo many vessels operated by different carrier will participate in the journey. An example preferably in c# and explanation will be appreciated very much. Thanks in advance
722779c3-c95a-fd21-d5f6-708e3b419de7
['750e84da-0b83-aa62-872f-eac80bd6c797']
Depending on what you're expecting, I'm not sure you'll be able to get what you want since JUnit uses reflection to "know" the names of your test cases while C++ doesn't have reflection. I think, once you start using CppUnit, it's pretty easy to use. In order to get automagic knowledge of test case names, in our build, I put together a script using typical command-line tools (eg perl, sed, awk, grep) so in practice, it's possible to get what JUnit affords, but it'll take a little more work. HTH, <PERSON>
84219182-7447-a162-2ab0-f28e9409bc16
['750e84da-0b83-aa62-872f-eac80bd6c797']
In our project, we typically inherit from one of our base exception types. It would be nice to be able to register with CppUnit an exception handler that would call a method on our exception. I've looked through the source and it doesn't look like this is possible currently. Can this be put on the TODO list, please? (I'll see if I can work on it if I get some free time). Thanks, <PERSON>
92a60d1e-03f2-7f1e-2a20-ff543c0b9cc9
['7525b137-5e02-fd72-31c5-c17a19605a36']
<PERSON>, I have found more problems with the compressed file stream classes. 1. There was an error in the previous changes you applied to FXGZFileStream, the extra readBlock line means input gets discarded. So it does not work at all. I've been using my patched 1.6.8 version until now and had not noticed this issue. I guess I must be the only one using this class! 2. There is a scenario during write at the end of stream where the input is all consumed but the compressor still holds data internally. Due to the while condition this was not handled so the final output was never flushed. My previous tests had not forced this scenario, it probably needs the right combination of volume and complexity of data. This problem only showed up in my application but I have not been able to replicate with test data. 3. There is the potential for a simliar situation during read with the file input all consumed but the decompressor still having more output to give. From my limited investigation it appears the libraries do not behave this way and will provide all output for a given input. I have however updated the test for readBlock result to only break on error not on zero bytes read. The code is more correct this way and removes the risk if the libraries change behaviour in the future. 4. With the above changes the flush at end of stream works but a call to flush mid stream would not result in all possible output being written. I have therefore reimplemented the flush methods to force a flush of the compressor. The appoach is simple but seems a bit clumsy, you may have a better way to do it. I did not need this feature but it seems logical to make this change for completeness. Context diff included below. Regards, <PERSON>.
0cee44cd-8a60-8f1e-e194-7ccd8f4e49e5
['7525b137-5e02-fd72-31c5-c17a19605a36']
<PERSON>, I have a set of objects that I am serialising to a file via save/load. The objects contain Icons. What I noticed was the transparency of the Icons changed after the objects were restored from the file. After some investigation I think I know what is going on. FXGIFIcon loads into an RGBA format and the original color table is not preserved, so must be reconstructed during save. It appears this is a 2 step process first determine unique colors, then determine which is transparent. The Icons I had chosen to test with, bigdoc.gif and minidoc.gif, contain 2 entries for white in their color tables, one for the transparent background, the other for the body of the document. So in this case I think the color table gets reconstructed with one white entry that is flagged as transparent. The result is that the body and background are transparent and only the border is displayed. If I use an Icon with a unique color for transparent, all works fine. Since this is probably an unusual case I doubt it is a high priority to fix. Editing the Icons to avoid the problem might be a good idea. Note that minapp.gif is a similar case. I have not done an exhaustive search for other cases though. I have a question also. While looking through the standard Icons I noticed that some do not have transparency set (as far as my image editor can tell) however they do get rendered with transparency and I could not see how this would be happening in the code. An example is dirupicon.gif. Regards, <PERSON>.
90461f22-b85c-33de-6895-01ec64e4ef38
['7536c120-51e8-103c-407f-98bc36747e94']
Dear <PERSON>, I posted about this recently, but it was lost in the noise. I was at the TCPM WG session where they were discussing a draft recommending TCP implementation practices. During the session, there was consensus to move the Eifel algorithm, to an experimental section of the document _because_ it has not been widely deployed and because of existing uncertainties regarding IPR. What was discussed in the session was the occurrance of an implementor being presented with a schedule of charges for licenses by an IPR holder. Eifel has been described as a way to handle packet duplication events experienced due to bicasting in the draft (section 10): http://www.ietf.org/internet-drafts/draft-elmalki-mobileip-bicasting-v6-06.txt While the IETF (in compliance with its policy: BCP 79) will not take a position on the validity of the IPR claims, please be aware that Eifel is not widely implemented, and isn't likely to be implemented by default in TCP stacks, due to its experimental nature. If bicasting is important, it may be worth investigating if another mechanism is available for reducing the effect of duplicated packets on upper layer protocol stacks. <PERSON>
c06d7229-2050-5ba1-b1f8-0ce7b8fa4c38
['7536c120-51e8-103c-407f-98bc36747e94']
Hi, I think there are some interesting discussions going on in a different thread, but I thought I'd start a new thread in order to talk about a contentious issue without polluting the other. Regarding draft-pashby-ipv6-network-discovery-00.txt, this provides a mechanism for devices to be made respond to queries from another device on the IPv6 network. This is not an existing capability. I'm concerned that if there is a way to find out all the nodes on a link, that this information may be used (by the querier, or another device) to cause remote flooding attacks onto a network, or to particular otherwise unmodified hosts. In IPv4 it is feasible/trivial to try all addresses in a subnet in order to find targets for attack, but in IPv6 >2^60 combinations may need to be tried. The anonymity of the present (but quiet) IPv6 node is probably useful in this case. There is no system, except MLD which can force response from unknown nodes in IPv6. With MLD, the reporters can be made to expose only one of their link-local source addresses. They are not required to expose global addresses. At the moment there's no security for MLD, but the risk is limited to link-local addresses which are not vulnerable to off-link attacks. I'm loath to introduce a more generic function like this which exposes global IPv6 addresses, unless there is verifiable trust available to the nodes, before they are forced to respond. <PERSON>
9f9f1227-3e97-3ba2-e382-5e6855200ebc
['7537adb0-3c10-b0fc-cc89-f06276469ba3']
Hi, I have written a JAPE for creating some annotation types (Anno A, Anno B , Anno C). I am running this application on a corpus where each document may or may not contain any of these annotation. I am trying to segregate the documents containing any of these annotations from those not containing. This segregation may occur in many ways- 1) either arrange documents in particular order in corpus (documents containing these annotations on top ) 2) or deleting documents not containing any of the above annotation from corpus. 3) Or creating a separate corpus for one of the category. Is there any in-built feature for this in GATE? Or will I have to do this using JAPE/Groovy? Any help will be appreciated. Thanks and Regards, <PERSON>
e4489357-c750-2cb0-6257-4926fdff8c8f
['7537adb0-3c10-b0fc-cc89-f06276469ba3']
Hi Users, I have recently added a new gazetteer list to my ANNIE gazetteer list. But I am unable to change the following *parameters of the list*- 1)Annotation Type--> I want it to be Lookup 2)Major 3)Minor Every time I add them manually in gazetteer list and then *save and initialize*/ *initialize* ANNIE gazetteer, it automatically deletes the parameters. Any help will be appreciated. Regards, <PERSON>
2e97550f-65ef-177f-ecb9-f3085eef94ea
['7542ba58-0e6a-b044-e83c-27b9b1b07d43']
hi thank you for reading my post is there any listbox like component for j2me applications ? I need something like a list box with following features : 1-it allows me to know when user navigate up/down in list box 2-when it somehow fired an event for user selection change , it let me know the current itemindex or item value. can some one help ? thanks
fdb7fe14-f0ca-a774-16ff-738f6c7a9df0
['7542ba58-0e6a-b044-e83c-27b9b1b07d43']
Hi Thank you for reading my post. I have a web application which i developed and works well on my tomcat 5.5.9 Now i purchase a hosting plane with tomcat 5.5.9 it is on linux and have cpanel. they said that i should upload all contents of my war file into http_doc folder , so i did I used ftp to upload them all , now that i try to access my web site it return HTTP Status 404 - /index.jsp meanwhile index.jsp in under that www folder. I test the server by uploading an html file to make sure that i put content of war file in correct place. nad it works ok with HTML file. can some one help me about this problem ? Thanks
17635a10-b4de-935c-d251-52647ce9f126
['75447476-9435-2ab8-078e-dabd4d16f6e7']
Hi all ! I want to get the caller method name which calls a particular api. So ideally i should get "testMethod" when ever i call "logData" method. There is another question related to this after earlier part is complete I want to use "aspectJLogging" method when ever "logData" method is called. Please suggest this as well. I had attached the files. I would be very thankful for your help. Many Thanks <PERSON>
37a253d0-08ae-15df-41a1-ce38468bc1a4
['75447476-9435-2ab8-078e-dabd4d16f6e7']
Hi all ! I want to get the caller method name which calls a particular api. So ideally i should get "testMethod" when ever i call "logData" method. There is another question related to this after earlier part is complete I want to use "aspectJLogging" method when ever "logData" method is called. Please suggest this as well. I had attached the files. I would be very thankful for your help. Many Thanks <PERSON>
58146e24-0ca5-c3aa-4750-6fdaf535945c
['75460ccf-2a50-0cd2-7d12-2e2d44af44ac']
Hello, i have a scenario in a Grails 1.1 application where a controller populates a domain object and passes it to a service for validation and saving (some business logic is also included in the service method being called). The service has *transactional=true* and i found that the object is saved when passed as a parameter in the call from the controller to the service method even calling *domainObj.discard()* inside the service method, doesn't discard (because the object is saved already but not necesarilly valid tough). Is this supposed to be that way? Where is the transaction created? In the controller or in the service? I also found that changing *transactional=false* makes the service work as expected. Here's a step-by-step if the former explanation was not clear enough (most probably due to my poor english writing): 1.- Controller finds domain object by *params.id* 2.- Controller populates domain object with data from *params* map 3.- Controller calls service.saveObj(domainObj) 3.1.- Domain object is saved (even if it's not valid) 3.2.- Service method is called 3.3.- A call to *domainObj.discard()* doesn't work!! Any help or clue will be appreciated. Regards, <PERSON>
b2988e37-c609-277e-047b-d21519be293c
['75460ccf-2a50-0cd2-7d12-2e2d44af44ac']
Hello, i changed my UrlMappings.groovy file so i can have error codes mapped to an ErrorController. Inside the controller i'm trying to access exceptions object in the same way that default error.gsp does but i get a null pointer exception. I'm mapping error codes to a controller so i can use withFormat and render the error description in html or json, depending on the request format. Any kind of help will be appreciated. Thanks
b8363a14-8cf7-5e46-c474-0d779297850b
['7552c9e7-62c5-1513-2ad4-f9d00c0bb9bc']
Can someone point me to where I can install a public file or where the documentation is?? I want to read a configuration file but every time I rebuild the application it wipes out the user data thereby making me have to rebuild the configuration all over again. During testing this is a pain. Secondly, I would like to have TWO applications share (read/write) the same file? In production? Thanks, <PERSON>
3b1a1c78-9446-d569-ba55-440e93adbbbf
['7552c9e7-62c5-1513-2ad4-f9d00c0bb9bc']
We've been getting a lot of pressure from management (sales) to support smart phones other than Windows. We started to look at MonoDroid and then Novell kicked (actually got sold) the project and staff to the curb so we stopped our plans. Now that Xamarin (I'd like to know what that means and how to pronounce it) is back and it seems to be a production ready solution, I've been asked to restart a proof of concept which I've been working on so we can put together a project plan. I've got most of the application figured out and broken down so we can design business layer, application functions, communications, and user interface. I am then hoping to be able to change out the user interface and then have a port to the iPad (which is where all the sales teams are saying the sales are). So here's the overall questions I'm stuck with. Here's my two design questions: 1. In our PC/Windows phone application we check our version against a service and if there's a new release, we download it and upgrade. So, going through this, I see Android and iOS use a sandbox concept differing from what I'm familiar with. So, how does one handle updates? Is it possible to download a program and once it gets downloaded it runs and replaces the application? 2. Secondly, one of our applications is a business application that runs on the sales staff's company owned phone/tablets. Company's own these devices and put them in the field and part of the business requirement is to make the sales application password protected so it cannot be uninstalled (which also deletes all the sales records). Is there a way to set a pin so an app cannot be removed? All I can come up with is a 'watch dog' approach where you have two applications each monitoring the broadcast message for package removal and then reloads it unless a pin is entered (but the sales data would be lost however I could resync from the server). But, this seems a little to much like malware. Can I somehow drop the application into the system/application store thereby making it a system application? Or how do you handle this business requirement? Thanks, <PERSON>
01d806e4-f521-2da4-a69d-717e444542de
['756674c0-c9a0-0f42-b1be-5aa5d9dc9149']
Hello, We have a table in HBase that has very large rows and it goes OOM when the table mapper attempts to read the entire row into a result. We would like to be able to map over each Cell in the table as a solution and it is what we are doing in the map anyway. Is this possible? Like the default behavior for Accumulo? We looked at the settings on Scan and didn't really see anything and the source code of Result looks like it wraps an array of cells so the data is already loaded at that point. We are using HBase .98.1 and Hadoop 2 APIs Thanks <PERSON>
ef919f07-4994-47d3-94cb-a7d10b99cf96
['756674c0-c9a0-0f42-b1be-5aa5d9dc9149']
Hello, We have a table in HBase that has very large rows and it goes OOM when the table mapper attempts to read the entire row into a result. We would like to be able to map over each Cell in the table as a solution and it is what we are doing in the map anyway. Is this possible? Like the default behavior for Accumulo? We looked at the settings on Scan and didn't really see anything and the source code of Result looks like it wraps an array of cells so the data is already loaded at that point. We are using HBase .98.1 and Hadoop 2 APIs Thanks <PERSON>
1833c928-5998-d226-ff4c-d8edd09b038f
['756f70c7-0571-c73b-1c36-6cc1268bc74e']
I have a couple of Cocoa applications that I want to make backwards compatible with 10.1.5. I am building on 10.2.4 with the Dec '02 tools. Most things work, but by experimentation I have found that I cannot link to NAN or use isinf(). __isnand() works, but I should not have to call a __ function. Printing is not working for me in 10.1.5. I get an empty, tiny page, but this may be a problem in my ad-hoc page setup code. Is there something I am missing? I can't find any documentation that cites NAN or isinf() as problematic in 10.1.5 compatibility. TIA, <PERSON>
5865ea77-f112-d6eb-8488-c6dc4e7cf843
['756f70c7-0571-c73b-1c36-6cc1268bc74e']
Hi <PERSON>, I will share with you a class factory approach that works for me. C++ is statically linked, so unless you want to jump through hoops you do not have the ability to add classes at runtime. If you can handle that, then all you need to do is add a ClassID -> class instance method to ClassFactory in order to use it. Cheers, <PERSON>
bdf8791d-0d27-0a1f-ef30-9c352ea310f5
['7588b9b1-2184-be61-b943-7f0447bc81f4']
About 95% of my contacts are CW and I operate an hour a day mobile CW with the FT100D. The 500 Hz filter is active almost all the time and it is excellent. I have a friend who bought the 300 Hz filter and I have used his FT100D. It is a nice tight filter but I doubt I ever spend the money to buy one. There are still a lot of guys out there who don't know how to get on frequency when answering a CQ. The more narrow the CW filter you must tune + or - to find the off frequency guys calling. You will not hear them at all unless you tune (good use of the RIT). I would say if you can get a GOOD deal on a 300 HZ filter go for it but you sure don't need one for most CW work. I know guys who have gotten rid of 250 Hz CW filters because they found them to be too tight. 73, Ron W8RJL
d9231c1f-f398-d8a4-2ba7-43b6c74e8773
['7588b9b1-2184-be61-b943-7f0447bc81f4']
I am contemplating re-mounting the FT-100D from the cab to the trunk of my Ford Escort. -SNIP- My reason for contemplating this change is to find a better solution for all of the cables this rig requires. Anyone have any thoughts as to the likelihood of success of this undertaking? Best Regards, <PERSON>, KB1FLR
5b76c9e7-641f-9811-9148-6b9e6c216828
['758ce57e-d0c5-fbfe-9a40-66f0dfe4c962']
However, my guess would be that you will encounter very heavy resistance for suggesting cmake, perhaps up to simply being totally ignored. I agree that's the likely outcome. My personal experience with cmake has been extremely bad. Mine has been great but I know not everyone's is and I'm admittedly a bit biased since I'm a part time CMake developer. From my experience, people tend to have either a love or hate relationship with CMake, but seldom in between. The need for me to deal with the Windows build is rare so I've got little motivation to drive the conversation either direction. Just that if it is something the dev community wanted, i.e. if the feeling in the group was "We don't like the way this scons business is going. Can somebody redo the build with CMake? " then I'd be willing to work on the implementation. <PERSON>
ececb204-aaba-4aef-9a92-d551141b06fa
['758ce57e-d0c5-fbfe-9a40-66f0dfe4c962']
Is there a logging infrastructure currently available to drivers in Mesa? I was looking to clean up some of swr's debug / info output and have it conditional on the MESA_DEBUG and LIBGL_DEBUG variables but then I realized that it's really something useful all across mesa so there may already be something there for it. If not though, I'd be interested in adding some very light weight functions for just that purpose could be used by any driver rather than just fprintf(stderr, ...); <PERSON>
7b31e1f9-c9bd-98d5-e438-199b89b6b43c
['7593c17f-5c4d-af6e-9aee-bd1d0b136936']
Hi, I'm creating a custom layer for my imx6q based board. I'm extending sysvinit_2.88dsf.bb by creating sysvinit_2.88dsf.bbappend in my custom layer. There is already a similar bbappend file in: meta-fsl-bsp-release/imx/meta-bsp/recipes-core/sysvinit/ bbappend file in my layer is not being used. I do not want to skip the bbappend file in meta-fsl-bsp-release and I want my bbappend file to be used as the last recipe for sysvinit. Is it possible to have multiple bbappend files? How can I include my bbappend file without any modification to the bbappend file in meta-fsl-bsp-release? Any inputs will be of help! Regards, <PERSON>
68b36f8c-1521-04a2-2da1-824f2e528170
['7593c17f-5c4d-af6e-9aee-bd1d0b136936']
Hi, I happen to return to imx6 mailing list after a long break. I vaguely recollect an old issue: on imx6q vivante graphics hardware rendering (for x11 build) did not work because the vivante libraries (x11 build) were not DRI2 compatible. As a result when using x11 then graphics is not hardware accelerated. Is this issue resolved? If yes, in which yocto release is this resolved? Any inputs will be of great help! Thanks & regards, <PERSON>
0546dc58-0b4e-6147-b851-412c229ee966
['75a1a516-9c3a-4d8e-2a75-be8eeb92861a']
I need to know if syswrite is a blocking call. I am using it to write to a UNIX Socket. Somewhere in my application, my code is getting blocked. I have narrowed it down to the data write code. So I am wondering if syswrite blocks. I am really in need of an answer. Any help is appreciated. Thanks <PERSON>
865c7e68-2182-1b82-6d67-1e39136f3d8c
['75a1a516-9c3a-4d8e-2a75-be8eeb92861a']
Ok. I have gotten the blocking() interface to set the blocking flag to 0, but unfortunately this has not resolved my problem with syswrite doing a blocking write. Below are some snippets of my perl code and a print out from my logging. The logging clearly shows that the sywrite is blocking. Anyone have any idea how I can resolve this problem? Are UNIX sockets able to work in a non-blocking mode? Any help greatly appreciated, kwo
bf177e27-11af-3ea1-0a77-cd0fe18eaf33
['75a4a6d0-c87b-0a0a-3906-0485ea4ea4d1']
Hi. In conjuction with DVD playback (under VDR) I am looking for the possibility to scale MPEG and SPU data from a DVD from 4:3 to 16:9. Some DVD menus are stored in PAN and SCAN mode, where the actual Video and SPU data is stored on the DVD as a 4:3 image. The decoding hardware is then supposed to re-scale it to 16:9 to fit the aspect ratio. Unfortunately I could not find such a feature in the DVB API. There is supposed to be an ioctl called VIDEO_SET_ATTRIBUTES. But it seems that it is not implemented, although it's documented in the API!?! In short terms: I am looking for a solution to display 4:3 Video and SPU Data as 16:9. The TV-Set should also accept the data as 16:9 and not try to rescale it back to 4:3 according to the information on PAL line 23. Thank you a lot in advance and happy holidays to everyone. <PERSON>.
5af0cd24-0c1c-9f73-0b8c-dbdfbe611bf7
['75a4a6d0-c87b-0a0a-3906-0485ea4ea4d1']
Hi, I'm trying to use the VIDEO_SET_ATTRIBUTES ioctl call to maybe fix some aspect ration problems under VDR. I'm using kernel 2.6.9 drivers with the firmware from CVS on Nov. 23rd with a Technotrend DVB-Sat FF card rev 1.5. I added my own method (SetDVDFormat()) into device.c dvbdevice.c of the vdr-1.3.17 sources and I tried to call this method from inside the DVD-Plugin. The ioctl for setting the VIDEO_SET_ATTRIBUTES unfortunately always returns -1. Now I'm stuck at this point and I have no idea on what else to try to make it work. I posted this problem to the VDR list before but I got no replys to it. Please take a look at the postings: http://www.linuxtv.org/mailinglists/vdr/2004/12-2004/msg00032.html http://www.linuxtv.org/mailinglists/vdr/2004/12-2004/msg00061.html Any help into the right direction will be greatly appreciated. Thank you a lot in advance. Andr.
9c9d84f7-ccfb-1484-7c6c-6f161b463617
['75cedbe9-2a77-5614-4899-8c28fba49538']
I'm currently working on a PHP+Apache2+SomeDB project at the company I work for. We're primarily a Windows shop, but Unix could make it also. The biggest issue is support for whatever platform is chosen. Personally, my experience is with Debian/Ubuntu, but as far as I know, both don't have major commercial backers (yes, I know Ubuntu has one, but how long have they been around)? I'm pretty much looking for everyones opinions on what experiences they have had with the following *nix/DBs. I'm leaning to Solaris 10/Postgres, but that is new ground for me. What does everyone think about this? Best support cost vs support offered is important also.. * Windows 2000/2003 * RHEL * SUSE Enterprise Linux * Solaris 10
6f993b09-d2bc-78ef-84b0-2df2779700a0
['75cedbe9-2a77-5614-4899-8c28fba49538']
Don't know if for-sale posts are allowed, so please don't flame me if they aren't... I've got 3 Cisco 2500s (2x 2501, 1x 2503) for sale. They are in slightly different flash configurations and RAM configurations. 2 are running IOS 11.x and one is running 12.x (I think). They all have rack ears, power cords, 10BT transceivers, and DCE-DTE cables. I'll also throw in a MicroAnnex XL terminal server with serial cables. The routers are in decent condition - the cases have some scratches and the occasional scuff. Everything is located in the Aurora, IL area and is pickup only. Let me know off-list if you're interested - I'm taking offers. If you'd like more information, let me know. Jon
988ee9a0-8df6-75ce-71c6-7daf5eddf940
['75d9a310-3c7b-5711-68a9-3d41f32034bc']
Hello I'm new to Debian, have used Slackware for years but latest release was a mess so I decided to move to Debian and give it a try, however I have some doubts and need to complete my configuration to keep working while learning to use this new distro, so these are my current doubts: 1) How can I disable the graphical login and/or avoid X11 to start automatically? I can't install the nVidia driver since it tells me to stop X11 before. And, is it possible to leave it that way, just with the simple text consoles, and start X11 manually when required? 2) How can I install KDE? Currently it is running with Gnome... I've just download all 5 DVDs plus the updates one... the gnome (un)install shows most KDE applications to install, but not the full window manager... and, once installed, is it safe to remove Gnome to free disk space? Just the window manager, not its libraries that should break some dependencies if deleted. 3) There are some programs/drivers that independently of the distro, I prefer to install them by hand from sources. Does it affect in some way if I do this? The most common ones are kernel, MySQL, msn, irc, some multimedia stuff, etc. but at least the kernel and MySQL that I don't like to use packages for them... 4) Is it possible to install LILO instead of GRUB? I know it is better and advanced, but I like LILO... would it be safe to install it without breaking some kind of dependence? Well, that's enough by now, today I need at least to setup the nVidia driver and hopefully the latest kernel (no matter if it is GRUB by now)... Thanks for your help, <PERSON>
7e99220f-6681-a57a-bf9f-296f99c2337a
['75d9a310-3c7b-5711-68a9-3d41f32034bc']
Hello First of all I want to thank all your kind responses, it will help me a lot to start using Debian. I will write a single message with comments for all those who replied to me, to avoid flooding the list: <PERSON>: Now I know that not all of the DVDs are necessary. Respect the 'expert' mode installation, when booted the DVD #1 it showed only "Install" and "Graphic Install" (I have tried both of them), but at any moment saw the expert or advanced installation, it just ran itself without asking details. I will try the "Debian way" with most packages, and see if they work for my needs, but at least kernel and mysql I need them from sources, although will leave the original shiping package to avoid breaking dependencies with httpd, etc. Matthias:
dcc00801-d301-fe93-7708-bd3cfab28e1b
['7608f8ea-cd1b-4c83-a3c2-4d8d9c07b167']
Hi! I'm writing a small test utility to evaluate some open source db's (MySQL, PostgreSQL, SAP DB and SQLite). This is my first time playing with the new DB-API 2.0 and I must say, it's *not* portable between db's (I think I'm not the first person who found out ;-) ). Besides of differences between SQL modes also the paramstyle is different. My first question is: why does sapdbapi only support 'named'? All other db's support 'pyformat'. Second question (not sapdbapi specific): why is there no 'BEGIN' for a transaction? As far as I found out a transaction is automatically started after a commit or rollback, but isn't 'BEGIN' specified by SQL-92? Best Regards, <PERSON>
b0cb2ac3-eb8e-178e-94a8-f9a932978e5d
['7608f8ea-cd1b-4c83-a3c2-4d8d9c07b167']
Hi, I'd like to add a micro navigation to my homepage. Actually there should be links to the title of the index.html pages. For example: Home / Products / Hardware / It's not a problem to generate a dict with the path as key and the title as value. But what is the easiest way to build the path? Is there a module to step up the path (html/products/hardware/index.html)? <PERSON>
dc5d6a18-eabd-8907-7e8f-eef94e1a69d2
['761da173-4f98-52aa-6830-5e92e351f616']
Hello, Could you please take a look at http://www.lilianne-blaze.net/wiki/bin/Java/HeaderPatternLayout ? Mostly inspired by http://juliusdavies.ca/logging.html (search for "A Log Rotation Best Practice"). The general idea is to write a line on top (and optionally end) of every log file with basic information about the current environment, like memory, Java runtime and VM info, operating system, etc. Any feedback, comments, suggestions, questions, would be most welcome. Greetings, <PERSON>
f1557c1d-7044-d472-6c27-ab575e43c593
['761da173-4f98-52aa-6830-5e92e351f616']
Hello, I have an InitializingBean, lazy-init=false, where afterPropertiesSet calls method xxx which is annotated with @Transactional(propagation=Propagation.REQUIRED, readOnly=false). The purpose is to populate the database with some initial data if it is empty. It fails, apparently @Transactional isn't used when called from afterPropertiesSet, is it a bug or is it by-design? For now I simply wrapped it in transactionManager.getTransaction(td); + transactionManager.commit(ts);, and it works as intended by me, but it seems a bit 'hackish'. Is there a way to make @Transactional work in such a scenario? Or are there any "best practices" applicable here? Greetings, <PERSON>
17a5d452-9cb8-08fe-1bfc-c906aa5792a6
['762a02b0-d328-1fc3-5ff5-380f682e7f57']
Hi, I want to discuss one particular topic in Wireless Sensor Networks. Several routing protocols have been proposed. Totally independent of this, several researchers have also focussed on developing energy efficient MAC protocols for sensor networks. To be more clear, the routing protocols assume 802.11 based MAC protocol. I want to know, if any research has been done in this direction. I am thinking if we can provide smooth integration of currently incompatible MAC protocols and routing protocols. Kindly let me know, if anybody in this group is working in this direction. I request the members to give me some references in this direction. Looking forward for your suggestions, Gokina.
5c5fdbdf-09aa-4f53-35ed-65bcea5fc8e3
['762a02b0-d328-1fc3-5ff5-380f682e7f57']
Hi, I am doing research Independently in Mobile Ad-hoc Networks and in Sensor Networks. I have some doubts in some of the research topics in these fields. 1. I have read some publications about trust management in Peer-to-Peer networks. The concept of trust has been extended to Mobile Ad-hoc Networks. In P2P networks the nodes are stationary where as in the second case the nodes can move. Assume that the trust management is not centralized but distributed across the network. Each node maintains trust information about the nodes which are in it's interest. Now, suppose one node is found guilty. The nodes in its surroundings can know about this. Immediately after the node is found guilty, this particular node can travel long distance where the nodes are not aware of his behaviour. I want to know if the existing publications take care of this kind of situation. 2. Is it possible to extend the concept of trust to Sensor Networks. Since, Security is also an aspect of Sensor Networks, Is there any mechanism so that one Sensor can choose the routing path that has trusted Sensors. 3. In Peer-to-Peer Networks there is a concept of Super Nodes. In this they exploit the property that some nodes are more powerful than other nodes. To the best of my knowledge, that main problem in Sensor Networks is energy. Is it realistic, if we assume that some nodes in Sensor Networks are more powerful than others and handle most of the traffic on the behalf other nodes which are less powerful. Will it be realistic, if we assume that these powerful nodes have no energy constraints. I mean can there be a situation where some of the Sensors in the Sensor Network have continuous power supply. I think if these assumptions are realistic, then there is a possibiltiy of giving new direction to research in Sensor Networks. I am not sure, if this can be done. I request the members in this group to clarify my ideas and if possible, kindly give me some references. Looking forward for your reply, Gokina.
9703cdde-2595-ec81-be1c-3c119e5a35c8
['7631d3d2-9e78-d786-91b0-b9eef5a8fc71']
Enter your vote today! A new poll has been created for the dpjudge group: What is your favourite method of playing Diplomacy? o dpJudge o Other Judges o Judge Tournaments o Casual Face-to-Face o Tournament Face-to-Face o E-mail o Postal To vote, please visit the following web page: http://groups.yahoo.com/group/dpjudge/surveys?id=2331633 Note: Please do not reply to this message. Poll votes are not collected via email. To vote, you must go to the Yahoo! Groups web site listed above. Thanks!
96159a51-0fe1-cfb0-b2f9-429588d8d5c6
['7631d3d2-9e78-d786-91b0-b9eef5a8fc71']
Enter your vote today! A new poll has been created for the dpjudge group: Three years after this question was already asked, it is asked again to see the change in taste over those three years: what is your favorite country to play in a standard game? o Austria o England o France o Germany o Italy o Russia o Turkey To vote, please visit the following web page: http://groups.yahoo.com/group/dpjudge/surveys?id=2317700 Note: Please do not reply to this message. Poll votes are not collected via email. To vote, you must go to the Yahoo! Groups web site listed above. Thanks!
342d529a-248f-4959-2d8f-c13a5e1f961b
['763d02ff-c90f-1d11-58c7-4ffb636d0a4a']
for last couple of weeks #master testresults are drifiting a lot. Yesterday I merged #libevent into master; <PERSON> was pressming me to do it And we will fix everything in master instead of working on branches. Now <PERSON> and me are fixing #master and test results. #master should be fixed today or tomorrow; please be patient. May be hold on to your commits unless it is absloutely essential? if compiler warnings hurt your eyes. this is good time to take break:) thanks for your understanding, -antony
42a792a0-e983-ad6b-a9cc-50198a8a328c
['763d02ff-c90f-1d11-58c7-4ffb636d0a4a']
Hi, Yesterday <PERSON> and I met with NSS guys and here are some notes from the meeting. NSPR threading: no need to use NSPR threading on Linux, because on Linux it is jut a wrapper around pthread. Don't open a NSS DB file simultaneously(not even one app writing and another re-reading). Due to the nature of in memory data structures things can go wrong. Close it completely and start again. Also use new format, not the old Berkeley DB. In the NSS code base, operations on X509, ASN 1 is not within the 'crypto boundary', only the crypto operations are. When it is running in FIPS more NSS will support non FIPS operations. It is upto applications not to decide. If we want to store PSK/Xauth/EAP secrets in NSS we could store the "secret" encrypted by the master key. And the NSS can decrypt it for libreswan. NSS can't talk to the Linux Kernel via Netlink; e.g to install ESP keys directly into XFRM module. However, they can talk to some of the TCP offload NIC. Human readable error strings in NSS is possible. They mentioned an application(libreswan) must initialize "error code tables" in NSPR to access it in NSS. It might be worth investigating. You may also have to install NSS utils. A quick googling shows libreswan use PR_GetError. However, libreswan seems to be missing initialization code, PR_ErrorInstallTable, nspr_InitializePRErrorTable. I haven't looked in detail. It seems prerr.h or prerr.c is a starting point. Google developers are actively working ARM and other non Intel Architectures. Cross compiling for MIPS Big Endian, for OpenWRT, should work! The buildroot Makefile seem to build MIPS, even though NSPR config files only seems to support Little Endian architectures. There is no OpenWRT Makefile yet. -antony
8915d8ae-d058-7eec-e8d7-f93a183a0142
['764fb61b-c11f-b0cb-2c8b-a9f66d4fa441']
Does anyone have any experience with using uIP to do UPnP setup or any other kind of Zero Configuration setup of the interface? Alternately, has anyone out there ever done a DHCP server under uIP? Dare I ask for some kind of simple DNS service? Any help or pointers would be appreciated. <PERSON>
af673adf-5e00-5dc8-ef52-ffcb7adec07e
['764fb61b-c11f-b0cb-2c8b-a9f66d4fa441']
This turned out to not be so hard. The hardest part was untangling some of the include recursion that occurs when creating a union to hold the application state structures. Otherwise, it was just a matter of initing both applications and creating a common appcall function to sort out which application/local-port is receiving data. <PERSON>
176ccacc-168c-53cd-1d3a-70065ce9779d
['765d27ff-9ff8-e8fa-06cc-9a3f63e250cd']
I have a ZFS-based NFS server (Solaris 10 U4 on x86) where I am seeing a weird performance degradation as the number of simultaneous sequential reads increases. Setup: NFS client -> Solaris NFS server -> iSCSI target machine There are 12 physical disks on the iSCSI target machine. Each of them is sliced up into 11 parts and the parts exported as individual LUNs to the Solaris server. The Solaris server uses each LUN as a separate ZFS pool (giving 132 pools in total) and exports them all to the NFS client. (The NFS client and the iSCSI target machine are both running Linux. The Solaris NFS server has 4 GB of RAM.) When the NFS client starts a sequential read against one filesystem from each physical disk, the iSCSI target machine and the NFS client both use the full network bandwidth and each individual read gets 1/12th of it (about 9.something MBytes/sec). Starting a second set of sequential reads against each disk (to a different pool) behaves the same, as does starting a third set. However, when I add a fourth set of reads thing change; while the NFS server continues to read from the iSCSI target at full speed, the data rate to the NFS client drops significantly. By the time I hit 9 reads per physical disk, the NFS client is getting a *total* of 8 MBytes/sec. In other words, it seems that ZFS on the NFS server is somehow discarding most of what it reads from the iSCSI disks, although I can't see any sign of this in 'vmstat' output on Solaris. Also, this may not be just an NFS issue; in limited testing with local IO on the Solaris machine it seems that I may be seeing the same effect with the same rough magnitude. (It is limited testing because it is harder to accurately measure what aggregate data rate I'm getting and harder to run that many simultaneous reads, as if I run too many of them the Solaris machine locks up due to overload.) Does anyone have any ideas of what might be going on here, and how I might be able to tune things on the Solaris machine so that it performs better in this situation (ideally without harming performance under smaller loads)? Would partitioning the physical disks on Solaris instead of splitting them up on the iSCSI target make a significant difference? Thanks in advance.
dc8a9bba-2ef5-3636-b8d7-f4c2794ccd6e
['765d27ff-9ff8-e8fa-06cc-9a3f63e250cd']
We have a somewhat complex NFS server environment with iSCSI backends[*], which holds /var/mail among other filesystems. Due to performance problems on /var/mail, we recently switched it to using SSD-based iSCSI backends. However, performance has failed to improve the way we expected and now we're trying to track things down and improve them. Low level raw disk performance seems to be fine; IO stats on the SSDs say that IO is fast even under load (with read and write wait times of only a few msec, low utilization, and so on), and under low load the fileserver can turn around IO through iSCSI in a msec or two. Under load what we see is relatively high system CPU utilization on the backends (on the order of 40% as reported by vmstat, significantly higher than seen on HD-based backends) and fileserver IO average service times of anywhere from 20 msecs to 80 msecs. We can't see any obvious saturation point; network bandwidth is moderate and the actual IO rate from the fileserver is surprisingly low (under 100 IOPs/sec, which is well below what the SSDs can sustain as far as we know). Right now we're using an old version of IET on an old kernel (<IP_ADDRESS>) because we started all of this in 2008 and qualifying new versions of anything for production use is a big deal. We're willing to upgrade either or both if either people think that it would be likely to improve the situation or the current IET versions have better stats. So, here's some questions: - have people seen anything like this before? - is there general advice on tuning iSCSI and network parameters for low latency on 128Kb random read IO requests? (We're using gigabit networking without jumbo frames and default iSCSI parameters.) - does IET have any support for getting iSCSI-level IO statistics, eg LUN-level 'disk' IO stats like the low-level disk IO stats? If IET doesn't have anything for this today, would people be interested in a patch to add them? - how unlikely is it that IET is adding IO delays under load (it's clearly not adding them all of the time)? Is this basically completely unlikely? Is there any relatively easy way to tell? (I ran a tcpdump packet capture through wireshark but it doesn't handily match up requests and responses for me.) Many thanks in advance for any advice and so on. I would be happy to give people more details if you think it would be useful. - cks
d387ad0f-eaab-229f-0fb6-691ab5e77135
['7660d094-796a-6db2-2b39-4077ac68533a']
<PERSON> I'm trying to work out the best way to use Puppet to control an existing estate, and then how to refactor the configurations. I'd like to use the appropriate tools where I can (eg taking distros to get consistent sets of packages, package managers for the bulk of deployment). During the adoption phase, I think that I want to identify configuration changes from a standard, known, installation image + kickstart file. The configuration is the details of the changed file contents and other metadata (eg existence, permissions). I can fairly easily identify changes from a standard build (eg find all packages and verifying them all, picking out the failed verifications - I'm not sure how to find those filesystem changes that are not part of any package, but leaving that aside for the moment). As I want to use standard builds, I'm basing the effort on fedora builds (puppet 25.5). What seems to me much harder is how to produce a manifest that can reliably reproduce the existing systems, and demonstrate the reliability of the manifest, for more than a couple of machines. I think that I need to do something like: 1- clone the filesystem of each host that needs to be put under control 2- find the files to put under puppet control (ie those filesystem changes that are not the result of the kickstart and subsequent update processes). 3- build a manifest for the changed files (I tried ralsh file for this and came seriously unstuck :-( ) 4- create a test VM and run the manifest against it 5- produce automated comparisons between the filesystem on the VM and the cloned filesystem from step 1: this comparison must be able to handle issues like 'the contents of file x should match the cloned contents, except where this IP1 address is present, when IP2 should be used.' This feels quite heavyweight, but when I tried a much lighterweight approach of running smaller parts of manifests I was finding that I'd need to clone the filesystem in any case, otherwise I couldn't identify any changes that Puppet had made, (nor revert back!!) Have I over-egged this? Are there any simpler approaches that I should try - I considered starting from scratch, but it's non-trivial identifying all of the IT services that are set up, let alone reproducing their behaviour from scratch? cheers <PERSON>
c4688e3e-203c-04dd-88a0-245f08832a49
['7660d094-796a-6db2-2b39-4077ac68533a']
<PERSON> I've not used Puppet yet, so this could be an obvious misunderstanding on my part. I've seen many manifest entries for files that write rules about file ownership and permissions. Surely these types of attributes should be managed through the package managers, with a possible override if there are any package errors. Otherwise, you've got to replicate an existing set of information in a new format. Am I missing something, or is there a simple way of working that just uses the ownership and permissions available in the package managers? cheers <PERSON>
1f35c596-f9a8-7689-4057-f635170b2b27
['766ff8a9-cc5d-18e2-a931-214e47a620b5']
Hi folks, I would like some information on buying a Linux supported wireless G card for my IBM T23 laptop. It supports PCMCIA type II cards. I want something supported under Linux obviously :-) but without requiring the use of ndiswrapper if possible currently. Can people give me some info on their experiences with any cards they would recommend. Please be specific as to exact make/model. Thanks, <PERSON>
56804f70-6b52-798f-30cd-87dca0d7a6dd
['766ff8a9-cc5d-18e2-a931-214e47a620b5']
Hi gang, Anyone know of any Zope/Plone web hosting providers in Toronto or Canada. For a decent rate of course :-) Also for .ca registration as well. I need to register a .ca domain for cheap as well. Thanks
896aa373-de54-1d02-ef5d-9bf89d53fb2c
['767cac0b-6676-963a-9f4e-ccd72df62d3e']
Hi Team, I am a Masters student at University of Florida. I am presenting Apache Jena to our class as part of research on RDF and semantic web. I have got material to speak about Jena but I am having trouble finding some use cases of Jena in current industry. I read that many Semantic web applications use Jena but I could not find any project which uses Jena. Can you help provide names of some projects which are currently using Jena and how Jena is being used there? That would be very helpful to generate interest among students. It would be helpful for future if we had a page on the Jena Homepage about Who is Using or something similar. Thanks & Regards <PERSON>
f2af1167-bc8f-0c2e-4acd-c2c3ba361b6d
['767cac0b-6676-963a-9f4e-ccd72df62d3e']
Hi, I am trying to implement text search in Jena via Fuseki. I have followed through the documentation and created assembler file. But after starting fuseki server using config parameter - there is no data in the dataset and thus returns no results for simple query or text query. What I have tried so far 1. Built the TDB dataset using java -cp $FUSEKI_HOME/fuseki-server.jar tdb.tdbloader --tdb=assembler_file data_file 2. Built the index using java -cp $FUSEKI_HOME/fuseki-server.jar jena.textindexer --desc=assembler_file 3. Started the fuseki server using fuseki-server --config ../assembler_file.ttl I tried the answer on Stackoverflow http://stackoverflow.com/questions/30447536/fuseki-indexed-lucene-text-search-returns-no-results but using --desc gives error no service name. Another user had similar issue a year ago as in this thread http://thread.gmane.org/gmane.comp.apache.jena.user/7892 but there are no solutions there as well. Can someone please help here? Thanks & Regards <PERSON>
1ef7169f-e4ea-8b24-7d7c-5c74713d3b5b
['768dc605-071e-4bfe-caea-7ec3072a017d']
Hopefully this is not too off topic. I have a server in which I'm creating an Exchange Information Store backup using ntbackup. The .bkf file it creates is 112GB. My goal was first to rsync it over to a local linux server, and then rsync just the changes over the internet to another linux server. I do this with other Exchange .bkf files (in the 4GB to 20GB range) and it seems to work well only transferring the changes over the internet. My usual method has been to map the Windows 2003 drive on the local linux server and then use rsync. When trying this with this system, the transfer speed is super slow from the beginning, at 6 mbit per second on a 10/100 mbit network. If I try the copy using smbclient and get, I actually get 60 mbit p/sec for the first 30gb, then it seems to slow to 6 mbit. A regular cp does basically the same thing as it starts out fast and then slows all of a sudden. This obviously will not be very practical as on the local copy I usually would copy the entire file each night. At that rate it will take days. Any suggestions on how to copy a 100gb file over a local LAN quickly? Or what copy/transfer command might keep the same quick throughput during the entire transfer? Or could there be something wrong on the network? Thanks, <PERSON>
11a14bb7-d32c-7bf5-e708-6b76e4de2493
['768dc605-071e-4bfe-caea-7ec3072a017d']
Hello, I've been trying to figure out how to control the space required for backups using rsync, large files, and an incremental backup scheme. In particular, I've got two customers in which I'm creating Exchange backups using the built-in MS backup, then rsyncing an exchange.bks file nightly. One customer has a 4.3GB exchanges.bks file, and although rsync works wonderfully by only sending 43mb worth of changes to this file across the line, using hard links obviously causes incremental directories of at least 4.3GB daily. With 30 days of incremental, this adds up. I've looked at rdiff-backup, but do not like it as it's so different than rsync. I've got a lot of scripting built around rsync and I like the way it handles file/dir selections, so I will not consider rdiff-backup. I've been experimenting with rdiff itself and trying to figure out a formula for creating delta files of only the changes to the Exchange.bks file and using these delta files in the incremental directories, instead of creating the hard links. I've gotten it to work, but not in a rotational scheme where you always save only the latest copy and the deltas, and can easily restore (patch) using just the latest and the delta. Has anybody found a solution for this? Or is there any other tool out there that automates this? Again, rsync works great for the transfer, but storing data incrementally is becoming the problem. Would be nice if a future version of rsync created rdiff incrmentals as an alternative option to 'cp- al' or 'link-destination'. Thanks, <PERSON>
c431220c-c2dc-4b50-116d-f63aeb51cce1
['76a4b395-71b1-2c2f-36f9-45af3fc0a87e']
All, I've been struggling to find a neat solution to this problem for a couple of weeks now, so I'd love to hear anyone else's opinion. I'm on Rails 3.2.13 and Ruby 1.9.3-p392. I have a table of train schedules, and a table of train schedule locations with a has_many association. Each schedule has a set of validity dates, e.g. each Friday between 1st January and 28th September. The times in the schedule locations table are all 'seconds since midnight on the day the train runs': where a train runs over midnight, the number of seconds is greater than 86400. Whilst this is reasonably 'clean', I want to re-engineer the schedule location model to return a Time object for each location, calculated on the fly based on the date I queried the schedule model for, i.e. Schedule.where(:runs_on => Date.parse('2013-01-01')) will return a Schedule model with an instance variable 'runs_on', and when I request the associated schedule locations, the 'arrival_time' and 'departure_time' fields are auto-calculated based on the instance variable in the schedule model. Is this a sensible way to do things? Is this even possible or am I over-engineering it? <PERSON>
ef40927c-5638-c465-48c2-0ff8d3b8cb6f
['76a4b395-71b1-2c2f-36f9-45af3fc0a87e']
Hello I have a user coming from a Vodafone /23 who's getting a 503 error when trying to access one of my sites over HTTP. I don't even see any traffic from this /23 when the errors happen, so I'm fairly sure the problem is at Voda's end. What's the best way, as a "not a VF customer", of reporting this to Vodafone? <PERSON>
cd27b880-eab9-721b-db77-2876dbbbd3ea
['76ae23c8-62a5-9c28-8578-15d52b24fcf8']
Hello Everyone, I am a returning XFCE user from several years ago. I've used KDE for years, but I've wanted something a bit more straight forward and remembered the feel of XFCE when I first transitioned to Linux almost 10 years ago. Awesome to see the updates that have been made since then. Great job to all! The only minor hurdle I encountered was the XFCE4 wiki. I was looking for window shortcuts, but the information on the wiki was totally incorrect. To the pages credit, it does mention that the information their needs to be brought current with the most recent version. Keybindings are usually the first thing I look for. They help me to keep general navigation and window management as efficient as possible. Not finding them, or getting inaccurate information on the very first search was a bit disconcerting. If possible I would like to help updates some of that information when time allows. <PERSON>,
9d146312-47db-d3ae-1666-686ae99f4362
['76ae23c8-62a5-9c28-8578-15d52b24fcf8']
I had an awful time today trying to install the HighLine gem on my machine at home. After several attempts, I tried disabling my firewall on a whim, and presto, the gem installed successfully. I don't recall having this problem with other gems in the past, so I am wondering if my trouble had to do with where the HighLine gem resides? Is there a specific port that gem installs communicate on? How should I setup my firewall to allow gem installs? Thanks <PERSON>
2dbcbcc4-c9d0-0742-f06c-2e6ff5569a54
['76b14da3-51e0-7aa6-0d3b-29673a73b91f']
Hi there, I am trying to set up Freevo in my Linux box. Because I have an DXR3 that I want to use as the TV output, I am following the Wiki documentation on how to configure Freevo to use DXR3. The part on configure DXR3 drivers(?) in the system went fine. But when I try to compile ffmeg I keep getting a segmentation fault error. Could any one help me? My dist is Mandrake 8.1 on a AMD K6II-500 CPU. ffmpeg version is the one mentioned in Wiki. Is there any other way to put freevo and DXR3 together? Thx
a03d9c71-3574-e7e7-dffe-411206ebe334
['76b14da3-51e0-7aa6-0d3b-29673a73b91f']
Hi there. I just have a few questions. 1: Divx with subtitles - I know that mplayer supports it. And Freevo? If so, how? If not, why not? 2: libcss - I have installed libcss. Is there anything else to be configured to allow me to view encrypted DVDs? 3: DVDs original menus - Is there a way to start a DVD like in a normal dvdplayer, starting with the dvd menu and choosing the audio, subs, chapters, etc from there? 4: Audio CDs - I can't play audio cds. When I insert an audio cd it will not start, and in the music menu, freevo tells me that no cd has been detected. Rafael M. Torres (<EMAIL_ADDRESS><PERSON> (oLi61mfjL3CYzPYX@example.com)
3d0f460d-0ebc-62bf-0e99-accbf7ba9fe6
['76b55302-ef9c-407e-8642-f62a4c3502a6']
When I try to setup smtp authentication with sasl on 9.3, my mail log tells me postfix is not compiled with sasl supprt. Huh??? I thought leaving out mysql lookup support was shortsighted, but his is worse. Does anyone know of prebuilt packages for SUSE 9.3 that supports mysql and sasl auth? Thanks <PERSON>
ad7cba17-4570-9a12-478b-8cd106693e25
['76b55302-ef9c-407e-8642-f62a4c3502a6']
Hi guys, I need to add an ASP site on a web server running Apache2 (SuSE 9.1 iirc). This server already has a large number of sites on, so I don't want to put something on it that will cause trouble. Of the APS implementations for Apache2, http://www.apache-asp.org/install.html appears to be the most popular. What are your experiences/recommendations? Thanks <PERSON>
d76b0e42-6675-bd22-deb2-15afaf47ee85
['76b89d92-d803-c1c3-f4df-79f4e415f8a9']
Hi, I was analysing whether one should go for Post-compile weaving or Load-time weaving. I read on the bea site (http://dev2dev.bea.com/pub/a/2005/08/jvm_aop_1.html) that Load-time weaving (using agents) affects both scalability and usability. Are these issues of scalability and usability applicable to post-compile weaving as well? Any help on the technical advantage of one over another would be highly appreciated. Thanks, <PERSON>,
eff152dc-538c-4e41-c8cc-ab79638224be
['76b89d92-d803-c1c3-f4df-79f4e415f8a9']
Hi <PERSON>, This is in context to the article on the bea site (http://dev2dev.bea.com/pub/a/2005/08/jvm_aop_1.html). They have talked of the problems with Bytecode Instrumentation-based Weaving, namely double bookkeeping: building a class database for the weaver, instrumentation is inefficient. I am not clear after reading the article whether these issues would be there with post-compile weaving as well. (I have posted the question on their site as well.) Thanks, <PERSON>,
4d66495f-b85e-6f53-c944-03b20e655601
['76ba4f10-ecaa-4950-bcf8-ff1a84ba28ae']
Hi <PERSON>, this looks promissing as we could always use the latest and maintained OneMediaHub app on our devices :) In using it this way, is also Notes and task supported? I use funambol in combination with SOGo (groupware). via a s4j SOGo connector synchronization takes place not against the funambol databases but (for contact, agenda and tasks) against the SOGo database. Will this work (in theory) using your connector? Thanks, <PERSON>.
9a1563bc-f175-a2b8-6175-192ce6ce8ab0
['76ba4f10-ecaa-4950-bcf8-ff1a84ba28ae']
Hi, last week received my new toy (Galaxy Tab 10.1) so finally have good use for funambol synchronisation (contacts, tasks and notes) with my SOGo server and android phone :) Ran into a small issue: the following two settings are not persistent (when you set them click [save] and re-enter the config screen they are disabled again): * Sync with server [Push client changes to the server] * Contacts [Always create new contacts in the funambol address book] Android version is 3.1 (Honeycomb) log file (trace level) attached. Hope this helps, regards, Ruud.
30001fa6-a842-1308-3c80-c09a5f9c688f
['76bb6608-a98a-c190-8229-8b79254e3d21']
Hi all, Thanks for your help so far, very useful! I ran into another thing: I compiled the same website using Visual Studio and using a nant script. I compared the dll and its satellite dll's with the Loetz Roeder's resource tool to see what resources are in there and what the names of the binary resource files in the DLL's are. The satellite assemblies are identical so that's fine. But there are differences in the main dll's: The names of the included binary resource file differ: The Visual Studio build uses the name of the class that the resource belongs to. Example: phw842_frm1.ascx.nl.resx belongs to phw842_frm1.ascx.cs (In Visual Studio, the is represented by the fact that this resx file 'hangs' under the ascx file and in the csproj file this is specified in the 'DependantUpon attribute. My nant build turns this resx file into phw842_frm1.Resources (resource file name is based on resx file name) and adds it to the dll). But the Visual Studio build turns this resx file into phw842_1.Resources (resource file name is based on the classname in the code file the resx belongs to (this class name differs from the name of the code file!)) and adds it to the .dll. The result is that resource managers in the code who try to load the binary resource file phw842_1.Resources by using the type of the class, won't find it. How can I change my CSC task so it includes resource files with names that are base on class names (types) instead of resx file names? The way I compile the files in Nant is by just listing all resx files in a resources node (see below example code), I guess I have to somehow specify the links between code files and resource files so Nant can figure out the names for the binary resource files? Note that for the satellite assemblies, I parsed the csproject files and code files with a program I wrote myself to get to the classenames (types) that must be used for the names of the .resource files. Thanks, Bert-Jan
ee736724-175b-f882-b0dc-0038995c09de
['76bb6608-a98a-c190-8229-8b79254e3d21']
Hi, I am running into the following problem: I use a nant script that has produces an exe. It has a references to Interop.TGXLib.dll and AxInterop.TGXLib.dll (wrapper dll's). Everything compiles fine. Now, when I run the program I just compiled, it is looking for dll's TGXLib.dll and AxTGXLib.dll, but it can't find them because these names are wrong, it should look for Interop.TGXLib.dll and AxIntererop.TGXLib.dll. If I rename my interop dll's to TGXLib.dll and AxTGXLib.dll, everything works fine. So the names of the references in the exe are wrong. The part 'interop.' has dropped off. Can I set these names somehow in the Nant script to force it to make references to 'Interop.TGXLib.dll' instead of 'TGXLib.dll" ? I don't want to rename files to get it to work. Maybe in the references tag for the CSC task or something? For example <includes name="Interop.TGXLib.dll" refid= Interop.TGXLib /> or something... Let me now, Bert-Jan Bert-Jan Lappenschaar
da8917ff-26a7-db52-9bca-1df3b55504d7
['76c2a5f0-8da4-0422-8679-f0e2fdee95e6']
I'm in the process of preparing a little lightning talk on Mondorescue ( I might post a draft here ) and while prepping for it I found there were not a lot of screenshots of mondorescue in action lying around. Anyway I ended up making a few for my talk and I thought other might find them useful. If possible could they be put on the website. The screenshots are my own work in a private capacity and are released into the public domain, or in case this is not legally possible I grant any entity the right to use this work for any purpose, without any conditions, unless such conditions are required by law. <PERSON>.
28e72f48-b739-7137-bfbc-4ec296199795
['76c2a5f0-8da4-0422-8679-f0e2fdee95e6']
A reminder to all list members that: 1. DNS related questions should usually be sent to more specific lists such as DNS operations: http://lists.oarci.net/mailman/listinfo/dns-operations 2. Discussion regarding the NANOG organisation and political issues surrounding it are off-topic for the main list and must only occur on the nanog-futures list http://mailman.nanog.org/mailman/listinfo/nanog-futures <PERSON> NANOG Mailing List Committee
49c756a9-8444-e447-efa7-bc980dd8cad6
['76c84d40-6c12-82fa-9555-95322ef6d385']
Hi, I am trying to use GnuGK 2.0.6 on Linux with OKI, Cisco ATA186, and MultiVoIP 800. I started with the simplest config file (actually empty, just 42 and allow). Endpoint registration seems fine, and call setup succeeds for any combination. Also, if I call from OKI to OKI, everything works fine. But if I call the Cisco or the MultiVoIP, I can only hear voice in one direction. This is not a NAT/firewall issue, because everything is just on my LAN. The codecs seem to match (723 in all boxes). Call setup seems fine, and I can not notice anything weird with the GK's "-ttt" output. Any pointers to the right direction where I should dig further? Boris
196e1329-2725-5ddc-6f9e-c55c566a4516
['76c84d40-6c12-82fa-9555-95322ef6d385']
<PERSON>, In such cases I would say it is time to attach the serial link and see what's going on with the booting. As you have already used the bootloader to put Familiar on your iPAQ, this should be already familiar to you: start minicom (make sure the parameters are right), stylus-hole-reset while holding the center of the joypad, connect iPAQ, use bootloader prompt. I always specify ttySA0 as my kernel console (I vaguely remember this was off by default?), that way I can see how far the kernel boot goes without any additional manipulations. <PERSON>
067fc220-81ed-669b-064d-49d680d41c68
['76ceeac9-f238-d9b7-e64b-24be41709401']
Hi <PERSON>, I have entries in rtrules to force traffic to a certain provider, i also have squid running on the firewall i want to force traffic for certain websites over a certain provider (using lo interface). This works however when testing if i simulate a failed link, LSM detects it but traffic that is defined in rtrules does not want to route, is there anyway to force traffic over provider1 and if provider1 fails traffic will go to provider2 Thanks
2bfaf219-5cdd-177b-9f63-bf7eeed42e20
['76ceeac9-f238-d9b7-e64b-24be41709401']
Hi i am using Squid 3.3.9 with Kerberos authentication on my network. we know have a requirement where we need to give guest users access on the same proxy, is it possible to run squid on a additional port and have different ACL's for those users connecting to that port? I know ideally having a different subnet is the best option Thanks!
eee2b868-d90f-c680-753f-9ec3a0cb549b
['7701653c-3db5-6543-b67c-b5027ff78a93']
Greetings, I’m recent to MIT Kerberos. Currently I’m working on a project at Faculty of Engineering University of Oporto implementing an OpenLDAP server with MIT Kerberos Authentication, using Cyrus SASL. The Linux distribution I’m forced to use is Slackware 9.1 I’ve configure, make and make install Kerberos v5 r1.3.1 in my system but the /etc/krb5.conf file is not created, not even the /usr/local/var/krb5kdc/ directory for the KDC database. My question is Have you found any problems with Slackware 9.1? I’ve already gave a full-read at the Installation Guide along with a look-up in the System Administration Guide, but haven’t found nothing regarding to this. If you know of any place where I can find information about Kerberos v5 installation in Slackware I would appreciate. Best Regards <PERSON>
4513afc6-e8e4-6b76-42f2-f23e1fe2d476
['7701653c-3db5-6543-b67c-b5027ff78a93']
Hi! I've a Tru64 v5.1b, with LDAP Module for System Authentication (from Internet Express 6.1). I've also installed OpenLDAP 2.0.27, which comes with Internet Express, and have no problems authenticating users stored in this OpenLDAP 2.0.27. After I've changed the LDAP Module for System Authentication to use another server (not on localhost) wich runs OpenLDAP 2.1.30. At first, the LDAP Module for System Authentication could'nt even pass the Test Configuration, then I've put the line "allow bind_v2" in the server and the Test Configuration was passed. But UNIX doens't authenticate users on that server. Even if I perform a "su username" or try to login with an account stored in that LDAP server, the attempt is rejected and there is no traffic in the server side. Is there any incompatibility with other versions beside 2.0.27? Another question I have is, is it possible to use Tru64 Unix to authenticate LDAP users trough SSL (port 636) or this isn't supported yet and must all be trough cleartext (port 389)?? Hope you can give me some light here. Thanks, <PERSON>