query_id stringlengths 4 64 | query_authorID stringlengths 6 40 | query_text stringlengths 66 72.1k | candidate_id stringlengths 5 64 | candidate_authorID stringlengths 6 40 | candidate_text stringlengths 9 101k |
|---|---|---|---|---|---|
6aa4f389-82dd-6b4a-fe4f-cb6a0321a7e0 | ['72066748-6c97-65a5-5540-3cdcef93d0b2'] | Don,
Is this the option you are looking for?
MQRO_PASS_DISCARD_AND_EXPIRY If this option is set on a message, and a
report or reply is generated because of it, the message descriptor of
the report inherits: v MQRO_DISCARD_MSG if it was set. v The remaining
expiry time of the message (if this is not an expiry report). If this is
an expiry report the expiry time is set to 60 seconds.
<PERSON>
| 34ee3d40-0166-4b3e-ee6e-27191d122c33 | ['72066748-6c97-65a5-5540-3cdcef93d0b2'] | Intra-Group Queuing will allow your QMGRs that are part of a QSG to send
messages to each other without using the channel initiator. The QMGR you
connect to would need still need to know where the local Queue resides
(just like when you use channels). This does require shared System
queues to be defined.
Gregg
|
3f6141e9-c867-bfdb-9afd-78a55541d965 | ['7207f195-5d2b-09bd-d516-25259960a714'] | Hi all,
I took a look on the exception classes generated by wsdl2java and got it to
work now: The point seems to be that the exception class needs to be
derivated from AxisFault, then the deserialization works.
I'm not happy at all to have to derivate my exceptions from AxisFault, as I
want to keep my business logic interfaces (and the exceptions which belong
to it) as independent from the underlying RPC technology as possible. Does
this really mean that it is not possible to use custom exceptions which are
not derivated from AxisFault?
Regards,
<PERSON>
| 6e885be1-0e6e-6983-f75c-781f0fdec5ae | ['7207f195-5d2b-09bd-d516-25259960a714'] | Hi,
Once again, as no one replied: Did really no one else face this problem up
to now? Don't you use exceptions in your service interfaces?
I took a look on the JAX-RPC specification 1.0 now: It defines that service
specific exceptions must extend java.lang.Exception - and not something
like RemoteException or AxisFault. This means that Axis 1.1 doesn't comply
with the specification in this point, unfortunately. Should I write a bug
report?
Regards,
<PERSON>
|
2b3fe6b8-8752-896a-767a-e4446d23071f | ['72261789-2813-75d9-5eae-73340fc58092'] | I'm noticing a RealVNC viewer issue when running VNC over an SSL VPN connection versus an SSH connection.
Under the SSL VPN that desktop connects and is displayed properly. The desktop window manager is Ubuntu Unity-2D. When I left click on the file browser icon, no problem that works fine. When I left click on Ubuntu Dash icon, the entire desktop session is disconnected and the VNC server is terminated.
I have a similar SSH VNC session that does not exhibit any problematic behavior.
Does anyone have any clues on what may be happening? I suspected something relative to the mouse and keyboard mappings, but I have been unable to identify anything specific that will correct the problem.
Kind regards, <PERSON>
| 3bc5d78a-6695-fcd1-1c7b-e785f2e01cfb | ['72261789-2813-75d9-5eae-73340fc58092'] | Sorry for inundating everyone with this new Jython questions!
I am trying to use the Xerces DOM parsers from Jython.
Please notice the object casting below. I have been working through
Chapter 10 in the Jython book to figure out how to accomplish this Java
object casting in Jython.
Can someone help me with how to do this?
Kind regards, <PERSON>
|
f5b2799a-2100-28ef-8404-dfd1657008b6 | ['72299629-052e-2c9b-0038-70c75a700cef'] | I have to implement a single sign on solution, but would like to know when we really use CAS.
When we say CAS as a single sign on solution for many appliactions, does this mean that all these applications as internal applications for an organization?
I mean, with CAS , we do the authentication for all these application against the SAME db/ldap/any-other-authentication-data-store .Is this idea correct?
Or if I need to have a single sign on solution just to have integrations with external applications like Salesforce,Siebel-on-demand etc. does CAS really help?
Pleas let me know.
Thanks
<PERSON>
| 55f80257-feea-675f-9aba-62287d4ac846 | ['72299629-052e-2c9b-0038-70c75a700cef'] | <PERSON>,
This works now for the IDP client , when they tried with another certificate
,which is a 1024bit/SHA1 certificate. :)
Mostly issue should have been with the wrong key as you guessed. And it
might have worked now, when they tried with a new setup , with another new
certificate. But they think,the issue was Our SP implementation(using
OpenSAML) , doesn't support 2048/sha256 certificate. But I had tested using
my test certificate which is of 2048/sha256, and it worked fine .
So just wanted to confirm once again that Opensaml doesn't have any
restriction on certificate key size/algorithm as 2048/sha256,?
Thanks
<PERSON>
|
20c214ba-9f1b-9b13-263c-3b8ea6dc1b01 | ['722acd2e-05e1-b81b-7d8d-7fd9a402a5e8'] | *This message was transferred with a trial version of CommuniGate(tm) Pro*
I am trying to install mysql.
I have completed 2 steps
1. running the installer
2. putting a my.cnf file in the /library/mysql/var directory
3. Installation has already made a mysql startup file and as per the
instructions below
Next we need to start the MySQL server before we can set the root
account password. The installer will install a Mac OS X Startup Item
into /Library/StartupItems/MySQL. This will make sure that the
MySQL server is started at boot. You can execute this startup item
manually to ensure that the MySQL daemon is properly started. You
can do so by executing the following command as root in the terminal:
shell> /Library/StartupItems/MySQL/MySQL &
i need to check this
when i run this it says you must be root to run ConsoleMessage
what do i do. I am aleready the admin user.
For your info I am trying to install on Macosx Jaguar 10.2.2
Bhavul
| 5f8249e8-400a-d6bc-c015-ddf28a441ae6 | ['722acd2e-05e1-b81b-7d8d-7fd9a402a5e8'] | *This message was transferred with a trial version of CommuniGate(tm) Pro*
I have recently loaded mysql on macosx 10.2
It has loaded it in the directroy /library/mysql
presently I have my user prompt on the terminal screen. I know that I need
the sql prompt in order to be able to issue sql commands. How can I get to
the sql prompt.
Please help
<PERSON> |
f16a89b2-07af-3b24-4954-c5cd25088dbd | ['7235ac7b-8c60-c9b4-3609-5abf552cf8b9'] | Today I found myself asking, why it is that I continue to read this list
and use this project.
And the answer I came up with is.... Beats the hell out of me.
Could it be the friendly, helpful people?
No that's not it. Because once again I find that instead of an answer to
a question, there is a
discussion on the "correct" way of asking the question. Once again I
find people being accosted
for having the nerve to ask a question in a way that someone finds to be
improper. And that
making sure that they are asking properly, is more important then them
getting an answer to their question.
Makes you feel all warm and fuzzy, doesn't it?
So I've asked myself again, why do I continue? Because it's my opinion,
that it's one of the best projects of it's type
to be found. BUT, and there is a but, ...
<PERSON> | 1955aa20-2581-1689-322e-67c0cd7b2706 | ['7235ac7b-8c60-c9b4-3609-5abf552cf8b9'] | For those of you that are using / have used IPCop
with a compact flash, What capicity are you using / would
you recomend? Looking at moveing from Smoothwall to IPCop
and using a smaller case.
And does IPCop come with lm_sensors, or something simmlar?
Have a micro-atx case, and want to do away with as many of the
fans as I can ( ie. the power supply fan, and the case fan.
Keeping just the cup fan, and a bit of case modding ( completely
diskless. ). So I will meed to monitor temps under load for a
while to see if it over heats.
Thanks,
<PERSON>
|
b44fc5a7-13ea-0c6d-8bb9-d2b182b592d4 | ['7243cfcc-86ca-e13b-a41c-bdd796fb395e'] | Hello,
I've noticed quite a lot of non-album track additions lately, and am a
little unsure how to handle them.
This is only for tracks that haven't been released in album form at
all, right? My default is currently to vote 'no', with ideally one
note to the moderator to add albums instead, and to vote 'yes' only if
it's clear that it is really a non-album track (by note, or suggestive
title). And the occasional 'abstain' when it's unclear. Is that ok?
Since I don't really see a way to add non-album tracks via the web
interface, it seems these are usually submitted through the tagger.
Does it encourage submitting unsorted MP3s as non-album tracks? And if
so, could it be made to warn users (more explicitly)?
Cheers
<PERSON> | 507fd8ac-c806-ac04-f8c1-69f921dd2c89 | ['7243cfcc-86ca-e13b-a41c-bdd796fb395e'] | Hello,
my flatmate and I just tried using the newest sarge-i386-netinst.iso
to install Debian on his Compaq notebook. Some incomplete notes on how
it went follow.
I realize that this is not released yet, and don't mind it not
working. I just wanted to give the new installer a test, and hope some
of this feedback is of use. Please CC: me on replies as I'm not
subscribed. Here goes:
The kernel booted flawlessly. Actually, hardware support didn't seem
to be a problem as far as we got.
The keyboard worked fine with English layout. Selecting the German
keymap didn't actually change the keymap. We tried both the PS/2 and
USB keyboard -- I don't really know how it is really connected, though
I'd guess PS/2. Not a big problem, though.
CD-ROM detection worked well. The installer module installtion seemed
to work fine, although there were some error messages I couldn't
place. I didn't write them down and don't have access to the machine
currently, but could try to get them tomorrow.
The network device was apparently probed successfully. However, the
static network setup failed because ifconfig didn't work -- busybox
printed its list of commands when calling ifconfig from the console.
We then tried the harddisk setup, using the all-in-one option. The
drive had already been partitioned with Partition Magic, which also
created filesystems. However, we didn't get the option to mount these
partitions without first choosing what filesystems to create. This
also happened on a second installation attempt, where the filesystems
had definitely been created correctly. After entering the mountpoints,
this module finished successfully.
As checked from the console, the partitions were mounted correctly.
However, /target/etc/fstab was not created, which caused the next
modules to fail. In particular, installing base from CD-ROM failed.
Cheers
<PERSON> |
7e50a9f6-b068-11aa-f0ed-5b7e28139924 | ['7258e62e-de0e-cdd1-2a82-3111dd8dfab0'] | Hi!
I've just installed OpenCMS 5.0.0 this week and started to create
a web site. Everything runs great, except I would like to
use another source code editor than the built-in one.
Is there any way to let OpenCMS call a .exe file when I select
"Edit sourcecode"?
Also, I tried installing "LEdit" just by copying LEDIT32.DLL into the
WINNT/SYSTEM32 folder, as the readme file suggests, but it won't
show up.
Thanks for helping,
<PERSON> | 0797348e-9fed-a487-0361-f86180e642bf | ['7258e62e-de0e-cdd1-2a82-3111dd8dfab0'] | Hi all,
I am using multiple content areas for selection in the WYSIWYG editor
as <PERSON> described in
http://mail.opencms.org/pipermail/opencms-dev/2003q2/005734.html
Is there any way to check whether one content area is empty (i.e. the
content editor did not enter anything in that particular body)?
The return type of CmsJspActionElement.include(...) is void, so I can't
directly check it. Is there any way to retrieve the content of a content
area
as, say, a String, where I could further process it in my JSP code?
Thanks for any hints,
<PERSON>
|
79f1449d-2306-c414-4933-3bcea8ed52ed | ['725fe1da-187a-6936-373b-8a7424bd3af1'] | Hi,
I'm working in the catalan localization of Sunbird. I see that ca is red in
the tindebox
http://tinderbox.mozilla.org/showlog.cgi?log=Mozilla-l10n-ca/1220715480.1220715552.11380.gz
But now I'm a little lost with all the recent changes. Today I have checked
out calendar project, from CVS TRUNK, MOZILLA_1_8_BRANCH and
SUNBIRD_0_9_BRANCH. And I see that dir calendar/locales/en_US/ is missing in
MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. Is that correct?
When I compare locales with TRUNK I see that everything is ok. So why is
tinderbox red ?
Thanks,
<PERSON><PHONE_NUMBER>.<PHONE_NUMBER>.11380.gz
But now I'm a little lost with all the recent changes. Today I have checked
out calendar project, from CVS TRUNK, MOZILLA_1_8_BRANCH and
SUNBIRD_0_9_BRANCH. And I see that dir calendar/locales/en_US/ is missing in
MOZILLA_1_8_BRANCH and SUNBIRD_0_9_BRANCH. Is that correct?
When I compare locales with TRUNK I see that everything is ok. So why is
tinderbox red ?
Thanks,
Quim Perez | 90230f70-5c2f-555a-ba79-de26e666c4f4 | ['725fe1da-187a-6936-373b-8a7424bd3af1'] | Hello,
I don't understand why Catalan locale is rejected.
https://l10n-stage-sj.mozilla.org/shipping/pushes?locale=ca&av=tb3.1
I see that my dashboard and tinderbox are green:
http://tinderbox.mozilla.org/showbuilds.cgi?tree=Mozilla-l10n-ca
When I finished and commited my work on Saturday 30th , I saw that
https://l10n-stage-sj.mozilla.org/ didn't get the update, as fast as
http://l10n.mozilla.org/dashboard/. So I couldn't validate
https://l10n-stage-sj.mozilla.org/. After that I've been busy until today. I
would like to sign in if it still possible. Sorry for the inconvenience
Regards
Quim. |
5858c363-abaa-be18-7248-383bcf3983b9 | ['729088ef-3e73-c250-f248-f690b97137b4'] | Hello,
we should have a mechanism to manage the state of the documentation. The
documentation is only helpful if the documents reflect the actual situation
of Qubes OS. For example, documents which are written for Qubes v1 or 2 are
outdated. Wikipedia displays certain meta information above articles which
are outdated and/or need to be reviewed to reflect the latest changes. My
idea is to have a similar mechanism. For example, it would be helpful to
look after every Qubes OS, Fedora or Debian release through the
documentation and mark the documents which need to be updated for the new
release. In addition, it would be helpful to create an issue for every
outdated document. This way it would be easy for the community to
contribute to the documentation.
Best regards
<PERSON>
| ca7e3543-3886-7ca6-8a1e-348ec4497676 | ['729088ef-3e73-c250-f248-f690b97137b4'] | Hello <PERSON>,
thank you very fast response.
You did not answer my question about Axt-Style vs. Non-Axt-Style.
What do you prefer or better what should be used Axt-Style or Non-Axt-Style?
I will create a script to remove all {% highlight trac-wiki %}
qvm-grow-private {% endhighlight %}, let this be my concern!
Best regards
<PERSON>
|
2cb80650-5752-7796-5c00-80a28d8717ed | ['72936592-6484-d638-ad5a-89ff677acb1c'] | Okay, I patched the doc/Makefile.am and got this. I'm a really dummy
with autotools, so I can't say I know what I'm doing. When I tried
adding aclocal -I m4 and automake to the compile script, I got the same
error as before I patched the doc/Makefile.am. Any suggestions?
<PERSON>
| a5ec4455-f719-22df-4838-59c61a94f72f | ['72936592-6484-d638-ad5a-89ff677acb1c'] | An updated 0.7.0 beta for Red Hat/Fedora systems (one that works this
time) has been uploaded to:
http://www.videolan.org/pub/testing/vlc-0.7.0-test1/rpm/redhat-fc/
If anyone is using Red Hat 9, let me know which dependancy files need to
be updated. Since I'm using Fedora now, I tried installing RH9 on my
old laptop, but it'll need a new keyboard to work right.
<PERSON>
|
38cd4dce-d078-6d85-1636-3e402927482b | ['729adf5c-1b23-0cfa-fd81-7cc7c09dce69'] | Hi,
First a warning, Io is new territory for me and I'm using VisualStudio 2005.
Could someone post an example of the IoBindingsInit() function for the
IoVM-2005-10-17-patched.tar.gz so I can try some OpenGL examples? The
MSVC solution IoFull-2006-05-26\projects\win32vc2005\Io.sln is out of
sync so I try to put the pieces together without using the _build.io.
If I'm completely off track some guidance is much appreciated.
Regards
<PERSON> | 8488663c-eb4a-5f6b-fae0-3f0acbdad8c3 | ['729adf5c-1b23-0cfa-fd81-7cc7c09dce69'] | Hi,
I have created Visual Studio 2008 solution and project files for the
new structure in the git repository.
It is not complete but includes Mesa (softpipe), glu and glut. I have
also added a demo application (gears) to see if it works.
I had to make some small modifications (cast from void pointer) to the
mipmap.c file in the glu project.
I have also included some c files generated by python when using the
scons build system.
This is my first post at mesa-users so be gentle with me ;-)
Regards,
<PERSON> |
6b1f86fe-050c-ebdd-398a-9518b28da266 | ['72b20c9b-ab19-fd41-58a7-2d5389aebc37'] | Hey folks,
as it turned out to be very hard to revert all patches from v3-2-test due
to conflicts, the following will happen:
On May 23 (release date for 3.2.0rc1), I will create a patch to remove
everything from v3-2-test which is not in v3-2-stable. If you are
interested in having a history, please revert your patches yourself before
that date.
Cheers,
<PERSON>
| e028897e-b647-da8d-4122-2e18dfb9c704 | ['72b20c9b-ab19-fd41-58a7-2d5389aebc37'] | Hello,
I have a question about winbindd design.
We noticed, that with "security = domain" winbindd
enumerates all members of a group plus all users
who have this group as their primary group.
"Security = ads" only enumerates the mebers of
the group, not the users which have it as their
primary group.
A Sniff shows, that the MMC works in the same way
as rpcclient / "security = domain". There ist a
second ldapsearch request for users having the group
as their primary group.
The question is whether winbindd should do the
second ldapsearch request, too. This means to
behave different to Linux, but like Windows.
With kind regards,
<PERSON>
|
8fd930bb-ef4a-58f3-9d2a-f875ac597791 | ['72b3f5ef-6781-3c81-2194-b31b6a632908'] | hello there
in our company we use two different trac. an internal and an external.
is it possible to put a prefix in front of the ticket id's so we can differentiate between internal and external tickets.
cause till now it's possible to have to tickets with the same id (intern, extern).
thanks for your help
regards
pascal
| af6e1351-d71a-35cb-7cc8-e01ea0f4baf5 | ['72b3f5ef-6781-3c81-2194-b31b6a632908'] | Hi all
I want to install trac on a real webserver. I made a subdomain trac.mydomain and got a new folder "trac" in the subdomains folder.
how can i install trac on my server? what source should i use? i'm confused about the infos on TracInstall site.
thanks for your help
regards
|
91b3ead4-20fb-b6ee-3196-b7dbe67514fd | ['72c3134a-6243-14df-62e0-d830268f5cf2'] | just as the topic describes.
I'm currently doing SNAT to force some destination tcp ports to be routed
through a specific route rather than the default route. To accomplish this
I mark thoes packets with iptables, use 'ip' to specify marked packets via
the specified route, and then use iptables to change their source address.
one problem i've noticed is that in netstat (and hence /proc/net/tcp) the
connections are shown as their original source. this causes problems with
some things such as identd, because identd won't see the SNATed to ip with
that port open. hence it will reply that it has no connection.
anyway i can get around this?
<PERSON> | a9ee4a2f-9e0e-1d14-b210-d2b211f6d31e | ['72c3134a-6243-14df-62e0-d830268f5cf2'] | I am trying to secure a machine that will have potentially untrusted
users. I would like to set a max memory usage per user though if this is
not possible then I can possibly work with setting a max memory per
process and then setting a max process, though that mix and match is not
so adjustable.
As the topic states, the question is if options that say 'in core memory
use' (memoryuse setting) or 'in core memory size' (memorylocked) are per
process or per user, such as the process limit is.
thanks
|
da52d5a0-e14b-35c4-19e4-7c989e157aaf | ['72c6b4be-0871-3182-df63-8bbb2f5dd874'] | I'm trying to compile in support for the 64-bit ifXTable counter support for the IF-MIB for Solaris 10 Intel (and Sparc ultimately). Based on bits I've read in the archives, I've
a) run ./configure with enable-mfd-rewrites
b) updated the source to 5.4.2.1 distro
What I'm running into is that 5.4.2.1 creates a package (my package installation is an outboard creation process that's been updated from many successful installs of the result for 5.4) that installs and starts up the agent successfully if I use the ./configure options *without* enable-mfd-rewrites (which is to say, the options I've used to date). If I reestablish the build environment with -enable-mfd-rewrites specified, I get an agent that starts up printing over and over to stdout
insert (-1) error on subcontainer 'interface container'
No other loggable output (I tried -Dread_config) gets a chance to print anything else out before the log/stdout gets flooded with this message over and over again in a loop.
Now, before I get into too many more specifics, should this work, based on experience anybody has had putting the -enable-mfd-rewrites option in per the instructions of README.solaris? I've read bits where folks suggested just going with the Sun-issued net-snmp binaries, but the reason I'm going to this trouble is that we have customizations (at least one of which has turned into a bugfix update) into our version of the agent.
Apologies in advance if the word on this has been brought up on the list or in the archives and I just didn't find it.
Regards,
<PERSON><IP_ADDRESS> distro
What I'm running into is that <IP_ADDRESS> creates a package (my package installation is an outboard creation process that's been updated from many successful installs of the result for 5.4) that installs and starts up the agent successfully if I use the ./configure options *without* enable-mfd-rewrites (which is to say, the options I've used to date). If I reestablish the build environment with -enable-mfd-rewrites specified, I get an agent that starts up printing over and over to stdout
insert (-1) error on subcontainer 'interface container'
No other loggable output (I tried -Dread_config) gets a chance to print anything else out before the log/stdout gets flooded with this message over and over again in a loop.
Now, before I get into too many more specifics, should this work, based on experience anybody has had putting the -enable-mfd-rewrites option in per the instructions of README.solaris? I've read bits where folks suggested just going with the Sun-issued net-snmp binaries, but the reason I'm going to this trouble is that we have customizations (at least one of which has turned into a bugfix update) into our version of the agent.
Apologies in advance if the word on this has been brought up on the list or in the archives and I just didn't find it.
Regards,
Wayne Tackabury
| 88303fe7-166f-c3a0-3b81-068bdecca38b | ['72c6b4be-0871-3182-df63-8bbb2f5dd874'] | Hi all/<PERSON>/Wes:
For all the pass scripts I've written in the past, I'm surprised I haven't run into this before, but I haven't.
I'm writing a pass script which will effectively synthesize a gauge and string object, to effectively detect a condition worth trapping on via a snmpd.conf 'monitor' directive. So the point is, my pass scripts need to GET other object in the management scope to make dynamic and conditional decisions on to drive the value of the OID's that the pass scripts themselves expose.
The problem is-I run the pass script no problem when directly invoking it with -g, and it works with the agent if I return static values without doing any snmp gets or get-nexts (in other words, the basic protocol through the pass script and its return is AOK).
The problem is-if I actually do a snmpget from the pass script, it times out. Now as it turns out, the get I need is from my application proxy agent, listening at an alternate port, so I simply updated the pass script to explicitly target the get at localhost:<that_port>. and all works fine. So I have a workaround, but I'm trolling for a general understanding of how to fix this the "next" time or when part of what I need is from, say, the ucd-snmp MIB or something whose instrumentation is built into the agent.
I'm suspecting the issue here is reentrancy. Now, there's a build option, which has had some reports of not working over the years, called "-enable-reentrant", which appears to affirmatively define (i.e., actually compile in) a fair degree of mutual exclusion primitive API code for the agent based upon what I can see from the #define hierarchy. On the other hand, I can't find in the code where this would actually affect reentry through the top of the GET PDU request processor.
OK, that's a coding question. Either way, after a rebuild with "-enable-reentrancy" in my target environment, the timing out on GETs issued from within my own pass scripts still is happening. And along with that, what my real questions are (I'm doing this on Solaris 10, gcc compilation from the 5.4.2.1 distro, etc.):
-- should the reentrancy support target the exact kind of thing I describe above, where a pass script servicing a GET request needs to issue another GET request for another item?
-- is this some kind of Bad Thing for pass script usage anyways? If so, can you folks suggest an alternate strategy?
Thanks in advance!
<PERSON><IP_ADDRESS> distro, etc.):
-- should the reentrancy support target the exact kind of thing I describe above, where a pass script servicing a GET request needs to issue another GET request for another item?
-- is this some kind of Bad Thing for pass script usage anyways? If so, can you folks suggest an alternate strategy?
Thanks in advance!
Wayne Tackabury
|
ba2b162b-dc9e-dbe8-205a-390d26d2ea46 | ['730d51b1-7247-ebd1-93e9-74f3ede44d20'] | I have down loaded disk1 of FC2 and cant get the loader to run.
the system boot till i get running /sbin/loader. then ti gose to the
blue screen and hangs for a while and the screen starts filling up the
random char's.
I tride it on 3 computers 2 pentiume II and p4. tryed to burn it with 2
diffrent burners each time I checked the md5sum before burning it.
the new computer I'm trying it on is a intel main board (D865PERL) with
intel p4 processor 256M DDR 333MZ ram 20 gig ide hard drive
<PERSON>
| db277683-be9f-33cf-19a5-8447c879a4a5 | ['730d51b1-7247-ebd1-93e9-74f3ede44d20'] | Hello
I have an IBM e-server x-series 335 with 2 xeon 2.6 GHz and 1 Gig of
ram. It uses a built in raid controller I have set up as a Raid 1 with 2 36
GIG hard drives. If I try to install using 64 bit version it tells me that
there are know 64 bit CPU and to use other version if I install 32 bit
version with the ACPI=off I can get it to install but it locks up all the
time. Most time I get a CPU 1 not responding error also when I check the
hardware it reports back 2 cpu I would have thought that it should have
shown 4 cpus since I have 2 dual core cpus running hyper threading. The
severe is an old one and the problem could be in the hardware, I have been
using Fedora since core 1 and never ran into so much trouble dose any one
have any experience install fedora on this server?
<PERSON>
|
918d946b-b05b-6e4a-99de-3a400458d189 | ['731a236b-9a23-545b-9987-d944d11d3988'] | Hi all.
What do I need to change so that Apache creates a world-readabe
access_log file?
I use a small tool (GeekTool - http://projects.tynsoe.org/en/geektool/)
to display the Apache log. Since I'm not normally logged in as
administrator, I need to change the read permissions of
/var/log/httpd/access_log to all. Only problem is - whenever the logs
get rotated, Apache creates a new access_log with the previous strict
permissions and I have to change them again ...
I thought this might be a FAQ but couldn't find anything at apache.org
or via a google search.
While this isn't a thing of life and death, changing the permissions
every week becomes tedious. So thanks in advance for any help. :)
Regards, <PERSON> | 26feda41-9b13-f43b-27f0-14d3bcb2b8d4 | ['731a236b-9a23-545b-9987-d944d11d3988'] | Hi all.
I think this is of potential interest to all Mac software developers,
so please excuse me in case it feels a bit off-topic to you.
_Introduction_
MacPAD is short for Macintosh Portable Application Description.
MacPAD files are used to get information about Macintosh software, like
prices, versions and updates.
Read more about MacPAD here: <http://macshareware.net/sdk.html>
or here: <http://macpad.sourceforge.net/>
_Discussion_
Some people, including <PERSON> and me, felt the need to add some missing
features to MacPAD. We started a private discussion and thought that it
is time for the community to join; since it is you, the developers,
that are supposed to benefit by using the format.
We set up a wiki at
<http://wiki.furrysoft.de/phpwiki/index.php/MacPAD>
to collect ideas and discuss possible implementations.
So please join us, add your thoughts and spread the word!
If you know other developer lists or forums that you think might be
interested, send them a copy of this message.
The topics we are discussing currently are:
1. Localization
2. Currency
3. Plugins
4. Private Extensions
5. A New Name
Hope to see you soon at the wiki. :)
Regard, <PERSON> |
f529d2fb-3d20-49e9-db66-014ec8163672 | ['7325a291-f672-ea55-1b78-6d7446ee523d'] | i have sent this to both lists in hopes that someone out of both
would have had experience with the framework.
I have read the docs in the source code download of wonder but I need
a little more guidance. I have a website (http://visionworxpro.com ,
not done yet) and would like to integrate paypal payments seamlessly
into the site. Can someone provide me with a little more guidance
than the docs provide. I don't like the deal about the Listener deal.
Can I just have a component that uses some sample paypal conneciton
code (api's) to connect to them to make the transaction and then
handle everything else my own way ??
Thanks,
<PERSON> | 69aa280f-ad49-d003-4e5e-a51eb96124d8 | ['7325a291-f672-ea55-1b78-6d7446ee523d'] | This is the first part of a two part series on Display Groups. Also
learn how to create a custom project template for XCode for code
reuse and to make your future projects a breeze. Covers To-Many
Relationships, Editing Context issues, API docs, learn to use
Jakarta's Commons Lang Utilities and create a fetch spec in code and
in the EOModel. This episode offers a lot of tips/tricks. Video is
available for $14 dollars and is nearly 2 hours long.
<PERSON> |
b16f4698-1640-6479-b8d1-5548b0867180 | ['73290adb-70fa-a2c0-eaed-4b839a9ff8b2'] | Hi,
I wanted to fresh up the 'Joining Active Directory as a Domain Member Server'
HowTo on the wiki for 4.x. So it was the first time, I had to compile samba
4.x as a normal member server (not as DC). And there a (maybe stupid) question
came up:
On 3.x I run the 'configure' from inside the source3 directory. Is this still
the same, if I only need a normal member file server (smbd/winbind) with
version 4.x?
For testing I configure/make/make install from the source3 directory and the
fileserver seems to work like expected on a first test.
I only wanted to know if this was the right 'configure' or if the one from the
top level of the sources should be used in this case too, like when compiling
for an DC.
Regards,
<PERSON> | 19454e71-71dd-2c0b-8a56-f1768063c1a4 | ['73290adb-70fa-a2c0-eaed-4b839a9ff8b2'] | Hello,
at work we are currently looking for something that allows to store
files on a Samba share. After a defined time, the files/directories
should be accessible only read-only on the share. So it should be a kind
of WORM feature (of course it wouldn't on Unix side. But that's OK).
So my thought was: Maybe this could be done through a VFS module (create
time + x days < current time: write access denied)
My questions:
- Is there already a module that does something similar and can be
easily adapted to provide a kind of WORM feature?
- Can this be done by some who is familiar with C? Or should it be
better done by someone who is Samba code experienced?
- Do you guess it's much work? 1, 2, 3 weeks,...?
If the total costs wouldn't be to high and we could realise it, the
final module should of course find back to the Samba project, as it
could be interesting for others, too.
Regards,
<PERSON> |
209740a0-cf72-dec4-a79b-3642db8fddd5 | ['735f1c2b-58a1-c8ac-a46b-4d423ef2bf0d'] | What small tasks can a relatively inexperienced person do to help along the
alpha port of redhat? I got KDE3 to work, and I thought that at the very
least I could test some packages and report where they are breaking.
Is this information useful? I understand that because of the lack of interest
(read: money) behind the alpha port it often lags behind. I don't think this
is really anyone's fault but I would like to help where I can (and as time
permits).
<PERSON> | 0f4b3a43-d281-a72d-c97b-653a09d494a3 | ['735f1c2b-58a1-c8ac-a46b-4d423ef2bf0d'] | Until the problem with the adaptec driver is tracked down (I understand these
things take time), I was considering using a symbios card (I have one lying
around) in the place of the adaptec one.
Are there any known issues with generic ultra-wide symbios chipsets in 3.2?
If not, I"ll proceed to upgrade!
Thanks all,
<PERSON> |
6f9cd0a2-dca6-a30c-50c2-02c98154212b | ['7367912f-c737-7aa1-0796-ba409e6229ba'] | Dear All
Can all of you who responded to this subject line share their mobile phone numbers and the area they live? If most of you are working in Hi-Tech area, we can plan for Lunch in Hitech Area and share the lunch bill.
I think working day will be good start with and then plan for next meetings.
Regards
USMurthy
| bf94178a-93a8-890b-1ed4-8a3eb1720f97 | ['7367912f-c737-7aa1-0796-ba409e6229ba'] | Dear All
I would like to have a Linux System dula bootable with Window XP
To start with I want the following functions in the Linux
- Internet browser
- DVD Writer in Linux
Can somebody suggest a place where I can get the PC and Linux CDs in Hyd?
Thanks
Murthy
|
0d3875c1-2896-dbeb-998a-9978a59b7225 | ['736ef59b-5a77-e5ee-25db-cf4e5f4a698d'] | Hey <PERSON>!
I've found it easy to hook up with a designer, even from Christchurch ;-)
Actually I've always worked with designers half way round the world, even
before
that from Munich. However, that was mostly doing icon designs.
Based on what design work you need your mileage might vary.. discussion
intensive topics like complete UIs could become tricky, but then again there's
always Skype and Co.
Just search the list archives for people recommending designers for various
topics, that's how I've found my guy!
cheers,
Jan
| fc9e2d60-a819-0cee-3a6a-a67569fc1dca | ['736ef59b-5a77-e5ee-25db-cf4e5f4a698d'] | Hi all -
can anybody suggest some good site, directory, online service, etc. where a Mac dev can find an experienced UI designer?
There's sites like 'http://builditwith.me/' and tons of places to find icon designers on the web - but these seem to be swamped with people who want to give it (UI design) a try but don't quite seem to have the experience.. at least that's what it looks like to me :-/
I'm currently looking for somebody re-imagining the UI of my new Cocoa app. It's a rewrite of an existing Carbon app with a healthy user base so I'm quite confident when it comes to the set of features to be included in the new app - however exposing that functionality in a way that won't overwhelm rather casual computer users and just making the UI super-awesome is where I'm in desperate need for somebody who has a designer mind set :-)
Any input appreciated!
Cheers,
<PERSON> |
ec11d111-8949-7114-3480-b7c29c4d2d3e | ['73719722-206e-cf53-ff16-8b992cac34ed'] | Hi all,
I wanted to share this proposal we arrived at after discussion with some AlterMundi hackers, so we can discuss it during our future IRC session. I've previously shared it with <PERSON>, who had some observations that he can better explain, but agreed on this initial assumption which triggered the proposal:
The scenario we see in our networks is that over a certain link quality which is considered acceptable, we want the clients to choose the gw with better bandwidth. So if for example this quality floor is TQ 100, then if a gw has 6Mbit/s advertised b/w and another has 3Mbit/s, the clients that see this gateways with a TQ above 100 will choose the faster one between them.
We observed that in the current implementation, advertized gateway throughput is used to modify the final gw selection by publishing unrealistic bandwidth. The proposal tries to fix this, as well as the "dynamic switching" for selection class 1. Looking at the current code involved we also believe it would allow to make the implementation simpler.
This would be the proposed options:
gw_sel_class [1,2]
1 will consider gw throughput, 2 will only consider TQ. When using selection class 1, clients will switch gateways if one with better throughput becomes available and reachable with a TQ above gw_tq_floor (see below). Defaults to 2.
gw_tq_floor
Only relevant for gw_sel_class 1. Above this TQ floor, the gw with the
best advertised throughput will be chosen.* Defaults to 100(?)
gw_tq_threshold
| bb41ebb4-4466-b78a-a65a-6ec1661efe8c | ['73719722-206e-cf53-ff16-8b992cac34ed'] | I'm curious as to what's the performance of current real world
batman-adv networks. Throughput, latency, concurrency behavior...
We have been conducting some tests on new hardware we plan to use as
replacement for our current recommended node setup and published the
results in our blog:
http://blog.altermundi.net/article/multiple-hop-mesh-performance-with-multi-band-dual/
We will publish a follow-up when these nodes are put to work in the real
networks.
Do others have data on their network performance? I'd also like to know
if anyone has some explicitly designed set of performance tests to share.
cheers,
NicoEchániz |
26c59d79-f214-5860-b9c6-85eef2090355 | ['737952ed-b0e4-fdc9-7d8b-4c7d38ffb07d'] | Hello
I'd like to know if there is a way to get the resulting coordinates
(i.e. x/y position in cm on the page) of a musical event in the compiled
ps/pdf file; what I want to do is to typeset a lilypond source, and
LATER post-process the resulting image. The typical case would be for
example highlighting a specific bar. Note that I do not want to do that
directly in lilypond, but as a completely separate step.
I tried looking at the documentation, but could not find anything.
Any hint?
Thanks in advance
<PERSON>
| f821e8ac-f12c-ca88-d438-dd6d7533248e | ['737952ed-b0e4-fdc9-7d8b-4c7d38ffb07d'] | Hello
I'd like to know if there is a way to get the resulting coordinates (i.e. x/y
position in cm on the page) of a musical event in the compiled ps/pdf file; what
I want to do is to typeset a lilypond source, and LATER post-process the
resulting image. The typical case would be for example highlighting a specific
bar. Note that I do not want to do that directly in lilypond, but as a
completely separate step.
I tried looking at the documentation, but could not find anything.
Any hint?
Thanks in advance
<PERSON>
|
b3e0caad-1568-2b93-52c4-7108f2b4ccf3 | ['73861ba3-95b1-dd5a-6e1f-6745af5cd13d'] | Check on the web interface to the Patton under: Telephony / Call-Router / ISDN Interface IF_PRI_1, then look at the lower section and see if the Accept/Emit button is checked in the Caller Name section. If it is, uncheck it and try the call again.
We had this issue - unfortunately, in our case, either caller-id with name will work, or we can use directed call pickup - in our scenario, they were mutually exclusive with the Patton. We opted for number only...
Nathaniel Watkins
IT Director
Garrett County Government
203 South 4th Street, Room 211
Oakland, MD 21550
Telephone: 301-334-5001
Fax: 301-334-5021
E-mail: <EMAIL_ADDRESS><mailto:<EMAIL_ADDRESS><PERSON>
IT Director
Garrett County Government
203 South 4th Street, Room 211
Oakland, MD 21550
Telephone: <PHONE_NUMBER>
Fax: <PHONE_NUMBER>
E-mail: h+eJ282PLSu8ObZd@example.com<mailto:h+eJ282PLSu8ObZd@example.com>
| 079ef70e-8972-d36a-d639-8e61cbe450c4 | ['73861ba3-95b1-dd5a-6e1f-6745af5cd13d'] | Hoping this is an easy question - I am connecting an older PBX with Sipx via a Patton PRI gateway. I'm pretty sure the call from the older PBX will come with a '6' in front of the 4 digit extension that needs to go to sipXecs (or whatver it's called...)
What is the best way to get rid of the '6' (does this happen at the Patton level or SipX).
<PERSON>
|
96865cdb-6e96-f9dc-18ad-988835b9d240 | ['7387c44b-1337-ccf9-b434-1ad77abff6ca'] | Hi,
I have a feature file with 3 scenarios. The first one is a setup scenario
and if it fails I don't want to run the other two. I can set a variable
like setupPassed = false if the first scenario fails. I can also check this
variable in a Before hook associated with the remaining 2 scenarios. My
question is, how could I force these 2 scenarios to fail immediately
(skipping all the steps) if setupPassed is seen to be false. Oh, and I'm
still on cuke4duke. Thanks.
<PERSON>
| 573665b4-97d9-df38-69a1-58555d878c1b | ['7387c44b-1337-ccf9-b434-1ad77abff6ca'] | Hi,
I'm actually running cuke4duke. Suppose my feature file hierarchy is rooted
at C:\cuke4duke\features
and that I have 2 subdirectories here A and B which contain A.feature and
B.feature and also A.txt and B.txt.
I want to have a step in A.feature that reads something like:
Then extract information from A.txt
The implementation behind this step would then look for A.txt in the same
directory as A.feature. How do I find this directory?
Any suggestions would be greatly appreciated.
Thanks, <PERSON>
|
48eb0cac-f60b-46f5-80d8-0705209be7f6 | ['739a9e87-fac7-c259-e689-4276f3ed9143'] | I have files I move around between several different environments, and in
each environment I have a config file (web.config) that is unique to each
environment.
I make changes on my machine, push them to a repository, move to a
development machine and "pull" them from the repository, then move them to
to a test machine and "pull" them from the repository, etc.
Through some trial and error I figured out that "most of the time" if I put
my web.config into my .gitignore file everything works fine, but not always.
Right now I'm spending my afternoon trying to figure out why whenever I
checkout the "master" branch it deletes my web.config file. It happens like
this. I will have my web.config file in my directory and then checkout a
branch into that directory (e.g. git checkout new-search) and everything is
fine. If I switch to another branch (e.g. git checkout fix-search)
everything is fine as well. But when I check out the master branch (e.g.
git checkout master) it deletes my web.config file.
Why is it deleting my config file and more importantly, how can I get it to
stop doing that so that I can spend my time working on code and not fussing
with git?
Thanks!
| 88b5880e-f56c-6e41-6af9-ecaee05cb764 | ['739a9e87-fac7-c259-e689-4276f3ed9143'] | I swear I'm going to figure out how to work with git at some point and stop
bothering everyone with these stupid beginner problems!
So here's my current problem. I'm trying to "checkout" (not sure if that's
the right term or not) files from my repository into an existing folder (a
folder created as a NetBeans project...something NetBeans recognizes as a
project folder). Clone doesn't work...tells me there is existing content.
When I deleted all the content and cloned it again, it put all the contents
into a sub-folder. So I tried using fetch. When I fetched, it took a very
long time and looked like it was doing work (it was showing me some kind of
progress), but when it got all done the folder was empty. I tried this a
couple of times and then tried pull, which did the same thing. At one point
after a fetch I typed "git status" and got a long message saying all my
files were deleted. Finally after searching around for clues decided to
clone into a separate directory then copy and paste all the contents of
that directory back over to my project directory.
I guess my question is, how do I get my files out of the repository and
into an existing folder on my computer? And why doesn't fetch or pull
actually fetch or pull any files down from the repository?
Thanks,
|
0039e97a-d5ca-08fe-dcee-e3736a954f9b | ['73a393f9-639b-ee2c-3742-03757588400d'] | Hi <PERSON>,
we need to troubleshoot a problem around excessive dynamic sampling and library cache locks at my client.
Probably we are facing a bug 14542720.
For the moment I'm looking for a way of how to match the sql_id's that have used dynamic sampling during parse
to the sql_ids of the dynamic sampling itself (those queries with the SELECT /* OPT_DYN_SAMP */ ...)
any idea?
<PERSON> | 87ee6caf-634e-003a-3409-b6557b6a7a91 | ['73a393f9-639b-ee2c-3742-03757588400d'] | Hi there,
i wonder if anybody of you already made some experiences with DBMoto from
Hit SW http://www.hitsw.com/products_services/dbmoto.html
It promises data replication between a lot of database systems
including oracle, db2, sql server. And it should be based on grabbing the logs
not on triggers? Can anybody confirm this? Experiences?
Actually Goldengate was my first idea for an upcoming project but
it does not support DB2 on i / AS400 ...
Thank you for any of your comments and thoughts..
<PERSON> |
61990ea4-10ab-bc41-21e4-adc953cdc1a7 | ['73baab60-8818-0f8f-14b5-58a6ccaa6694'] | The insert-at-cursor signal in GtkTextView doesn't seem to be emitted.
There are a few references to a similar problem with GtkEntry, for which
using a signal for the GtkEditable ancestor worked around, but I cannot
see any suggestions for GtkTextView. Anybody know what is going on?
The key-press-event is working fine, but that leaves you with a lot more
work since you have to decide whether a character would be inserted...
<PERSON> | 2acb4854-683c-4451-cbd3-ee5d1a674e1d | ['73baab60-8818-0f8f-14b5-58a6ccaa6694'] | I am displaying a structure that seems to be circular, and it shows
(x_LIST . #-3#)
x_LIST is my own symbol but what is the meaning of the other member of the pair?
And how does one track it down this notation in the guile manual? I am
sure I came across this before, but cannot locate it in the manual
(again?)
<PERSON> |
08bf5834-943c-63cc-3bc1-2b8f1674d704 | ['73c34315-a0aa-c795-3d41-376beadbf6b6'] | Hi,
I'm trying to use OpenSSL RSA routines into my custom GINA for authentication with USB token in Microsoft Windows XP.
I've used the RSA_new() method in a normal application and it works very well (Encrypt a random challenge with OpenSSL RSA public key and then Decrypt with PKCS11 hardware method to retrieve the response).
But when I use the code in my GINA, the system crashs. I've discovered that the problem is the RSA_new() function.
Probably, there is a problem in memory allocation...
Is it possible to allocate the memory for a RSA key in other ways??
Thanks | 0554026d-ec42-35b5-65c7-61a00884f681 | ['73c34315-a0aa-c795-3d41-376beadbf6b6'] | Hi,
I\'d like to export a public key (generated on-board with a USB
crypto-token)
to verify a sign with OpenSSL functions. (I don\'t use certificate)
The data are signed with USB token private key and I must verify the sign on
the computer (it\'s a challenge-response authentication).
I have the Modulus & Public Exponent on a file (user database on the host)
and they\'re in HEX base (byte..no PEM or DER format).
Here an example of public key modulus on the token:
äkò(jÖ o0¡áÛÑ↔Ü[ÉGþ╦;▀$█☼7á|Ú¯¿/¦9¾Û¿úÍúÕ
¿ÁÌ►╣├│┼Im*©:ÿcÍý¶wW
Can I export the public key from the token in DER/PEM format and use it
with OpenSSL code
Verify function? Must I use the OpenSC pkcs11 engine to export the key?
It seems impossible to use a public key generated with an hardware token...
Thanks
|
2798942b-15b6-0b27-9759-48d113a1fa60 | ['73d7abb9-26af-b150-75e2-b980c55b4a29'] | Hi,
I successfully managed to work with ceph jewel. Want to try luminous.
I also set experimental bluestore while creating osds. Problem is, I have
20x3TB hdd in two nodes and i would expect 55TB usable (as on jewel) on
luminous but i see 200GB. Ceph thinks I have only 200GB space available in
total. I see all osds are up and in.
20 osd up; 20 osd in. 0 down.
Ceph -s shows HEALTH_OK. I have only one monitor and one mds. (1/1/1) and it
is up:active.
ceph osd tree gave me all OSDs in nodes are up and results are 1.0000... I
checked via df -h but all disks ahows 2.7TB. Basically something is wrong.
Same settings and followed schema on jewel is successful except luminous.
What might it be?
What do you need to know to solve this problem? Why ceph thinks I have 200GB
space only?
Thanks,
| ffea3930-2d20-3eb0-81af-48d91e69f26e | ['73d7abb9-26af-b150-75e2-b980c55b4a29'] | I already tried 512MB but re-try again now and results are the same. Both without tuning;
Stripe 2 replica 2: dd performs 250~ mb/s but shard gives 77mb.
I attached two logs (shard and stripe logs)
Note: I also noticed that you said “order”. Do you mean when we create via volume set we have to make an order for bricks? I thought gluster handles (and do the math) itself.
Gencer
|
480c18cb-c6b9-9d30-1a95-9b39262c2b8e | ['73d7dc3d-6820-9628-f651-35764675d65f'] | Since I installed Observium a few weeks ago I have noticed that the
overview timestamp for syslog and event log delta is wrong for most (not
all) devices, this is prevalent on the main overview and the specific
device overview page. By wrong I mean "-1y -1d -xh." etc..
Hovering over the incorrect entry pops up the correct timestamp so I think
it's just the logic in calculating the delta between current time and that
event that's broken, also viewing the syslogs and entry logs directly with
the original timestamp is fine as well. I checked my local timezone and php
timezone just to be sure and they are reporting correctly, I would imagine
this would break the original timestamp anyway which is not the problem
here.
Hover over example:
I find it particularly weird as not every device is doing this, the ones
that are have no real pattern either, mixture of N7K,N5K,ASA etc and some
even have a mixture of working and broken entries (individual device
overview):
[image: Inline image 1]
Anyone else seen this, any ideas?
Cheers.
| 5fbf0076-4afd-756d-3c29-e8ac1359543b | ['73d7dc3d-6820-9628-f651-35764675d65f'] | Hi,
I have quite a few alerts stuck in the "Other" state, any idea on what
causes this?
I had this problem a while ago and then it seemed to disappear after
creating a few more alerts and several svn updates so i'm not really sure
what cleared it but after adding more devices it's back again. I've tried
regenerating the alerts a few times but that makes no difference, is there
some debug that can be turned on to work out what is going on here?
It's across multiple alert checkers and even something as simple as "device
status eq 0" for a device down check has multiple alerts stuck in "other"
state. Hovering over the alert shows an empty status:
[image: Inline image 2]
[image: Inline image 1]
Cheers
|
5bd3ca72-9a34-89e3-0ac1-5aa992a404ac | ['73e591da-1aee-8d22-a028-0a64728a3148'] | Issue #2669 has been updated by <PERSON>.
Regarding the ums0 issues: How was ums0 doing with the LATEST-ISO? It seemed to fail with 3.6.2 but I didn't see any mention of it in your report using LATEST-ISO. If it works better with the latest ISO, it could well be that it is related to the mouse not working with our old USB stack but working better with our new USB stack (which is default for LATEST but not for RELEASE).
Regarding the keyboard issues: If it is atkbd(4) (i.e., PS/2), then it could be related to the removal the KB_CONF_FAIL_IF_NO_KBD flag from our kernel configs in master (it's still present in RELEASE). It was originally done to fix issues seen on Lenovo laptops but it might well be happening on other machines.
Regarding booting off mfi(4): You said you could install to it, so that's a good thing. :) The message you quoted seems to be from your box itself and not from DragonFly. Are you sure there is nothing else to care about for booting from it? I remember some RAID controllers have some option in their BIOS to enable booting. I'm not sure what's the cause here.
<PERSON> | ae7802e0-fb1d-4aec-ebd1-5a2f129a6549 | ['73e591da-1aee-8d22-a028-0a64728a3148'] | Issue #2828 has been updated by <PERSON>.
Hmm. Did you measure the actual power consumption? The default (2) is supposed to be a hybrid approach that falls into ACPI. But looking at it, it is quite possible that the code in kern/lwkt_thread.c that is responsible for resetting gd_idle_repeat to 0 is breaking the concept.
-Matt
|
bd73ff8b-6b82-67c9-3819-67e80c83c4cc | ['7400d89a-66be-3cdd-bd84-0928efdae52c'] | ""hey everyone, i need some big time help, and fast! one of my songs has
some how become corrupt, and
i can't figure out what to do to get it back. when i ever i try and open it
up and causes logic to
freeze up and die. all my other files seem to be working correct. i even
went so far as to reinstall
my entire system, and that didn't work! (i didn't do the whole system just
because of that, it's been
acting very unstable lately, this was the final excuse). so please please
please help me. thanks in
advance.
-mark""
| 198f5511-df5b-8bd3-e0a8-22c1b91eb118 | ['7400d89a-66be-3cdd-bd84-0928efdae52c'] | I know this is a bit of a dumb question...but how do I buy an unlock key for
the logic instruments?
My demos expired and I want the exs24 and es2, but emagic doesn't really say
how to buy them.
I've ssen them at audiomidi.com, but they look like they are vst versions.
Anyone know the best, fastest way? I'd prefer to get an unlock code, with
the physical materials shipping later.
Thanks,
<PERSON> |
b3c55654-3745-e8a7-01a0-ae9def63d899 | ['7402cc2a-d41b-1286-06fa-8588c3df9b1e'] | Hi
I'm trying to set up courier on a FreeBSD based virtual dedicated server
(from johncompanies). Installation and configuration has gone well but I
can't seem to get esmtpd to run. All the other services (including
authmysql, pop3d, pop3d-ssl, imapd and imapd-ssl) are running perfectly
on the machine. The tests that I have run on the system worked
perfectly. Mail is delivered locally and to virtual users from mysql
without a hitch.
From looking at my logs I see no reference to esmptd being started (from
either my startup script or from the esmptd script itself). When run
(certainly with "emstpd start") then nothing appears to run and nothing
is returned. I have tried esmtpd-ssl to check if the same thing happens.
Something slightly different happens here with "ll_daemon_resetio:
Permission denied" being returned when run.
I wonder if the fact that the virtual dedicated server has no loopback
address has any relation to this. Something else I have noticed, if I
use webadmin to configure any of the esmtpd conf then the esmtpd config
are made into 0 byte files (all contents deleted). I have since stopped
using webadmin ;)
Any ideas on how to get esmtpd working? Any help would be greatly
appreciated (this one of those. had to be done yesterday kind of
projects)
<PERSON> | 23d3f0a0-03d4-3eda-fda8-6844d31fedcf | ['7402cc2a-d41b-1286-06fa-8588c3df9b1e'] | Hi
I'm trying to set up courier on a FreeBSD based virtual dedicated server
(from johncompanies). Installation and configuration has gone well but I
can't seem to get esmtpd to run. All the other services (including
authmysql, pop3d, pop3d-ssl, imapd and imapd-ssl) are running perfectly
on the machine. The tests that I have run on the system worked
perfectly. Mail is delivered locally and to virtual users from mysql
without a hitch.
From looking at my logs I see no reference to esmptd being started (from
either my startup script or from the esmptd script itself). When run
(certainly with "emstpd start") then nothing appears to run and nothing
is returned. I have tried esmtpd-ssl to check if the same thing happens.
Something slightly different happens here with "ll_daemon_resetio:
Permission denied" being returned when run.
I wonder if the fact that the virtual dedicated server has no loopback
address has any relation to this. Something else I have noticed, if I
use webadmin to configure any of the esmtpd conf then the esmtpd config
are made into 0 byte files (all contents delted). I have since stopped
using webadmin ;)
Any ideas on how to get esmtpd working? Any help would be greatly
appreciated (this one of those. had to be done yesterday kind of
projects)
<PERSON> |
477b39f6-d685-f09b-9772-d124912ebff0 | ['7415b05f-6aa5-4f30-c427-479890841b07'] | I had gotten the DDR core to work in simulation, and it synthesized,
but did not work in hardware. Has anyone been able to get it working
on a real board?
I'm using an Avnet Virtex-II evaluation board, with an XCV2P7ff896,
speed grade 5. It has two Micron mt46v16m16 chips. I haven't been
able to figure out what the problem was (and have since moved on to
trying to get some of the controllers from Xilinx working). Does
anyone have any suggestions on what it might take to get the core from
OpenCores working?
Thanks,
<PERSON> | 574ef423-9461-dd30-384f-7c1cc3ad2eb8 | ['7415b05f-6aa5-4f30-c427-479890841b07'] | Hi,
I have a Radio Shack USB to serial adapter, but I can't get the driver to
show up for the device in /proc/bus/usb/devices, and I saw a warning
message at least once in /var/log/syslog that said something like the
driver dosen't know what it is. I did try using insmod specifing the
vendor and product id's, but that didn't help as well. Does anyone have
any suggestions, or should I return this and buy one that's supported?
thanks,
<PERSON> |
c252a45e-d891-71e6-a2a1-36f1a4cd529a | ['7420f44e-fe75-7326-f0ec-4950d161d44b'] | Hi,
Given a certain data.frame, the lattice xyplot function will plot the
data as.is and join the data point in the order of the data frame. It is
my (probably flawed) understanding that, using the same data frame,
ggplot orders the data by increasing order of the x-axis variable. Can
one control this behavior?
Thanks
<PERSON>
| 338d223b-a5c9-c68f-eb77-4d88104a3874 | ['7420f44e-fe75-7326-f0ec-4950d161d44b'] | Dear R-users,
[Sorry to have to send this again but my former posts were trimmed
because of some embedded HTML code]
readline is a nice function to get user input when one is working with
the terminal. This function does not allow any interaction with the user
when a script in batch mode (I guess by design)... Unfortunately, that
is exactly what I am trying to do.
I would greatly appreciate if any R-user could:
- tell me if this is possible
- indicate a function I could look at, that would allow this type of
interaction during a batch run.
Thank you in advance,
|
d8b956a6-1957-7f4f-9ac5-b52ae284e6ef | ['7430a7db-e9e3-dfd8-7081-facefab1632e'] | Hello,
Is there a way to build omniORB entirely in its own namespace? For
example, references to CORBA::string_free would have to be qualified as
omniOrb::CORBA::string_free, etc.
The problem we are dealing with is that a single process happens to use
two CORBA implementations. We are providing a shard library that uses
omniORB, but the process we are linked with at run-time uses Visibroker.
If we could isolate ourselves from the global CORBA namespace, we would
be able to make it work, I think.
Thanks for any suggestions,
<PERSON><IP_ADDRESS>string_free would have to be qualified as
omniOrb<IP_ADDRESS>CORBA<IP_ADDRESS>string_free, etc.
The problem we are dealing with is that a single process happens to use
two CORBA implementations. We are providing a shard library that uses
omniORB, but the process we are linked with at run-time uses Visibroker.
If we could isolate ourselves from the global CORBA namespace, we would
be able to make it work, I think.
Thanks for any suggestions,
Slawomir
| 47978e97-0332-6820-d9aa-a19adab375cb | ['7430a7db-e9e3-dfd8-7081-facefab1632e'] | Hello,
Why does the value of JSESSIONID cookie change per requested
portType within a WebService?
I used the same Service instance, with the maintainSession
property set to true, when obtaining different port stubs via
the Locator and still calls on those stubs had a different
JSESSIONID sent out.
JSESSIONID value was the same when making calls on the same
portType only.
Any idea?
<PERSON>
|
ff688274-5d30-f0d1-bec8-23b323cc9fa7 | ['74359334-9bbf-5bfd-3c60-0f1ff7a2bc2c'] | Hi,
I just want to say thank everyone who contributed to this project for
all the hard work that was put into making this powerful system work.
I have already figured out and implemented a simple windows XP/2000
PXE-based installation with partially auto-generated unattend.txt, and
I look forward to configuring various post-install scripts to install
various software and hotfixes. I thought that the declarative
structure of config.pl was pretty clever until I learned about how
todo.pl works. Then I was really blown away! Another one of the
coolest things I like about this Unattended system is how dosemu is
used to run Windows Setup inside a Linux OS!
Thanks.
| 5d9560ee-c00d-5a73-bccc-912e5dff0447 | ['74359334-9bbf-5bfd-3c60-0f1ff7a2bc2c'] | I understand that you guys are busy with GSoC as it is, plus you are
already volunteering your time to this project, and I would like to
ask whoever can to please comment on my Google application for this
testing project. I asked a question about the "abstract" text/section
and I would like to get feedback on it as soon as possible. That way I
can edit it before the deadline after which the abstract becomes
read-only.
Thanks for your time.
|
1d02b81b-a0e0-2aa7-7b61-33eb800f7133 | ['7438cdb3-2b7f-0227-d66d-60ef2dc8405c'] | Hi
I've installed version 2.2 without problems.
Then, I ve changed line
Listen 80
to Listen 8080
and restarted the server.
Nevertheless, my server is still responding on port 80 !!
This is not the first time I install apache on windows or Linux but is the
first time on Vista and I did it as administrator.
Which is the way to change the configuritaion on vista ?
Thanks in advance
J
| 74bbad26-79cd-27d2-911a-66fb65fc6ca2 | ['7438cdb3-2b7f-0227-d66d-60ef2dc8405c'] | Hi
After a big fight with proxies I finally got maven working.
I've generated a sample project and an eclipse project without problems.
When I tried to get an ant build I realized the generated build file
ignored my local repository and put a lot of lines trying to download
jar and install them in my profile folder.
I wonder why maven did it and how could I fix it ?
Also, I'm curious about how to customize goals. I've seen pom files
from some software I've downloaded to use it as examples but all of
them use jellu scripts and are made with maven 1. I've been looking
maven 2 docs but I couldn´t found a I guide step by step to customize
goals or use ant tasks as they did in maven 1.
Could somebody give me a clue in this matter ?
Thanks in advance
J |
ab3ef331-a7ca-56ef-4b21-f4f26caa5e58 | ['74511451-6989-c377-a926-df9fecba8339'] | Thanks for all the advice, I learned a lot today about making
presumptions when writing my scripts. He was running my script from the
Windows desktop, which was why the window was closing after he finished
entering the search query. Once he started the script from the DOS
prompt, it worked great. Thanks for pointing that out to me. I won't
make that rookie mistake again. :)
<PERSON>
| c096ef6f-8cb9-3007-fe2e-43ff41a55221 | ['74511451-6989-c377-a926-df9fecba8339'] | I've written a python script for a fantasy hockey league I'm in, and I'm
using Python 2.1 on Mandrake Linux 8.2. I have a friend who downloaded
Python 2.2 on his Windows 2000 and when he runs the script, it 'crashes'
after he enters a search criteria and hits enter. The whole screen
disappears. I'm hoping if I include the section of code I have
questions about that someone could offer advice. The following script is
stored in the same directory as two text files, "fwfha_skaters.txt" and
"fwfha_goalies.txt" and the script works well on my computer. If
someone could tell my why it isn't working on Windows I'd be grateful.
<PERSON>
|
dda5dc8c-6eca-9808-ce8d-98b26d646749 | ['74546d01-a587-ccbe-c2e5-817e38529dbc'] | I have installed octave on an old Linux system that I cannot upgrade. I
tried installing 3.03 but the gnuplot version was not recent enough so I
went to 2.1.73. This version compiles but I need to specify the
--disable-readline flag in ./configure. Once at the octave command line, a
minor inconvenience is that Up-Arrow does not recall the most recent
commands. Instead it displays [[^A. I was wondering if this is a readline
issue? Is there a way to modify the key bindings in an Octave setup file?
Thanks for your help!
| 77d82e7d-82b6-42cf-ff20-2bc70b504279 | ['74546d01-a587-ccbe-c2e5-817e38529dbc'] | Hi all,
I would like to carry octave around with me on a USB stick. I've
compiled octave 3.05 (on Suse 11.4) and it runs fine on that machine.
The problem I am having is how to copy it over to a USB stick. I
copied the entire build directory and launch with ./run-octave.
I type ldd octave to see the dependencies and copy all the .so files
into a directory on the USB stick "lib" inside the build directory
then add this to LD_LIBRARY_PATH.
However when I try to ./run-octave from the USB stick on a new
machine, it doesn't work. It looks like maybe octave is looking for
the .so files using absolute paths for the original machine rather
than in my "lib" directory.
I apologize if this message is unclear/naive - I know I'm hacking
without a clue and really need some "big picture" help how to
accomplish what I want. Thanks.
<PERSON> |
ff194594-e839-f85a-f0ba-62ee2d9c1086 | ['7461e39c-0f8a-a5d6-9bbc-5b851d8fd12c'] | Hey all,
I'd like to phase out cwiid [1] (as it's super old and unmaintained).
Currently it's an optdepends for supercollider, but its use will be
deprecated in the next version [2], so I will remove it there
eventually.
The only other packages depending on it are kodi (make) and
kodi-eventclients.
However, is it really required there?
For now, I rebuilt cwiid, but I guess it could be dropped midterm, if
it's not a hard requirement for kodi et al.
Best,
<PERSON> | 12d91acc-8005-4e4b-1398-dd1c10616b11 | ['7461e39c-0f8a-a5d6-9bbc-5b851d8fd12c'] | Hey all,
I was wondering, what the current upstream status is.
I'd like to package the latest tuna for Arch Linux, but I'm a little
confused about the current state of things:
While in the past I used [1], it seems that <PERSON> repository [2]
is currently holding the latest updates.
Will the latter eventually be merged into the former, or is it time for
me to switch?
Additionally: What can I use as the current upstream website for tuna?
Currently I have: https://rt.wiki.kernel.org/index.php/Tuna
Best,
<PERSON> |
bffb942a-7591-b047-becf-c2f67e3ff77a | ['7465d810-50bc-992a-752b-8170e6602650'] | Hi,
We are new to packetfence and trying to setup packetfence with Meraki
access points. Stuck at the point where to add switch on Packetfence config
to add the new AP. How do I go about this setup, what is the IP of Meraki
cloud controller and do I need more setting to get this working. Anybody
tried Meraki with packetfence?
Kind Regards,
| d30d6f63-5f5c-eb8d-281a-020734eeb6cd | ['7465d810-50bc-992a-752b-8170e6602650'] | Hello,
I am setting up PF Zen 5.5 in inline mode. I have configured two interfaces
one management and another inline. Everything is working OK except that I
do not get a registration portal on the inline network. The default gateway
is the IP for PF inline interface, I can ping it but even if I input the IP
on a browser I don't get a Portal/Registration page. Users on the LAN are
not prompted to register when they get an DHcp lease from PF.
Any ideas will be appreciated.
Thanks and Kind Regards,
<PERSON> . S
|
17d1c316-6efd-b6b2-e644-45cf3eb6047b | ['74799ebb-7031-8603-322a-d0eb87b30a6b'] | I was trying to run the Wizard tutorial and the instructions produce code that doesn't compile. It appears to be due to differences between version 7.1 which was used in the tutorial and 7.3 which I have. Can I install version 7.1 and still have version 7.3? Is version 7.1 available?
Thanks | 38282cfc-8313-adb7-c9a9-1221456dcfb5 | ['74799ebb-7031-8603-322a-d0eb87b30a6b'] | I frequently see reference to layer.xml files when I lookup tips on how to organize my netbeans platform application. The latest version of Netbeans doesn't have that. Is there a way to accomplish what one used to do with the layer file? Specifically I want to control the startup window sizes of TopComponents.
Thanks |
284bad9d-247c-4a5f-212d-19678230e00d | ['74879d9c-3405-714d-3291-77c861dbbdfb'] | It is true that Jews and Christians share only the Hebrew Bible (Old
Testament). However, I would suggest that if you look through the BS
class, you will find texts of the Bible edited and published by both
Jewish and Christians, as well as Jewish and Christian commentaries on
these books. I think it would be unfortunate for readers of any
persuasion not to have editions and commentaries on the same literature in
the same place. It should remind us all of our common heritage.
Just to give you an example: the Biblia Hebraica Stuttgartensia is the
standard text of the Hebrew Bible (slowly being replaced by the Biblia
Hebraica Quinta, a joint project of the German Bible Society and the
Hebrew University of Jerusalem). The Stuttgartensia was based on the work
of <PERSON>, a German Protestant, who based his work primarily on the
St. Petersburg Codex (or Leningrad Codex, if you must). The St.
Petersburg and Aleppo codices were both copied by Karaite Jews, and not
accepted by orthodox Jews as authoritative until <PERSON> declared
them to be the most accurate manuscripts of the Bible.
Michael S. Borries
| efd3d2cc-9343-322b-9abd-94f8566a65db | ['74879d9c-3405-714d-3291-77c861dbbdfb'] | For those with the title Shui hu zhuan (translated into English under the
titles Outlaws of the marsh, All men are brothers, <PERSON>, and
Marshes of Mount Liang), LC has added the two authors to whom this work is
sometimes ascribed (<PERSON>, <PERSON> and <PERSON>). The correction should
appear in OCLC in the next few days. You may want to check you authority
records and your bib records.
<PERSON>
|
df58ac5c-1a78-e55c-7da7-8faad59a7a6e | ['7493753d-67dc-9888-38be-919a5bb2d943'] | Hi,
Have 2.1.5 with 2.6.9 (Debian kernel-source) and hibernate package
0.99-1 from cp.yi.org
Hibernate doesn't run :
Your kernel does not have any recent Software Suspend 2 support compiled in.
Please follow the HOWTO linked from http://swsusp.sourceforge.net/ for
instructions on how compile Software Suspend into your kernel.
hibernate: Aborting.
Have all as module, as in the wiki here:
http://swsusp.sourceforge.net/wiki/DebianInitrd
What can I do?
Thanks
<PERSON> | 27f6b299-354b-f09c-ba12-101da8dab597 | ['7493753d-67dc-9888-38be-919a5bb2d943'] | Hello everybody,
Hi run Tomcat 4.1.18 Standalone on Linux debian(sid)and use J2sdk1.4.1_01.
I have a servlet which manipulate exif informations in jpg images.
it works well but when I try to run it on tomcat it fails (error 500)
I have the following error:
"Can't connect to X11 window server using '0:0' as the value of the display"
What can I do?
<PERSON> |
a9b6efb7-9e12-628f-4b81-10301695675e | ['7495048c-aa95-0911-35c7-06b35d673593'] | This patch adds the possibility to enter the prompt at boot time by
typing the sequence of buttons "123" rather than a single button:
http://www.freebsd.org/~attilio/Sandvine/STABLE_8/boot123/boot123.diff
This is useful in the cases where a serial console is likely going to
be used which can carry on spourious character, leding to the prompt
erroneously.
This option is wrappered into the BOOT_PROMPT_123 option, in order to
maintain the current POLA.
This patch has been contributed back by Sandvine Incorporated.
Please review.
Thanks,
<PERSON> | 5c2f802f-abd5-ad93-110c-2e837e29ec87 | ['7495048c-aa95-0911-35c7-06b35d673593'] | Hi all,
first of all, I have to say I'm not too much familiar with
DragonFlyBSD kernel (so I have no idea if this is completely new for
the project), but I contributed some for FreeBSD kernel and I'm rather
experienced with IA32.
What I would like to work on are some x86 'improvements' to DFLY
kernel. For improvements I mean adding different versions of some
critical functions for different versions of the CPU (i.e.: in P4 you
could just implement an atomic memory barrier using mov + *fence
instructions against using xchg or stubs like these).
First of all, a good discussion point is about optimizations
activation: do you prefer compile-times stubs or run-time patching?
Actually, I'm planning to improve and add a run-time patching concept
inherited from Linux to FreeBSD, and maybe you would be interested in
its port. Run-time patching is very useful, but on the other side it
faces a lot of problems (inlined functions can't fit and it deals with
dimensions problems), if we don't want to loose performance matters.
Compile time stubs are simpler and possibly quicker but the kernel
needs re-compilation in order to get benefits, that is not so nice.
So, some feedbacks about this is very appreciated.
Once that we have choicen a method in order to apply changes, the
first thing I would like to add (BTW, I don't know if it exists
alredy) is sysenter/sysexit support replacing interrupt 0x80 (I have
an item in the FreeBSD list for volounteers about it, since I think I
would like to add it there too) and possibly evaluating the usage of
FPU/MMX/XMM registers in order to gather syscalls parameters.
Feedbacks about it are very appreciated too. I have other ongoing
projects we can discuss later, but they are very architecture
dependant so, since maybe they can fit in FreeBSD, they cannot in
DragonflyBSD.
Thanks for your time,
<PERSON> |
395667df-0e5a-adba-fd14-d13096909bae | ['749c496b-fdd7-be6e-63d5-17edc279c84d'] | A week ago, my car refused to start in a parking lot after a routine drive to the store. Now it won't start or charge. I am stumped and could use some advice where to start looking.
The drivetrain is an Azure Dynamics DMOC445 with the AC50 motor. It has run perfectly for two years. The wiring and relays are from Electro Automotive and are wired according to the book. Nothing obvious has disconnected.
There is pack voltage at the controller (267v - should be 288v, but 267 should be enough). My cells all seem to be OK (I can't reach every one, but the individual banks sum up to the expected amounts). The key switch relay clicks when the key is turned, but the DMOC contactor does not. I tried to connect my laptop to the DMOC, but nothing was received - it seems to not be turning on at all. When I run the charger, it cycles on and off (the fans spin, but I hear a relay clicking every 2-5 seconds and no charge is being applied to the pack). It never did this before.
I am suspecting a 12v issue, possibly one of the charging relays, but not sure how to figure it out or where to start. I am learning why non-engineers should not do conversions L
Has anyone else seen an issue like this, or can give me some ideas of how to start troubleshooting it? If anyone in the Portland area is willing to help in person, there would be beer and/or cash involved.
Thanks!
<PERSON>
| 6fd81dde-2829-19ae-0495-3e8d78f39f4f | ['749c496b-fdd7-be6e-63d5-17edc279c84d'] | I used a heater from CANEV.com in my car and it was very simple to install.
This heater already has an over-temp switch built-in, and comes with a
contactor and fan switch. A very simplified wiring diagram can be seen
here:
http://www.evsource.com/datasheets/heaters/heater_wiring_diagram_v2.pdf
CANEV has a more detailed diagram, but you'll need to email them to request
it.
The fan switch is connected to the 12V+ side of the contactor with a one-way
diode (29 cents at Radio Shack) on each of the fan switch posts for
Hi-Med-Low. So you have three 12V+ wires from your switch, each passing
though a diode, then all three connected to the contactor 12V+. 12V- on
switch and contactor go to chassis ground. Otherwise it is like the
drawing.
The heater comes on whenever the fan is switched on. And you can't have
heat without the fan, so fire risk is almost nil.
If you don't have an over-temp switch in your heater, you can install one
between the three combined leads from the fan switch diodes and the 12V+ on
the contactor. Over-temp then shuts off the contactor. Another one-way
diode is needed between the fan switch leads and the over-temp switch.
Good luck!
<PERSON>
Electric '73 VW Fastback
Portland, OR |
c80fb4dd-6a84-6158-284b-7f81e50e66a8 | ['749c8193-94ed-0a40-3c33-b177aeea72c2'] | Hi,
I tried posting before but something must've gone wrong as I can't find a
trace of it on the list.
Recently I started using Leo on a documentation project so I'm mainly using
Leo to edit and get rst output.
A bit of context. I am using rst mostly because it's easy to write and it
provides structured output in the form of XML. I convert this XML to DITA
which is a topic-based documentation format. I chose Leo for it's outlining
capabilities and cloned nodes. The outline structure in .leo files can be
converted to DITA maps (kinda like table of contents) and the node content
to DITA topics which are pulled into DITA maps to form a manual.
What I am looking for is a way to get the unique ids that Leo gives each
node and use them to link topics to maps. This fits in very well with topic
re-use as cloned nodes all have the same id. I looked a bit at the rst3
plugin code but I wondered if there is a way to get the node ids into the
rst output without patching this plugin code.
I am curious if there is an easy way to do this.
Thanks
<PERSON>
| 3dfa12a8-92e5-470e-3e6d-80b8bdbe4584 | ['749c8193-94ed-0a40-3c33-b177aeea72c2'] | Hi,
Recently I started using Leo for a documentation project at work. I mainly
use it to generate rst files (using @rst nodes and rst3). These files are
converted to XML and then to DITA which is a topic based documentation
format.
DITA based documentation consists of small re-usable topics which are
linked together via so-called maps (similar to a TOC). Re-use of topics is
perfectly captured via cloned nodes. Great. But I somehow need to be able
to give each topic a unique id. The id's that Leo generates inside the .leo
files are just fine for this purpose. However I need to get these id's (I
mean the v element's t attribute) into the generated rst file (I think this
is like sentinels?).
This way I can generate the DITA maps from the .leo XML file and reference
the individual topics based on the same id. I can then write out a big
collection of topics as an rst document and use other @rst nodes to capture
the hierarchy of topics containing only cloned nodes.
It doesn't matter much if these id's are written out as rst fields or
comments under the heading as long as it's in the generated file.
Is there a way to achieve this in <PERSON> without coding or patching the rst3
plugin?
Thanks a lot.
Cheers,
<PERSON> |
0c2a24ad-619d-66bf-bdcc-c7051897ef76 | ['74a46c09-a3dc-db75-8ba7-a3f4574fbf11'] | Hi everyone,
As part of the Groovy planning process, I've posted a review for multi-echo
laser rangefinders:
http://www.ros.org/wiki/sensor_msgs/Reviews/2012-08-01_MultiEchoLaser_API_review_API_Review
This review focuses on the ROS API for scanners that return multiple ranges
(and intensities) for a single beam. Example lasers with this feature are
the SICK LMS151 and LMS511, as well as the Hokuyo UTM-30LX-EW.
Please take a look and contribute your thoughts.
Thanks,
- <PERSON> | a1892a63-0a43-2dcc-9e8a-599a5f4bd878 | ['74a46c09-a3dc-db75-8ba7-a3f4574fbf11'] | Hi ROS Community,
I've been working on a driver that connects the sensors in Android devices
to the ROS environment. At this time, it only publishes
sensor_msgs/NavSatFix messages, but I will soon introduce sensor_msgs/Imu
and sensor_msgs/Image to publish data from accelerometers, gyroscopes,
magnetometers, and front/rear cameras.
To get more information and to install, please see the following:
http://www.ros.org/wiki/android_sensors_driver/
https://market.android.com/details?id=org.ros.android.sensors_driver
To file bugs, request features, view source, or contribute UI, translation,
or other improvements, please see the Google Code project:
http://code.google.com/p/android-sensors-driver/
I hope everyone finds this useful and I look forward to hearing your
feedback and seeing cool uses for Android devices in robotics.
Thanks,
- <PERSON> |
a5bfe4e4-776e-368e-bb66-eeeef48825ac | ['74b20919-1950-f5ae-2902-b2b3abdc126b'] | Hello -
I've noticed that converting a PICT to TIFF (grayscale 16-bit with alpha
channel) in IM 5.4.7 - RH 9.0 is resulting in a change in the computed
signature obtained with "identify -verbose". The command I am using is
"convert foo.pict foo.tiff" - no options are involved. The reason I am
concerned is that this apparently means the underlying image data is being
modified, and that the exact same command in a previous version did *not*
result in a signature change. Is there a way to prevent this?
BTW - if I remove the alpha channel first - "convert foo.pict +matte
foo_nomatte.pict" followed by "convert foo_nomatte.pict foo_nomatte.tif" -
the signature does not change between the two "nomatte" images. Anyone know
how to force preservation of the data or prevent mucking around when the
alpha channel is present?
Pls forgive me if this topic has been discussed - I didn't find it in the
archives.
Best regards -
Ken
| 129d3689-0734-0cae-e557-38be57f0cabd | ['74b20919-1950-f5ae-2902-b2b3abdc126b'] | Hello -
I upgraded my version of IM to the latest downloadable on the
imagemagick.org website - however converting PICTS to TIFFs with an
alpha or opacity channel is still failing. When viewed in display, only
one half of the original image remains in the TIFF - the other half is
black.
Again the problem doesn't occur if I remove the alpha channel first. The
conversion goes fine and the image signatures are preserved - which
isn't the case when I perform the conversion described above.
It appears that the presence of an opacity or alpha channel is
corrupting the conversion of PICTs to TIFFs in IM. This did not occur in
the pre-IM 5.4.7 version. Has anyone else experienced this problem? Is
it a bug and is there a workaround?
Pls forgive me if this topic has been discussed - I didn't find it in
the archives.
Best regards -
<PERSON> |
7f6ae547-25a8-d314-bd54-5b0a58e927fc | ['74b6b66f-e005-71d9-dcd9-6775d46aabe5'] | Hi all,
(please feel free to copy this mail to your local list)
Version 406 (June 2004) of the MozOO.org CD has been released. The
MozOO.org CD contains a simple set of productivity tools for Windows.
The CD ISO image is small enough (110MB) to fit on a MiniCD.
The main components Mozilla and OpenOffice.org, with language files and
other stuff for both, and a nice installer that takes care of all
necessary things in the proper order. It's very simple, targeted.
The primary target audience is home and small office users. Non-geeks.
They won't move to Linux yet, but this offers an easy and appropriate
first step with the tools everybody uses.
It allows people to not worry about upgrading their office suite, having
a good browser, a bayesian spamfilter built into their email program,
etc...
Plus it's an excellent answer to the "open source is losing the desktop
battle" nonsense that's appeared in the press ;-)
For more info as well as the ISO image, see http://mozoo.org/.
Yes it would be excellent to see this CD with a little booklet available
at news agents around the country and beyond!
Feel free to talk to me about this: <EMAIL_ADDRESS>
(you can also ask me for my phone# there, if you don't already have it -
if you do, feel free to call to discuss ideas.)
Ahyes, the design allows for easy localisation, so that's another thing
we'll be looking into now that the glitches in the earlier release have
been ironed out. If you can help with that, or anything else, please
also contact me. All assistance and ideas will be much appreciated.
Thanks.
Regards,
<PERSON>.
| e7e6c19e-0177-08ff-43ba-cce52f542b3f | ['74b6b66f-e005-71d9-dcd9-6775d46aabe5'] | Hi,
We're restarting a MySQL User Group in Sydney.
<PERSON> is sorting out a venue, but if you have offices or another
suitable place near the CBD, do let us know.
We're aiming for a first meeting Tuesday 7 Feb 7pm, since I'm actually
in Sydney next week for a MySQL training course.
If you're interested, please sign up at http://mysql.meetup.com/142/ so
you will receive further announcements about the venue etc.
It would also give us a better idea of scale.
Hope to see you next week!
Thanks.
Regards,
<PERSON>.
|
d21add2f-e214-b13c-4433-14dca63c0d01 | ['74bc1b43-4fb7-52d9-ecd9-ff31a58a7873'] | Hi,
I am using NB6.1Beta on Windows XP (Japanese edition) with JDK 1.6.0_5.
My project is encoding as UTF-8 (project --> properties --> java source
--> encoding ) and comments are wrote in Japanese.
After updated the project using CVS, I try to resolve conflicts using
Versioning --> CVS --> Resolve conflicts, however the comment in the
source files, both local and remote source file, are changed to
unreadable characters in *compare editor*. Yes, only in the compare
editor. n the Java source editor, they are displayed normally .
I guess the compare editor doesn't show the source code with the same
encoding type as the project setting.
Is it a already known bug? I hope it can be resolved when NB 6.1 releases.
Regards
<PERSON> | 549ff6fe-f243-19d2-a1c4-9cc20c965aac | ['74bc1b43-4fb7-52d9-ecd9-ff31a58a7873'] | Hello everyone,
I made a patch to rdoc -- just 3 lines -- which makes the rdoc(generating
html) run faster and use less memory.
I listed the test result in the below.
(Tested on 32 bits Linux, Ruby1.9.2-trunk, 4GB memory, Intel Quad 2.6GHz)
Using rdoc to process whole lib directory:
original : 15 minutes, 1.3GB memory
patched: 7 minutes, 65MB memory
I compared the generated results(patched and unpatched) using "diff -r"
command,
and confirmed that the patch does not break anything.
I attached the patch file, please check it out.
However, I only profiled rdoc with html output and found the problem.
So I am not sure whether or not other output format code has similar
problem.
BTW, I found the *bug* using a memory profiler I created for Ruby1.9.
Right now, the memory profiler can show how many live instances for each
class, size and
age(survived from how many times of GC) of each instance, alos can show the
allocation call trace of each instance.
However, I am still working on it.
So, if you have interest on the memory profiler, you can contact me.
Best regards,
-Tetsu |
3d692c95-0189-cf64-f55b-131bbfe187e3 | ['74d5811d-bdbc-fab3-7c95-fd1b533dd699'] | if you are referencing Tapose it's metaphor is really based o. The defunct MS Courier project. I do see some similarities to Newton.
I will be checking out the Notes Plus app as it seems to give a more Newtonish approach of HWR (instant and delayed) and drawing smoothing. Those were things I used a lot with my Newtons.
Sincerely,
<PERSON> | 001025ca-a103-d5e1-05a3-604d78f3ad56 | ['74d5811d-bdbc-fab3-7c95-fd1b533dd699'] | I agree with your statement <PERSON>. "Regarding Mossberg's review, I think it's wrong to assess who is and who is
not a plausible user of the iPad. This product will define a new way of
using mobile devices; way beyond what the iPhone did. "
With the apps more and more users will flock to the iPad because of the overwhelming useful of the product. At BestBuy today the demographics of people buying the iPad ranged from 20's to 80's, males and females. They all had different uses and were all happy. We had to activate some just so they could play with them on the drive home. (the passengers I hope).
My MacBook Air server us perfectly fine and you cannot beat the weight!
Sincerely,
G.
|
52cbce79-bc9c-0b6a-ef78-82e734c66e00 | ['74e1a09e-946d-176c-f3f8-9f910119d600'] | Hello everyone, I am <PERSON>, a third year computer science undergrad with an
interest in Information Retrieval and Natural Language Processing. I am
interested in doing the "Improvements to Debian search" project with
Debian. I've been involved with open source development for some time now
and have also implemented a couple of state of the art weighting schemes
for Xapian , Given my experience with open source development , Xapian and
Information retrieval ,I really think I'll be able to justify the project
by implementing state of the art weighting algorithms , query expansion
using relevance feedback , highly efficient query expansion algorithms,
spell checking ,resistance to keyword spamming and many features in Debian
search to drastically improve the search results and the search experience
for the Debian community. The search on search.debian.org, mailing lists ,
Debian wiki and the packages site is the gateway for the huge Debian
community to gain information about Debian and hence, improving search will
definitely aid both Debian developers and users a lot. I'm already in touch
with <PERSON> ,the mentor of this project and am working on my proposal
which I'll put up on the wiki in a couple of days.
-Regards
-Aarsh | 32bf3e27-4257-a80a-035f-872389d2b70a | ['74e1a09e-946d-176c-f3f8-9f910119d600'] | Hello admins. I apologize for not being able to do any work in the past
week as I was busy packing for debcamp and debconf and with my visa
interview and visa procedure which took place in a separate city as the
embassy is located there. However, I am way ahead of my schedule already
and the admins need not be worried. :) I shall resume work today itself. :)
Can't wait for debcamp ! :)
-Regards
-Aarsh |
392774dc-bf42-1d01-a389-699ee5726bb4 | ['74e84488-3a66-c025-3b45-30e88fa43cff'] | This reverts commit c51aa112f53adc87250177002aa3e008305e0777.
There are several serious issues with the stated rationale for and
intended use of the "Adult" Additional Category that have yet to be
resolved.
- The menu specification is designed for worldwide application, and
thus must be culturally impartial. Different cultures have
different criteria for something that is "adult-only". Ambiguous
examples provided on the XDG list include "Bible" and "Art Gallery"
applications.
- Given a set of criteria for what constitutes "Adult" content, the
categorisation is binary, but whether an application is "Adult" is
not. Ambiguous examples provided on the XDG list include a
"Breastfeeding Tutorial" application; the "PornView" application,
which does not in fact include or provide access to any pornography;
web browsers, which provide access to a wealth of pornography on the
Internet; and text adventure games that include the possibility of
violent death.
| 7fa74e9f-647d-c1f4-ba18-2cd14c3bf31b | ['74e84488-3a66-c025-3b45-30e88fa43cff'] | pathconf(_PC_NAME_MAX) was implemented on top of statfs(). The 32bit
version therefore fails EOVERFLOW if the filesystem blockcount is
sufficiently large.
Most pathconf() queries use statvfs64(), which avoids this issue. This
patch modifies pathconf(_PC_NAME_MAX) to do likewise.
<PERSON>
|
e2119d0c-9992-d0ea-e04f-df803514a3a3 | ['74f7656a-54c2-0a86-53cc-cb8a83410a94'] | Hi,
As I just subscribed to this list, I would like to introduce me first.
I'm an electronic engineer, writing from Brazil. I'm 27 years
old, work for the auto parts industry, and I have been a Spectrum
user since 1987 (actually, it was the TK90X, a Brazilian clone of the
Spectrum +).
Back in 87, my dream was to have a Sam. I know that some of
them came to Brazil at that time by the hands of some game pirates.
We could not import those machines at that time because there was a
law that does not allow us to buy imported computers, so most Amigas,
Macs, Commodores and even some Ataris wer bought by "not so legal"
ways (this includes my Amiga 1200)
I've downloaded SimCoupe for my Mac, and it was really nice
to see the dream machine emulated on my monitor.
Two things then came up in my mind: First, I don't know how
to use the Sam. The basic is different from the Speccy, and I could
not find any manual.. Second, I still want a Sam !!
So, if anybody has a second Sam to sell, or knows were I can
get one, I would appreciate very much.
Best Regards,
Marcelo | 0154d87a-6a28-3432-51e4-6bb8a56d747e | ['74f7656a-54c2-0a86-53cc-cb8a83410a94'] | Hi Folks,
Take a look at this. Just amazing... I'm really considering
to purchase one
http://virtuals.atlant.ru/peters/e-index.htm
Unfortunately, their english is even worst than mine :-)
Seems to be they used software translators
<PERSON>,
<PERSON> |
902052b5-71c8-a35b-8b64-205a3835b9b2 | ['74fb815c-f364-813f-50d2-da2fd6cbc3b9'] | Hi,
I've tried to set up the biblio.frameworkcode at import time by setting
up a link between the 099$e subfield and biblio.frameworkcode in koha.
But it seems this fields in not updated in either way.
Even if i create a new non-default notice and that I export the notice,
the 099$e subfield is not created in the export files (marcxml, ...).
Do i miss something? I'm using koha-3.00.06.
Best Regards,
Frédéric | cafc21b8-081b-e4f5-9afa-ed63fa7dfe3e | ['74fb815c-f364-813f-50d2-da2fd6cbc3b9'] | Hi,
I'm trying to switch our old ILS to Koha but I'm stuck with how I'll
import Serials in the system. The information I have in the old database
is rather sparse but I have a value I don't know how to map. I have a
publication date, which is normally linked to one item notice (and not a
serial notice). The problem is there is no mapping in the koha standard
(sample?) UNIMARC definition. Moreover, I don't know which koha field to
map to? I've found "timestamp" and "dateaccessioned" but do not know
which one to choose and if i'm in the right direction. I plan to add a
995$w field to map to the koha field.
I have to admit i'm computer scientist only ... and that the person who
manage the old ILS is not a librarian, so I'm on my own to handle the
issues. Hope you can help!
Thanks a lot!
Regards,
Frédéric |
54d572fc-5be6-8e9c-6c38-cfdc86e18cf3 | ['7509fcb2-51ed-91bf-3a2a-538473468a47'] | I am very new to DDD. I have just started reading the book Domain-Driven Design: Tackling Complexity in the Heart of Software(by <PERSON>).
Following the Cargo example in the book, I am trying to figure out how one can keep track of which Vessel (Truck or Ship etc) carried cargo from one location to another. I am not sure whether a Vessel should be an attribute of a Voyage or Carrier movement(Registration No etc ), or model the vessel as an entity on its own.
I am assuming that in the process of transporting cargo many vessels operated by different carriers will participate in the journey. Pleas help.
Thanks in advance
<PERSON>
| d23fd4b1-4c17-fbee-c013-a2820131a0ed | ['7509fcb2-51ed-91bf-3a2a-538473468a47'] | I am very new to DDD. I have just started reading the the domain driven design bible. someone please help?
I need to keep track of which Vessel (Truck or Ship etc) carried cargo from one location to another. I am not sure whether a Vessel Registration No should be an attribute of a Voyage or Carrier movement.
I am assuming that in the process of transporting cargo many vessels operated by different carrier will participate in the journey. An example preferably in c# and explanation will be appreciated very much.
Thanks in advance
|
722779c3-c95a-fd21-d5f6-708e3b419de7 | ['750e84da-0b83-aa62-872f-eac80bd6c797'] | Depending on what you're expecting, I'm not sure you'll be able to get what you want since JUnit uses reflection to "know" the names of your test cases while C++ doesn't have reflection.
I think, once you start using CppUnit, it's pretty easy to use. In order to get automagic knowledge of test case names, in our build, I put together a script using typical command-line tools (eg perl, sed, awk, grep) so in practice, it's possible to get
what JUnit affords, but it'll take a little more work.
HTH,
<PERSON>
| 84219182-7447-a162-2ab0-f28e9409bc16 | ['750e84da-0b83-aa62-872f-eac80bd6c797'] | In our project, we typically inherit from one of our base exception
types. It would be nice to be able to register with CppUnit an
exception handler that would call a method on our exception.
I've looked through the source and it doesn't look like this is possible
currently. Can this be put on the TODO list, please? (I'll see if I
can work on it if I get some free time).
Thanks,
<PERSON>
|
92a60d1e-03f2-7f1e-2a20-ff543c0b9cc9 | ['7525b137-5e02-fd72-31c5-c17a19605a36'] | <PERSON>,
I have found more problems with the compressed file stream
classes.
1. There was an error in the previous changes you applied to
FXGZFileStream, the extra readBlock line means input gets discarded. So
it does not work at all.
I've been using my patched 1.6.8 version until now and had not noticed
this issue.
I guess I must be the only one using this class!
2. There is a scenario during write at the end of stream where the input
is all consumed but the compressor still holds data internally. Due to
the while condition this was not handled so the final output was never
flushed. My previous tests had not forced this scenario, it probably
needs the right combination of volume and complexity of data. This
problem only showed up in my application but I have not been able to
replicate with test data.
3. There is the potential for a simliar situation during read with the
file input all consumed but the decompressor still having more output to
give. From my limited investigation it appears the libraries do not
behave this way and will provide all output for a given input. I have
however updated the test for readBlock result to only break on error not
on zero bytes read. The code is more correct this way and removes the
risk if the libraries change behaviour in the future.
4. With the above changes the flush at end of stream works but a call to
flush mid stream would not result in all possible output being written.
I have therefore reimplemented the flush methods to force a flush of the
compressor. The appoach is simple but seems a bit clumsy, you may have a
better way to do it.
I did not need this feature but it seems logical to make this change for
completeness.
Context diff included below.
Regards,
<PERSON>.
| 0cee44cd-8a60-8f1e-e194-7ccd8f4e49e5 | ['7525b137-5e02-fd72-31c5-c17a19605a36'] | <PERSON>,
I have a set of objects that I am serialising to a file via
save/load. The objects contain Icons.
What I noticed was the transparency of the Icons changed after the
objects were restored from the file.
After some investigation I think I know what is going on. FXGIFIcon
loads into an RGBA format and the original color table is not preserved,
so must be reconstructed during save. It appears this is a 2 step
process first determine unique colors, then determine which is
transparent.
The Icons I had chosen to test with, bigdoc.gif and minidoc.gif, contain
2 entries for white in their color tables, one for the transparent
background, the other for the body of the document. So in this case I
think the color table gets reconstructed with one white entry that is
flagged as transparent. The result is that the body and background are
transparent and only the border is displayed.
If I use an Icon with a unique color for transparent, all works fine.
Since this is probably an unusual case I doubt it is a high priority to
fix.
Editing the Icons to avoid the problem might be a good idea. Note that
minapp.gif is a similar case. I have not done an exhaustive search for
other cases though.
I have a question also. While looking through the standard Icons I
noticed that some do not have transparency set (as far as my image
editor can tell) however they do get rendered with transparency and I
could not see how this would be happening in the code. An example is
dirupicon.gif.
Regards,
<PERSON>.
|
90461f22-b85c-33de-6895-01ec64e4ef38 | ['7536c120-51e8-103c-407f-98bc36747e94'] | Dear <PERSON>,
I posted about this recently, but it was lost in the noise.
I was at the TCPM WG session where they were discussing
a draft recommending TCP implementation practices.
During the session, there was consensus to move the Eifel
algorithm, to an experimental section of the document
_because_ it has not been widely deployed and because
of existing uncertainties regarding IPR.
What was discussed in the session was the occurrance of
an implementor being presented with a schedule of
charges for licenses by an IPR holder.
Eifel has been described as a way to handle
packet duplication events experienced due to bicasting
in the draft (section 10):
http://www.ietf.org/internet-drafts/draft-elmalki-mobileip-bicasting-v6-06.txt
While the IETF (in compliance with its policy: BCP 79)
will not take a position on the validity of the IPR claims,
please be aware that Eifel is not widely implemented,
and isn't likely to be implemented by default in TCP
stacks, due to its experimental nature.
If bicasting is important, it may be worth investigating if
another mechanism is available for reducing the effect of
duplicated packets on upper layer protocol stacks.
<PERSON> | c06d7229-2050-5ba1-b1f8-0ce7b8fa4c38 | ['7536c120-51e8-103c-407f-98bc36747e94'] | Hi,
I think there are some interesting discussions going on
in a different thread, but I thought I'd start a new thread
in order to talk about a contentious issue without polluting
the other.
Regarding draft-pashby-ipv6-network-discovery-00.txt,
this provides a mechanism for devices to be made respond
to queries from another device on the IPv6 network.
This is not an existing capability.
I'm concerned that if there is a way to find out all the
nodes on a link, that this information may be used
(by the querier, or another device) to cause remote flooding
attacks onto a network, or to particular otherwise unmodified
hosts.
In IPv4 it is feasible/trivial to try all addresses in a subnet
in order to find targets for attack, but in IPv6 >2^60
combinations may need to be tried.
The anonymity of the present (but quiet) IPv6 node is
probably useful in this case.
There is no system, except MLD which can force response from
unknown nodes in IPv6. With MLD, the reporters can be made to
expose only one of their link-local source addresses.
They are not required to expose global addresses.
At the moment there's no security for MLD, but the risk is
limited to link-local addresses which are not vulnerable to
off-link attacks.
I'm loath to introduce a more generic function like
this which exposes global IPv6 addresses, unless there is
verifiable trust available to the nodes, before they are
forced to respond.
<PERSON> |
9f9f1227-3e97-3ba2-e382-5e6855200ebc | ['7537adb0-3c10-b0fc-cc89-f06276469ba3'] | Hi,
I have written a JAPE for creating some annotation types (Anno A, Anno B ,
Anno C). I am running this application on a corpus where each document may
or may not contain any of these annotation.
I am trying to segregate the documents containing any of these annotations
from those not containing.
This segregation may occur in many ways-
1) either arrange documents in particular order in corpus (documents
containing these annotations on top )
2) or deleting documents not containing any of the above annotation from
corpus.
3) Or creating a separate corpus for one of the category.
Is there any in-built feature for this in GATE? Or will I have to do this
using JAPE/Groovy?
Any help will be appreciated.
Thanks and Regards,
<PERSON>
| e4489357-c750-2cb0-6257-4926fdff8c8f | ['7537adb0-3c10-b0fc-cc89-f06276469ba3'] | Hi Users,
I have recently added a new gazetteer list to my ANNIE gazetteer list. But
I am unable to change the following *parameters of the list*-
1)Annotation Type--> I want it to be Lookup
2)Major
3)Minor
Every time I add them manually in gazetteer list and then *save and
initialize*/ *initialize* ANNIE gazetteer, it automatically deletes the
parameters.
Any help will be appreciated.
Regards,
<PERSON>
|
2e97550f-65ef-177f-ecb9-f3085eef94ea | ['7542ba58-0e6a-b044-e83c-27b9b1b07d43'] | hi
thank you for reading my post
is there any listbox like component for j2me applications ?
I need something like a list box with following features :
1-it allows me to know when user navigate up/down in list box
2-when it somehow fired an event for user selection change , it let me
know the current itemindex or item value.
can some one help ?
thanks | fdb7fe14-f0ca-a774-16ff-738f6c7a9df0 | ['7542ba58-0e6a-b044-e83c-27b9b1b07d43'] | Hi
Thank you for reading my post.
I have a web application which i developed and works well on my tomcat 5.5.9
Now i purchase a hosting plane with tomcat 5.5.9 it is on linux and have
cpanel.
they said that i should upload all contents of my war file into http_doc
folder , so i did
I used ftp to upload them all , now that i try to access my web site it
return
HTTP Status 404 - /index.jsp
meanwhile index.jsp in under that www folder.
I test the server by uploading an html file to make sure that i put
content of war file in correct place. nad it works ok with HTML file.
can some one help me about this problem ?
Thanks |
17635a10-b4de-935c-d251-52647ce9f126 | ['75447476-9435-2ab8-078e-dabd4d16f6e7'] | Hi all !
I want to get the caller method name which calls a particular api.
So ideally i should get "testMethod" when ever i call "logData" method.
There is another question related to this after earlier part is complete I want to use "aspectJLogging" method when ever "logData" method is called.
Please suggest this as well.
I had attached the files.
I would be very thankful for your help.
Many Thanks
<PERSON> | 37a253d0-08ae-15df-41a1-ce38468bc1a4 | ['75447476-9435-2ab8-078e-dabd4d16f6e7'] | Hi all !
I want to get the caller method name which calls a particular api.
So ideally i should get "testMethod" when ever i call "logData" method.
There is another question related to this after earlier part is complete I want to use "aspectJLogging" method when ever "logData" method is called.
Please suggest this as well.
I had attached the files.
I would be very thankful for your help.
Many Thanks
<PERSON> |
58146e24-0ca5-c3aa-4750-6fdaf535945c | ['75460ccf-2a50-0cd2-7d12-2e2d44af44ac'] | Hello,
i have a scenario in a Grails 1.1 application where a controller populates a
domain object and passes it to a service for validation and saving (some
business logic is also included in the service method being called). The
service has *transactional=true* and i found that the object is saved when
passed as a parameter in the call from the controller to the service method
even calling *domainObj.discard()* inside the service method, doesn't
discard (because the object is saved already but not necesarilly valid
tough). Is this supposed to be that way? Where is the transaction created?
In the controller or in the service?
I also found that changing *transactional=false* makes the service work as
expected.
Here's a step-by-step if the former explanation was not clear enough (most
probably due to my poor english writing):
1.- Controller finds domain object by *params.id*
2.- Controller populates domain object with data from *params* map
3.- Controller calls service.saveObj(domainObj)
3.1.- Domain object is saved (even if it's not valid)
3.2.- Service method is called
3.3.- A call to *domainObj.discard()* doesn't work!!
Any help or clue will be appreciated.
Regards,
<PERSON> | b2988e37-c609-277e-047b-d21519be293c | ['75460ccf-2a50-0cd2-7d12-2e2d44af44ac'] | Hello, i changed my UrlMappings.groovy file so i can have error codes mapped
to an ErrorController. Inside the controller i'm trying to access exceptions
object in the same way that default error.gsp does but i get a null pointer
exception.
I'm mapping error codes to a controller so i can use withFormat and render
the error description in html or json, depending on the request format.
Any kind of help will be appreciated.
Thanks
|
b8363a14-8cf7-5e46-c474-0d779297850b | ['7552c9e7-62c5-1513-2ad4-f9d00c0bb9bc'] | Can someone point me to where I can install a public file or where the
documentation is??
I want to read a configuration file but every time I rebuild the
application it wipes out the user data thereby making me have to rebuild
the configuration all over again. During testing this is a pain.
Secondly, I would like to have TWO applications share (read/write) the
same file? In production?
Thanks,
<PERSON>
| 3b1a1c78-9446-d569-ba55-440e93adbbbf | ['7552c9e7-62c5-1513-2ad4-f9d00c0bb9bc'] | We've been getting a lot of pressure from management (sales) to support
smart phones other than Windows. We started to look at MonoDroid and
then Novell kicked (actually got sold) the project and staff to the curb
so we stopped our plans. Now that Xamarin (I'd like to know what that
means and how to pronounce it) is back and it seems to be a production
ready solution, I've been asked to restart a proof of concept which I've
been working on so we can put together a project plan.
I've got most of the application figured out and broken down so we can
design business layer, application functions, communications, and user
interface. I am then hoping to be able to change out the user interface
and then have a port to the iPad (which is where all the sales teams are
saying the sales are).
So here's the overall questions I'm stuck with.
Here's my two design questions:
1. In our PC/Windows phone application we check our version against a
service and if there's a new release, we download it and upgrade. So,
going through this, I see Android and iOS use a sandbox concept
differing from what I'm familiar with. So, how does one handle
updates? Is it possible to download a program and once it gets
downloaded it runs and replaces the application?
2. Secondly, one of our applications is a business application that
runs on the sales staff's company owned phone/tablets. Company's own
these devices and put them in the field and part of the business
requirement is to make the sales application password protected so it
cannot be uninstalled (which also deletes all the sales records). Is
there a way to set a pin so an app cannot be removed? All I can come up
with is a 'watch dog' approach where you have two applications each
monitoring the broadcast message for package removal and then reloads it
unless a pin is entered (but the sales data would be lost however I
could resync from the server). But, this seems a little to much like
malware. Can I somehow drop the application into the system/application
store thereby making it a system application? Or how do you handle this
business requirement?
Thanks,
<PERSON>
|
01d806e4-f521-2da4-a69d-717e444542de | ['756674c0-c9a0-0f42-b1be-5aa5d9dc9149'] | Hello,
We have a table in HBase that has very large rows and it goes OOM when the table mapper attempts to read the entire row into a result.
We would like to be able to map over each Cell in the table as a solution and it is what we are doing in the map anyway.
Is this possible? Like the default behavior for Accumulo?
We looked at the settings on Scan and didn't really see anything and the source code of Result looks like it wraps an array of cells so the data is already loaded at that point.
We are using HBase .98.1 and Hadoop 2 APIs
Thanks
<PERSON>
| ef919f07-4994-47d3-94cb-a7d10b99cf96 | ['756674c0-c9a0-0f42-b1be-5aa5d9dc9149'] | Hello,
We have a table in HBase that has very large rows and it goes OOM when the table mapper attempts to read the entire row into a result.
We would like to be able to map over each Cell in the table as a solution and it is what we are doing in the map anyway.
Is this possible? Like the default behavior for Accumulo?
We looked at the settings on Scan and didn't really see anything and the source code of Result looks like it wraps an array of cells so the data is already loaded at that point.
We are using HBase .98.1 and Hadoop 2 APIs
Thanks
<PERSON> |
1833c928-5998-d226-ff4c-d8edd09b038f | ['756f70c7-0571-c73b-1c36-6cc1268bc74e'] | I have a couple of Cocoa applications that I want to make backwards
compatible with 10.1.5. I am building on 10.2.4 with the Dec '02
tools. Most things work, but by experimentation I have found that I
cannot link to NAN or use isinf(). __isnand() works, but I should not
have to call a __ function.
Printing is not working for me in 10.1.5. I get an empty, tiny page,
but this may be a problem in my ad-hoc page setup code.
Is there something I am missing? I can't find any documentation that
cites NAN or isinf() as problematic in 10.1.5 compatibility.
TIA,
<PERSON> | 5865ea77-f112-d6eb-8488-c6dc4e7cf843 | ['756f70c7-0571-c73b-1c36-6cc1268bc74e'] | Hi <PERSON>,
I will share with you a class factory approach that works for me. C++
is statically linked, so unless you want to jump through hoops you do
not have the ability to add classes at runtime. If you can handle
that, then all you need to do is add a ClassID -> class instance method
to ClassFactory in order to use it.
Cheers,
<PERSON>
|
bdf8791d-0d27-0a1f-ef30-9c352ea310f5 | ['7588b9b1-2184-be61-b943-7f0447bc81f4'] | About 95% of my contacts are CW and I operate an hour a day mobile CW
with the FT100D. The 500 Hz filter is active almost all the time and it
is excellent. I have a friend who bought the 300 Hz filter and I have
used his FT100D. It is a nice tight filter but I doubt I ever spend the
money to buy one. There are still a lot of guys out there who don't know
how to get on frequency when answering a CQ. The more narrow the CW
filter you must tune + or - to find the off frequency guys calling. You
will not hear them at all unless you tune (good use of the RIT). I
would say if you can get a GOOD deal on a 300 HZ filter go for it but
you sure don't need one for most CW work. I know guys who have gotten
rid of 250 Hz CW filters because they found them to be too tight. 73,
Ron W8RJL
| d9231c1f-f398-d8a4-2ba7-43b6c74e8773 | ['7588b9b1-2184-be61-b943-7f0447bc81f4'] | I am contemplating re-mounting the FT-100D from the cab to the trunk
of my Ford Escort. -SNIP- My reason for contemplating this change is to
find a better solution for all of the cables this rig requires.
Anyone have any thoughts as to the likelihood of success of this
undertaking?
Best Regards,
<PERSON>, KB1FLR
|
5b76c9e7-641f-9811-9148-6b9e6c216828 | ['758ce57e-d0c5-fbfe-9a40-66f0dfe4c962'] | However, my guess would be that you will encounter very heavy
resistance for suggesting cmake, perhaps up to simply being totally
ignored.
I agree that's the likely outcome.
My personal experience with cmake has been extremely bad.
Mine has been great but I know not everyone's is and I'm admittedly a bit
biased since I'm a part time CMake developer. From my experience, people
tend to have either a love or hate relationship with CMake, but seldom in
between.
The need for me to deal with the Windows build is rare so I've got little
motivation to drive the conversation either direction. Just that if it is
something the dev community wanted, i.e. if the feeling in the group was
"We don't like the way this scons business is going. Can somebody redo the
build with CMake? " then I'd be willing to work on the implementation.
<PERSON> | ececb204-aaba-4aef-9a92-d551141b06fa | ['758ce57e-d0c5-fbfe-9a40-66f0dfe4c962'] | Is there a logging infrastructure currently available to drivers in Mesa?
I was looking to clean up some of swr's debug / info output and have it
conditional on the MESA_DEBUG and LIBGL_DEBUG variables but then I realized
that it's really something useful all across mesa so there may already be
something there for it. If not though, I'd be interested in adding some
very light weight functions for just that purpose could be used by any
driver rather than just fprintf(stderr, ...);
<PERSON> |
7b31e1f9-c9bd-98d5-e438-199b89b6b43c | ['7593c17f-5c4d-af6e-9aee-bd1d0b136936'] | Hi,
I'm creating a custom layer for my imx6q based board.
I'm extending sysvinit_2.88dsf.bb by creating
sysvinit_2.88dsf.bbappend in my custom layer.
There is already a similar bbappend file in:
meta-fsl-bsp-release/imx/meta-bsp/recipes-core/sysvinit/
bbappend file in my layer is not being used. I do not want to skip the
bbappend file in meta-fsl-bsp-release and I want my bbappend file to
be used as the last recipe for sysvinit.
Is it possible to have multiple bbappend files?
How can I include my bbappend file without any modification to the
bbappend file in meta-fsl-bsp-release?
Any inputs will be of help!
Regards, <PERSON>
| 68b36f8c-1521-04a2-2da1-824f2e528170 | ['7593c17f-5c4d-af6e-9aee-bd1d0b136936'] | Hi,
I happen to return to imx6 mailing list after a long break.
I vaguely recollect an old issue: on imx6q vivante graphics hardware
rendering (for x11 build) did not work because the vivante libraries
(x11 build) were not DRI2 compatible. As a result when using x11 then
graphics is not hardware accelerated.
Is this issue resolved? If yes, in which yocto release is this resolved?
Any inputs will be of great help!
Thanks & regards, <PERSON>
|
0546dc58-0b4e-6147-b851-412c229ee966 | ['75a1a516-9c3a-4d8e-2a75-be8eeb92861a'] | I need to know if syswrite is a blocking call. I am using it to write to a UNIX Socket. Somewhere in my application, my code is getting blocked. I have narrowed it down to the data write code.
So I am wondering if syswrite blocks.
I am really in need of an answer. Any help is appreciated.
Thanks
<PERSON> | 865c7e68-2182-1b82-6d67-1e39136f3d8c | ['75a1a516-9c3a-4d8e-2a75-be8eeb92861a'] | Ok. I have gotten the blocking() interface to set the blocking flag to 0, but unfortunately this has not resolved my problem with syswrite doing a blocking write. Below are some snippets of my perl code and a print out from my logging. The logging clearly shows that the sywrite is blocking.
Anyone have any idea how I can resolve this problem?
Are UNIX sockets able to work in a non-blocking mode?
Any help greatly appreciated,
kwo
|
bf177e27-11af-3ea1-0a77-cd0fe18eaf33 | ['75a4a6d0-c87b-0a0a-3906-0485ea4ea4d1'] | Hi.
In conjuction with DVD playback (under VDR) I am looking for the possibility to scale MPEG and SPU
data from a DVD from 4:3 to 16:9.
Some DVD menus are stored in PAN and SCAN mode, where the actual Video and SPU data is stored on the
DVD as a 4:3 image. The decoding hardware is then supposed to re-scale it to 16:9 to fit the aspect
ratio.
Unfortunately I could not find such a feature in the DVB API. There is supposed to be an ioctl
called VIDEO_SET_ATTRIBUTES. But it seems that it is not implemented, although it's documented in
the API!?!
In short terms:
I am looking for a solution to display 4:3 Video and SPU Data as 16:9.
The TV-Set should also accept the data as 16:9 and not try to rescale it back to 4:3 according to
the information on PAL line 23.
Thank you a lot in advance and happy holidays to everyone.
<PERSON>. | 5af0cd24-0c1c-9f73-0b8c-dbdfbe611bf7 | ['75a4a6d0-c87b-0a0a-3906-0485ea4ea4d1'] | Hi,
I'm trying to use the VIDEO_SET_ATTRIBUTES ioctl call to maybe fix some aspect ration problems under
VDR. I'm using kernel 2.6.9 drivers with the firmware from CVS on Nov. 23rd with a Technotrend
DVB-Sat FF card rev 1.5.
I added my own method (SetDVDFormat()) into device.c dvbdevice.c of the vdr-1.3.17 sources and I
tried to call this method from inside the DVD-Plugin.
The ioctl for setting the VIDEO_SET_ATTRIBUTES unfortunately always returns -1.
Now I'm stuck at this point and I have no idea on what else to try to make it work.
I posted this problem to the VDR list before but I got no replys to it. Please take a look at the
postings:
http://www.linuxtv.org/mailinglists/vdr/2004/12-2004/msg00032.html
http://www.linuxtv.org/mailinglists/vdr/2004/12-2004/msg00061.html
Any help into the right direction will be greatly appreciated.
Thank you a lot in advance.
Andr. |
9c9d84f7-ccfb-1484-7c6c-6f161b463617 | ['75cedbe9-2a77-5614-4899-8c28fba49538'] | I'm currently working on a PHP+Apache2+SomeDB project at the company I work
for. We're primarily a Windows shop, but Unix could make it also. The
biggest issue is support for whatever platform is chosen. Personally, my
experience is with Debian/Ubuntu, but as far as I know, both don't have
major commercial backers (yes, I know Ubuntu has one, but how long have they
been around)?
I'm pretty much looking for everyones opinions on what experiences they have
had with the following *nix/DBs. I'm leaning to Solaris 10/Postgres, but
that is new ground for me. What does everyone think about this? Best support
cost vs support offered is important also..
* Windows 2000/2003
* RHEL
* SUSE Enterprise Linux
* Solaris 10
| 6f993b09-d2bc-78ef-84b0-2df2779700a0 | ['75cedbe9-2a77-5614-4899-8c28fba49538'] | Don't know if for-sale posts are allowed, so please don't flame me if they
aren't...
I've got 3 Cisco 2500s (2x 2501, 1x 2503) for sale. They are in slightly
different flash configurations and RAM configurations. 2 are running IOS
11.x and one is running 12.x (I think). They all have rack ears, power
cords, 10BT transceivers, and DCE-DTE cables. I'll also throw in a
MicroAnnex XL terminal server with serial cables. The routers are in
decent condition - the cases have some scratches and the occasional scuff.
Everything is located in the Aurora, IL area and is pickup only. Let me
know off-list if you're interested - I'm taking offers. If you'd like more
information, let me know.
Jon
|
988ee9a0-8df6-75ce-71c6-7daf5eddf940 | ['75d9a310-3c7b-5711-68a9-3d41f32034bc'] | Hello
I'm new to Debian, have used Slackware for years but latest release was a
mess so I decided to move to Debian and give it a try, however I have some
doubts and need to complete my configuration to keep working while
learning to use this new distro, so these are my current doubts:
1) How can I disable the graphical login and/or avoid X11 to start
automatically? I can't install the nVidia driver since it tells me to stop
X11 before. And, is it possible to leave it that way, just with the simple
text consoles, and start X11 manually when required?
2) How can I install KDE? Currently it is running with Gnome... I've just
download all 5 DVDs plus the updates one... the gnome (un)install shows
most KDE applications to install, but not the full window manager... and,
once installed, is it safe to remove Gnome to free disk space? Just the
window manager, not its libraries that should break some dependencies if
deleted.
3) There are some programs/drivers that independently of the distro, I
prefer to install them by hand from sources. Does it affect in some way if
I do this? The most common ones are kernel, MySQL, msn, irc, some
multimedia stuff, etc. but at least the kernel and MySQL that I don't like
to use packages for them...
4) Is it possible to install LILO instead of GRUB? I know it is better and
advanced, but I like LILO... would it be safe to install it without
breaking some kind of dependence?
Well, that's enough by now, today I need at least to setup the nVidia
driver and hopefully the latest kernel (no matter if it is GRUB by now)...
Thanks for your help,
<PERSON> | 7e99220f-6681-a57a-bf9f-296f99c2337a | ['75d9a310-3c7b-5711-68a9-3d41f32034bc'] | Hello
First of all I want to thank all your kind responses, it will help me a
lot to start using Debian. I will write a single message with comments for
all those who replied to me, to avoid flooding the list:
<PERSON>:
Now I know that not all of the DVDs are necessary. Respect the 'expert'
mode installation, when booted the DVD #1 it showed only "Install" and
"Graphic Install" (I have tried both of them), but at any moment saw the
expert or advanced installation, it just ran itself without asking
details. I will try the "Debian way" with most packages, and see if they
work for my needs, but at least kernel and mysql I need them from sources,
although will leave the original shiping package to avoid breaking
dependencies with httpd, etc.
Matthias:
|
dcc00801-d301-fe93-7708-bd3cfab28e1b | ['7608f8ea-cd1b-4c83-a3c2-4d8d9c07b167'] | Hi!
I'm writing a small test utility to evaluate some open source db's
(MySQL, PostgreSQL, SAP DB and SQLite). This is my first time playing
with the new DB-API 2.0 and I must say, it's *not* portable between db's
(I think I'm not the first person who found out ;-) ). Besides of
differences between SQL modes also the paramstyle is different.
My first question is: why does sapdbapi only support 'named'? All other
db's support 'pyformat'.
Second question (not sapdbapi specific): why is there no 'BEGIN' for a
transaction? As far as I found out a transaction is automatically
started after a commit or rollback, but isn't 'BEGIN' specified by SQL-92?
Best Regards,
<PERSON>
| b0cb2ac3-eb8e-178e-94a8-f9a932978e5d | ['7608f8ea-cd1b-4c83-a3c2-4d8d9c07b167'] | Hi,
I'd like to add a micro navigation to my homepage. Actually there should
be links to the title of the index.html pages. For example:
Home / Products / Hardware /
It's not a problem to generate a dict with the path as key and the title
as value. But what is the easiest way to build the path? Is there a
module to step up the path (html/products/hardware/index.html)?
<PERSON> |
dc5d6a18-eabd-8907-7e8f-eef94e1a69d2 | ['761da173-4f98-52aa-6830-5e92e351f616'] | Hello,
Could you please take a look at
http://www.lilianne-blaze.net/wiki/bin/Java/HeaderPatternLayout ?
Mostly inspired by http://juliusdavies.ca/logging.html (search for "A
Log Rotation Best Practice"). The general idea is to write a line on top
(and optionally end) of every log file with basic information about the
current environment, like memory, Java runtime and VM info, operating
system, etc.
Any feedback, comments, suggestions, questions, would be most welcome.
Greetings, <PERSON> | f1557c1d-7044-d472-6c27-ab575e43c593 | ['761da173-4f98-52aa-6830-5e92e351f616'] | Hello,
I have an InitializingBean, lazy-init=false, where afterPropertiesSet
calls method xxx which is annotated with
@Transactional(propagation=Propagation.REQUIRED, readOnly=false). The
purpose is to populate the database with some initial data if it is
empty. It fails, apparently @Transactional isn't used when called from
afterPropertiesSet, is it a bug or is it by-design?
For now I simply wrapped it in transactionManager.getTransaction(td); +
transactionManager.commit(ts);, and it works as intended by me, but it
seems a bit 'hackish'. Is there a way to make @Transactional work in
such a scenario? Or are there any "best practices" applicable here?
Greetings, <PERSON>
|
17a5d452-9cb8-08fe-1bfc-c906aa5792a6 | ['762a02b0-d328-1fc3-5ff5-380f682e7f57'] | Hi,
I want to discuss one particular topic in Wireless Sensor Networks. Several routing protocols have been proposed. Totally independent of this, several researchers have also focussed on developing energy efficient MAC protocols for sensor networks. To be more clear, the routing protocols assume 802.11 based MAC protocol. I want to know, if any research has been done in this direction. I am thinking if we can provide smooth integration of currently incompatible MAC protocols and routing protocols.
Kindly let me know, if anybody in this group is working in this direction. I request the members to give me some references in this direction.
Looking forward for your suggestions,
Gokina.
| 5c5fdbdf-09aa-4f53-35ed-65bcea5fc8e3 | ['762a02b0-d328-1fc3-5ff5-380f682e7f57'] | Hi,
I am doing research Independently in Mobile Ad-hoc Networks and in Sensor Networks. I have some doubts in some of the research topics in these fields.
1. I have read some publications about trust management in Peer-to-Peer networks. The concept of trust has been extended to Mobile Ad-hoc Networks. In P2P networks the nodes are stationary where as in the second case the nodes can move. Assume that the trust management is not centralized but distributed across the network. Each node maintains trust information about the nodes which are in it's interest. Now, suppose one node is found guilty. The nodes in its surroundings can know about this. Immediately after the node is found guilty, this particular node can travel long distance where the nodes are not aware of his behaviour. I want to know if the existing publications take care of this kind of situation.
2. Is it possible to extend the concept of trust to Sensor Networks. Since, Security is also an aspect of Sensor Networks, Is there any mechanism so that one Sensor can choose the routing path that has trusted Sensors.
3. In Peer-to-Peer Networks there is a concept of Super Nodes. In this they exploit the property that some nodes are more powerful than other nodes. To the best of my knowledge, that main problem in Sensor Networks is energy. Is it realistic, if we assume that some nodes in Sensor Networks are more powerful than others and handle most of the traffic on the behalf other nodes which are less powerful. Will it be realistic, if we assume that these powerful nodes have no energy constraints. I mean can there be a situation where some of the Sensors in the Sensor Network have continuous power supply.
I think if these assumptions are realistic, then there is a possibiltiy of giving new direction to research in Sensor Networks. I am not sure, if this can be done. I request the members in this group to clarify my ideas and if possible, kindly give me some references.
Looking forward for your reply,
Gokina.
|
9703cdde-2595-ec81-be1c-3c119e5a35c8 | ['7631d3d2-9e78-d786-91b0-b9eef5a8fc71'] | Enter your vote today! A new poll has been created for the
dpjudge group:
What is your favourite method of playing Diplomacy?
o dpJudge
o Other Judges
o Judge Tournaments
o Casual Face-to-Face
o Tournament Face-to-Face
o E-mail
o Postal
To vote, please visit the following web page:
http://groups.yahoo.com/group/dpjudge/surveys?id=2331633
Note: Please do not reply to this message. Poll votes are
not collected via email. To vote, you must go to the Yahoo! Groups
web site listed above.
Thanks!
| 96159a51-0fe1-cfb0-b2f9-429588d8d5c6 | ['7631d3d2-9e78-d786-91b0-b9eef5a8fc71'] | Enter your vote today! A new poll has been created for the
dpjudge group:
Three years after this question was already asked, it is asked again to see the
change in taste over those three years: what is your favorite country to play in a
standard game?
o Austria
o England
o France
o Germany
o Italy
o Russia
o Turkey
To vote, please visit the following web page:
http://groups.yahoo.com/group/dpjudge/surveys?id=2317700
Note: Please do not reply to this message. Poll votes are
not collected via email. To vote, you must go to the Yahoo! Groups
web site listed above.
Thanks! |
342d529a-248f-4959-2d8f-c13a5e1f961b | ['763d02ff-c90f-1d11-58c7-4ffb636d0a4a'] | for last couple of weeks #master testresults are drifiting a lot.
Yesterday I merged #libevent into master; <PERSON> was pressming me to do it
And we will fix everything in master instead of working on branches.
Now <PERSON> and me are fixing #master and test results. #master should be fixed today or tomorrow; please be patient. May be hold on to your commits unless it is absloutely essential?
if compiler warnings hurt your eyes. this is good time to take break:)
thanks for your understanding,
-antony | 42a792a0-e983-ad6b-a9cc-50198a8a328c | ['763d02ff-c90f-1d11-58c7-4ffb636d0a4a'] | Hi,
Yesterday <PERSON> and I met with NSS guys and here are some notes from the meeting.
NSPR threading: no need to use NSPR threading on Linux, because on Linux it is jut a wrapper around pthread.
Don't open a NSS DB file simultaneously(not even one app writing and another re-reading). Due to the nature of in memory data structures things can go wrong. Close it completely and start again. Also use new format, not the old Berkeley DB.
In the NSS code base, operations on X509, ASN 1 is not within the 'crypto boundary', only the crypto operations are. When it is running in FIPS more NSS will support non FIPS operations. It is upto applications not to decide.
If we want to store PSK/Xauth/EAP secrets in NSS we could store the "secret" encrypted by the master key. And the NSS can decrypt it for libreswan.
NSS can't talk to the Linux Kernel via Netlink; e.g to install ESP keys directly into XFRM module. However, they can talk to some of the TCP offload NIC.
Human readable error strings in NSS is possible. They mentioned an application(libreswan) must initialize "error code tables" in NSPR to access it in NSS. It might be worth investigating. You may also have to install NSS utils.
A quick googling shows libreswan use PR_GetError. However, libreswan seems to be missing initialization code, PR_ErrorInstallTable, nspr_InitializePRErrorTable. I haven't looked in detail. It seems prerr.h or prerr.c is a starting point.
Google developers are actively working ARM and other non Intel Architectures.
Cross compiling for MIPS Big Endian, for OpenWRT, should work! The buildroot Makefile seem to build MIPS, even though NSPR config files only seems to support Little Endian architectures.
There is no OpenWRT Makefile yet.
-antony |
8915d8ae-d058-7eec-e8d7-f93a183a0142 | ['764fb61b-c11f-b0cb-2c8b-a9f66d4fa441'] | Does anyone have any experience with using uIP to do UPnP setup or any other
kind of Zero Configuration setup of the interface? Alternately, has anyone
out there ever done a DHCP server under uIP? Dare I ask for some kind of
simple DNS service?
Any help or pointers would be appreciated.
<PERSON>
| af673adf-5e00-5dc8-ef52-ffcb7adec07e | ['764fb61b-c11f-b0cb-2c8b-a9f66d4fa441'] | This turned out to not be so hard. The hardest part was untangling some of
the include recursion that occurs when creating a union to hold the
application state structures. Otherwise, it was just a matter of initing
both applications and creating a common appcall function to sort out which
application/local-port is receiving data.
<PERSON>
|
176ccacc-168c-53cd-1d3a-70065ce9779d | ['765d27ff-9ff8-e8fa-06cc-9a3f63e250cd'] | I have a ZFS-based NFS server (Solaris 10 U4 on x86) where I am seeing
a weird performance degradation as the number of simultaneous sequential
reads increases.
Setup:
NFS client -> Solaris NFS server -> iSCSI target machine
There are 12 physical disks on the iSCSI target machine. Each of them
is sliced up into 11 parts and the parts exported as individual LUNs to
the Solaris server. The Solaris server uses each LUN as a separate ZFS
pool (giving 132 pools in total) and exports them all to the NFS client.
(The NFS client and the iSCSI target machine are both running Linux.
The Solaris NFS server has 4 GB of RAM.)
When the NFS client starts a sequential read against one filesystem
from each physical disk, the iSCSI target machine and the NFS client
both use the full network bandwidth and each individual read gets
1/12th of it (about 9.something MBytes/sec). Starting a second set of
sequential reads against each disk (to a different pool) behaves the
same, as does starting a third set.
However, when I add a fourth set of reads thing change; while the
NFS server continues to read from the iSCSI target at full speed, the
data rate to the NFS client drops significantly. By the time I hit
9 reads per physical disk, the NFS client is getting a *total* of 8
MBytes/sec. In other words, it seems that ZFS on the NFS server is
somehow discarding most of what it reads from the iSCSI disks, although
I can't see any sign of this in 'vmstat' output on Solaris.
Also, this may not be just an NFS issue; in limited testing with local
IO on the Solaris machine it seems that I may be seeing the same effect
with the same rough magnitude.
(It is limited testing because it is harder to accurately measure what
aggregate data rate I'm getting and harder to run that many simultaneous
reads, as if I run too many of them the Solaris machine locks up due to
overload.)
Does anyone have any ideas of what might be going on here, and how I
might be able to tune things on the Solaris machine so that it performs
better in this situation (ideally without harming performance under
smaller loads)? Would partitioning the physical disks on Solaris instead
of splitting them up on the iSCSI target make a significant difference?
Thanks in advance.
| dc8a9bba-2ef5-3636-b8d7-f4c2794ccd6e | ['765d27ff-9ff8-e8fa-06cc-9a3f63e250cd'] | We have a somewhat complex NFS server environment with iSCSI
backends[*], which holds /var/mail among other filesystems. Due to
performance problems on /var/mail, we recently switched it to using
SSD-based iSCSI backends. However, performance has failed to improve the
way we expected and now we're trying to track things down and improve
them.
Low level raw disk performance seems to be fine; IO stats on the SSDs
say that IO is fast even under load (with read and write wait times of
only a few msec, low utilization, and so on), and under low load the
fileserver can turn around IO through iSCSI in a msec or two. Under load
what we see is relatively high system CPU utilization on the backends
(on the order of 40% as reported by vmstat, significantly higher than
seen on HD-based backends) and fileserver IO average service times of
anywhere from 20 msecs to 80 msecs. We can't see any obvious saturation
point; network bandwidth is moderate and the actual IO rate from the
fileserver is surprisingly low (under 100 IOPs/sec, which is well below
what the SSDs can sustain as far as we know).
Right now we're using an old version of IET on an old kernel (<IP_ADDRESS>)
because we started all of this in 2008 and qualifying new versions of
anything for production use is a big deal. We're willing to upgrade
either or both if either people think that it would be likely to improve
the situation or the current IET versions have better stats.
So, here's some questions:
- have people seen anything like this before?
- is there general advice on tuning iSCSI and network parameters for
low latency on 128Kb random read IO requests?
(We're using gigabit networking without jumbo frames and default
iSCSI parameters.)
- does IET have any support for getting iSCSI-level IO statistics, eg
LUN-level 'disk' IO stats like the low-level disk IO stats?
If IET doesn't have anything for this today, would people be interested
in a patch to add them?
- how unlikely is it that IET is adding IO delays under load (it's
clearly not adding them all of the time)? Is this basically completely
unlikely? Is there any relatively easy way to tell?
(I ran a tcpdump packet capture through wireshark but it doesn't
handily match up requests and responses for me.)
Many thanks in advance for any advice and so on. I would be happy
to give people more details if you think it would be useful.
- cks
|
d387ad0f-eaab-229f-0fb6-691ab5e77135 | ['7660d094-796a-6db2-2b39-4077ac68533a'] | <PERSON>
I'm trying to work out the best way to use Puppet to control an
existing estate, and then how to refactor the configurations. I'd like
to use the appropriate tools where I can (eg taking distros to get
consistent sets of packages, package managers for the bulk of
deployment). During the adoption phase, I think that I want to
identify configuration changes from a standard, known, installation
image + kickstart file. The configuration is the details of the
changed file contents and other metadata (eg existence, permissions).
I can fairly easily identify changes from a standard build (eg find
all packages and verifying them all, picking out the failed
verifications - I'm not sure how to find those filesystem changes that
are not part of any package, but leaving that aside for the moment).
As I want to use standard builds, I'm basing the effort on fedora
builds (puppet 25.5). What seems to me much harder is how to produce
a manifest that can reliably reproduce the existing systems, and
demonstrate the reliability of the manifest, for more than a couple of
machines.
I think that I need to do something like:
1- clone the filesystem of each host that needs to be put under
control
2- find the files to put under puppet control (ie those filesystem
changes that are not the result of the kickstart and subsequent update
processes).
3- build a manifest for the changed files (I tried ralsh file for
this and came seriously unstuck :-( )
4- create a test VM and run the manifest against it
5- produce automated comparisons between the filesystem on the VM and
the cloned filesystem from step 1: this comparison must be able to
handle issues like 'the contents of file x should match the cloned
contents, except where this IP1 address is present, when IP2 should be
used.'
This feels quite heavyweight, but when I tried a much lighterweight
approach of running smaller parts of manifests I was finding that I'd
need to clone the filesystem in any case, otherwise I couldn't
identify any changes that Puppet had made, (nor revert back!!)
Have I over-egged this? Are there any simpler approaches that I should
try - I considered starting from scratch, but it's non-trivial
identifying all of the IT services that are set up, let alone
reproducing their behaviour from scratch?
cheers
<PERSON> | c4688e3e-203c-04dd-88a0-245f08832a49 | ['7660d094-796a-6db2-2b39-4077ac68533a'] | <PERSON>
I've not used Puppet yet, so this could be an obvious misunderstanding
on my part. I've seen many manifest entries for files that write rules
about file ownership and permissions. Surely these types of attributes
should be managed through the package managers, with a possible
override if there are any package errors. Otherwise, you've got to
replicate an existing set of information in a new format.
Am I missing something, or is there a simple way of working that just
uses the ownership and permissions available in the package managers?
cheers
<PERSON>
|
1f35c596-f9a8-7689-4057-f635170b2b27 | ['766ff8a9-cc5d-18e2-a931-214e47a620b5'] | Hi folks,
I would like some information on buying a Linux supported wireless G card for my IBM T23 laptop. It supports PCMCIA type II cards. I want something supported under Linux obviously :-) but without requiring the use of ndiswrapper if possible currently. Can people give me some info on their experiences with any cards they would recommend. Please be specific as to exact make/model.
Thanks,
<PERSON> | 56804f70-6b52-798f-30cd-87dca0d7a6dd | ['766ff8a9-cc5d-18e2-a931-214e47a620b5'] | Hi gang,
Anyone know of any Zope/Plone web hosting providers in Toronto
or Canada. For a decent rate of course :-) Also for .ca registration as
well. I need to register a .ca domain for cheap as well.
Thanks |
896aa373-de54-1d02-ef5d-9bf89d53fb2c | ['767cac0b-6676-963a-9f4e-ccd72df62d3e'] | Hi Team,
I am a Masters student at University of Florida. I am presenting Apache Jena to our class as part of research on RDF and semantic web. I have got material to speak about Jena but I am having trouble finding some use cases of Jena in current industry.
I read that many Semantic web applications use Jena but I could not find any project which uses Jena. Can you help provide names of some projects which are currently using Jena and how Jena is being used there? That would be very helpful to generate interest among students.
It would be helpful for future if we had a page on the Jena Homepage about Who is Using or something similar.
Thanks & Regards
<PERSON> | f2af1167-bc8f-0c2e-4acd-c2c3ba361b6d | ['767cac0b-6676-963a-9f4e-ccd72df62d3e'] | Hi,
I am trying to implement text search in Jena via Fuseki. I have followed through the documentation and created assembler file.
But after starting fuseki server using config parameter - there is no data in the dataset and thus returns no results for simple query or text query.
What I have tried so far
1. Built the TDB dataset using java -cp $FUSEKI_HOME/fuseki-server.jar tdb.tdbloader --tdb=assembler_file data_file
2. Built the index using java -cp $FUSEKI_HOME/fuseki-server.jar jena.textindexer --desc=assembler_file
3. Started the fuseki server using fuseki-server --config ../assembler_file.ttl
I tried the answer on Stackoverflow http://stackoverflow.com/questions/30447536/fuseki-indexed-lucene-text-search-returns-no-results
but using --desc gives error no service name.
Another user had similar issue a year ago as in this thread http://thread.gmane.org/gmane.comp.apache.jena.user/7892 but there are no solutions there as well.
Can someone please help here?
Thanks & Regards
<PERSON> |
1ef7169f-e4ea-8b24-7d7c-5c74713d3b5b | ['768dc605-071e-4bfe-caea-7ec3072a017d'] | Hopefully this is not too off topic.
I have a server in which I'm creating an Exchange Information Store
backup using ntbackup. The .bkf file it creates is 112GB. My goal was
first to rsync it over to a local linux server, and then rsync just the
changes over the internet to another linux server. I do this with other
Exchange .bkf files (in the 4GB to 20GB range) and it seems to work well
only transferring the changes over the internet.
My usual method has been to map the Windows 2003 drive on the local
linux server and then use rsync. When trying this with this system, the
transfer speed is super slow from the beginning, at 6 mbit per second on
a 10/100 mbit network. If I try the copy using smbclient and get, I
actually get 60 mbit p/sec for the first 30gb, then it seems to slow to
6 mbit. A regular cp does basically the same thing as it starts out fast
and then slows all of a sudden. This obviously will not be very
practical as on the local copy I usually would copy the entire file each
night. At that rate it will take days.
Any suggestions on how to copy a 100gb file over a local LAN quickly? Or
what copy/transfer command might keep the same quick throughput during
the entire transfer? Or could there be something wrong on the network?
Thanks,
<PERSON>
| 11a14bb7-d32c-7bf5-e708-6b76e4de2493 | ['768dc605-071e-4bfe-caea-7ec3072a017d'] | Hello,
I've been trying to figure out how to control the space required for
backups using rsync, large files, and an incremental backup scheme. In
particular, I've got two customers in which I'm creating Exchange
backups using the built-in MS backup, then rsyncing an exchange.bks file
nightly. One customer has a 4.3GB exchanges.bks file, and although rsync
works wonderfully by only sending 43mb worth of changes to this file
across the line, using hard links obviously causes incremental
directories of at least 4.3GB daily. With 30 days of incremental, this
adds up.
I've looked at rdiff-backup, but do not like it as it's so different
than rsync. I've got a lot of scripting built around rsync and I like
the way it handles file/dir selections, so I will not consider
rdiff-backup.
I've been experimenting with rdiff itself and trying to figure out a
formula for creating delta files of only the changes to the Exchange.bks
file and using these delta files in the incremental directories, instead
of creating the hard links. I've gotten it to work, but not in a
rotational scheme where you always save only the latest copy and the
deltas, and can easily restore (patch) using just the latest and the
delta.
Has anybody found a solution for this? Or is there any other tool out
there that automates this? Again, rsync works great for the transfer,
but storing data incrementally is becoming the problem.
Would be nice if a future version of rsync created rdiff incrmentals as
an alternative option to 'cp- al' or 'link-destination'.
Thanks,
<PERSON>
|
c431220c-c2dc-4b50-116d-f63aeb51cce1 | ['76a4b395-71b1-2c2f-36f9-45af3fc0a87e'] | All,
I've been struggling to find a neat solution to this problem for a couple
of weeks now, so I'd love to hear anyone else's opinion. I'm on Rails
3.2.13 and Ruby 1.9.3-p392.
I have a table of train schedules, and a table of train schedule locations
with a has_many association. Each schedule has a set of validity dates,
e.g. each Friday between 1st January and 28th September. The times in the
schedule locations table are all 'seconds since midnight on the day the
train runs': where a train runs over midnight, the number of seconds is
greater than 86400.
Whilst this is reasonably 'clean', I want to re-engineer the schedule
location model to return a Time object for each location, calculated on the
fly based on the date I queried the schedule model for, i.e.
Schedule.where(:runs_on => Date.parse('2013-01-01')) will return a Schedule
model with an instance variable 'runs_on', and when I request the
associated schedule locations, the 'arrival_time' and 'departure_time'
fields are auto-calculated based on the instance variable in the schedule
model.
Is this a sensible way to do things? Is this even possible or am I
over-engineering it?
<PERSON>
| ef40927c-5638-c465-48c2-0ff8d3b8cb6f | ['76a4b395-71b1-2c2f-36f9-45af3fc0a87e'] | Hello
I have a user coming from a Vodafone /23 who's getting a 503 error when
trying to access one of my sites over HTTP. I don't even see any traffic
from this /23 when the errors happen, so I'm fairly sure the problem is at
Voda's end.
What's the best way, as a "not a VF customer", of reporting this to
Vodafone?
<PERSON> |
cd27b880-eab9-721b-db77-2876dbbbd3ea | ['76ae23c8-62a5-9c28-8578-15d52b24fcf8'] | Hello Everyone,
I am a returning XFCE user from several years ago. I've used KDE for
years, but I've wanted something a bit more straight forward and
remembered the feel of XFCE when I first transitioned to Linux almost 10
years ago. Awesome to see the updates that have been made since then.
Great job to all! The only minor hurdle I encountered was the XFCE4
wiki. I was looking for window shortcuts, but the information on the
wiki was totally incorrect. To the pages credit, it does mention that
the information their needs to be brought current with the most recent
version. Keybindings are usually the first thing I look for. They help
me to keep general navigation and window management as efficient as
possible. Not finding them, or getting inaccurate information on the
very first search was a bit disconcerting. If possible I would like to
help updates some of that information when time allows.
<PERSON>,
| 9d146312-47db-d3ae-1666-686ae99f4362 | ['76ae23c8-62a5-9c28-8578-15d52b24fcf8'] | I had an awful time today trying to install the HighLine gem on my
machine at home. After several attempts, I tried disabling my firewall
on a whim, and presto, the gem installed successfully. I don't recall
having this problem with other gems in the past, so I am wondering if
my trouble had to do with where the HighLine gem resides? Is there a
specific port that gem installs communicate on? How should I setup my
firewall to allow gem installs?
Thanks
<PERSON> |
2dbcbcc4-c9d0-0742-f06c-2e6ff5569a54 | ['76b14da3-51e0-7aa6-0d3b-29673a73b91f'] | Hi there,
I am trying to set up Freevo in my Linux box.
Because I have an DXR3 that I want to use as the TV output, I am following the Wiki documentation on how to configure Freevo to use DXR3.
The part on configure DXR3 drivers(?) in the system went fine.
But when I try to compile ffmeg I keep getting a segmentation fault error.
Could any one help me?
My dist is Mandrake 8.1 on a AMD K6II-500 CPU.
ffmpeg version is the one mentioned in Wiki.
Is there any other way to put freevo and DXR3 together?
Thx
| a03d9c71-3574-e7e7-dffe-411206ebe334 | ['76b14da3-51e0-7aa6-0d3b-29673a73b91f'] | Hi there.
I just have a few questions.
1: Divx with subtitles - I know that mplayer supports it. And Freevo? If so, how? If not, why not?
2: libcss - I have installed libcss. Is there anything else to be configured to allow me to view encrypted DVDs?
3: DVDs original menus - Is there a way to start a DVD like in a normal dvdplayer, starting with the dvd menu and choosing the audio, subs, chapters, etc from there?
4: Audio CDs - I can't play audio cds. When I insert an audio cd it will not start, and in the music menu, freevo tells me that no cd has been detected.
Rafael M. Torres (<EMAIL_ADDRESS><PERSON> (oLi61mfjL3CYzPYX@example.com)
|
3d0f460d-0ebc-62bf-0e99-accbf7ba9fe6 | ['76b55302-ef9c-407e-8642-f62a4c3502a6'] | When I try to setup smtp authentication with sasl on 9.3, my mail log
tells me postfix is not compiled with sasl supprt. Huh??? I thought
leaving out mysql lookup support was shortsighted, but his is worse.
Does anyone know of prebuilt packages for SUSE 9.3 that supports mysql
and sasl auth?
Thanks
<PERSON>
| ad7cba17-4570-9a12-478b-8cd106693e25 | ['76b55302-ef9c-407e-8642-f62a4c3502a6'] | Hi guys,
I need to add an ASP site on a web server running Apache2 (SuSE 9.1
iirc). This server already has a large number of sites on, so I don't
want to put something on it that will cause trouble. Of the APS
implementations for Apache2, http://www.apache-asp.org/install.html
appears to be the most popular.
What are your experiences/recommendations?
Thanks
<PERSON>
|
d76b0e42-6675-bd22-deb2-15afaf47ee85 | ['76b89d92-d803-c1c3-f4df-79f4e415f8a9'] | Hi,
I was analysing whether one should go for Post-compile weaving or
Load-time weaving. I read on the bea site
(http://dev2dev.bea.com/pub/a/2005/08/jvm_aop_1.html) that Load-time
weaving (using agents) affects both scalability and usability. Are these
issues of scalability and usability applicable to post-compile weaving as
well?
Any help on the technical advantage of one over another would be highly
appreciated.
Thanks,
<PERSON>,
| eff152dc-538c-4e41-c8cc-ab79638224be | ['76b89d92-d803-c1c3-f4df-79f4e415f8a9'] | Hi <PERSON>,
This is in context to the article on the bea site
(http://dev2dev.bea.com/pub/a/2005/08/jvm_aop_1.html). They have talked of
the problems with Bytecode Instrumentation-based Weaving, namely double
bookkeeping: building a class database for the weaver, instrumentation is
inefficient. I am not clear after reading the article whether these issues
would be there with post-compile weaving as well. (I have posted the
question on their site as well.)
Thanks,
<PERSON>,
|
4d66495f-b85e-6f53-c944-03b20e655601 | ['76ba4f10-ecaa-4950-bcf8-ff1a84ba28ae'] | Hi <PERSON>,
this looks promissing as we could always use the latest and maintained OneMediaHub app on our devices :)
In using it this way, is also Notes and task supported?
I use funambol in combination with SOGo (groupware). via a s4j SOGo connector synchronization takes place not against the funambol databases but (for contact, agenda and tasks) against the SOGo database.
Will this work (in theory) using your connector?
Thanks,
<PERSON>. | 9a1563bc-f175-a2b8-6175-192ce6ce8ab0 | ['76ba4f10-ecaa-4950-bcf8-ff1a84ba28ae'] | Hi,
last week received my new toy (Galaxy Tab 10.1) so finally have good use for funambol synchronisation (contacts, tasks and notes) with my SOGo server and android phone :)
Ran into a small issue:
the following two settings are not persistent (when you set them click [save] and re-enter the config screen they are disabled again):
* Sync with server [Push client changes to the server]
* Contacts [Always create new contacts in the funambol address book]
Android version is 3.1 (Honeycomb)
log file (trace level) attached.
Hope this helps,
regards,
Ruud. |
30001fa6-a842-1308-3c80-c09a5f9c688f | ['76bb6608-a98a-c190-8229-8b79254e3d21'] | Hi all,
Thanks for your help so far, very useful!
I ran into another thing:
I compiled the same website using Visual Studio and using a nant script.
I compared the dll and its satellite dll's with the Loetz Roeder's
resource tool to see what resources are in there and what the names of
the binary resource files in the DLL's are.
The satellite assemblies are identical so that's fine.
But there are differences in the main dll's:
The names of the included binary resource file differ:
The Visual Studio build uses the name of the class that the resource
belongs to.
Example:
phw842_frm1.ascx.nl.resx belongs to phw842_frm1.ascx.cs (In Visual
Studio, the is represented by the fact that this resx file 'hangs' under
the ascx file and in the csproj file this is specified in the
'DependantUpon attribute.
My nant build turns this resx file into phw842_frm1.Resources (resource
file name is based on resx file name) and adds it to the dll).
But the Visual Studio build turns this resx file into phw842_1.Resources
(resource file name is based on the classname in the code file the resx
belongs to (this class name differs from the name of the code file!))
and adds it to the .dll.
The result is that resource managers in the code who try to load the
binary resource file phw842_1.Resources by using the type of the class,
won't find it.
How can I change my CSC task so it includes resource files with names
that are base on class names (types) instead of resx file names?
The way I compile the files in Nant is by just listing all resx files in
a resources node (see below example code), I guess I have to somehow
specify the links between code files and resource files so Nant can
figure out the names for the binary resource files?
Note that for the satellite assemblies, I parsed the csproject files and
code files with a program I wrote myself to get to the classenames
(types) that must be used for the names of the .resource files.
Thanks,
Bert-Jan
| ee736724-175b-f882-b0dc-0038995c09de | ['76bb6608-a98a-c190-8229-8b79254e3d21'] | Hi,
I am running into the following problem:
I use a nant script that has produces an exe. It has a references to
Interop.TGXLib.dll and AxInterop.TGXLib.dll (wrapper dll's).
Everything compiles fine.
Now, when I run the program I just compiled, it is looking for dll's
TGXLib.dll and AxTGXLib.dll, but it can't find them because these names
are wrong, it should look for Interop.TGXLib.dll and
AxIntererop.TGXLib.dll.
If I rename my interop dll's to TGXLib.dll and AxTGXLib.dll, everything
works fine.
So the names of the references in the exe are wrong. The part 'interop.'
has dropped off.
Can I set these names somehow in the Nant script to force it to make
references to 'Interop.TGXLib.dll' instead of 'TGXLib.dll" ? I don't
want to rename files to get it to work.
Maybe in the references tag for the CSC task or something? For example
<includes name="Interop.TGXLib.dll" refid= Interop.TGXLib /> or
something...
Let me now,
Bert-Jan
Bert-Jan Lappenschaar
|
da8917ff-26a7-db52-9bca-1df3b55504d7 | ['76c2a5f0-8da4-0422-8679-f0e2fdee95e6'] | I'm in the process of preparing a little lightning talk on Mondorescue
( I might post a draft here ) and while prepping for it I found there
were not a lot of screenshots of mondorescue in action lying around.
Anyway I ended up making a few for my talk and I thought other might find
them useful. If possible could they be put on the website.
The screenshots are my own work in a private capacity and are released
into the public domain, or in case this is not legally possible I grant
any entity the right to use this work for any purpose, without any
conditions, unless such conditions are required by law.
<PERSON>.
| 28e72f48-b739-7137-bfbc-4ec296199795 | ['76c2a5f0-8da4-0422-8679-f0e2fdee95e6'] | A reminder to all list members that:
1. DNS related questions should usually be sent to more specific lists
such as DNS operations:
http://lists.oarci.net/mailman/listinfo/dns-operations
2. Discussion regarding the NANOG organisation and political issues
surrounding it are off-topic for the main list and must only occur on the
nanog-futures list
http://mailman.nanog.org/mailman/listinfo/nanog-futures
<PERSON>
NANOG Mailing List Committee
|
49c756a9-8444-e447-efa7-bc980dd8cad6 | ['76c84d40-6c12-82fa-9555-95322ef6d385'] | Hi,
I am trying to use GnuGK 2.0.6 on Linux with OKI, Cisco ATA186, and
MultiVoIP 800. I started with the simplest config file (actually empty,
just 42 and allow). Endpoint registration seems fine, and call setup
succeeds for any combination. Also, if I call from OKI to OKI,
everything works fine. But if I call the Cisco or the MultiVoIP, I can
only hear voice in one direction. This is not a NAT/firewall issue,
because everything is just on my LAN. The codecs seem to match (723 in
all boxes). Call setup seems fine, and I can not notice anything weird
with the GK's "-ttt" output. Any pointers to the right direction where
I should dig further?
Boris
| 196e1329-2725-5ddc-6f9e-c55c566a4516 | ['76c84d40-6c12-82fa-9555-95322ef6d385'] | <PERSON>,
In such cases I would say it is time to attach the serial link and see
what's going on with the booting. As you have already used the
bootloader to put Familiar on your iPAQ, this should be already familiar
to you: start minicom (make sure the parameters are right),
stylus-hole-reset while holding the center of the joypad, connect iPAQ,
use bootloader prompt.
I always specify ttySA0 as my kernel console (I vaguely remember this
was off by default?), that way I can see how far the kernel boot goes
without any additional manipulations.
<PERSON>
|
067fc220-81ed-669b-064d-49d680d41c68 | ['76ceeac9-f238-d9b7-e64b-24be41709401'] | Hi <PERSON>,
I have entries in rtrules to force traffic to a certain provider, i also
have squid running on the firewall i want to force traffic for certain
websites over a certain provider (using lo interface).
This works however when testing if i simulate a failed link, LSM detects it
but traffic that is defined in rtrules does not want to route, is there
anyway to force traffic over provider1 and if provider1 fails traffic will
go to provider2
Thanks
| 2bfaf219-5cdd-177b-9f63-bf7eeed42e20 | ['76ceeac9-f238-d9b7-e64b-24be41709401'] | Hi i am using Squid 3.3.9 with Kerberos authentication on my network.
we know have a requirement where we need to give guest users access on
the same proxy, is it possible to run squid on a additional port and
have different ACL's for those users connecting to that port?
I know ideally having a different subnet is the best option
Thanks! |
eee2b868-d90f-c680-753f-9ec3a0cb549b | ['7701653c-3db5-6543-b67c-b5027ff78a93'] | Greetings,
Im recent to MIT Kerberos. Currently Im working on a project at Faculty of
Engineering University of Oporto implementing an OpenLDAP server with MIT
Kerberos Authentication, using Cyrus SASL.
The Linux distribution Im forced to use is Slackware 9.1
Ive configure, make and make install Kerberos v5 r1.3.1 in my system but
the /etc/krb5.conf file is not created, not even the /usr/local/var/krb5kdc/
directory for the KDC database.
My question is Have you found any problems with Slackware 9.1?
Ive already gave a full-read at the Installation Guide along with a look-up
in the System Administration Guide, but havent found nothing regarding to
this.
If you know of any place where I can find information about Kerberos v5
installation in Slackware I would appreciate.
Best Regards
<PERSON> | 4513afc6-e8e4-6b76-42f2-f23e1fe2d476 | ['7701653c-3db5-6543-b67c-b5027ff78a93'] | Hi!
I've a Tru64 v5.1b, with LDAP Module for System Authentication (from Internet
Express 6.1).
I've also installed OpenLDAP 2.0.27, which comes with Internet Express, and have
no problems authenticating users stored in this OpenLDAP 2.0.27.
After I've changed the LDAP Module for System Authentication to use another
server (not on localhost) wich runs OpenLDAP 2.1.30.
At first, the LDAP Module for System Authentication could'nt even pass the Test
Configuration, then I've put the line "allow bind_v2" in the server and the Test
Configuration was passed. But UNIX doens't authenticate users on that server.
Even if I perform a "su username" or try to login with an account stored in that
LDAP server, the attempt is rejected and there is no traffic in the server side.
Is there any incompatibility with other versions beside 2.0.27?
Another question I have is, is it possible to use Tru64 Unix to authenticate
LDAP users trough SSL (port 636) or this isn't supported yet and must all be
trough cleartext (port 389)??
Hope you can give me some light here.
Thanks,
<PERSON>
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.