query_id stringlengths 4 64 | query_authorID stringlengths 6 40 | query_text stringlengths 66 72.1k | candidate_id stringlengths 5 64 | candidate_authorID stringlengths 6 40 | candidate_text stringlengths 9 101k |
|---|---|---|---|---|---|
85fd551d-0b5c-1b30-31be-6d5ef404181b | ['58949159-68f3-525c-2433-51a30f62ebb3'] | .
Hi, fellow 'Ham-sters'
I hope everyone had a great Christmas, and that <PERSON> was good to you.
I arrived back in UK yesterday after spending 7 days with my family in the Gambia, and happily got away before the hold-ups and bad weather here, returning to find all the snow had gone !!
Back to reality for another year!!
Great news!!! I ordered a UT106 from USA for 98 dollars (56 pounds)two weeks ago, and it got to this address yesterday just after me. I was amazed that I had no duty, VAT or anything to pay on it!! Cheapest here was <PERSON> at £99. The guy is still advertising them on US Ebay, so if you want one - jump in quick !! (RF Components - Walter Zapata).
I have not heard much about the UT106 in use, so was anxious to get it fitted (a doddle!), and try it out in the PCR 1000.
When installed, it automatically places another 'rack' on the ICOM component receiver which is toggled on the tool bar at the top with an Icon called 'DSP' - which even if you don't use the Icom software, is a good check that you have installed it correctly. I didn't the first time, as there is a little 'ribbon' connector that you have to push in to a connector on the UT106 and the IC1000 main board, I got it in the wrong way round!!! (Well, I am 75 with only one brain cell!). The scanty 'destructions' are not too clear on this - but after opening up the box again (8 screws), I then tried changing all the connections with the case open until finally working.
I use 'Talk PCR' as my software interface, and usually MixW as my PSK 31 program, or Ham Radio Deluxe, with MMSTV for Slow Scan TV, and my ears for 'Phone' SSB or CW (Morse)... if the morse goes over about 25wpm, I use CWget. When the UT 106 is installed, with 'TalkPCR' you get a little window on the screen away from the receiver, which allows you to switch it on and off (to use DSP), switch the Noise Limiter on and off, and vary its effect with a little slider, and to switch in and out the notch filter.
Hopefully thinking logically, I decided that to test this little Gizmo, I couldn't rely on my ears - nor did I have any of the sophisticated test equipment required to give a practical and subjective appraisal. What better than the 'waterfall' or 'Spectrum' viewer on a PSK31 program? .... I could SEE the results of clicking the DSP (Digital Signal Processing) in and out, as well as the effect on the 'noise' shown on the waterfall when varying the
gain on the Noise Limiter.
VERY IMPRESSIVE ... on 7.035 MHz at 9pm last night (Wednesday)... PSK 31.. pulled in stations that were hidden in the noise prior to switching in the UT106. Noise level at about half way. I then went to bed, as had been travelling and awake for the last 36 Hours - not good for this old fart !! Also wife not best pleased that I had done nothing else since getting out of the Taxi home, but to install and mess with radio gear, and eventually going off to bed without a 'Good Night' even, let alone the usual little 'peck' !!
Thursday Morning - 9am - 11.30am - (Other half still in the land of dreams), have been on 20M (14Mhz) Amateur Band. Listened to some RTTY and PSK, (still impressed, and can HEAR the difference when I switch on and off), but now listening to 'Phone' USB - again... Wow!! the UT106 really sorts out the wheat from the chaff, quiet signals, just 'leap out' - Compared all the time, by the way, to switching to the same frequencies with my Kenwood TS520SE Transciever and my 'Frog' Yaesu FG7700 ...... this Receiver totally out performs them both.
Ooooops !! - 'She that must be obeyed' has been activated !! Have you made coffee?, have you fed and walked YOUR dog? Did you go to bed? Are you still playing with your toys?
Oh dear ... I had better switch to 'wife pleasing' mode! and I havn't listened/tried CW yet!! Better go folks - or I won't be having a 'Happy New Year' !!!
I really hope you all do ... (Have a Healthy, Happy and Sunspottted New Year)
Further impressions on the PCR1000 with UT106 will have to wait until next year, I guess. (Its only 2 days!!).
73's <PERSON> ....... or,
Old Fart John
| fe8c0df9-7800-af26-8400-06ae443ec482 | ['58949159-68f3-525c-2433-51a30f62ebb3'] | .
I have just uploaded two (printable - 20-odd pages!!) Files into the Files Section. Firstly a complete list of Amateur Call Sign Prefixes, listing the Associated Country.
Secondly, a file (1 Page!) of Amateur Mode Frequencies for each Band for the various Modes, - CW, PSK31, RTTY, SSTV, etc for each Band.
I hope these are useful to someone !
Regards,
<PERSON> |
bb5ee638-d37a-3923-924c-9d56a90cef2a | ['5894d137-2412-fb94-a876-65fbb2b66a58'] | Hey,
I got the whole system to work. Also connected some cpes to genieacs.
However, now i would like to directly call the rest api without the gui,
but i cant find how the urls are built up and what
functions/functionalities there are. I was looking in api-functions.js and
nbi.js since that looked like the places to find this. Could you give me
any directions as to where i can find this information. If i can find and
use the possible calls, im also willing to document these for you.
Kind regards,
<PERSON>
| 67476c9b-e7c3-983c-697c-1b3f2dc60bb1 | ['5894d137-2412-fb94-a876-65fbb2b66a58'] | Hey,
I am running GenieACS with different CPEs, so far I had no issues
connecting them to the ACS. However, now I am trying to connect an Icotera
device to the ACS. The ACS does not seem to like that. When the CPE sends
an inform request the CPE returns a 500 Internal Server Error.
I have attached a packet dump file taken at the ACS. Does anyone know how
to solve this? Might this be a bug in the server?
Kind regards,
<PERSON>
|
bdba24c9-9a6b-da0e-9038-b9cf2aa818f4 | ['58a7baf2-5218-51a4-e3e1-f468e6e8be91'] | Hi,
Can anyone please tell me what does "permutation_test_score"(and the p_value it returns) do in scikit-learn? I am assuming it outputs the statistical significance of the performance of regression models. I am planning on comparing the performance of various regression models if the performance measure they are reporting is statistically significant. To this end, I wanna output the p-value of the prediction first, and if it was smaller than a certain cut-off, I would then report the performance metrics, such as r2 and MSE.
Do p-value and score outputs from "permutation-test-score" not provide me with what I want?
<PERSON>
| 522e0a6a-89c6-2ed3-13df-6390560d5354 | ['58a7baf2-5218-51a4-e3e1-f468e6e8be91'] | Hi folks,
I have a question regarding how to use permutation_test_Score. Given data X (predictor) and Y (target), I hold aside 20% of my data for testing (Xtest and Ytest) and would then Perform hyperparameter-tuning on the rest (using Xtrain and Ytrain).
This way I can get the best parameters via RandomizedSearchCV. I now want to call permutation_test_score to compute the score, as well as the p-value of the model prediction. But the question is what X and Y should I send as input arguments to this function? I could send in X and Y but then my hyperparameter parameters were already tuned to Xtrain and Ytrain, which are a part of X and Y and that would bias the output values. Any help would be greatly appreciated.
Thanks,
<PERSON> |
3945767e-1a75-7d49-5c44-af17f682935e | ['58a9c86e-4812-b84d-c515-847bd8c6d502'] | Hi folks.
I'm just wondering if there is any empirical or analyst data on the
adoption and effectiveness of Agile methodologies in various types of
organization. It's clear that there are advocates, detractors, critics
and zealots on all sides, so I'm thinking empirically here.
Specifically, I'd be interested in data or stories of Agile in the
development of enterprise products with a large, diverse customer base
and/or the development of product suites that must share common user
experience attributes.
Feel free to contact me offline, to point to references or to engage the
group.
Thanks much!
| 88b1d699-f383-8f7b-9312-599318f2c82f | ['58a9c86e-4812-b84d-c515-847bd8c6d502'] | Hi there folks -
Boy, who could resist a goad like that?
I'm a usability engineer at a Northwest software company with about 400
people. We have several dev teams that use different methodologies.
One of these teams uses "Agile", and I put that in quotes because I
can't say that it's "real" Agile. Rather, I have three assertions:
1. "Agile" is like soup stock - it's a set of tools/practices with
properties that are beneficial in some cases. As with any ingredient it
can be modified, spiced differently and made to be what is useful within
an organization. Dogmatists begone!
2. Agile is not a frictionless surface. It is not product
development nirvana. Agile can suck mightily, or simply be ill-suited
for the project to which it is applied. There's a higher-level debate
in business software about the locus of control between IT groups and
Business groups. I think it applies to product development
methodologies as well, and Agile seems to fall on the side of the
Dev/Test side of the house.
3. Sadly, I think Agile can also be used as way to suggest "We can
do more - quicker! Better!". This is a "mom and apple pie" argument
that upper managements loooove to hear. But there's a cost, I think, in
larger projects that end up being done haphazardly.
Of course, none of this applies to my specific situation.
Thoughts?
Thomas
|
a331ac0a-03a3-72ce-0110-50e77515efbe | ['58a9d7bc-403c-c3a3-7d97-f0dd47b75b6b'] | Hi,
i built my first kernel the Debian way and filed
a kernel bug. To Debian for now:
https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=798300
I decided not to mention ISO producer programs by "affects".
They work fine with the kernel.
How long should i wait before trying at upstream ?
(Urgency is low. But when it begins to smell ...)
Have a nice day :)
<PERSON> | 77c4242c-9477-f95e-ae60-dd5b4b696f63 | ['58a9d7bc-403c-c3a3-7d97-f0dd47b75b6b'] | Hi,
i can reproduce the SIGBUS only when gcc got option -O2.
(I removed all -O2 from ./configure, ran make clean and make.)
I can silence the error, with -O2 enabled, by replacing
a->u = *(union w_list_data *)data;
with
memcpy(&(a->u), data, sizeof(union w_list_data));
So this seems to be a bug of gcc, unless somebody can point
me to C prescriptions which would declare the "a->u =" variant
undefined or elsewise ambiguous.
I will test the error preventing code change in upstream now
and will apply it if no obstacles appear.
But we have to expect more such pitfalls if -O2 stays enabled
in the Debian compilation of libburn, libisofs, and libisoburn.
I really would had thought that -O2 was safe with gcc.
Now i have to reconsider.
I propose to try whether the strcmp() problem of genisoimage
in bug 731806 is related to -O2 too.
Currently i am waiting for <PERSON> to provide me with the
input directory tree which on the first hand caused this bug
report. Just to be sure that no other bugs pester this use case.
Have a nice day :)
<PERSON> |
67752b0a-b3ac-85fc-14ce-cc957f2de62c | ['58c8a046-8aa9-ac58-b015-db643d42e501'] | Hi There,
The discussion about speex triggered the following question:
Is there a way to use other encoders, to stream with icecast / oddcast.
Is there an format of witch the encoder shoul be standarized to (ACM or
something like that), if so would it be possible to use open license
encoders and commercail encoders with this. like FHG mp3, mp3pro, speex and
so on.
regards,
<PERSON> | 24d10a43-049b-386a-9971-2afd92d4648b | ['58c8a046-8aa9-ac58-b015-db643d42e501'] | Hi There,
Is there a nice helpfile to telle me how to set up:
* Multi-Level Fallbacks
Multi-level fallbacks allow for specifications of a series of fallback
mounts that you could use, for instance, to set up a series of
progressively lower bitrate streams that would be cascaded through. For
instance, a modem user would connect to the highest bitrate stream, and
then could be cascaded down to a progressively lower stream until they
reach a value they can handle.
When i put up a faalback, do i have to remove other normal settings in my
configuration of icecast 2.1.0?
Regards,
<PERSON> |
fb7ba67a-3416-171d-b847-e6c712b8d7af | ['58d1b7af-5705-6c3c-43ea-ae3a9d70f2a5'] | Hi all,
I'm looking for a way to serve random error pages for 504/502 errors.
In addition i also want the error pages to be served based on the
$request_uri.
Something like this:
1) When this URL is requested: http://mydomain.com/search?type=ABC
and if the proxy returns a 504 or 502 and when the GET parameter 'type' has
'ABC' as value then serve
any random document from "/path/to/custom/error/typeABC/"
2) and the same applies to http://mydomain.com/search?type=DEF just that the
directory to be used to serve
error document should be "/path/to/custom/error/typeDEF"
I'm sure this can be achieved with named locations and probably try_files.
But I'm just not able to get anything working.
Would appreciate some pointers.
Regards
Harish
| 317ba6ed-7167-fb78-8280-65b45b10f9db | ['58d1b7af-5705-6c3c-43ea-ae3a9d70f2a5'] | Hello All,
I have a few questions on backup servers and proxy_next_upstream in general:
For the backup server the doc says: "only uses this server if the
non-backup servers are all down or busy"
1) So what exactly constitutes down or busy? Is it just response codes
- 502,503 and 504 or anything else?
2) Will the same request which received such an error be sent to a
backup server or is it only subsequent requests?
3) Can I say 504s happen when proxy_connect_timeout/proxy_read_timeout
is reached? Is there a way to find out which among these is happening?
4) Also, In proxy_next_upstream :
error — an error has occurred while connecting to the server, sending
a request to it, or reading its response;
timeout — occurred timeout during the connection with the server,
transfer the request or while reading response from the server;
I'm not able to clearly distinguish between both of these. What kind
of scenarios explain these?
5) In a case when proxy_next_upstream is off and when a backup server
is provided, will nginx still try all non-backup servers before
sending it to backup servers?
Thanks
<PERSON> |
96863daf-8f3f-5bb5-f079-7841429d11cd | ['58f61491-993b-0f58-a7bb-d3d5188fe2ee'] | I'm trying to use both these packages and wondering whether they are possible...
To make this simple, my ultimate goal is determine long stretches of
1s, but I want to do this within groups (hence using the data.table as
I use the "set key" option. However, I'm I'm not having much luck
making this possible.
For example, for simplistic sake, I have the following data:
Dad Mum Child Group
AA RR RA A
AA RR RR A
AA AA AA B
AA AA AA B
RA AA RR B
RR AA RR B
AA AA AA B
AA AA RA C
| dceadccd-a18b-1e97-b701-1637d8524336 | ['58f61491-993b-0f58-a7bb-d3d5188fe2ee'] | Hi All,
I'm trying to merge two files together using:
combinedfiles <- merge(comb1,comb2,by=c("Place","Stall","Menu"))
comb1 is about 2 million + rows (158MB) and comb2 is about 600K+ rows (52MB).
When I try to merge using the above syntax I get the error:
Error in merge.data.frame(comb1, comb2, by = c("Place","Stall","Menu")) :
negative length vectors are not allowed
Is there is something that I'm doing wrong? I've merged larger files
together in the past without a problem so am curious what might be the
problem here...
Thanks in advance!
~K |
5c9dfe0e-18c3-55a0-0092-4927f2833f19 | ['58f8f2b2-66a5-f8b6-3ef6-5ff211086bb7'] | Hi everybody,
it has been a long time since i've been toying around with my wiimote,
but i remember that it was'n hard to make it run.
Now, i tried to make it functional again (on an Arch Linux system).
Wmgui, hcitool sca, everything works fine!
BUT, the wiimote external doesn't (i downloaded version 0.3.2 from
puredata.info). It compiles fine, but discovery
doesn't work (again, no problem with wmgui!). Ok, i can connect with
explicit address. But still, only button data is
processed, i can't see any acceleration data or IR data ... (i can see
it in wmgui).
A year ago or so, everything worked quite well ...
Any hints ?
Regards,
<PERSON>
| 44e6093f-c72c-ae32-e3f9-c90c39425c3c | ['58f8f2b2-66a5-f8b6-3ef6-5ff211086bb7'] | Hi everybody ...
i've been trying to make haskore-realtime (and, in the long run, haskore-supercollider) work in the year 2014 (OS: Arch Linux, 64Bit), but had no success compiling the package so far, even in a sandbox (directory-<IP_ADDRESS> fails to build ...).
Would it be hard to upgrade to recent library versions ? Though i have some knowledge of Haskell as a language, i'm pretty new to the development system ...
Regards,
n
|
98c7dae1-a654-831c-be40-cc7df6df8219 | ['59132748-20f1-280b-c3e3-7ab439ae7d9e'] | RH 9 Intel
I just got an NCD Explora X Terminal from Ebay - a whopping $15 for the
unit. I wanted to experiment using my RH 9 installation as a platform
for learning this technology.
I figured I'd have a cheap X-Terminal solution - I'm new to this.
However, to my dismay, I've found that I've got to get the NcdWare
software in order to run this unit.
The problem is that the software has a whopping $995 list price!
Does anybody have any ideas where I can get this stuff cheaper, used,
second-hand, etc???
Thanks all.
| 25ff35fa-7e82-3535-7249-353fdd726f0f | ['59132748-20f1-280b-c3e3-7ab439ae7d9e'] | RH 9 Intel
I've moved my hard drives from the built-in IDE controller to a PCI IDE
card. What do I change to indicate to the OS to boot from the new
partitions?
Previously, the OS booted from /dev/hda and there were partitions on
/dev/hdb also. When it boots now, I see it trying to find drives on
hda, hdb, hdc, etc. Finally it finds the OS stuff on /dev/hde. Is
there a way to avoid this and indicate to the process to go directly to
/dev/hd?? to boot.
TIA.
|
587c7e70-476f-9d3b-a174-34a1194d5fd6 | ['5929378a-662c-3889-eda3-59b80acaa31a'] | I agree with others in that the O'Reilly Python books are very good. But if
you want to get up to speed quickly or as a good refresher if you've been
away from Python for a while, I think there is no better book than <PERSON>'s "Python", part of the Visual Quickstart Guide series from Peachpit
Press. Not only is it up to date, having been published this year, but it
includes plenty of quick examples and is organized in a manner that makes it
very easy to find the topics that you are looking for.
-Barron
| 524519f3-da0e-c269-1c30-b47b86317d42 | ['5929378a-662c-3889-eda3-59b80acaa31a'] | Hello,
I'd like to access a database (specifically Oracle) from my Python code and
found cx_Oracle through the Vaults of Parnassus:
http://www.computronix.com/utilities/
They don't offer a package for OS X (surprise, surprise) but one can
download the source:
http://www.computronix.com/download/cx_Oracle-2.4.tar.gz
How hard would this be to compile for OS X? There is no visual interface for
this which I think would make the port much easier. If it's not too hard,
would someone tell me the basic steps (or point me to a quick tutorial)?
Thanks,
<PERSON>
|
ac4d1d0c-bd2e-8d5a-bb60-7a03b72680ae | ['5935edf8-1bc2-992a-bc95-eba55704e664'] | Hello,
I have difficulty with the search method of the re module, when applied to an
object retrieved from urllib.In the following program, f3 is read once and the
search is effective. But when I try to read it again, it acts as if the index of
a file had to be "rewinded" (e.g. as would do a "seek" on a "true" file). I
thought that f3.read was to be considered as a string. Could someone help me
understand my mistake and give me a nice solution (means not storing data into a
file).
Thank you in advance.
| 8cb9ce0a-6ad0-cbcc-5046-4f52009de85d | ['5935edf8-1bc2-992a-bc95-eba55704e664'] | Hi <PERSON>,
Thank you for your investigation. It helps me understand that my python code is
correct. I did not forget any character, while the name of the url comes from
regex extract of the page returned by the server itself. But you're right on the
fact that the name has something wrong, when used in this way.
I tested the web site "manually", that is pasting the entire name of the web
site directly into my Netscape. Same result : "Netscape can't locate the server.
No host specified" (this message translated from French)
Therefore the question is : how does it come that a Javascript function using
'document.location="http://blablabla"' is able to redirect my browser to the
said web site, and not 'urllib.urlopen("http://blablabla")' ?
More, I do not understand the error message. Does it come from the server which
denies the request *or* from the dns which cannot translate the name into an
address *or* from Netscape which seems not to receive/understand the host name ?
Regards
|
61c5aaa2-c5d3-9569-c801-4feb19e252ee | ['59732b8f-68cb-0d5c-3698-7dd2e150d7d1'] | Hi <PERSON>,
I've set up some solutions like this for some customers.
What we have done is the following <IP_ADDRESS>
A web server in DMZ (Apache or IIS, doesn't matter technically, but don't want to restars religious war..)
Exchange server in Lan, with IIS.
Because IIS as to speak RPC with Exchange, and we don't want to see this protocol trough the firewall, Exchange 2000 permit to put an other web server in Front, and requesting only HTTP (A kind of Proxy)
Users have to use a 128 Bit SSL Link from client to Proxy in DMZ. Then HTTP from DMZ to Inside.
Concerning Anti-virus : I hardely recommend Sybari's ANTIGEN. A very good and stable AV for Exchange, using 2 different Engine (List is long. See www.sybari.com)
So POP 3 was not needed.
Concerning Incoming mails, you can add a layer with installing a Mail relay in DMZ. We use sometime an IIS with GFI Mail essential.. Very good and stable. (Win32 shop)
Or you can put some *nix mail, with anti-spamming rule, etc...
A good option is to mix the OS's : *Nix in DMZ, and Win 32 in Lan, in this case.
You can contact me off-list, should you require more detailed info and Schema we already have done...
Useless to say that all DMZ and Intern server have to be fully patchen downstripped, etc..
Sometime, a good IDS in DMZ (Snort... works on *NIX and win32) add some more security.
Good luck
Max
| 95b99663-7230-9cbb-8a23-7771b07dc87c | ['59732b8f-68cb-0d5c-3698-7dd2e150d7d1'] | Hi <PERSON>
Just my 0.02€..
I tried using Nessus on my laptop, within vmware (version 3 fow Win 2000, demo 30 days).
I had no problem with it. Unless with a wireless card, that vmware can't support. In demo version.
It has worked like a charm with a Free BSD minimal version, and with some other distro (RH7.2 and other i don't remember).
I really think about buying a full version of VMware.
So I can stay using Win 2000 for all common stuff, using a win32 client for Nessus, and Nessus runing in a VMWare session.
Very fine and worked like a charm. No need for dual-boot, nor to change the OS, etc... And I can use a graphical client, which i feel more comfortable with.
Definitely a very good solution, perhaps the best, For me at least.
Just my experience.
Max
|
4a6ca489-de14-6c04-8b5d-3fdf18fa8715 | ['599247cf-6555-e84d-6623-cadc982bb269'] | I've got a K2000 that I've created a drum fill program. your basic snare, toms, floor tom
and crash cymbal, to use on transitioning from verse to chorus etc.
the snare is what I want to edit. at the end of the sample, it's sounds gated, like the end is
being sucked down. this I don't like and have played with different alg., env. LFO's...with
no luck. anyone know how to rid the sample of this?
When I access the snare from EditProg*KEYMAP, the snare does not possess this
characteristic.
thx. in advance.
<PERSON> | aa7a015b-1e8d-177d-f618-9c88be8dec8a | ['599247cf-6555-e84d-6623-cadc982bb269'] | my original hard drive started intermittently working, then not working. so I bought a new used hard drive off eBay. I was able to format it and save a couple of files to it, when it spun down. the disk page says it's asleep. when I try to access it, to load or save, it says "problem mounting disk". I'm beginning to think my original hard drive is still good.
If I power down, wait, and power up, the disk spins up, but I get this clicking sound, kind of a knocking and the disk page reads asleep, even though it is spinning.
anyone experienced this one before?
thx.
<PERSON> |
29b6dcf4-7e65-a339-f7a1-0c68065a594f | ['59955b2d-6552-26ac-6cf2-a31c02c62eb4'] | When we run a test suite, some tests might fail. We get a report in
the test-output folder. Now, if we want to re-run the failed tests,
we get a different report.
Is there a way to re-run failed tests and have the original report
updated?
Basically, I have a condition like this: I am doing UI Functional
Automation using TestNG. We have 200 test cases right now. At times,
a few tests will fail. I can re-run the failed test cases, but I then
do not have a full report to send out. I can send out the report with
the failures, or I can send out the report with the re-ran tests. I
would like those combined so, to those looking at the report, it
appears there was just one test run.
Any way to do this in TestNG?
<PERSON> | a5972352-af59-8c33-d050-ef405a2c962d | ['59955b2d-6552-26ac-6cf2-a31c02c62eb4'] | Hi, I am also new to this community and to TestNG. So far, I think
TestNG is much nicer than JUnit, but I am also doing functional
testing rather than API testing.
I noticed the same thing that <PERSON> mentioned, the full class names
are displayed rather than the SimpleName.
I also noticed that JUnit includes the time it takes each test to run
(in seconds). TestNG is only displaying the total execution time in
milliseconds. Having the time for each test I found to be usefull
since I am often asked what part(s) are taking the longest amount of
time. With TestNG, I can not answer that.
Just my two-cents for an improvement to go along with the SimpleName
display.
Thanks,
<PERSON>
|
c4c1b454-5bf5-6341-3e33-614d8aa7b96b | ['59962112-16c1-d942-aeb7-851536308183'] | Hi,
Ok so i have managed to get my html template working in a file. So the
basic examples are like this
e.g. main.go ---> template e.g view.html (section) ----->section is read
from .txt file e.g. test.txt
Now i want the following
e.g. main.go--->template e.g. view.html(page section 1) ----- section 1
read from test.txt
(page section
2)-----section 2 read from read.txt
How would i do that?
regards,
Will
| e075cf9a-2d29-5bbe-694b-3cd0b1f33e41 | ['59962112-16c1-d942-aeb7-851536308183'] | Howdy <PERSON>,
Is there anyway of getting web access statistics from a pure golang server
(built in a way as described in this article
http://golang.org/doc/articles/wiki/)? e.g. number of unique visits, per
day statistics, region statistics. How would one go about building this
feature in a golang webserver?
Note that there is no front-end apache server nor nginx server.
regards,
<PERSON> |
22d07d92-5025-3c0b-e13e-bd677b20ba99 | ['59ad3b12-1bd0-d6a1-8456-d16fcb1bebff'] | I should like to use SQLite from within a VC 2003 NET program. Is this
documented somewhere?
I do not know which parts to dowload, to start with. I downloaded the .dll
but it comes with no .h file and now I am confused. Is there somewhere on
the net a sort of tutorial or at least a C++ example?
TIA,
jb | d468b9e4-8ecc-cb2a-7785-2a74e61a3489 | ['59ad3b12-1bd0-d6a1-8456-d16fcb1bebff'] | For list-related administrative tasks:
http://list.cs.brown.edu/mailman/listinfo/plt-scheme
I have started playing with plt mzscheme. I should like to extend mzscheme
(and later mred) with a C++ file that uses a C++ library.
Now when I compile and link the first "hello world" program from "Inside
MzScheme" it is compiled but not linked as __gc_personality_v0 is missing.
I called the file hw.cpp instead of hw.c.
I have had this error before and it is a C++ problem. What can I do? I use
Cygwin (and gcc3.2).
TIA,
|
d74f729b-b6c5-26dc-80ce-514b074dc034 | ['59b034c1-c23a-c2f8-ca69-40767f4633d4'] | <PERSON>,
I too built up on an open usage platform and yes, when the subscribers
logged into their PowerCode portals and viewed usage charts I got plenty of
calls. We have not yet implemented metered billing because the pipe is
still not capable of delivery, but soon.
What I told the concerned callers was pretty much what I explained
previously, that a small percentage of subscribers are utilizing the
majority of the systems resources and that it was effecting everyone. I
went on to explain how the goal was to charge those that use more services
for their usage, and assure resources remain available for low volume users.
I also add that based on FCC regulations I can not restrict any specific
type of traffic, so this is the only fair way to assure everyone gets what
they want.
I tell them that our pricing model will not change cost to about 80% of our
subscribers, and the other 20% will see increases based on actual usage.
Many are fearful because they see the abusive rates charged by cellular
carriers for small packages and immediately thing we are going to start
hammering them for $150 per month. Like much of what I have read here, I
too am looking at about 30-50 GB of transfer as a base with a small per GB
cost.
The real value to the upgrade for me will be once we demonstrate we can
deliver a solid stream that people that are trying to pull multiple streams
will have the option to doing so by upgrading to a higher bandwidth package.
And that is the point I was making before, that the amount of transfer has
little to do with the pipe size, but that size does impact the subscribers
ability to have concurrent streams.
So we are really focusing on three things; first, we are separating the
basic and power subscribers, then we are offering those power subscribers
the option to get whatever they want, providing they are paying for it.
Sure a few will be pissed because they have this entitlement to unlimited
service. Tell them you will start the day the power and gas company remove
their meters.
In the long run, the decisions made will provide maximum benefit to all
subscribers. Perhaps we will see a few that refuse to pay and leave, but we
will increase significantly as word gets out about our new capabilities.
Remember, all those smart televisions need a pipe to connect to these
streaming services. And that is the simplest answer, your changes in
billing are to accommodate a market that did not exist when you deployed.
When you and I put our systems in place Netflix was not streaming. So we
absolutely must accommodate these new high demand users, while acknowledging
the long time basic users. Just remember that many of them will move to the
other side over the next few years and be very glad you were able to
accommodate their new requirements.
<PERSON>
| 95feb1c5-ae5f-4088-1f58-628be07f7c7a | ['59b034c1-c23a-c2f8-ca69-40767f4633d4'] | I believe <PERSON> to be correct. Packages based on speed are not the answer.
We call our connection a pipe, so lets use a related analogy;
You can have two homes with water service. One is an older home that has a
½ inch water main, the other is new construction and has a 1 inch service
main.
House number 1 has the original fixtures, so the toilet uses 6 gallons per
flush, the shower flow is 7 gallons per minute and the clothes washer uses
40-55 gallons per load.
House number two, being built under new codes that promote conservation has
a low flow toilet that will use 1.6 2 gallons per flush, a low flow shower
head that restricts flow to 2.5 gallons per minute and a new clothes washer
that uses 20 gallons per load.
With a family of 5 in each house, its easy to see that , despite the
smaller service pipe, that house number 1 will have many times the water
usage as house number 2. A smaller pipe did nothing to control the flow
because the flow limit of the pipe was not reached.
Those two pipes are exactly like a 3 meg and 5 meg Internet connection.
Within reason, the size of the pipe will do little to limit heavy bandwidth
usage. It only serves to spread it out, creating a longer period of time
that it puts a demand on our networks.
Like most, we saw our network performance begin to deteriorate as Netflix
switched from a physical to a digital delivery system. The others since
then have continued to slow our once speedy connections. Now we, as an
industry, are faced with a continued rebuild to meet a voracious demand for
bandwidth to deliver content that we never intended, or anticipated. Worse
yet, we are being positioned to provide these improvements to support the
business model of companies that barely acknowledge our existence.
And they are getting smarter in their use of our pipes. There was a time
when if you didnt have a good 4.5 meg flow, Netflix would not stream. They
have gone to much more advanced encoding that will adjust to feeds of less
than 2 megs, rendering a 3 meg rate limit useless in defending against them.
The issue of Net Neutrality somehow became synonymous with no caps. It
appears we are the only service that is viewed by consumers and governments
that should be given away. Services like water, natural gas and electricity
are each brought to a home and metered for actual usage, because it is the
only fair way for those that use these services to pay their fair share. In
most locals, the billing is specifically broken down into two parts. The
first addresses the base cost of the connection to the property, and the
second reflects the cost of the metered usage.
How is Internet different? We are a service that delivers a commodity to be
used and never recovered. The bits of data we move for our subscribers are
no different than the kilowatt, gallon or therm moved by the others. Could
you imagine if consumers demanded there be no metering on these services?
We are being restricted by network limits from delivering the full pipe to
subscribers. This limitation is a function of cost. Under our current
structures we cannot justify the cost of building a large pipe to each
subscriber. After all, we are an industry built on contention. This
sharing of bandwidth was the impetus of the WISP business for many years,
but that concept has outlived its usefulness. Our subscribers no longer
want to surf the web or check email. Most now do that on their smart
phones. No, our pipe has become an unwilling player on the next pervasive
shift in the paradigm, as subscription video shifts to a digital delivery
medium.
Just as VoIP has been disruptive to POTS, and satellite was to cable, we are
on the cusp of the next trend in consumer electronics. Televisions today
are being built with Ethernet ports and wireless networking. They are
coming with built in apps for all the streaming services. And they want all
this to work over OUR pipes.
So we need to face reality and understand that if we dont provide these
services, we have become useless to our subscribers. Our failure to respond
to this trend will throw the doors open for someone to come in to our
markets and pluck each of our hard earned subscribers until we are decimated
and a faint memory. If you think subscribers are satisfied with basic
Internet services today, you are in denial.
The answer is we need to build out robust networks that can deliver copious
amounts of bandwidth to our subscribers. Our repayment will come by
employing the time proven practice of metering for usage.
We can divide our subscribers into two groups. The cutting edge-tech savvy
type that is creating our issues, and the rest who will be joining them. I
am sure that most of us have similar network statistics. If I look at one
of my network segments I have the top 4 users consuming 25% of all
bandwidth. I hit 50% at the 13th subscriber. This is a change in trend.
It used to be just 8 that consumed 50%. And yes, bandwidth consumption has
increased accordingly. This change from 8 to 13 subscribers being in the
top 50% indicates my high usage subscribers have increased by 120% in
roughly the past 6 months. Post holiday season I expect to see at least a
300% increase in my high usage subscribers, which without changes to my
network, will bring data flow to a standstill.
So build and meter. Dont ignore the elephant in the room referenced
earlier in this discussion. Just look at copper phone lines that peaked at
186 million in 2004 which today number about 84 million. In just 9 years,
pureplay VoIP, cable VoIP and cellular technologies caused a 55% shift in a
once-thought untouchable market.
Joe
|
9cd3bca3-c9a3-687b-4fef-e7d521bb15e3 | ['59c7d96a-cbab-d583-854f-465a3da235eb'] | [Qemu-devel] [PATCH 0/3 v5] Live migration without shared storage
I tested the live migration without shared storage for the patch of
the block migration. It succeeds for several times but sometimes it
fails. I used the qemu-kvm-0.12.3 and the kvm 2.6.33 module on the
ubuntu latest version.
In the failure case, i can see the shell on the destination site
showed the correct content for the migrated VM, but it just hanged
there without any response to the mouse and the keyboard. I can see it
prints "migration completes" twice. Is it a bug or did I do anything
wrong?
Another question is about the command line for the destination site.
For memory
migration, i will start the vm based on the shared image. For block
migration, there is no shared image. What i did is creating an empty
image with the same format (e.g. raw or qcow2) as the source migrated
VM. Is it correct or it doesn't matter which format it is in the
destination site?
For the memory migration, it emphasizes that the path for the VM
should be the same as the source path. But for the block migration, i
change the destination path to a different one and it also works. Does
that mean path is not important now?
The command line must be "migrate -d -b ..". If it is "-b -d", the
qemu-monitor doesn't understand. Does that intend to be designed like
this or will it be modified in the future?
Thanks!
-Jie | bc7c583d-c65e-355e-12bd-2e575f7196a3 | ['59c7d96a-cbab-d583-854f-465a3da235eb'] | Hi,
I am confused about the main thread, monitor thread and migration
thread. Hope somebody can give me a help.
I observe that when a VM (smp=1, which is a file server) boots up,
there are two threads for qemu (e.g. i use "top" and "ps" to monitor).
I think one thread is for the vcpu. The other thread is for the main
thread. Is the main thread the same as monitor thread? When clients
connect to the server and start running workload, one more thread is
generated. I think this one is I/O thread.
Am I right for understanding these 3 threads?
When migration (memory and block) starts , 3 more threads are
generated. What are those 3 threads? From the code, it seems there is
no new thread generated for migration. In the mail list, there is a
series of discussion about whether migration-tcp should be threaded.
The conclusion is no. Is migration using the same thread as monitor?
Is qemu-kvm or linux kernel responsible for thread scheduling? Does
KVM has its own algorithm or it only let Linux kernel to handle? If I
want to understand the thread creation details, where is the code? I
search in the qemu-git tree and only see the VCPU thread is created in
qemu-kvm.c. Any other portion that i should look into?
Thanks a lot!
-Jie |
ee4f7708-a134-0ec1-5228-ec2574a47564 | ['59c7fff9-56e3-dc91-212c-e20a2ee90fb7'] | Hi all
I tried to install a few extensions to my mozilla on debian testing,
especially mycroft and google bar, but did not succeed. It looked as if
it were installing something, but then nothing changed. I very much
guess that it could have something to do with permissions. Although I
don't see a reason why it should not be possible to install an extension
just for a single user in his home directory.
And since I did not find any hint on the internet I now ask you
directly: Is there a howto somewhere for installing mozilla extensions
in debian?
Thanks a lot in advance,
<PERSON> | d0563edd-b562-6bfc-8f01-b38aad5d19ad | ['59c7fff9-56e3-dc91-212c-e20a2ee90fb7'] | Hi
As <PERSON> said before, if you're running unstable, the official version is surely the best.
If you're running testing though, you're probably better off using the apt sources listed in http://cran.us.r-project.org/bin/linux/debian/ReadMe since there are some dependency hassles with installing the unstable version on testing (at least I got tired of going through all the dependency stuff).
Cheers,
<PERSON>
|
9d9eb052-196d-1007-d2c6-63d7a50f8647 | ['59cb8ef0-b5cd-2617-4632-9aac5f26f849'] | I'm helping start a robot lab and we are considering using aibo + pyro for
our first jaunt. One issue is that we will definately need microphone input
from the aibo, so I'm wondering if anyone has ever got that working in
Python/Pyro before? One goal of the lab is to maintain abstraction from the
hardware details as much as possible (e.g. buying off the shelf robots), so
I'm hoping it would not be difficult for me to get the microphone working.
Any help is greatly appreciated,
<PERSON>
| a2d66c98-ad26-84db-d063-b2f9e6bb9591 | ['59cb8ef0-b5cd-2617-4632-9aac5f26f849'] | Hello,
I want to port Mono to a 2-issue MIPS-like processor and for starters I'm
quite happy if only one instruction slot is useful (and the other is a nop)
in every 2-instruction-bundle. I need these instructions to be aligned in
64-bit as not all instructions are allowed in each slot. I was going to
start with the current MIPS port as a starting point (thanks <PERSON>!) but
wanted to find out if the alignment issue is easy to satisfy or if it will
be very difficult to enforce the instruction alignment.
Thanks in advance,
<PERSON>
|
6d4829c1-cae8-3d23-925c-4561dca963c1 | ['59e0fd0f-1f29-a997-9e17-24dceb860f32'] | Greetings,
My LTSP users authenticate to a Mac OS X Open Directory server, and
it works great. However, we allow our student users to have
relatively simple passwords. The problem is when we try to change
passwords from within the fedora core 4/ltsp environment. Fedora
requires them to create a strong password. Is there any way I can
avoid the strong password, and allow them to choose whatever they want?
Thanks,
<PERSON>
| b91f6785-5436-0f11-9d0c-086ac6b217e9 | ['59e0fd0f-1f29-a997-9e17-24dceb860f32'] | Greetings,
Does anyone have any experience using the newer dual-core opterons in
their ltsp servers? My most recent servers have all been dual
processor opterons, but as opposed to spending $180 for a single core
opteron, I can spend roughly $330 for the dual-core versions. I'm
thinking that two dual-core opterons would offer incredible
performance in our lab of approx 30 clients. Any thoughts or advice?
Thanks,
|
c453c289-e1b0-2d47-71df-3addfa0776b3 | ['59ed2a33-2ece-6997-89bc-a13c38efd95e'] | Hi folks!
I have a Newton 2100, good condition for sale.Everything works fine, no
scratches on the screen, the only thing is, that the contrast button is
pretty much on the left side. I also have a Newton leather case,the one
where together with the keyboard, you work like a laptop.... new Newton
keyboard, dongle, 2 serial cables, one for mac, one for pc, 56k modem card,
2 wifi wlan cards, lan card, 2 power adaptors, one with european+traveller
adapters, a bunch of software on cd,memory flash cards,etc.
I also have an apple stylewriter II in very good condition, which is perfect
for Newton.
If You are interested in all or part, please write, so I can send a detailed
list and pictures.
I also have an Apple Powerbook 2400c with Newertech G3 processor, Mac OS 9.2
installed, and a second Powerbook 2400c, original, for parts or so. And some
RAM to it too. Please reply to: <EMAIL_ADDRESS><PERSON>.
If You are interested in all or part, please write, so I can send a detailed
list and pictures.
I also have an Apple Powerbook 2400c with Newertech G3 processor, Mac OS 9.2
installed, and a second Powerbook 2400c, original, for parts or so. And some
RAM to it too. Please reply to: pcMJ+PhCL0DwC8k6@example.com
Thanks: <PERSON>
| 7534df36-8ac3-6558-b75b-f2b46561914d | ['59ed2a33-2ece-6997-89bc-a13c38efd95e'] | Hi!
My newly accquainted eMate is working fine, the only problem is, with
different position of the screen and moving backward and forward, the
backlight is coming on and off, and sometimes it queeks (makes strange
quikee noise and some lines appear on the screen. I am new to eMate, but
this is the "hinge problem", right???
I do have a repair guide from the internet already, so I will try. Was just
a question to make sure.
Thanks: <PERSON>
|
f148f4d3-3a14-2699-4a4f-b40e3b9dba05 | ['59f3e545-2f5a-5802-feec-a3c979a7a273'] | Greetings all!
I seem to be picking up on a couple of different ways to create worlds for
croquet.
In following through some of the tutorials for squeak and croquet I see
worlds as morphic/teapot projects of some sort; though in certain
discussions I've been involved in there was a concern for the ability to
employ 3DSMAX or similar tools to author the world and export it as an
ase-type file.
Would anyone care to clear this up for me? I was rather under the
impression that world building/artifacting tools would somehow be integral
to the croquet environment. Is this a misconception?
Thanks in advance!
| 37311d46-fc75-5bff-26f0-f4161ef5c9b7 | ['59f3e545-2f5a-5802-feec-a3c979a7a273'] | I have three different (squeak) installations right now- one a browser
plugin, one that came with the croquet project package, and the last is
the free-standing (read no croquet) 3.7 squeak release.
Is there any way I can load croquet up independent of underlying squeak?
say, for instance, load croquet into my squeak 3.7 context?
Best Regards
Twitch
|
3db62ec3-e561-d1be-861c-096826f2e828 | ['59f5a87d-5491-f202-97ec-6b209d0c8c14'] | Is there any work being done to create a TreeTable like component for
Qooxdoo? This is a primary component of our interface and if it does
not exist or is not in progress already for Qooxdoo, I may have to
take a stab at implementing myself.
If there is no such component, would it make more sense to build on
QxTree or QxListView?
Thanks!
<PERSON>
| 33e0bdb8-0c48-7138-be74-70c1971b4284 | ['59f5a87d-5491-f202-97ec-6b209d0c8c14'] | I am curious about some "construction" decisions. Is there reasoning behind
making QxListView from "div" components rather than "table"? My main
curiosity is we are evaluating QooxDoo for some major internal applications,
and want to make sure we understand as much as possible before making a
decision.
One reason I ask this is we have had to deal with large tables and sorting
here, and have had to do some interesting stuff to make them fast - we have
at least one app with a requirement for tables containing 5000 rows with
sorting capability - with IE support a must. For QooxDoo, if you load
QxListView with 5k rows in the demo, it dies.. it has actually crashed my
IE, and when it dosnt, it just falls asleep.. even with only a few k rows,
sorting is unacceptably slow...
Thanks
<PERSON> |
33bb99b2-f583-3ded-1b7c-c25c4aba868d | ['5a02e3bf-2fc2-bf42-02c3-8b11c5800ffa'] | We need to change the groupID/artifactID in our POMs but we want to continue to add to the historical data that our project has created so far.
Currently when we change it. Sonar thinks is a new project and starts a new history.
How can we rename the existing data in Sonar to appear to be from the new groupID/artifactID ?
<PERSON> | 0cb739a1-f1a5-ad4b-3b6e-3d9ebd2e722c | ['5a02e3bf-2fc2-bf42-02c3-8b11c5800ffa'] | I was wondering if there was a way to identify newly added classes ? I noticed that some developer checked in about 30 new ones and I don't know where they are.
I didn't really see anything in the documentation, though i suppose I can check the Hudson logs for that day :-)
<PERSON> |
0f1c54c3-0e05-6e3f-67a3-496c02fcf1e1 | ['5a0cb908-ac51-e8bd-289a-a25d65f3644f'] | Dear Fail2ban users,
running fail2ban v.0.10.1 on an opensuse box.
currently looking into the recidive jail to ban persistent abusers.
From what i understand the bans are stored in the persistent database
storage so the bans can be added on restart without re-scanning the logs
files.
If i set a bantime of 1w in recidive jail the jail.conf informs me that
i should increase the dbpurgeage to 7.5 days
so the bans with 1w can live long enough before getting purged
but if i do a permanent bantime -1 what value should I set the dbpurgeage?
what's the relation between bantime, persistent storage and dbpurgeage?
would be nice if someone could perhaps enlighten me on the topic ;)
thanks & greetings
<PERSON>
| 413a7ad5-ba9f-3e8f-b151-ff0fe7857d34 | ['5a0cb908-ac51-e8bd-289a-a25d65f3644f'] | hello again,
my current setup is varnish 5 in front of apache web server with 30
different domains / websites
in my vlc i can switch between sites and decide to do varnish caching or
just passing to the backend with out caching.
what's the overhead / delay with varnish when i just pass everything to
the backend and back to the client without varnish doing any caching?
i mean there must be some overhead as varnish is in between ...
... but how big is it? time delay?
does anyone here have any experience with this?
thanks & greetings
<PERSON> |
93357ec4-b35a-a268-5055-d08aa82170c0 | ['5a45f8f8-1b41-f2ec-6403-7f9c455480d0'] | Hello folks,
I'm running into a hard time trying to add new OIDs in the certificates
issued by OpenCA. We already have a nice structure running, with an
online RA and an offline CA, but now we are trying to add more
information in the certificates, specially some info needed for the
ICP-Brasil (Brazil PKI). The new OIDs should be in the Subject
Alternative Name section, and all in otherName format.
The OIDs are the following:
<IP_ADDRESS>.3.1 - 51 charactes max
<IP_ADDRESS>.3.5 - 41 charactes max
<IP_ADDRESS><PHONE_NUMBER> charactes max
<PHONE_NUMBER> charactes max
<PHONE_NUMBER> characters
Has anyone made this work? Or even anyone already customized OpenCA to
be ICP-Brasil compliant?
Thanks,
| 97abccbe-f446-a63a-f575-6dd919ea6648 | ['5a45f8f8-1b41-f2ec-6403-7f9c455480d0'] | Hello all,
Here at my company we use the full set of TrendMicro products, but we're a
little disappointed with some facts of the solution, so, we're thinking
about changing the product. We've been hearing good things about Kaspersky
lately, then we would like to know from you guys if you have any experience
with it, went through a migration, what about day-to-day work and, most
important of all, if you had any outbreak while using it.
Well, Kaspersky is one of the solutions we are looking after, but any other
anti-malware product that you recommend for a 10000+ user company would be
nice to hear.
Best regards,
|
5f494790-3d68-a18f-3b5d-5b7250b090b2 | ['5a45ffae-2fa8-0d04-806f-2b15e75e1439'] | Hi,
The current implementation of netbeans binds many file extensions to
mime types in the openide file filesystems/FileUtil.java. Most of these
bindings are reasonable. However, The extensions "c", "cc", "c++", and "h"
are bound to the mime type "text/plain". This makes it imposible for C and
C++ files to be recognised with the correct mime type unless they use
nonstandard file extensions. The extensions "C" and "H" cannot be registered
as C++ extensions either, because case is changed during the binding.
These extensions should not be registered in the openide but should be
left for registration by modules adding C and C++ support. An alternate
approach would be to bind these extensions to something like "text/x-c"
and "text/x-c++". I don't like this approach as well, but would be willing
to generate a new patch if thats what the netbeans.org consensus preferred.
There is a second mime extension binding problem which could be more
controversial. All extensions bound in netbeans have their case
changed to lower case. I suppose this could be considered a ``feature''
because windows is very case insensitive. However, this causes problems
because one of the less popular file suffixes for C++ is ".C". When I tried
to bind ".C" to the C++ mime type it was converted to ".c" and generated an
exception saying c was being registered twice (even the error message used
the lower case c rather than the upper case C which I was in the code).
I've attached 2 possible patches to this email. The full-mime-fix
fixes both problems mentioned above. It removes the binding for c, cc, c++,
and h. It also removes the forced case conversion and binds the extension
in the case it was typeed in. The 2nd patch, partial-mime-fix, could be used
if the netbeans comunity objected to leaving the case as its typed. I
strongly prefer the full fix but can certainly live with the partial as
its not one of the more popular extensions for C++.
The netbeans.org/contributing.html page says the way to get a patch
adopted is to email it to this list. Please consider this a request for
one of these patches (preferably full-mime-fix) to be applied to the code.
The end result would be the ability for a C/C++ module to bind the proper
extensions to C and C++ mime types. The result with the partial fix would
also allow this but would restrict the file extensions able to be used by
C++ developers.
Thanks,
<PERSON>
| 44168483-c9ff-9dde-6e20-8ce446439b10 | ['5a45ffae-2fa8-0d04-806f-2b15e75e1439'] | Hi,
There was a recent change in the module test mode. Previously, loading a
module in test mode loaded from the manifest and a jar file wasn't needed. In
the new scheme the test module is loaded from a jar.
What I haven't been able to figure out is how to reload a new test module
without restarting netbeans. Is this still doable? If not then the test module
scheme isn't nearly as useful as it was previously. If it IS still possible to
reinstall an updated module in test mode then somebody should update the tips
document (http://www.netbeans.org/download/apis/org/openide/doc-files/tips.html).
If it is no longer possible to reinstall a test module then I'm very curious
why this feature was removed. I often work remotely and having to restart a
new netbeans can take 1 - 2 minutes!
<PERSON> |
9bf5e026-f449-04b3-8c4d-f6ed72ac14af | ['5a49ba18-3ac7-77f0-ff3f-49645b28cd74'] | Hi all, I am having a problem I hope someone can help me with. I am
pretty much a complete Tivoli newbie
so please forgive me if this is a stupid question. I've searched the
manuals, Google, etc. and cannot
find a resolution to this.
I have a small Tivoli test environment I am using to learn this stuff. I
have one machine, that is the TMR server,
TEC server, Endpoint Gateway, etc. Basically, all of the "Server stuff"
is on one box. Additionally I have
one other machine running an endpoint. The server machine is running
win2k server, the endpoint is
running Windows XP Professional. The TEC version is 3.9, and ITM is
5.1.2, TMF is 4.1.1 if that matters.
The problem I am having is this: events generated by the heartbeat
monitor, are apparently not being
sent to TEC. That is, if I take down the endpoint machine that I'm
monitoring, the heartbeat monitor
detects that situation (wdmmngcache -l shows the endpoint as "unreachable"
) and a notice is sent
to the monitoring notice group. However, nothing appears in the TEC
console. Additionally, if I run wtdumprl,
I see nothing related to a heartbeat event.
However, I know that TEC is (more or less) working OK, as events from the
endpoint (for example, a ProcessHoggingCPU event when
the virus scan was running) show up in my TEC console.
I have run wdconfig -D heartbeat.send_events_to_tec=true
and wdconfig -D heartbeat.tec_server=EventServer to configure the
heartbeat to send
events to TEC. I have also run the task to import the baroc files, and the
active rulebase
in my EventServer does have the hearbeat events rules loaded.
Also, FWIW, I can run wpostemsg and manually send a message, and I am able
to see that
when I run wtdumprl.
Any advice regarding how to resolve this would be greatly appreciated.
Thanks,
| b51f2472-00f1-b932-ca5c-33b5737c166b | ['5a49ba18-3ac7-77f0-ff3f-49645b28cd74'] | Hi all, I have a question that I'm hoping someone can help me with. I've
recently created a logfile adapter (using TEC 3.9) to monitor a WebSphere
Application Server logfile. In order to allow it to co-exist with other
LFA's
I used the "Identifier Name" field to give my adapter a unique name.
Now, I find out that the group who needs to use this adapter is on TEC 3.8
which apparently does not support the "Identifier Name" to distinguish
amongst TME adapters.
So my question is: given that they are using TEC 3.8, is there any way to
run multiple TME adapters? For that matter, is it possible to run
multiple
non-TME adapters? And if so, is manually installing and configuring the
adapter(s) the only way to achieve this?
Thanks,
|
0a457e9c-ac6a-f28f-ab80-ed07168ef12c | ['5a90d7aa-2674-ed3d-6769-c7231cd0db94'] | BTW: PediaPress has a stand on the Frankfurter Buchmesse (Frankfurt
Book Fair), booth E427 in hall 4.2. We'd be really happy to meet
people from the community to talk about all kinds of MediaWiki related
stuff.
So, if some of you are there and can make it... we're looking forward
to meet you!
<PERSON> | d7636074-0738-e075-bbf2-8f68fd191e5f | ['5a90d7aa-2674-ed3d-6769-c7231cd0db94'] | We're heavily using the MediaWiki API in our opensource project mwlib (http://code.pediapress.com/
), so first of all: Thanks to you all for implementing this
functionality to MediaWiki!
Maybe you're following the discussion initiated by <PERSON> on
Foundation-l about appropriate attribution. As there is yet a consesus
to be found, we plan to include all authors (minus minor edits, minus
bots) after each article in documents (PDFs, ODFs) rendered from
article collections.
Currently we're using an API query with prop=revisions, requesting
rvprop=user|ids|flags. Afterwards we're filtering out minor edits,
anonymous/IP edits and bot edits (via regular expression on username
and comment) and combine edits by the same author. To retrieve the
data for all revisions for heavily edited articles (e.g.
[[en:Physics]]), this requires lots of API requests with rvlimit=500.
Is there a way (or a plan to implement one) to retrieve the list of
unique contributors for a given article (from a given revision down to
the first one)? Ideally this would accept parameters for the mentioned
filtering. I guess inside of MediaWiki code this can be handled very
efficiently (using appropriate database queries) and would eliminate
the need to transfer lots of redundant data over the socket.
<PERSON> |
68249009-6b6f-d030-8340-50ef220b1f6c | ['5aaf79e9-aec5-25b1-6d9c-c6365a712fde'] | Hi all,
I am implementing a power mac protocol and for this I want to send a
packet with variable energy i.e. the transmittor node can change its
transmission energy during the transmission after some fixed time and
that should reflect in the packet being sent.For example if a packet is
sent at 0.28W , i would like the same packet to be sent at some lower
power level after sometime.Hope I have made myself clear.
Any help or even a small hint would help me in giving a direction to
proceed.I searched the net too but could not find anything relevant.
Plz help...
Warm regards,
<PERSON> | 688ecfda-adce-a1b8-cc57-6af0f8afb4a2 | ['5aaf79e9-aec5-25b1-6d9c-c6365a712fde'] | Hi all,
Its a simple question and I hope somebody will answer it.I have just
started using ns2 and simulating some wireless scenarios and comparing
the results for IEEE 802.11. the topology consists of 31 nodes in a
straight line with each node transmitting cbr packets to the next node.
Data delivered per joule should be constant with respect to distance
between the nodes but I am not getting the exact result. Rest of the
parameters are the same(defaults) in ns2. I dont understand why am I not
getting the exact results.
Somebody plz help...
Thanks and Regards,
<PERSON> |
2aeb2d5c-341c-c598-84a9-a9c45d4150c7 | ['5ab5a2bc-1bcb-98ba-7abd-a5182fa89ee7'] | Hi <PERSON>,
If you're going to be using PostgreSQL 7.2.2 (the latest recommended
version), then the port number is a setting you can change in your
postgresql.conf file. This file is created inside the PostgreSQL "data"
directory when you first initialise the database with "initdb".
This file is also very well commented and easy to figure out, so you'll
be fine to just open it with a text editor and make the required change,
then restart PostgreSQL. As a precaution, it's a good idea to make a
backup of your postgresql.conf file too, just in case something goes
wrong while you're editing (unlikely though).
:-)
Regards and best wishes,
<PERSON>
| 4a8ca5c8-9a80-6c93-d8bc-de3de716d5a2 | ['5ab5a2bc-1bcb-98ba-7abd-a5182fa89ee7'] | Hi everyone,
<PERSON>, a journalist with IDG, is looking to do a brief story on
PostgreSQL and has a couple of questions for us.
So, being newly motivated to set up the collaborative area for the
Advocacy site, it's now ready for general usage
(http://advocacy.postgresql.org/documents/)
<PERSON>'s questions (from below) have been added into a page in the new
collaborative area, so people can make notes/comments/etc directly, and
we should be able to get the results to her by about this time tomorrow
if it's as effective as I'm hoping...
http://advocacy.postgresql.org/documents/NadiaCameron
:-)
Regards and best wishes,
<PERSON>
|
e4b2f6c7-c927-7700-47e9-6b359480fe92 | ['5ab6e7cc-39c1-9230-cd79-a73d991b3e09'] | I'm a new Squeaker, so I don't know if people even say "source code."
I'm making an editor with some new features I like. I'm making it in
Squeak. That seems to be going well, but after I'm finished, as well as
running standalone, I'd like it to run integrated with such environments as
MIcrosoft's Visual Studio.
Microsoft's stuff expects a C dll. I'd like to translate my Smalltalk
program into C.
SPiCE seems to be the ideal thing. It expresses a good part of the
Smalltalk environment in C, which is a tough bullet to bite, but once it's
bitten, almost all of one's Smalltalk program can be expressed in C without
change. The implementation seems to be pretty complete, based on <PERSON> paper.
Is this facility available somewhere for me to use?
Thanks for any help.
Regards,
Rick
| 1d13b7ae-1c85-93e9-d335-1b49934cbb5b | ['5ab6e7cc-39c1-9230-cd79-a73d991b3e09'] | I asked a few days ago whether the SPiCE program for converting a Smalltalk
program to C was available. Apparently, it's not. Thanks, <PERSON>,
for the info.
The other, less complete but still interesting, solution is Producer, by
<PERSON> and <PERSON>. Is that one available somewhere to be adopted
and perhaps extended?
Regards,
<PERSON> |
893617db-c266-ac7d-38a6-86fc1f1e44c8 | ['5aba385e-0947-f1e3-7e61-4f5652bce8e5'] | Hello,
A sky mote was able to process 10 KB of sensor data using wget-shell
example to other Telosb's. It's approximately taking 25 sec for the TX.
Currently, wget-shell application in Contiki-OS is using HTTP. Is there any
other approach to reduce the end-to-end latency for a large size of file
transmission. Will it help if I modify the ContikiMAC (explicitly
radio-duty cycle)? Any suggestions.
Best regards,
<PERSON>
| a73e5ef9-f7e0-e0a6-970c-24baad924926 | ['5aba385e-0947-f1e3-7e61-4f5652bce8e5'] | Dear all,
I am trying to solve a problem regarding file download from coffee file
system. After successful transmission, I was able to write the file into
coffee file system for Sky mote. After login to the mote, it shows the
size of the file from sky-shell terminal. How can I download the file to my
local machine from CFS? There is a (tools/sky/serialdump.c) file can be
used to trace the serial port regarding this; but it can't determine the
stating and ending of the file. Just print out the hex and binary output to
another terminal. Is there a way that I can rebuild the original file using
this or directly downloads from CFS. Any ideas.
Thanks.
<PERSON>
|
5061aad0-c999-5c56-4ad4-ea884d5a5234 | ['5abdc934-c8d6-8403-ed9f-f215fe7f5b5b'] | Hi there,
I'd like to use <asminfo> and <mkiisdir> in my build. So I have a
recentish build of NAnt head. And I copied the NAnt.Contrib.Tasks.dll
from NAnt 0.8.3 final into my compiled NAnt bin folder. But If I try to
use <mkiisdir> I get 'Unknown task <mkiisdir>' Am I being dumb?
Thanks,
<PERSON>
| 914baa44-8c53-9f65-fb34-f003a1af8bb3 | ['5abdc934-c8d6-8403-ed9f-f215fe7f5b5b'] | Hi there,
I've hit a problem (bug?) using <solution> . If it can't find a
reference, then NAnt throws an internal error, rather than this getting
reported more nicely. This causes me a problem since Internal Errors
don't get sent to the logger, and so I don't see the error appear in a
CruiseControl.NET log (in CCNet we use the Logger to capture NAnt output.)
I think this would be fixed if <solution> used the <csc> task internally
rather calling the csc.exe compiler directly, but that may be too much
effort.
Cheers,
<PERSON>
|
f38b8328-0b66-9688-8456-889c9a019a3d | ['5ac9d506-942a-325a-d787-13ad2c8f6a3a'] | I have noticed that the cvs interface in this product, when running the
CVSNT version 2.5.02.2088 cvs.exe executable, fails on a sandbox update.
However, the same application running with the cvs.exe from the
TortoiseCVS folks ( 2.0.62.1817) operates properly. So, if you are trying
to resolve this issue, which manifests itself as a 'Checksum failure '
message, use the Tortoise version with EA. A future version of EA will
address this issue, apparently.
<PERSON>
| d641c7b0-2d18-0b53-445e-0edfe33c66bb | ['5ac9d506-942a-325a-d787-13ad2c8f6a3a'] | There is no limit on the # of repositories that I know of. The only trick
is to make sure you enter the right repository name in CVSROOT and that
you grant appropriate privileges in the repository project folders. Also,
if you are doing special wizardry in your CVSROOT directory like email
notification and so on, I think you have to replicate all this to the
other repository as well. They are, as far as I know, logically distinct
from each other. That means dup'ing all the commit scripts, verify
messages, etc. etc. Anybody solve it another way, like with links?
Allen Marshall
Cambridge Systematics, Inc.
100 CambridgePark Drive, Suite 400
Cambridge, MA 02140
tel 617 354 0167
fax 617 354 1542
e-mail <EMAIL_ADDRESS><PHONE_NUMBER>
fax <PHONE_NUMBER>
e-mail MVbnGMODyAiXPpbR@example.com
www.camsys.com
|
a291c23a-4ae2-8d4d-cd82-090cc7244528 | ['5aca5db5-1f3f-bd79-6577-e4b600265e08'] | Another dumb question here.....
All of the information I've found so far related to parallel rendering using VTK involves the use of multiple individual computers and dates from five or six years ago.
Is there such a thing as using one of the new multi-core chip PCs for parallel rendering applications? I notice that MicroCenter is selling a P7 desktop for around $800 now and you'd think the idea would occur to somebody.
<PERSON>
| 23a92084-3b9e-97ff-ba4f-46e7df4c75bb | ['5aca5db5-1f3f-bd79-6577-e4b600265e08'] | This one finally dawned on me..... the wrapper libs basically don't contain real code, just wrapping code; the real vtkio.dll has to get put in the path for anything new in it to get found, duh..... The problem vanishes when I do that.
<PERSON>
|
d4043cb2-8e3b-ce3d-fb61-bb58c4ef678c | ['5accfee4-49e6-f446-79ff-24477c8738a5'] | "Is this still around ? I filed a bug several months ago. Which version
are you using ?" -- <PERSON>
I don't have my laptop handy right now but I'm running the grub that
came with Gentoo 1.4. I also preformed an "emerge system" so it should
have updated it with something new off the portage tree (this was last
week).
If this is a bug in grub (sorry about the rhyme), what version was it
fixed in?
I'll re-emerge grub and try with: grub-0.93.20040125.ebuild
http://www.gentoo.org/cgi-bin/viewcvs.cgi/sys-boot/grub/
Sincerely,
<PERSON><PHONE_NUMBER>.ebuild
http://www.gentoo.org/cgi-bin/viewcvs.cgi/sys-boot/grub/
Sincerely,
Rob Elm
| db099d09-6ecc-796c-9ad7-d2c1f2621b61 | ['5accfee4-49e6-f446-79ff-24477c8738a5'] | Most modern email clients support multipart MIME which encodes both the
text and HTML messages within the body of one email. Besides that, the
Bug-Grub listserv does not have a charter which would prohibit the
sending of HTML messages (at least as far as I know). If you're truly
offended by html formatted email, maybe you should consider using a mail
client which is intelligent enough to ignore the HTML encoding.
Sincerely,
<PERSON>
|
258d1510-6ae7-3c4e-10b7-e9ab5316ca87 | ['5af3d5bd-6c32-8417-1b24-1f9b7a120201'] | <PERSON>,
mam dobrou zkusenost s amd64 na serveru, a ted <PERSON> na
IBM x226 s Xeonem s em64t. em64t ma byt kopie amd64, ale <PERSON>, <PERSON> to "obšlehli". Chci na tom valit 64bit debian sarge. Ma nekdo
nejakou zkusenost? Google nasel vety typu "xeon slows down a little in
general when you enable 64bit mode. AMDs speed up. Given the AMD is
usually faster at mosts tasks already in 32 bit mode, the situation is
not improved
(for intel) in 64bit mode.", coz <PERSON>.
Petr | c0afb033-a57c-899d-19be-19c09f09d455 | ['5af3d5bd-6c32-8417-1b24-1f9b7a120201'] | <PERSON>, this is on onboard Geforce 6150 using proprietary nVidia
driver.
BTW, a workaround for this issue is to try out Alt+Ctrl+F7/8/9/10/.. at
step 5) when there is the normal login prompt on screen and user '<PERSON>'
wants to get into his/her session. If the user 'B' has really been
logged in already then his/her console is running somewhere at F7-Fx.
What I am reporting here is that gdm (or who) does not switch to that
console automatically.
IMHO it should work as follows: whenever any user logs out the gdm
should try to locate any other running console. It normally works when
second or subsequently logged in user logs out but it doesn't seem to
work when the user-that-logged-in-_first_ logs out.
Thanks.
|
d1093ab1-785d-05ae-7b30-717b659159ae | ['5b070074-4e93-b4a1-454e-2f8a6cac71f8'] | Hi, My name is <PERSON>, I would like to ask you if it is possible to develop an OS based on Debian (similar to SkoleLinux/DebianEdu) specific for psychologist (both applied psychologist and researcher psychologist) built with a set of predetermine tools such as pspp (data analysis), neurodebian, zotero (citation manager), bibus (connection to data base provider) pgp tools for confidential reports about patients,...
I am psychologist and professor at University. I don't know about programming, but I would help to search for the appropiate tools for an specific OS for psychologist (asking partners, testing tools in debian packages...).
Psychologists as other researchers usually depend on privative software to research and to improve the knowledge about human behavior and mind. We would love a new Free Software OS based on Debian.
Currently I am user of Debian Stretch and I think this would be a very useful OS to built an specific OS for psyhologist. I offer myself to contribute of topics related to psychology during the development process and the spread of the new OS if this is released.
Please, let me know your opinion about my proposal.
Thank you very much for yor attention.
Kind Regrads,
<PERSON>
| 5312e527-1942-fa7c-6519-45a28f6a3c1f | ['5b070074-4e93-b4a1-454e-2f8a6cac71f8'] | Dears, I am <PERSON>.
I am a new member of this list. I would like to ask you an important question for me:
Using TAILS OS (which forces all Internet connections through Tor Network), I need to get access to data bases related to academic and scientific journals, but connections through Tor Network are refused by those servers, so I must use other OS or other web browser, but I would prefer don't do it.
Is it possible to manage Tor browser settings in order to get into those data bases using Tails?
Sorry for the annoyances and thank you very much for your attention.
Kind regards,
<PERSON>
|
1845bd89-cb81-878f-957f-b2c3b6c06436 | ['5b263b0b-2537-a33c-ca45-43fb6ebf03e6'] | Hi all,
I've committed myself to learning Python and have started reading
"Learning Python" (<PERSON>) and looking through various online
resources.
My career so far has involved a little high school teaching and about
10 years of translating Japanese to English, but no programming or
coding.
I've also joined this list today and this is my first post.
My aim is to be able to create Python programs, specifically
"activities" that work on the OLPC's XO laptops and SoaS (Sugar on a
Stick).
My questions are: how long would you estimate it to take and how
complicated would it be to create the following as Python programs? (I
know it varies depending on the person, but for example, how long
would it take _you_?)
(i) a simple guitar tuning program involving an image of a guitar and
the playing of each of the standard strings of a guitar (E, A, D, G,
B, E) upon key input by the user
(something similar to this:
http://www.gieson.com/Library/projects/utilities/tuner/ (page is 782kb
to open))
and
(ii) a very basic turtle art program with an intentionally limited set
of commands and on-screen display words (say, a total of 30 to 50
specific strings), wherein the entire set of strings is offered to the
user (perhaps at first use of the program) in a format that enables
easy and full localization of the program so long as each of the
strings is translated appropriately and inputted to the program. I
know of turtle.py and xturtle.py, but I'm thinking of starting
something from scratch. It's the easy localization I'm interested in.
Hope these questions are appropriate. I'm grateful to be able to ask them here.
<PERSON> (in Australia)
| e24df326-1b91-c717-7b41-f2a3fe1bf810 | ['5b263b0b-2537-a33c-ca45-43fb6ebf03e6'] | Hi all,
I was just wondering if anyone here can recommend any freeware program
that has been written in python and for which the source code is
available.
Basically I just want to see a program that does something relatively
simple and straightforward, but something that is "real world," I mean
something that people actually use, hopefully something that comes
with it's own installer and GUI.
I'd like to try out using the program as an ordinary user and then I
want to look at the source code to see how it's been achieved.
Any ideas or suggestions?
<PERSON>
|
959330c3-48e7-3c6b-7543-dc5a11713e5e | ['5b313f77-759d-e278-92a0-6e11d937b18b'] | Hi,
<PERSON>, thanks for your reply. I'll test the patch and use
memory_cache_shared set to OFF.
Sorry, I was wrong. Objects bigger than maximum_object_size_in_memory
are not cached on disk. Although objects smaller than
maximum_object_size_in_memory but bigger than 32KB were written to
disk, I guess they got a HIT because Squid keeps a copy in memory of
hot and in-transit objects. That explains why the UFS store was
"ignored" when Squid was restarted.
Thanks. | 92ed2058-1331-f56c-338e-4686ef30797d | ['5b313f77-759d-e278-92a0-6e11d937b18b'] | Hi!
Thanks <PERSON> and <PERSON> for your replies, together they pointed me
into the right direction!
Based on the the URLs sent by <PERSON>, I edited the
"/etc/php5/cli/php.ini" file and tested different values for
"max_execution_time" and "max_input_time" but none changed the PHP's
script behavior. Then, I remembered <PERSON> mentioned a 60sec timeout. I
saw my cache.log and yes there was an exactly 60sec delay after
starting squid and the first Warning. So, I searched the "php.ini" for
a similar value and found this directive: "default_socket_timeout". I
changed it to 300sec and the Warnings started to show up accordingly.
Then I changed it's value to "-1" and the warnings haven't shown up
again!
Squid doesn't complain anymore about my PHP-scripts, but I don't know
if this change has secondary effects or any other consequences. I'll
be monitoring them, but in any case I have the backup Perl-scripts.
Thanks again! |
53ea4e5c-82dd-a9d1-7bab-3fe0ecf15e41 | ['5b36bb38-744d-dc81-95ab-063948edbf42'] | <PERSON>,
How can one configure LyX to use another pdf viewer when clicking the
"view pdf (pdf-latex) button?
I could not find out. It depends what is set automatically during
installation.
Actually I would like to disburden Adobe Acrobat Reader from this
task because too slow and change to Skim (or PDFView).
joachim
| 5c0bbbc8-d215-c1e4-6629-fac4d7e92a2b | ['5b36bb38-744d-dc81-95ab-063948edbf42'] | Moin Lyxers,
not for the first time I got today an LyX when clicking "show pdf" an
error message:
Die Datei existiert nicht:
(I suppose in english: The file doesn't exist:)
/tmp/lyx_tmpdir.L12205/lyx_tmpbuf1/Lyra-Pracher-odt-teil.pdf
That was after adding an TeX code for another "footnoteB" (from the -
loaded - manyfoot package), because I'm using for that the "handout
(tufte)" layout because of the many sidenotes I need.
But because - at a certain point, when LaTeX (via LyX) said "too many
floats" - I changed to the TeX-box \footnoteB command to get a real
footnote. Which normally gives the planned result.
If the error message "The file doesn't exist ... .pdf" is a known
problem, I ask:
==> how to come out of this situation best?
I found out by trial and error, that I deleted the last TeX inset
\footnoteB{...}, (but after many other trials, even after clothing
and reopening LyX, after deleting the LyX - /tmp ... xyz-files and so
on).
Regained functionality of LyX in making and showing a pdf result, I
reintroduced the TeX-inset \footnoteB and it worked.
Fine for the project. But Question:
=> are there reasons for that behaviour of LyX and/or other ways than
just trying out, deleting the last inset(s) and remaking them?
==> What might be the reason of the LaTeX-error about "too many
flouts" when inserting "footnotes" = sidenotes in the "handout
(tufte)" class, even if on the page where I am working and on the
horizontal space on the margin there is place enough for another
"footnote"=sidenote?
Goutgaun! <PERSON> |
dea71aa6-7b2b-0b1f-5e7c-d4b634a916ff | ['5b4c2917-71e6-c87b-821d-320d55a85abe'] | Hi all,
Following a recent conversation with <PERSON> (and contrary to what I
initially claimed here) there's value in adding a third storage type to
the Quota API: Session storage.
Contrary to temporary storage which might not get wiped across UA
sessions, Session storage MUST get wiped when the session is closed.
Happy to provide patch if there's agreement this is a valuable addition to
the spec.
Best,
<PERSON> | f09153b0-c7d5-9371-179c-14f0ea41d75c | ['5b4c2917-71e6-c87b-821d-320d55a85abe'] | Hi,
I recently stumbled upon a number of use case and requirements docs (such
as MediaStream Capture Scenarios[1] or HTML Speech XG[2]) that were
published as officially looking W3C documents (for whatever that means, at
least, it's not a page on a Wiki).
I think that's tremendously useful, especially for authors who can have a
much better understanding of the purpose of a specification that way (and
therefore use it the right way and for the right purpose).
It's also a smart way to get authors involved without corrupting them into
thinking like spec writers or implementors.
What are the WebApps WG's plans with regards to that (if any)?
Thanks,
|
56d645bc-4778-33dc-3f7f-dd0c0392bccc | ['5b4f5fac-1bdb-770f-6cf7-a17b6dff94c7'] | Hi, devs
By watching recent changes in trunk, I feel the release of Trac 0.8.1
will be done in a few days future.
But changes in [1250] breaks the existing wiki pages which has Japanese
(also korean, chinese, etc.) only heading string by excetpion. I've
reported in #1224. It brings us broken page and we cannot view nor modify
the page any more except modifying code of WikiFormatter.py by himself.
I strongly wish this issue is fixed before release.
--- Regards,
| ba7c01f8-032a-6674-bc0a-d25d8ba03981 | ['5b4f5fac-1bdb-770f-6cf7-a17b6dff94c7'] | Hi,
I got source code archive: TortoiseCVS-1.6.10-src.zip from SourceForge.
http://sourceforge.net/project/showfiles.php?group_id=48103&package_id=41678&release_id=233635
But I cannot expand it with so many claims like this:
skipping: TortoiseCVS-1.6.10/xxx/xxxxxx.xx unsupported compression method 12
The bytes of downloaded zip file and displayed bytes on web pages are same.
I tried unzip 5.51 and older unzip, but both tools failed.
What's happen?
Was anybody able to expand it successfully?
What tool is used to make zip file?
P.S.
|
dcc96484-1a57-d3d8-baa6-910807510843 | ['5b56581a-9e13-a4fc-a122-336a088c23e7'] | I'm looking to setup a file server using RAID 5 (or the equivilant
RAID-Z) and I'm interested in using ZFS.
It looks like my primary options are Solaris or FreeBSD and since I'm
beginning to really dislike Solaris I'm leaning towards using FreeBSD
however I've heard that there are some issues with ZFS on FreeBSD
specifically regarding some "kmap_mem" or something like that.
Can anybody tell me if ZFS is safe to use for home users? I'm just
looking to setup RAID-Z with 3x 300 GB and another pool for 1x80GB (I
guess).
Thanks for any help!
| 9f0d093c-1008-4b3c-df93-25601c3c5d9c | ['5b56581a-9e13-a4fc-a122-336a088c23e7'] | In case nobody saw my last e-mail (from the "Is lua thread safe?"), I
have some more questions... (on top of the questions asked in my last
e-mail)
1) If I create a lua thread, is the global data shared with the main
lua state? The manual seems to contrict itself and says that the data
IS shared but then it says it's not. From doing some tests, it seems
like the data IS in fact shared.
For example, if I load a script (using lua_dofile) and run it on my
thread, the script's global functions show up in my main lua states
global table.
My main goal is this...
In my game I will have tons of scripts being run but I need Lua to
know the current user (or NPC or object) running the script. I can
create a global "CurrentUser" userdata but since it's global it is
shared and gets overwritten. I tried creating a thread but that
doesn't seem to help me at all...
2) Once a script is run via lua_dofile(), all functions that were
declared stay in the global table, right? So I can access them later?
3) How can I load a script and NOT execute it but only run a function
contained inside of that script?
Also, if anybody could answer any of the questions from my last e-mail
(about lua beiing thread safe) I would greatly appriciate it.
Thanks! |
fc0d5a7a-8172-2ecc-b741-3eaf1dfb3c3c | ['5b60aa76-04d8-67db-380c-2eef6c5bf707'] | G'day iFolders
The download page (http://www.ifolder.com/index.php/Download) for the 3.4 iFolder clients does not list clients for Windows or Mac. And locating the 3.5 client, whilst not rocket science would definitely be easier by having a direct link.
For example:
http://forgeftp.novell.com/ifolder/client/3.5/current
http://forgeftp.novell.com/ifolder/client/3.4/current
However, for 3.4 if you backtrack thru the various commits you can eventually find the last Windows & Mac release here (there might be newer, but I cannot find them).
http://forgeftp.novell.com/ifolder/client/3.4/20060616-0101
Which raises the question, why have there been no current builds of the iFolder client for Windows and Mac? And why is there no 3.5 clients at all for Windows and Mac?
Wanting to be helpful, I tried to edit the iFolder Wiki download page to include the updated download links, but this is a "protected wiki page", so can someone with the correct permissions please attend to this.
Thanks for listening!
| 27fa6814-63a1-0aff-f6fc-34b0507650f1 | ['5b60aa76-04d8-67db-380c-2eef6c5bf707'] | G'day
We are actively looking to employ two people in response to the growing demand for FOSS services. Ideally, I'd like to support new local talent, preferrably Uni. or College graduates who are smart/keen and just need a break to get into the IT industry.
The work will be a mixture of LAMP development, Linux administration (desktop & server) and FOSS system installs/ maintenance/ customisation. So I expect that you will have some existing skills in LAMP or Linux (if not, why are you on this mailing list?). Having some good graphic design skills would be a real bonus.
If you are currently employed but want to change jobs (perhaps working for a Microsoft-monoculture-only-shop?) then feel free to apply as well.
So if you are interested or know someone that might be please email the resume to <EMAIL_ADDRESS> I will respond to all applications.
Cheers
<PERSON>
|
e9fcad53-48ac-8062-0544-f82ff44a19f6 | ['5b6423f7-7347-6910-6039-49e864817bc9'] | Hi,
I just had to do this myself, and I figured some of you may be in the
same position, so I'll share it (for Linux / UNIX):
We had a remote server vacuum our data, probably seeding a cache and
using our box as a free rendering service. Which is fine, but this was
very aggressive. So while GeoServer now has the ability to log requests,
you may not always want to restart or reconfigure the service,
especially if it's transient.
A handy tool is ngrep ( http://ngrep.sourceforge.net ), and the command
to use is ngrep -qd eth0 'GET' tcp port 8080
If this doesn't reveal the culprit you can try filtering for POST
instead, and make sure that 'eth0' is really what you want.
If GeoServer is running on a local machine you have plenty of tools to
choose from (like Wireshark), but on remote machines where the capture
may require more bandwidth than you have) it's trickier.
-Arne
| 6691a87d-8f97-0d05-a8a7-88366f7d0355 | ['5b6423f7-7347-6910-6039-49e864817bc9'] | Hi,
I've installed the latest versions of everything, including DIMP 1.1.1,
on top of a Courier IMAP server.
The problem is that I cannot find a way to sort by message date, only
arrival date (which really appears to be message number). This is a
problem when I move stuff from my inbox into subfolders, the messages
are listed in the order they got moved into the subfolder, not according
to when they were received or the timestamp set by the sender.
Ideally I'd like to expose both the Message date and the message number,
and be able to sort by clicking on the appropriate column. Is it
possible to achieve this without too much hacking ?
(I realize I can still set this through the defaults in message options,
but that's not really something I can tell <PERSON> to do)
-Arne
|
39165f6e-1e94-3eb2-3411-0a013ee0b4af | ['5b73c8a5-bd8b-d40a-7ab9-045f59c6fbb1'] | I believe that I have made some progress on this, but I'm not confident
that my solution is the correct one. There were several factors playing
into this that pulled my attention away from the real issue.
I originally reported that this problem was CGI specific; it turns out
that this is not the case. The way I was reproducing what appeared to be
a major leak was by sending a number of parallel requests to CGIs that
generate a lot of output. The result of this was that Apache needed a lot
of temporary buffers at the same time. After digging around for a while,
I discovered that Apache on Windows never calls
apr_allocator_max_free_set(), and the default behaviour is to hold on to
all memory ever allocated, and allocate memory future requests out of that
block. When I spawned the parallel requests, Apache grabbed a whole bunch
of memory, and then kept reusing it. After setting a limit here, I was
able to move on to the bigger problem.
After making the change above, I realized that every request was leaking a
little bit, not just the CGI requests. To confirm this, I made a simple
configuration file that redirects all requests to a static HTML file.
Sure enough, it was leaking.
The problem, as best as I can tell, is that mpm_winnt calls
apr_bucket_alloc_create() once for each thread, and registers it in the
pchild pool. This bucket allocator is then passed through to
core_create_conn(), and used for all apr_bucket_XXX() routines during the
request. The pchild pool is not cleared until the server shuts down, so
the memory used here grows and grows. To solve this problem I changed
mpm_winnt so that it creates the bucket allocator using the ptrans pool,
which gets cleared after every connection is finished.
After making this change, the system behaved much better. Just to check,
I then undid my first change related to the maximum memory to hold on to,
and the system continued to function corrrectly. So in the end, I only
needed the one fix.
The following patch shows the changes I made. My question now for the
experts is whether this will break anything.
Thanks,
<PERSON>.
| bf7abc6f-bcd7-5c2b-564b-73708e0b03da | ['5b73c8a5-bd8b-d40a-7ab9-045f59c6fbb1'] | Hi there,
I am using Apache 2.0.43 on Windows, and am having some memory issues.
As far as I can tell, Apache is slowly growing larger and larger during
its lifetime. In stress testing, we've seen it grow as large as 2
gigabytes. I have tried running Apache within Purify, and unfortunately
it looks like the memory is properly cleaned up when the program exits,
which makes it tricky to try to figure out where things are going wrong.
Searching through the Bugzilla database, bug 11427
(http://nagoya.apache.org/bugzilla/show_bug.cgi?id=11427) sounds like it
may be similar to what I'm experiencing.
I should mention that I do have my own module loaded, as well as some of
the standard Apache modules. There is a chance that the leak is in my
code, but I don't think this is the case, as my handlers should clean up
after themselves each time they are called. Any traditional memory leaks
should show up within Purify.
The memory growth seems to happen most often if a number requests hit the
server simultaneously.
Has anybody else seen this happening? Does anybody have any suggestions
of techniques I can use to diagnose this?
Thanks,
<PERSON>.
|
d00c9397-7d79-f1c4-3ccf-21e66af7ba5d | ['5b961b8a-b402-8806-567f-e1026d3e3c95'] | Hi,
I m encountering problems to compile the IPAQ sleeve drivers. Drivers sources
exist in the linux kernel but there is no possibility to compile it.
During make menuconfig, it is not possible to activate the option IPAQ_SLEEVE
and IPAQ_PCMCIA_SLEEVE, they do not appear in the menus. If I make a search
during menuconfig, I can see that these options exist but:
IPAQ_SLEEVE is set to IPAQ_SLEEVE (??)
IPAQ_PCMCIA_SLEEVE is set to n
If I modify directly the .config then the options are not taken into account
(.config seems to be overwritten when we do make).
If I modify the h3900_defconfig then the option are not taken into account
either when doing a make h3900_defconfig
Someone here could help me compile these drivers ?
Thx in advance,
<PERSON> | 13e9841e-15af-6046-8cdd-8e9d52394262 | ['5b961b8a-b402-8806-567f-e1026d3e3c95'] | Hi,
I m trying to boot angstrom on a ipaq h3900. Since there are no images
available for this device, i build the angstrom environment using the
following guide
http://www.angstrom-distribution.org/building-angstrom
Once finished, i got jffs2 images. I tried both minimalist and console
images but i got the following error during the boot :
"Kernel panic -not syncing: Unable to mount root fs on unknown-block(2,0)"
Anyone knows where it could come from ?
Regards,
|
f8393953-5166-497e-c448-5bfdd1fb8fe6 | ['5b97d700-3902-c5a8-b42f-cc5d4a9052aa'] | This is one of the places where IE happens to be a better dev env than FF.
FF has a known bug with its caching. If the main page hasn't changed, then
it assumes that dependencies haven't either (JS, CSS, XAP, etc). Thus, it
never even checks whether they're stale and always uses the cached version.
So, if you're doing any web app development in FF, you have to remember to
either Ctrl+F5 after each change, or, when that isn't working, close the
browser without saving tabs re-open it, and nav back to your page (saving
tabs saves the cache info, so you come back to a stale page). I end up
setting a lot of shortcut buttons, so that I can single-click to whatever
page I'm working on. Also many FF instances, with one page per browser,
because using tabs means losing more tabs when I have to shut down the whole
browser. Tabbed browsing? Not in my code windows.
I do mostly straight JS + CSS dev, not Silverlight dev, but run into this
bug there all the time. It's a browser thing, and affects all web app
development on FF, regardless of the rest of the technology stack. Sucks,
but eventually you just get used to "why is that test failing only on FF? Oh
yeah, I have to go hard-refresh all the pages again."
<PERSON> | d18cba51-d45b-835d-4748-cf7f340f7b7a | ['5b97d700-3902-c5a8-b42f-cc5d4a9052aa'] | My apologies, but I announced the wrong date for <PERSON> <PERSON> talk in
January.
This talk is on Jan 14th, the second Wed in Jan. It is not at our usual
third Wed (Jan 21), as I previously announced. I have updated wiki
(http://www.xpdx.org) & calagator, so they shouldn't continue to sow
confusion.
I apologize for any confusion. I hope you all can still make it; this should
be an excellent talk.
Arlo
|
0fef63dd-92e9-0e75-0f46-d17693b6681b | ['5b9ae334-22b5-5539-fe2d-7b8eadf7361b'] | Hello everybody,
I had to migrate from debian to centOS and I would like to build rpm packages
of opensips on a CentOS dev server.
I have retrieved the opensips.spec-4.1 file from the packaging folder in
opensips source code and I'm trying to adapt it to opensips-4.4.
I would be very grateful if someone that have already done that could help
me...
Best Regards,
| e304dfcd-a214-c07f-b9f2-9603569cd6b0 | ['5b9ae334-22b5-5539-fe2d-7b8eadf7361b'] | Hello everybody,
I would be interested, and I'm sure not to be the only one, to have a kind of daily opensips log analyser report each day of:
- blacklisted IPs by the pike module
- total number of calls/transaction (opensipsctl fifo get_statistics tm: ,......)
- flapping or failed gateways of the dispatcher module
- ...
I used to have a logwatch running on my servers, and before trying to add an opensips file into
/usr/share/logwatch/scripts/services I would like to know if you have a better way to do it?
Thanks
|
94699cc8-b948-684b-b7fd-d450632d9785 | ['5b9b8c4f-30c3-d874-4649-6748c81d6097'] | Hi out there!
I got a question concerning xterm and rxvt with cygwin.
When I run xterm under LinuX, all function keys are passed to the
application running (e.g. Midnight Commander). When I try doing the same
with cygwin, only F1 to F4 are passed through, F5 and higher result in
some key messages on the command line. With rxvt none of the function
keys work.
When I log into a linux box and start xterm on the remote session (still
display is the cygwin box) erverything works (all function keys and even
mouse support).
What did I miss with the cygwin installation of xterm? Did I make a
serious configuration mistake?
Regards,
<PERSON> | f63c12cb-bf73-99cd-97f9-10126197eb7d | ['5b9b8c4f-30c3-d874-4649-6748c81d6097'] | Hello!
I am sorry to (maybe) destroy your enthusiasm but from my point of view
there are some more reasons for not running either gnome or kde on cygwin:
The main thing is: they are very large packages and even on my linux box
they are that slow, that I prefer just using icewm as window manager and
only have kde and gnome installed in case some application needs the
libraries. If I start to compare cygwin running under win98 I really can
not advise you to run kde or gnome - maybe on NT kernel based machines
that is different.
Cygwin is (and here the developers have made a real good job - Thanks!)
a very good tool if you want to run applications like nedit (which is my
favourite editor) and other basic unix routines I really miss under
windows - but as any Operating system running on top of another one - it
has it`s speed limitations.
Another great usage for cygwin is to use it for running xdm sessions or
if you have to do admin work on a unix box over the network (it has ssh,
sftp and all the other nice tools which windooze does not have). So for
me (personal opinion) it is a great (and long expected) extension to
that "operating systems" from Redmond which I am forced to use some times...
By the way: I would not even use kde or gnome with linux - but that is
MY personal opinion about it. Please do not kill me for it!
So again: Thanks to all the developers!
Regards,
Uwe |
b92ab33f-2dee-b475-a9ab-0a6483b16e01 | ['5ba658e8-184d-99c3-f4ac-c4bd3ceeec17'] | As it came out, as usual, I unpacked the new revision of the debian linux
kernel sources, copied over the .config file from the previous one, made a
make oldconfig and built a kernel image using make-kpkg. Everything went
smoothly, as far as I can tell. I then installed the new image it on my
laptop, and rebooted into it, to find out that _any_ 32 bit executable would
segfault! Even the ld-linux linker itself!
I reinstalled the old package, rebooted, and I had 32 bit compatibility
back.
Does anybody know what the heck I did wrong? Or what changed between the -6
and the -7 version of linux-source-2.6.32 that might have caused this?
Thanks in advance
<PERSON>
| 6833737e-1831-7172-f6f9-d05b00fa81da | ['5ba658e8-184d-99c3-f4ac-c4bd3ceeec17'] | Hello.
This may be trivial or a known problem, if so please feel free to just point
me to where the answer is.
For some time now, in debian sid it has become a problem to use the nonfree
fglrx driver with gnome 3, and with the latest xorg. So I finally bit the
bullet and switched to the free driver on my laptop (sporting a mobility
radeon hd 3400 series). The good: xorg appears to work fine, even if quite a
bit slower, but since I am not after high-end 3D graphics, it's quite ok for
me. The bad: as soon as kernel mode setting is enabled at boot time, the
console disappears completely and is not visible anymore.
If I disable kernel mode setting, then the console stays visible, but gnome
3 will not work. Is there a known configuration, setting, workaround with
which I can have a working text console *and* working xorg with kernel mode
setting enabled (to run gnome 3 efficiently)? Since shit occasionally
happens, especially following sid, I am used to sometimes drop to a plain
text mode console to fix things, and having a laptop which is completely
unusable until the graphic login manager fires up is, to say the least,
disturbing...
Thanks in advance, bye
Giacomo
|
adfe2d3e-713f-a54e-61d7-a495c7622d5a | ['5bc86e17-50ac-a947-a831-0d52434cfedd'] | <PERSON>,
I agree. IMHO default tuning for specific benchmarks doesn't often benefit
the average user.
Perhaps the default should be the size of the receive buffer modulo blocksize
and an alternate default (pre-set) available to revert to the current
behaviour, (even then I'd prefer to see the existing limit tunable with a
defined safe range, while that preset default was selected in /etc/system).
<PERSON>
| 0623e155-56cd-a822-744d-2326b5cd2c09 | ['5bc86e17-50ac-a947-a831-0d52434cfedd'] | Hi <PERSON>..
you don't in the end say whether you are binding to port 0 or not,
but the fact that you are seeing all your pkts in 'sctpOutOfBlue'
indicates you are probably not binding at all.
Here's how it plays.
The incoming pkt is checked to see if it belongs to a specific connect
via the Solaris sctp stack.
If it's not found in the relevant hash, it is assumed to either be for a
raw IP socket or an "out of the blue" packet.
It is checked against a hash of raw binds, if there is a specific bind it
goes there, if there is a wildcard bind (port 0) all unclaimed packets
go there. If there is no bind in the hash then the packet is deemed
to be "out of the blue".
Either do a specific bind to a port and get only the packets for that port
or bind to port 0 and get all the rest.
The other web page you refer to may have been written prior to wildcard
bind being integrated in the stack.
HTH
-George
|
89c139ea-cc46-8061-b479-23c32285ae3f | ['5bd79d7b-30d3-ab19-6d79-0d7054755d24'] | Hi folks!
I'm sure there's an easy way to do this, but what I need to do is calculate
weekly returns from daily OHLC data, but I want my weekly return equal
log(Friday Close) - log(Monday Open). I'm using timeSeries from RMetrics
and I don't see a quick easy way to do this. Any ideas?
Best,
<PERSON>
| cf89d7ab-b321-09b6-5b99-dd7e03fe7256 | ['5bd79d7b-30d3-ab19-6d79-0d7054755d24'] | Hi all! Does anyone know if a vector autoregression package is avaialable
that allows binary variables as part of the endogenous system? I'm looking
for something along the lines of what is implemented in "Dynamic Forecasts
of Qualitative Variables: A Qual VAR Model of US Recessions" by <PERSON>, 2003, Fed Reserve Bank of St. Louis. Another possibility is the
autoregressive conditional hazard model by <PERSON>. Thx!
Best,
<PERSON> |
3934f51b-49c9-95e5-7263-cad6fdd1e229 | ['5c1750e0-1ab7-88f5-1eae-b61ca6bd0aff'] | Hi,
I don't seem to be able to find any way to trap/handle OS signals in
erlang (TERM, INT etc). Am I missing something or is this just "not
the erlang way", and if the latter then is there some alternative?
I'd like to be able to trap a TERM signal and shutdown cleanly The
best I could come up with was to start my erlang process from a shell
script which itself trapped signals and then sent the erlang node a
rpc message to stop cleanly. Is there a better way?
Thanks,
<PERSON> | 1d6473dd-a377-3a97-3b1a-b58a0043977f | ['5c1750e0-1ab7-88f5-1eae-b61ca6bd0aff'] | Hi,
I was trying to run a php script I obtained as a CGI using a php4
binary with miniserv.pl but kept running into the "No input file
specified" error message.
After a bit of Googling I found that a common problem withr running
php CGIs in this way is that the SCRIPT_FILENAME enviroment variable
may not have been defined - however I see that this is defined in
miniserv.pl after some discussion of this I found in the archives [1].
Anyway - that wasn't my problem, just a bit of backgroud.
I traced the C code of the php4 binary when running my php CGI until I
found that it was failing when doing a fstat() on a filename it had
opened which turned out to be a directory and not a file as expected.
It turns out it is basing this on the PATH_TRANSLATED environment
variable which miniserv.pl had defined as the same as my document
root.
I had a look at the CGI spec [2] which stated the following for
PATH_TRANSLATED :
PATH_TRANSLATED is derived by taking any path-info component
of the request URI (see section 6.1.6), decoding it (see
section 3.1), parsing it as a URI in its own right, and
performing any virtual-to-physical translation appropriate to
map it onto the server's document repository structure. If the
request URI includes no path-info component, the
PATH_TRANSLATED metavariable SHOULD NOT be defined.
I found the last bit interesting so I tried editing miniserv.pl to not
define PATH_TRANSLATED and suddenly everything started working. So -
is it possible that PATH_TRANSLATED is not being defined correctly
here, or iare there cases where it should not be defined as seems to
be being suggested here?
Thanks,
<PERSON>
|
b45f0528-d8b7-48a4-1da6-48c4fb867ef8 | ['5c251618-3fd5-1e09-274d-106473a90f73'] | Thank you very much for your information!
Couple more questions.
What do you mean by ideal/quiet environment? I do have other traffics
currently in the same channel. But the maximum, so far, I have got is only
12.8 Mbits/sec for a signal strength taking from -34 to -43 dBm and 7.26 dBm
for -67 to -77 dBm. What do you think will be the ideal parameters so that I
can tune/setup my mesh network to get the maximum throughput?
Why is that 802.11s doesn't support 11n rates? Is there any thing like we
can configure to achieve this?
Thank you,
<PERSON>
| cbbbe3f2-be42-fefd-a919-4b22c3b9bfd8 | ['5c251618-3fd5-1e09-274d-106473a90f73'] | Hello Friends,
I am studying the mesh networking (using ath9k driver, openwrt backfire
10.03 on a TPLINK1043ND device). Currently am running with six mesh nodes
and doing basic iperf experiment to find the throughput between mesh nodes.
It will be really helpful if you people can tell me what is the throughput
expected out of a mesh network.
For a single hop (direct link) throughput I get is 11.0 Mbits/sec on avg and
for a two hop link am seeing 6.5 Mbits/sec. Is this throughput normal?
Thank you,
<PERSON>
|
3d123bbb-f7e0-44b6-d094-4b950f1d4512 | ['5c2782ba-5fae-0834-c326-1b4b17e20205'] | Several weeks ago I set up a server with Koha 3 beta2 in preparation for
migrating to Koha once koha 3.0 final was released. I have now set up a
brand new server running Debian and Koha 3.00.00 and I am attempting to set
up the same configuration that I had on the original server. As part of
this I need to import two large marc record files. When I attempt to manage
the staged records the progress quickly hangs, and makes no progress for
several hours.
After attempting to import file1, I could look at the results and see that
biblios 1-149 showed as imported, while 150 onwards remained as staged.
Fighting with this for awhile, I got file1 to import fine using
bulkmarcimport.pl, and attempted again to use the web interface for file2.
This again hung, although biblios 1-549 had successfully imported this time,
after that again shows as staged. I would rather not use bulkmarcimport.pl
again, as there are duplicate biblios that I would like to merge, and from
my understanding this is only possible with the web interface.
Both of these exact files imported just fine on Koha 3 beta2. Other than
upgrading to koha 3.0 final the configuration should be the same. Is there
anything that could have broken this functionality in koha 3 final?
Watching top after the import hangs, there is a process "background-job-"
which appears to respawn every few seconds with a new pid. Nothing
interesting seems to show up in any of the log files. I am using zebra, and
the koha-zebra and koha-zebraqueue daemons are both running.
Thanks,
<PERSON>
| 7543f423-7caf-18a3-1a7b-09a280515a80 | ['5c2782ba-5fae-0834-c326-1b4b17e20205'] | Hi,
My library is migrating to koha from Winnebago. I have used
bulkmarcimport.pl to import our entire collection into koha, which looks to
have been successfull. I am running into trouble scanning in the barcodes
on our books however, because winnebago stored only the 5 digit material
number in the 852p barcode field, but the barcodes on the books contain much
more information, in what I understand to be the codabar format.
After the import, Koha now has for instance that a book's barcode is 12345;
But the barcode on the book, which we would like to keep is 3055500012345x,
where 3 means it is a book, 0555 is our location code and x is a check
digit. In the old software typing in either the short version, 12345, or
the long version 3055500012345x, would get the same results. Koha works
fine with the short version, but reports an invalid barcode if you use the
barcode scanner to scan in the book, since it is in the long format. Is
there any way to get koha to interpret these barcodes correctly?
Thanks,
|
9f0581ff-517b-856d-fd61-cfdc57e8f82d | ['5c31f905-b85f-2f0f-ffd7-dcc0077e219f'] | Dear mailing list,
I try to get a final solution to the following problem: I have two color
microarray data (common reference, dye-swap, multi-factor, multi-group) of
which I want to extract differentially expressed genes (DEGs) using LIMMA.
Those DEGs I want to analyse for the over-representation of GO-Terms using
GOstats.
I got as far as a list with GO-Terms and their respective P-values and even
could construct a nice direct-acyclic graph out of this data.
*However, the output I get does not account for the following:
Are single genes/transcripts up- or down-regulated?
What about the up- and down-regulation of whole over-represented GO-Terms?
*
*Could I get there using clustering approaches?
*
I'm sure somebody on the list has already tried such an approach, *finally
yielding over-represented GO-Terms, which are either up- or down-regulated.*
Thanks in advance,
<PERSON>
| 8c782d47-6a32-5e1f-ec7f-9d4ba7dcada4 | ['5c31f905-b85f-2f0f-ffd7-dcc0077e219f'] | Hi bioconductor mailing list,
I have severall datasets and all of which share 2 reference samples. I want
to do a normalization that takes this into account, so that each dataset is
normalized against its own reference. In the end I would like to have all
datasets normalized in this way, so that when thrown into a PCA the
reference samples cluster together.
Your help is greatly appreciated! If possible, please guide me to
BioC-packages or functions as I am pretty new to BioC.
Thanks in advance, <PERSON>
|
d7b5811c-16ab-e5a2-b9ce-4125c6da3181 | ['5c41a08f-5950-45e0-9b86-9f107a088e64'] | Hi!
I've just pushed five changes from Ubuntu xorg to alioth xorg.git..
Hopefully they all are useful, and if not, easily reverted :)
There are a bunch of other changes that could prove useful. I'll try to
either put them in a git branch or otherwise separately available for a
review. I've done that for an older version back in March, and they are
still available here:
http://users.tkk.fi/~tjaalton/dpkg/xorg/patches
I'm not sure how radically the postinst is going to change in the future,
so maybe some of the quirks could be dropped now.
comments?
t | 8dcfd402-925c-ba7c-3406-c91665880f32 | ['5c41a08f-5950-45e0-9b86-9f107a088e64'] | Hi
I've now even tested the patch against 1.5.3, and it works like a charm.
Ok to commit? The rationale is pretty clear I think. "Some" distributions
might have used 'vmmouse' instead of 'mouse' by default, and those who
upgrade to a newer release could get upset when they can't single-click :)
t |
fa968a95-cdd0-04b9-96c8-250588f86c82 | ['5c445b60-086f-b700-2397-7c67790e6186'] | FWIW, after the first day of WWDC this year, I decided to ship MoneyWell 2.0 as a Lion-only app. There has been some negative flashback from our customers, but not too much. We'll see if that increases after we (finally) ship it. We will continue to offer our 1.7 release on our website for those people still running Leopard and Snow Leopard, but the development advantages of moving to Lion were too great to ignore. We were able to ditch a huge amount of custom code and add most of the Lion-specific features.
Additionally, there is the marketing angle. Apple does tend to promote apps that take advantage of their latest OS more than older apps.
Peace,
<PERSON>
| 3094b853-41cc-c027-2692-0a03a9f77fff | ['5c445b60-086f-b700-2397-7c67790e6186'] | I'm not sure what the issues are with the Lion UX, but I love them. I think that some people will always struggle with change, but if you give Lion a chance you'll see how important the changes in it are. For the average user, not having to deal with opening documents or restoring apps is huge. Snow Leopard was a mess of the old Aqua and modern variations of controls. Scroll bars have needed to die for a long time and gestures are a huge time saver for us trackpad users. :)
Lion adoption will be much faster than Leopard and Snow Leopard. I think you'll be doing your customers an injustice by clinging to the past and infecting them with negative comments. Like it or not, Lion is the future of OS X and you should embrace it.
Peace,
<PERSON>
|
54bbc8b0-64cf-bfee-9d71-d32e5c1f95ac | ['5c44f1bf-f561-6803-b571-f165fc86a639'] | Without going into a sales pitch (you can get the detail off our
website) we do offer such things...
Home Phone Wireless is a fixed-line service running over our mobile
network (so you get a landline number).
Home Zone tells your phone you're in the house when you arrive home and
converts it from a mobile to your landline number (which I think is what
you're talking about in the US).
And just to clarify a point about mobile termination rates - MTR and
Fixed Termination Rates (FTR) are exactly the same in NZ. We don't
differentiate. Currently they're both 15c/minute (due to decline by
another cent in April next year).
Cheers
<PERSON>
<PERSON>
| 52fd35ac-1750-4026-b4a5-47e5425b1df5 | ['5c44f1bf-f561-6803-b571-f165fc86a639'] | Hi all. We can't make any public statement about this interference until after the court hearing, early next week, but after that I'm happy to get one of the network guys to come in and chat about it with you if a: that would help, b: that's of interest and c: they're not already here under their own steam.
Cheers
<PERSON>
|
a1324183-cc8c-bcb2-2b4d-8dbe94038d99 | ['5c491191-c639-91a3-02de-53525c3da746'] | With TortoiseSVN, labels are decorated and svn operations are available on
any folder/file in a subversion repository regardless of its depth in the
filesystem tree. I think that this is a great feature.
I'm pretty sure that with subclipse, you need to have the root of the
project checked out from a svn repo. I was wondering if it is possible to
have subclipse recognize any subtree not rooted at the project root as being
under subversion control?
Thanks,
<PERSON>
| 47af42bd-5b7c-8dc1-ea30-76c924130c15 | ['5c491191-c639-91a3-02de-53525c3da746'] | That comment line in Tix.py seems to be specifying information for Emacs
python-mode. Is it?
If so, should PyDev even be looking at this at all?
I would venture to say that the warning should not be suppressed. If the
Emacs specific settings are ignored entirely then the error won't happen in
the first place. Is this a worthwhile approach?
Regards,
<PERSON>
|
441c9800-c90f-5ef3-7518-9a31f22be44a | ['5c54a7f0-779c-e2f0-8f05-beb84a00725e'] | Dear all
During the last weeks I have learned quite a lot about the QoS
mechanisms available on the Cat6500 platform.
Now I would like to confirm somehow whether I understand one thing
correctly:
As soon as I have a connection with less bandwidth than the interface
line rate connected to the Catalyst there is no way to achieve decent
QoS functionality involving class based and/or priority queueing
according to the low bandwidth available behind a
(Fast)Ethernet-interface, right?
One can use the PFC to globally police the output rate of the port. But
although it is possible to use different classes with different policies
here, one cannot do any flexible shaping/priority- or bandwidth-queueing
at this level.
The actual cos-based queueing will be done on port level and its
mechanism always works in relation to the actual port line rate (using
wrr bandwidth RATIOS).
Thus there is no way to achieve something like a shaping down to 5MBit/s
AND considering the dscp/cos-classes at the same time.
And it does not matter which queueing capabilities the linecard actually
provides. I tried this with 6348, 6548 and 6748 together with Sup720-3BXL.
Did I get this right or have I overlooked some (more or less obvious)
solution?
Regards,
<PERSON>
| b9c17713-7178-158c-d41a-fb62847caec4 | ['5c54a7f0-779c-e2f0-8f05-beb84a00725e'] | Hi all
Today we were facing a ddos attack with approx. 1.5 GBit/s and 3
Mpackets/s. The traffic came in on a port of a 6704 and went out on
another port of the same linecard.
We have two annoying phenomena to deal with:
1. The 6509 is equipped with a Sup720-3BXL and DCEF on all linecards.
During the attack it dropped all of its LDP neighbourships (actually
they were flapping) although the CPU usage remained below 40% for most
of the time and only went up over 50% at a very short peak not related
to the LDP drops.
What might have caused the dropping of the LDP neighbourships and what
would one do to protect them?
2. On the outgoing interface we recorded a huge amount of output drops
up to 60kpackets/s. The receiving router here is another Sup720 with
6704 but without DCEF.
From my understanding 1.5GBit/s and 3Mpackets/s should be no reason to
drop packets on a 10G link.
So what might have caused these packets to get dropped?
Thanks for any advice or hint,
<PERSON>
|
4c20e305-9db7-dc0c-5378-5e00953e0fb7 | ['5c5b047b-a330-bfc1-e3c9-4f499b564592'] | Hi all,
I'm using Apache 2.2 on LE for several years now and i use intensively the
virtual host mechanism (300 virtual hosts per http instance).
As soon as you start sharing resources between applications, you want to
protect your platform and prevent an application from taking all the
threads and cause a resource starvation for the others.
To do so i used the third party mod_qos module to limit the number of
simultaneous connexions per virtual hosts. Now that i intend to migrate to
Apache 2.4, my first tests reveal that this module is not compatible with
this version of Apache, that's also the warning that developper raised for
his module.
It appears that i don't find any real alternative to substitute the mod_qos
module with something else. Here are my questions :
- Can you confirm that Apache does not provide any mecanism allowing to
limit the number of connection per virtual host, just to be sure that i'm
not missing something ?
- Why Apache doesn't provide such a functionality ? :) From my very humble
user perspective, i'm surprised that this is not a native functionality as
it seems to me that my need is probably shared by many users. Moreover
Apache provides many other complex functionalities, the one i'm describing
would be probably something simple to implement compared to other
functionalities already available.
Thanks in advance for your feedbacks.
<PERSON> | d93af485-669c-09d2-bd33-7f98f4e98a28 | ['5c5b047b-a330-bfc1-e3c9-4f499b564592'] | Hi everyone,
I'm actually using Tomcat on my environment platform (Tomcat 5.5 / Tomcat 6
and soon Tomcat 7). I have a frontend Apache http Server using the jk
connector to communicate with Tomcat instance.
I'd like to change this connector and use the mod_proxy one for several
reasons. The main difficulty to handle is relative to the remote-user
information. Indeed the jk connector automatically transmits the information
so that the application can retrieve it using a request.getRemoteUser()
method call.
If i'm not using the ajp connector anymore, i need to handle something on
the tomcat side to set the remote user in the request object. I thought i
could use a valve to do this. And that's where the road ends, i have watched
the ajp conenctor code in order to see how the remote user is set in the
request but i can't find it.
Could you please tell me how i can do this ?
Best regards.
<PERSON> |
381cf7d1-0322-4479-4ded-628b115646ec | ['5c662a1a-b2d8-6a9d-71f5-9e6f498d3deb'] | Hi,
This is probably off topic, but out of curiosity I wouldn't mind asking.
Does anybody experience system hang ups after upgrading to 8.0?
IBM T43 8.0 generic (thought that something went wrong during upgrade
from 7.2 to 8.0. I did fresh install, but the problem persists).
Which would be the correct mailing-list to post the message to?
Regards,
<PERSON>
| 0d0d8b83-cbfe-8403-3258-e3f9d342f4da | ['5c662a1a-b2d8-6a9d-71f5-9e6f498d3deb'] | Hi,
Does anyone experience sluggish kmail when scrolling a folder with emails
containing attachments? Nothing crushes or hangs up, but it takes +- 30
seconds to open an email.
Same goes when sending an email with an attachment (ie. with jpegs).
It takes a good couple of minutes to send it (ie. progress bar stops and 99%).
Regards, <PERSON>
|
fa1bd3b5-ea63-590e-b3b0-9d36bad2d9c9 | ['5c668d29-4e78-b5c8-ce42-31c77f2b6f95'] | Hi,
I have successfully compiled XMMS 1.2.5 for OpenWindows on Solaris
2.6.But...
The problem is that when I play MP3 files they are broken up due to what
would seem to be the MP3 decoder not decoding the data stream fast enough.
The CPU is only ever about 12-15% utilized but it is almost like something
is preventing XMMS from using more CPU in order to cleanly/smoothly decode
MP3 files.
I have tried turning on "Use Realtime priority when available" but this does
not seem to have any effect.
Also tried to re-compile after changing the -O2 flags for the gcc compiler
to be just -O and still the same result.
I am using GCC 2.95.3
If I use a command line MP3 player (mpeg3play) all the MP3s I play decode
and play perfectly. Only problem is that being command line it is not very
robust. ie. No play lists or user interface etc.
Any help or suggestion would be appreciated.
Cheers,
<PERSON>
| caad85db-905b-6044-dfb3-bf9738b9b8f2 | ['5c668d29-4e78-b5c8-ce42-31c77f2b6f95'] | I have what would appear to be a similiar if not the same problem on SUN
Solaris 2.6.
mpg123 on it's own compiles and runs fine ( takes 8-10% CPU when playing
MP3s).
XMMS plays the MP3s but the audio is broken up, almost as if it does not get
enough CPU time or something.
<PERSON>.
|
f147b2f0-2cc0-a0a7-7dcc-d264ff6652b1 | ['5c6c99a8-cfbb-76d0-5494-72863bbe5761'] | Thank You : )
your post gave me some comfort : )
Yes, the problem is related to paging, and not necessarily activity/places.
as for your opinion of master/detail,
I think it is a shame to not take advantage of the fact that we can
display master/detail side by side, without a full page reload,
something that was not possible with traditional webframework,
and is nearly possible now with AJAX/GWT, only if there comes a solution
to implement the paging correctly.
regarding your suggestion on using "continuous scrolling",
I am going to investigate that now, and I believe it will work,
however my concern is this,
lets say we have a list of employees with continuous scrolling,
and next to it, we show detail of employee, when one employee is selected
from the list.
if we bookmark place employee/777
when navigating to place employee/777 for the first time from bookmark,
how can the master activity which contains a continuous scrolling
table/list,
highlight employee with id 777 in the list, when in fact the continuous
scrolling list
presents only an X number of items initially, and load the rest on demand ?
do you think this is possible ?
Thank You
| ef536cf2-6daf-5808-29c8-c4d43ed6e0b7 | ['5c6c99a8-cfbb-76d0-5494-72863bbe5761'] | Thank you very much for sharing your experience.
on this comment,
" I also pay close attention to widget lifecycles which is critical to
successful implementations of widgets that get attached, detached and
reattached and so on and so on. "
why widget lifecycle is important ?
why is it important when they are attached/detached ?
and how can you monitor it ? (logging your java code or viewing resulting
html markup gwt creates ?)
how can you tell in what state a widget is: attached/detached ?
what are the states ? only attached and detached ?
Thank You
|
39001400-8a8f-5e78-a790-0ff65060bcd6 | ['5c6d97f1-5fa2-fe0a-5ce2-148c8fb43c9b'] | Hello,
I guess this is the wrong place about asking about qmailrocks and
bitdefender..
It looks like it can be done following qmailrocks's install guide, just
replace clamav with bitdefender.
Howto install bitdefender, ask/read on bitdefender website.
QmailScanner will check for virusscanner and set apropriate settings - on
compileing/installing, else follow the guide.
http://qmail-scanner.sourceforge.net/
But I would suggest you to try clamav if there are no reasons not to do so.
better knowledge about it in the forums ( on qmailrocks ;)
B/R
Ole J
| 3ee95420-19ae-5c98-17c8-c6d87828b528 | ['5c6d97f1-5fa2-fe0a-5ce2-148c8fb43c9b'] | Obvious the table are missing,
For what I have experienced, newer mysql db's does not support . ( dots ) in
table name, like vpopmail.mydomain.com
Anyway I am not a MySQL wizard, but this is something you can give some
thoughts until some of the specialists here on the mailinglist wakes up
Try to migrate the data from old table to a new table without dots.
Navicat software is a good thing for handling MySQL - transfer, backup and
more.
B/R
Ole J
|
276b635b-eaab-2268-cc72-87eca90c21e1 | ['5c701c64-8e8a-ce3d-1d3e-6f4c0b0ff75e'] | Hi
I am having two problems with my fedora 10 system.
1 Once I tried modifying
/usr/share/PolicyKit/policy/org.freedesktop.hal.storage.policy to make
it mount removable pendrives automatically always
without any authentication. Now I dont know what went wrong, I am
not able to mount my removable drives with deskbar applet or from
nautilus. I have to manually mount it using mount command with root
priviliges. I tried replacing the package that provides this
file(brand new file) still no succes.
2 I dont get a shutdown in my system menu. for that matter no where in
my GUI, I am able to use shutdown or restart (including GDM). I always
hv to shutdown or reboot from command line these days.
Please help me if somebody already faced this problem and worked around it.
thanks in advance
<PERSON>
| ebe11f45-ef03-cdc8-517c-68d96ab172b5 | ['5c701c64-8e8a-ce3d-1d3e-6f4c0b0ff75e'] | Hi
I am writing a graphicsitem that displays a lot of thumbnails on screen
in a big window and then
depending on navigation keys browse through them and some action on
selecting any of them.
I have done most of it using pixmap items and grid layout. But now i am
confused whether usign QTableview will give me any advantage other than
scalability or it will be overkill for such a small task.
Thanks
|
6e785f67-f01f-7b91-49f8-d7fceef59ba6 | ['5c83777e-f957-b42f-5c08-b7be0f479299'] | Some time ago, i was getting an error while trying to use RTTI
functions in a class descending directly of TObject
Later i discovered that i needed to descend from TPersistent due to
$M+ directive (http://www.freepascal.org/docs-html/prog/progsu44.html),
so i did.
Now, i decided to go back and descend directly from TObject. To allow
RTTI usage i would put M+ around my own class
I tested again and using RTTI functions (SetOrdProp, GetPropInfo)
worked even without declaring {$M+}
So, is it really necessary to put $M+ to use RTTI?
Luiz
| ca612a5c-82ee-fc12-3089-236896021859 | ['5c83777e-f957-b42f-5c08-b7be0f479299'] | Hi,
I've managed to set a COM interface property (tkInterface) through RTTI
using SetInterfaceProp.
But i could not find a way to set a CORBA interface property
(tkInterfaceRaw) since SetInterfaceProp expects IUnknown.
In the attached example, when i try to use SetInterfaceProp i get the
message "Incompatible type for arg no. 3: Got "IMyIntf", expected
"IUnknown"".
Is there a way to set a CORBA interface property through RTTI?
<PERSON> |
9293628e-a431-09fd-3adc-01f71c3ee922 | ['5c94ccd4-b89f-6cde-92b8-c980abe11c42'] | Mifos 2.1.5 is now available. This is a hotfix release that includes the
following fixes:
* changes to mandatory savings amounts are now accurately reflected in the
UI and appear correctly in the collection sheet report
* issues/withdrawals amount is now displaying correctly in the collection
sheet entry pipeline
* Improvements to Question Group query to improve loading speeds of
center, group, and client pages
* Other improvements to the batch job the generates meetings for savings
* A removed fee is no longer showing on the 'Review installments' and
'Preview Loan account information' pages
* We fixed the collection sheet error: "The amount specified for <Loan
product> of customer <customer name> is invalid. Please specify a
valid amount."
For the download and documentation, please see
http://mifos.org/product/download-mifos.
List of issues fixed can be found here: http://tinyurl.com/3et7u55 and in
the release notes: http://tinyurl.com/43sga7p.
Please note the upgrade instructions as well when upgrading to this release.
Cheers,
<PERSON>
| 67170ccc-bb6b-7936-0c58-d880d0130278 | ['5c94ccd4-b89f-6cde-92b8-c980abe11c42'] | Mifos 2.1.6 is now available. This is a hotfix release that includes the
following fixes:
· Stack trace appears is no longer appearing after submitting
Question Groups with response which has more than 200 characters
· Client details page are no longer loading too slowly for some MFIs
For the download and documentation, please see
http://mifos.org/product/download-mifos
List of issues fixed can be found here: http://tinyurl.com/427kfya and in
the release notes: http://tinyurl.com/43sga7p
Please note the upgrade instructions as well when upgrading to this release.
Cheers,
<PERSON> |
3b56dde5-2183-bad8-e626-5b0be52ba20f | ['5c96b32f-a765-a2fc-74e3-f23b4c773c64'] | Hi <PERSON>,
Occasionally I receive strange error report like the one below. You see
that the AND part is missing the right side of the "iss_id<>"
expression. This is not the only SQL code I get this strange behavior.
Seems to me the original caller may had the value passed in the call
chain but somehow the value got lost. Reviewing the code this should not
happen, but then I see the result like this below.
I experienced at another case that, for instance, de-associating a
custom field from a project (which has project code less than 10) will
delete the custom field values from ALL projects accidently - due to the
same malformed SQL as like above. However if I de-associate from project
where the project code is equal or greater than 10 the value removal
will work fine and removes only custom field values belonging to that
particular project.
Do you have idea whats going on here? What kind of Eventum/etc logging I
could enable to track this down?
I suspect the step when a value is taken through the web interface.
Thanks
<PERSON>
| 076dc399-af4d-dd6f-6397-e8a357e7b81a | ['5c96b32f-a765-a2fc-74e3-f23b4c773c64'] | All,
We experienced sever data loss recently: while de-associating a custom
field from a project all these custom field values have been deleted
*across all* the projects. I would expect that only those custom field
values would be deleted which belong to the project de-associated from.
Say have a custom field "notes". This is associated with several
projects (A, B & C) and the projects' issues actually store values in
this field. When we de-associate this "notes" field from project A we
will see all "notes" values disappear across projects A and B as well.
However I would expect "notes" values for project A & B would remain in
place.
My small investigation shows that there is a table issue_custom_field
which stores all the actual custom field values for all projects. During
the de-association step I see that all the records belonging to the
custom field number (icf_fld_id) are deleted, not taking into
consideration the project specific issues (icf_iss_id).
I guess on the code level this hits the removeIssueAssociation() at
class.custom_field.php. This function makes an attempt to collect
together the list of project specific issues which would be used a few
lines down for the actual delete statement as part of an "IN" clause. In
my case this list gets long (1000+ entries) so I guess this hits the
limits of an "IN" list and somehow executes ignoring the list and
resulting in the data loss. All this is guess - I need to go ahead and
re-work the IN statement to a JOIN to confirm this. There is no error
msg logged during the operation.
Anyone else had this problem before? If the above right then better off
not to de-associate any custom field until this issue gets fixed...
<PERSON> |
25c9562f-a844-bed0-68ed-e04730b7d094 | ['5c9ba563-e2ac-6118-316e-694edff9c5a7'] | Hi,
One of my local users needs to be able to receive specific emails from
one particular user, however, the sender is sending attached files which
are zipped .exe files. At the moment, I' having to unquaratine these
mails daily, but ideally I would like to set MD so that the extensions
check is skipped for this sender.
If I put in a filter_sender section, where it matches the sender, that
returns a ACCEPT_AND_NO_MORE_FILTERING, will this be executed before the
extension check, or will I need to do something different here ?
Thanks,
<PERSON> | 6d8c0443-6068-6ca0-aa3b-17d4133aba1d | ['5c9ba563-e2ac-6118-316e-694edff9c5a7'] | Hi,
Could some please let me know what files and the best way to copy an
existing AWL and Bayes DB on a system running MD 2.43, SA 2.63, to a
totally new machine running MD 2.51 & SA 3.0.3. All the path names,
configs etc are in the same place.
The procedures that are in the notes only show upgrading the same
system, not physically a new machine.
Thanks in advance
<PERSON> |
03d5f8c2-b341-6ef9-31e5-64c4cdbe25f4 | ['5cba6505-a934-79d5-e952-370699eb49cd'] | I'm not that experienced with Maven, or Eclipse.. So forgive me if this is a
stupid question.
I have a class that relies on a properties file. I place that in ./conf for
Maven to pick up and add to the War.
However, if I run my test files in Eclipse, those file haven't "moved"
yet... So the path for the app is different.
Can anyone give me some suggestions for how to deal with that?
<PERSON>
| 1a6a6478-00c6-1456-ed9d-62dc345d4d5c | ['5cba6505-a934-79d5-e952-370699eb49cd'] | Using maven 1.0.2 deploying to Tomcat 5.5.x
and creating struts app with maven genapp struts
I can't seem to figure out where to put the context.xml so it get's placed
in the correct spot during deploy.
I'd normally create webapp/META-INF and put context.xml there... But it
seems like xdoclet would want to handle this rather than just letting the
file structure of the app do it.
Any thoughts?
Greg Akins
STP Systems
<PHONE_NUMBER>
tD+8vlTcEam3Gh/D@example.com |
434c0e0e-43f2-5daf-842a-88a17b012044 | ['5cc16554-a78d-6bd3-a7f1-08b38f34b4e7'] | According to
http://developer.mozilla.org/en/docs/Code_snippets:Post_data_to_window#POSTing_data_to_a_new_window
I've tried to post form data to a new window.
Providing aReferrer with null, the new window opens up but does not
show anything (not even the URL I've provided). If I also set the
aPostData to null, at least the URL (and hence the page without
processed parameters) is displayed.
So, obviously, posting data according to the provided example does
not work. Can anyone confirm this or give more hints on what could
have gone wrong? Thanks for feedback!
<PERSON>. | 15370618-5d2b-0d38-881f-3b7d6ad07466 | ['5cc16554-a78d-6bd3-a7f1-08b38f34b4e7'] | Hello all,
I've written a Firefox extension and it's doing some
initialization stuff when the browser window is opened.
Now I'm looking for the easiest/best way to assure that
the initialization routine is only run once, ie. if
the user opens a new browser window, the extension
should run again. Any ideas on how to achieve this?
Thanks in advance,
<PERSON>. |
abe07e37-314e-c288-c05b-8bbe179ef197 | ['5cc1c17f-6668-72c8-3f7e-2704826b69b6'] | Please enter the details of your bug report here
The ECB methods buffer is not displaying class members in my C++
source files. However, it does display the non-class methods that
exist in the same file. It also is not jumping to the non-class
method when pressing return or clicking the middle mouse button in the
methods buffer.
These things work with the older version of semantic.
Thanks for looking into this,
<PERSON> | d69a5530-7276-6906-3689-7cf483c68e0b | ['5cc1c17f-6668-72c8-3f7e-2704826b69b6'] | Hi,
If I have an edit window with source code in it, ecb populates a
methods buffer. When I start VM (M-x vm) this opens a new frame. I
read my mail and then quit VM, which closes the frame, At this point
the ecb methods buffer is cleared, forcing me to rebuild the methods
buffer.
This seems to be new behavior in the 2.20 release.
It also seems to be related to VM in some way. I can open new frames
manually, or use other packages, like the woman manual reader which
opens and closes new frames, and not see this behavior. I tried using
the (defadvice fix as described in the conflicts section of the manual
and it did not fix the problem.
Thanks,
<PERSON>
|
7024aa90-0e74-63c2-2d83-e165db7603a9 | ['5cc335a4-a5d5-d9ac-f8bd-83a4104458b1'] | Hi,
Can i have variable think times for different page.
Some of my pages people will be spending maybe 30-40 seconds without initiating the next request. some they might do it in 5 seconds, some maybe 10-15 seconds.
can i have such variable think times. Does a random timer do something simlar. Or it randomly assigns think time to any request and doesnt take care of these special pagesa that user spends more idle time on.
regards,
<PERSON> | 15299c80-aaf7-3e02-9631-2670fae4449e | ['5cc335a4-a5d5-d9ac-f8bd-83a4104458b1'] | Hi,
Can I assign a different IP address with a users sessions ?
I got a load balancer thats got sticky sessions associated to source IP. Most request go from 1 ip as they are being generated from couple of machines only, eventhough i got huge number of users.
AS all requests come form 1 ip all sessions get routed to just one app server. I got 4 app servers with this load balances, remining 3 dont get any requests at all.
Is there any way to mimic varying ips?
regards,
<PERSON>
|
bad7cb20-7880-c086-1640-2f710dcc580f | ['5cc53c6a-d543-5488-8c80-95910976965a'] | Hi <PERSON>,
Here's the promised update with the big multi-BSSID changes,
which are related to HE (802.11ax) for which we also have some
updates.
Note that I took the liberty of including <PERSON>'s rhashtable
API removal, I merged net-next for that and verified that the
API he removed is also not used in anything new, but if you
apply/applied any patches after I pulled you should probably
double check that again.
Please pull and let me know if there's any problem.
Thanks,
<PERSON> | 383f59f0-1d18-a2ee-e74a-7dd8b404f776 | ['5cc53c6a-d543-5488-8c80-95910976965a'] | I collected everything on
http://johannes.sipsolutions.net/Projects/mem1394
It's tested now and works for reading, writing isn't implemented yet.
There were some bugs in the previous patch for the dynamic device
registration (wrong checking of return value of cdev_add) and in the
node_interface registration (misnamed function) that have both been
corrected now.
If you want, I'll post the patches for a wider audience to lkml.
johannes |
0e52ee4f-f8e6-02fb-251b-3493a65cf057 | ['5cc67edc-89b0-5c6f-16d3-dac981c4c3d0'] | Not a complaint about the Tapestry list (or PMC members) but I though
I voice my opinion about the policy to see if there are others like
me. I think I understand the reasoning behind the Apache mailing list
policy but I think it's outdated. To me it has two problems.
1. I cannot participate in the list from my work email and I cannot
read my personal email (easily) while I'm at work. I suspect there are
others like me.
2. There is no real built in archive system which makes web searches a
problem. Nabble is considered a hosting site and hosting sites are
blocked by my work firewall.
These two things make it difficult for me to participate in the email
list and make the archives useless to searches because when I click on
a result 9 times out of 10 I get a firewall violation. It's things
like this that make open source adoption problematic in the corporate
world.
I realize you are bound by the Apache rules but if no one complains
about them they will never change.
Thanks for your hard work
<PERSON>
| def04487-890c-70a8-98a4-c297bed5fb57 | ['5cc67edc-89b0-5c6f-16d3-dac981c4c3d0'] | I migrated 2 sites from 3.x to 4.0.x. I just put in the new libraries
and fixed everything that did not compile. For the most part the 3.x
html and page files will work. Both took a couple of days to get a
site that mostly worked. The biggest problems were things like PDF
pages, all the plain HTML stuff was pretty easy.
But as you say, you end up the something odd after this. On one
project I just put up with the fact that it's an in transition
project. Anything new I do the 4.x way. When I have to mess with old
code I update things. On the other project I took the opportunity to
clean things up and started over. I needed to make some schema
changes, I wrote a bunch of hivemind services and pretty much rewrote
the entire site. That took about 6 weeks but the new site is better,
has test cases and is much easier to work on now.
I would try and covert the site, get some experience with 4 and when
the time is right rework the site to take advantage of the 4.x
features
<PERSON>
|
14633768-e909-92c6-9b47-966e27aa0acf | ['5ccc1537-f5f8-6b99-d5a6-777cdb9cde64'] | Hi,
I am using the YAHOO.widget.simpledialog to show a dialog to confirm if they
want to go ahead and submit .But i found that the dialog would show and
would not wait for the answer .the code would proceed further and submit .I
know that the answer to the dialog can be handled by functions .but since i
am going top down on my code to check and submit what do you think would be
best possible way to achieve this.
<PERSON>
| f8fe10d2-b31e-30b9-ae16-35c0f2fee080 | ['5ccc1537-f5f8-6b99-d5a6-777cdb9cde64'] | Hi Everybody,
I tried changing the property of the calendar after it is rendered.I think
it has to done using the method queueProperty(I am not sure if I should be
using setProperty) of config object.After which I am firing the queue using
fireQueue of the config object.As per the documentation this sets the
property and fires the event.what I found was it was setting the property
properly .but would not fire the event .
Here is a link to what I mean
http://chinnakaruppan.googlepages.com/fireQueue.htm
.Try hitting on the next of the calendar .it would throw an alert.But when
you click on the button it supposed to throw on alert which it is not
doing...
thanks for you time....
THnks
Chinna
|
600625f8-74aa-4f1a-715b-a978b22295cf | ['5cd4a2f1-5c8f-57c3-d37c-64f8ba119dca'] | I downloaded the V9.5 back two months ago, we have NOT applied any of the
V9.5 yet, so all my developers are still running V9.0.
Do I need to get a fresh download of the base V9.5 or should I continue to
use the one I downloaded two months ago?
Thanks,
| c2549f2d-c736-90a4-a385-dd945b751757 | ['5cd4a2f1-5c8f-57c3-d37c-64f8ba119dca'] | I do not understand what all the hoopla has been, to be totally honest. I
just updated my RDi V9.1.1.1 to RDi <IP_ADDRESS> by just simply changing my
repository from 91 to 95 in the IBM Installation Manager. To IBM's
credit, the Install Manager A) updated my IBM IM from 1.8.1 to 1.8.3,
then updated my Version of RDi, and took care of loading up the latest
Fixpack. This all happened with just running the Install.
Maybe it was my misinterpretation of the Blog over the last several months
but this was really nice. I thought I had to download the new version,
then download the fixpack, etc. but after all was said and done I did not
have to do any of that. AM I just losing it or did I just get great
service from IBM?
Jerry Roedersheimer
Rumpke Consolidated Companies, Inc.
Systems & Programming Manager
<PHONE_NUMBER>
|
dd57f6b1-6a44-966c-f07f-55faaa6627af | ['5cd8e141-6ea9-c034-1719-18b038bf2f0f'] | Hi -
I'm using Plone 1.0.1 + CMFCTRAddons 0.1 on two servers: one FreeBSD and
one WinNT.
I'm uploading content via WebDAV using Windows 2000 Web Folders.
- I can upload News Items to both servers
- I can upload Plone Events to the Windows NT server
- When I try to upload Plone Events to the FreeBSD box, I get an error
that says "An error occurred copying some or all of the selected files"
- If I uninstall CMFCTRAddons from the FreeBSD server, I can upload Plone
Events (Of course, they get turned into Documents in the process)
Does anyone know what's going on here? Thanks.
<PERSON> | fa6ed431-91a7-ac6b-5fc6-759e957b8583 | ['5cd8e141-6ea9-c034-1719-18b038bf2f0f'] | Hi-
I recently posted two problems with Plone Events/Calendar
1. Events on March 31 didn't show up on the slot calendar
2. Could not upload Events to Plone via WebDAV (with CMFCTRAddons)
The really odd thing was that I observed these problems with
Plone 1.0.1 on FreeBSD, but not on Win32!
The answer: Apparently, the Win32 Plone 1.01 installer was
built with a patched version of CMFCalendar 1.3
I've installed CMFCalendar 1.3.1 on my FreeBSD site now,
and the problems have gone away.
Thanks to <PERSON> and others for getting me started with
Zope debugging, so I could track this down.
<PERSON> |
a6a2ea6c-a2a6-6c51-e441-e38b0c08cd5e | ['5cf49ef7-b50a-2bac-5863-36053e18670b'] | Hi
I am having a problem executing following query:
cursor.execute("select * from projects where projectid between '%s'
and '%s'"%(id1,id2))
But say if i have projectid's from 1 to 20 in database & if i give the
project range 1 to 2 then it gives me the result 1,2,11,12,13,14.....
When I execute this query in mysql i.e
select * from projects where projectid between 1 and 2,I get the
correct result.
I also tried doing
cursor.execute("select * from projects where projectid between '%s'
and '%s'"%(int(id1),int(id2))
but no success
What should I do??
Regards,
<PERSON>
| cfb87f8f-ef80-b1a3-4e6b-bd711200356f | ['5cf49ef7-b50a-2bac-5863-36053e18670b'] | Hi
I am drawing a vertical barchart for project vs start date
The vertical bars are for projects.
Now, at the top of each bar I also want to display the number of projects.
Also I want to show the color convention i.e
Say I have different bars for H.R projects,etc...
Now,if HR project bars are of blue colors then at the side of graph I want to show a small rectangle of blue color with "H.R project" beside it.
How should I do this?
Regards,
<PERSON> |
4cd6279e-3d75-e509-450f-d108bd9043a9 | ['5cf565f7-7d5c-0504-c76c-483aa3f133dd'] | I added a node last week which appeared to work fine. But, after a
while, it loses its default route, and so cannot access the Internet.
It can talk to other nodes fine though, and uses what should also be
its default route in order to get to all of the other nodes on a ping.
The ETX is around 300.
This is a two hop node. I use to have this problem with my one hop
nodes, but that was fixed as of rev:3790. Part of the problem with
this new node is that it should be a one hop node, but it is
preferring to communicate with an antenna I didn't intend it to.
Regardless, it is connected to the mesh, so it should be able to find
out about the Internet access at the Internet-connected node. And so
far it only knows about that for a short period of time right after
boot.
In HSLS, how often does a node get updated routing metrics about a
node two hops away? So far I haven't seen the default route come and
go. I see it only immediately after boot. What sort of logical
messages would cause a node to lose its default route, and why are
they being sent to this node?
-John | 0b77f02c-3151-7cbd-e3db-209f415b60f4 | ['5cf565f7-7d5c-0504-c76c-483aa3f133dd'] | Is there a known problem with the web server in 0.5.5? I haven't been
able to pull up an interface through a web browser using that release.
Reverting to 0.5.4 gives me the interface back.
Has the way IP addresses become assigned changed as well. I get a
different IP on the wired NIC everytime I boot with 0.5.5.
Regards,
<PERSON>
|
fc3326f7-109a-e493-11ac-c138b54c32b8 | ['5cf581ab-a8bd-bfe4-54e0-a4c288e62b54'] | Hi Guys,
Sorry to bother. But we are really anxious to resolve this issue.
We're having a very weird problem with the i915 driver on Opensuse 13.1
and wanted to see if one of you Intel experts can point us in the right
direction. We're experiencing this issue on several hundred of our
machines, and we're dying for some troubleshooting help.
Here is the issue:
Randomly, our screen display will change to a solid color. It looks like
it takes what we think is the upper left pixel and display that color over
the entire screen. We can start a virtual x session into the screen and
still see the entire screen just fine, but our customer standing in front
of the screen can only see one color. If we reboot the PC everything comes
back as it should. However, if we just restart the X server it does not
fix the problem.
This occurs randomly, but we're seeing it 2x per day with 67 machines
running the same setup below. Interestingly, our machines running on a
later model Lenovo M71e do not experience this behavior, and I believe
that model has the Intel integrated HD Graphics 2000.
After some experimenting, it also has those behavior
1. When we keep power to the display but power cycle the PC it DOES fix
the problem.
2. When we keep power to the display and just restart X it DOES NOT fix
the issue.
Do you guys have this similar issue?
Thank you
<PERSON>
| d81e9bff-01dc-5389-466a-461e87ac07c6 | ['5cf581ab-a8bd-bfe4-54e0-a4c288e62b54'] | Hello,
We have an Eaton 3S 550 UPS using NUT tool to send power loss event. By
default the power voltage threshold is set to 85 volts. If the voltage is
less than 85 volts, it will trigger ONBATT event from NUT. Is there a way
in NUT we can change the threshold from 85 volts to a lower voltage like 40
volts?
We were trying to change ups.conf to modify the output of upsc ..., but
looks like it never pickup the new settings.
Thank you so much for your help
Regards
<PERSON> |
426e59b6-d34c-5902-3dd3-2c4226d7002d | ['5d05c270-917d-2799-8ff1-5d80d4f588b3'] | Hello,
RubyCorner has been slowly (but continually) growing as bloggers
writing about Ruby and Rails join it. Rigth now we have 286 blogs, and
379 users registered at RubyCorner.
We must confess, that a big work load took us apart from the
RubyCorner development for couple of months, but we are back. We are
fixing some issues reported by our users, and filling some gaps (adding
some missing feeds).
If you have Rails related pictures at Flickr, or Rails related
videos at YouTube, just tag them with the 'rubyonrails' tag and they
will be displayed at RubyCorner.
Best regards from Venezuela,
<PERSON>
| 7932fee2-0fba-7ccb-73cc-bb00c02bc44c | ['5d05c270-917d-2799-8ff1-5d80d4f588b3'] | Hello,
I want to get the production environment configuration from the
database.yml in the /shared folder of a deployed rails application. I
could use scp to get the file, write it to temporary file, read it and
open it with YAML, but is there a better way to do this?
Best regards,
|
61d41e8c-401d-b0af-0ad9-fa38ad464b65 | ['5d10ad99-8189-8672-303e-49321e112834'] | Hello,
In Djnago 1.7 Manager class extends BaseManager (actually, it extends its
proxy, created using from_queryset). BaseManager is not mentioned in
official documentation, so it is not part of official ("stable") api, right?
"All the public APIs (everything in this documentation) " (from here
https://docs.djangoproject.com/en/dev/misc/api-stability/)
But it does not start with underscore and its module
(django.db.models.manager) is public module aswell. So, it should be
considered as public class according to python codestyle.
I believe this fact confuses user. May one use BaseManager or minor api
change may brake such usage?
The other problem is there is no official Manager reference: there is a
good manual, but no reference where all methods are listed (like we have
for QuerySet, for example).
So, what is the official, stable Manager API? Most developers know answer
to this question intuitively, but there is no formal way to answer it,
right?
I believe that in Ideal Project there should be something like this:
1) Move all "non public" classes to "_non_public" packages (or make classes
_NonPublic)
2) For each public class list all members in doc or directly in python
documentation.
What do you think?
Ilya.
| 4ef10115-b553-1bcf-8f6a-02e638597147 | ['5d10ad99-8189-8672-303e-49321e112834'] | Hello,
I have server CentOS 5.7. It has mysql-community installed.
I need to run perl-based tool (mysqlreport) and it requires perl-DBD-MySQL
(not a surprise)
When I am trying to install perl-DBD-MySQL via yum -- it says it requires
libmysqlclient.*15 *and it conflicts with libmysqlclient.*16 *which was
installed as a part of mysql-community.
I tried to make symbolic link and force DBD to work with 16. But it failed
with error. Looks like libmysqlclient provides it version or has different
API (which is strange for me)
The only one idea I have is to fetch libmysqlclient.15 from some packet and
put it to my LD folder directly.
But what has happened? Where is DBD for new version (16)? What is the best
("official") way to solve it?
Thanks. |
8be394ab-76fc-40ca-8d4e-e84744f89aa5 | ['5d1ea9eb-3d97-dced-7615-61dd3a0c49f2'] | I am considering starting a project that would design and implement an
open-source DRM solution for Ogg Vorbis media. Ideally this DRM
implementation could be extended to other formats of media, but I believe Ogg
is a good starting point because of its openness.
To avoid duplicating someone else's efforts, does anyone know of any current
DRM projects for Ogg Vorbis? If not, I would like to start creating a
high-level architecture document to get this project underway.
I realize that the idea of DRM contradicts with the philosophies of many Ogg
Vorbis developers and users; however I believe it would be in the best
interests of Vorbis to have an open, independent DRM solution.
<PERSON>
| 4c9eaa11-543b-19be-7710-41fcdc6f09e0 | ['5d1ea9eb-3d97-dced-7615-61dd3a0c49f2'] | I am wondering if there are any active projects to produce Java
wrappers for the MusicBrainz libraries? If not, I would be interested
in creating these wrappers. From looking at the MusicBrainz code I
believe creating JNI wrappers would be relatively straightforward, but
am wondering if other developers are aware of any issues I should be
aware of.
I would like to create MusicBrainz wrappers so that application
developers can easily embed MusicBrainz functionality in existing Java
applications. Of course all work I perform will be released under the
LGPL.
Regards,
RS |
d98f51ca-dbb6-db3e-dcf3-b5a2a84e9f42 | ['5d1f1f51-fff9-655e-4cf0-82a28869b7ba'] | Hello,
As of the 0.9x Alsa releases, my SB16 ISA PnP won't work (when loading
modules, they cannot find the card). It works perfectly in the 0.5x
releases. In addition, the kernel modules work fine. If have read the
website and my modules.conf is fine. ISAPNP is working fine on my
system ( allthough I have ./configured the drivers with and without
--with-isapnp=yes and the same inability to find the sb device is
given). Suse's 8.0 rpms as well as their updates rpms have the exact
same problem as when I compile my own. Any suggestions?
Thanks!
| 48f6a148-6eb2-cd64-e184-ace1134c90da | ['5d1f1f51-fff9-655e-4cf0-82a28869b7ba'] | ALSA Project Team,
i386 system with a SoundBlaster 16 PnP in the only isa card slot. All other slots are PCI. Linux kernel 2.4.18 with sound support as kernel module only (SuSE 8.0). ALSA 0.5x works great when using the --with-isapnp=yes configure option. In addition, SuSE's ALSA 0.5x RPMs work great also. alsaconf modifies /etc/modules.conf correctly; and no irq, dma, etc parameters are required.
Problem: ALSA0.9x wont work at all. I've read all documentation and I am aware of the syntax changes (snd-card-sbawe vs. snd-sbawe for instance). I have not altered my kernel that works with ALSA 0.5x (stock SuSE 7.3 and 8.0; I've tried both!). SuSE's included ALSA rpms dont work. Compiling and installing ALSA 0.9rc3 goes well, but modules wont load; device not found errors are given for snd-sbawe and snd-sb16. I even removed the isapnp.h file before configuring. The kernel sb.o module still works, so it is not a bad card; it has worked for 2 years previously.
Lets solve this problem together. My machine is at your disposal. If you can give advice, I will try them and report results back. BTW, I do not have a /etc/isapnp.conf file as my only isa card (this card) was set up satisfactorily by ALSA 0.5x (compiled using --with-isapnp=yes did the trick).
Thank you!
<PERSON>
|
6253aaef-3ef4-6ee9-4b64-194a8aaab1d7 | ['5d203361-f49a-cd7b-7c4d-ac25a35a0b58'] | Hi
I'm interested in the group's feedback on the most accepted way to deliver a
final PT report to a client. Best practices indicate that reports are only
sent to a select group of people in each of the Red/White/blue teams, and
docs are sent via encrypted email and/or the document itself encrypted with
public/private keys exchanged at the start of the engagement. I've even
heard that sending electronic copies of the report is a no-no and only a
hardcopy should be couried. Could someone weight in on caveats and/or
industry standards for report delivery?
Also how would report delivery best practices from an internal pesting team
differ (if at all) from that of a third party consulting outfit.
Many thanks.
| 0873556f-4ff9-87f6-0504-8dc5f30c0be4 | ['5d203361-f49a-cd7b-7c4d-ac25a35a0b58'] | ok we all know that in addition to good network, host and application
security skills, programming in C is a pre-requisite for a decent pen tester
or at least one who wants to write their own security tools or simply audit
the open source code they use. My question is, despite their similarities
should a pen tester be concentrating on C or C++ ? That's it!
Thanks.
|
bf934c22-7dd6-4c69-e16c-5012669014aa | ['5d2323a8-15fa-3910-f2e1-767c7f07dbb5'] | Hi folks,
I have installed the ovirt OS on 3 dl380's.
It has been deployed with gluster and all is fine up to a point.
When I try to install the engine it is trying to get out to the internet
but unfortunately I am behind a proxy with strict rules, is there a way to
point to a local copy of the engine using the GUI install or can I scp the
iso over to the first node for local install, is there a particular
partition where the iso should be stored.
Any help appreciated,
BR/<PERSON> | f2cf3f82-c4b3-15b6-77da-c8856ca016fe | ['5d2323a8-15fa-3910-f2e1-767c7f07dbb5'] | Hi There,
Our 3 node Data center is up and running and I am populating it with the
required vm's.
It's a test lab so I want to provide pre fabricated environments for
different users.
I have already set up a CentOS box for nested virtualisation which works
quite well but I would really like to group multiple machines together in
to one template so that when a user deploys or chooses the template they
get all the vm's together as one. Is there a way to do this in oVirt
without nested virtualisation?
I particularly want to provide a ceph set up with 3 nodes 1 mgmt server and
a couple of clients, which I had planned to do through nested virt but
believe that multiple vm's would be cleaner.
Any suggestions as to how to achieve this?
Any help or hints would be appreciated,
BR/David
|
1229d780-01f4-054b-191b-e63a2f452466 | ['5d31e210-1c45-f37d-5ca7-b91c4ccabb93'] | Hi everyone,
I encounter a strange problem. My computer is 64bit CPU, and found the
visual studio 2008 expression does not contain the 64 bit compiler, so i
use the SDK v6.1 to compile the 64bit itk.
I firstly open this sdk console, and run the cmake 2.8.6, in its
confuguration UI, I choose to output the makefile of itk. Finally, i build
all itk examples sucessfully by using "nmake" command.
However, when i run the deformableRegistration15 example (the fixed image
and moving images are 50 MB around) and try to output the deformation
field, there is a memory allocation problem in the computation of
deformation field. The program terminates.
I found the size of elements it trys to allocate is 28537730, it is not
large. I also remove the code of applying the transformation parameters on
the moving image to obtain the registration result (namely, use the
resampling function).
Then, the progam is still in crash. its very strage.
I think the program is not running in 64 bits.
Can anyone tell me why?
thank you in advance!
best regards,
| c9d1e3b3-c57d-82b3-eb76-63c1f6c42b1a | ['5d31e210-1c45-f37d-5ca7-b91c4ccabb93'] | Hi everyone,
Now, I try to use the ITK3.2 to do registration work. when I use
ImageRegistraion2 example to do simple translation registraion test, and
found that the registration process is very very slowly. For a 2D CT slice
image with size 350*270, it costs me half an hour, can anyone tell me whats
the problem? The time consumption of the ImageRegistration4 is simiIar. I
did not modify any code in the example.
But when i use DeformableRegistraion15 to do 3D registration, the time
consumption is normal.
Any suggestion is appreciate.
thx
|
1ce158fd-8066-b220-1fb8-e424da0444a4 | ['5d36c8b1-efa0-d8db-3f9b-e8290d5f2f42'] | Dear all,
I would like to crop a particular contiguous area of country polygons to a rectangular bounding box, and return new polygons, from which I can work out the centroids. I would like to do this so that I can use the centroids of the cropped country polygons to work out nice label placements for a map. I am having trouble cropping a SpatialPolygonsDataFrame object by another polygon using gIntersection (I suspect this is not for this purpose?).
Can anyone help edit the following to make it work? The gIntersection command fails. I think I need to find an alternative, but I am not sure what to do yet. If anyone can suggest a command I'd be most grateful.
Many thanks in advance,
<PERSON>
| b114566e-8f4f-22fc-64d6-77d6181e1119 | ['5d36c8b1-efa0-d8db-3f9b-e8290d5f2f42'] | Hi <PERSON>,
I think you can try spatstat::nncross. You will need to turn your sp objects into spatstat objects. So coerce your locations into a point pattern of class 'ppp', which I believe is done easily enough with maptools::as.ppp and your polygons into a spatial lines pattern of class 'psp' using maptools::as.psp.
Then it should be as simple as:
dists <-nncross(locations.ppp,sea.psp)
HTH.
<PERSON><IP_ADDRESS>nncross. You will need to turn your sp objects into spatstat objects. So coerce your locations into a point pattern of class 'ppp', which I believe is done easily enough with maptools<IP_ADDRESS>as.ppp and your polygons into a spatial lines pattern of class 'psp' using maptools<IP_ADDRESS>as.psp.
Then it should be as simple as:
dists <-nncross(locations.ppp,sea.psp)
HTH.
Simon
|
dfa7ec13-ffb7-1845-f6ea-087371f61cbd | ['5d37cd6d-b4e0-2cd8-df4e-71596b3439f7'] | Hi folks,
I had some trouble getting the stackless mailing list
repaired, after one message was repeated thousands of
times. This was a message from October, but
interestingly, the thousands of messages appeared
in both the October and the November archives.
Looking forther at one instance of the thousands of
copies, I recognized that the message had no "Date:"
entry, just a Resent-Date, for some unknown reason.
After clearing the whole pipermail archive and recreating
it with the bin/arch command, the message again appeared in
the November archive.
I took one of the fields with valid date enties and added
it as "Date:" field by hand. Then the archiving worked
as expected.
My question: When a message happens to have no date, messages
appear to be archived into the wrong archive file. There
are a couple of other date fields available, like the usual
"from" - line on top of the message, fields like Resent-Date
or dates from the message passing through other servers.
Wouldn't it make sense to do a more exhaustive search for
a proper date, instead of accepting a missing date?
cheers - <PERSON>
| 4d1845a4-7b1b-2ad3-3827-2f1ccdf05e85 | ['5d37cd6d-b4e0-2cd8-df4e-71596b3439f7'] | Hi all,
as promised, I've uploaded the slides from my Stackless
talk to the stackless site. It can be downloaded here:
http://www.stackless.com/#EuroSlides
The page has also been updated to reflect the recent
evalution. Some code concerning thread pickling
will be added, soon.
EuroPython was just -- great!
I'm sure this will be repeated next year
or even earlier.
cheers - <PERSON>
|
8926bd59-45b4-d482-b597-532413707a15 | ['5d455d3a-a927-c7b7-0aba-135bd19bcd5a'] | I realize that this might be problematic/devastating if you had an AI PK and
did this, however in my case there is no problem as there is no related data
in the database yet lol.
My question is, how can I reset AI? For example I deleted several rows and
now my AI starts with the AI number after the last row that was deleted. Ie.
1,2,3,4,9,10,11 when I want to start numbering at 5 not 9!
Even if I insert/update with non-nulls then the sequence is still pushed
ahead.
Is there a way to reset this?
Thanks!
| 51e05ab6-de90-b9d3-3616-5ff27c0e1a8a | ['5d455d3a-a927-c7b7-0aba-135bd19bcd5a'] | Unless I am misinformed, subscription to the DIGEST format of this list
should cause all messages to come in one daily e-mail? Can someone else who
is subscribed this way please at least let me know that this is at least the
case for them?
I continue to get every single post as an individual e-mail. I have tried
canceling and re-subscribing with the DIGEST option selected, to no avail. I
have also not received any response from admin.
Thanks!
|
b4994fcd-b83d-35d8-6013-dd23867d9824 | ['5d5525a5-cb14-9d08-8b41-ddc571001a40'] | Hi, <PERSON>,
Thanks. I've updated the Turmvilla - Nova as 64bit and now I could start
the Noux subsystem.
i'm still facing hang issue in starting up the Linux subsystem.
This time I could capture the Logs from the /shared/log file.
A few questions :
1. For the wifi configuration, I've tried to configure it from within Noux
under the /config/wlan_configuration.
But it seems to points to WPA_SUPPLICANT file. i have configured both but
not sure if it is correct as I checked the /reports for wireless state and
it seems to be not working.
How can I made the configuration permanent as the files seem to be
overwritten everytime NOUX startup.
Is there a manual on how to configure NOUX ?
2 . For the Linux subsystem, from the /shared/log file, it seems to
indicate some issues with the virtualbox configuration. Writing to
non-existent port. Any advise on how to move on.
Regards.
| 9b927e9d-49fc-5545-e169-2ee5b7abde66 | ['5d5525a5-cb14-9d08-8b41-ddc571001a40'] | Hi,
With the previous Turmvilla scenario, I wish to setup two virtualbox
sessions, one running LInux process and the other running Windows,
concurrently. I created a new Win7 subsystem, emulating the LInux
subsystem, with new vm_win7.vdi.
However, I could startup only one of each sessions from NOUX, either
Windows or Linux, but not both sessions at the same time. From CLI monitor,
I could see that RAM is allocated separately for both sessions. I suspect
some concurrency issue but could not find the issue.
I've attached the log from captured from NOUX.
Another issue I encountered is that the virtual box mouse pointer in the
WIndows session and the Nitpicker mouse pointer can't seem to merge
together unlike the Linux session where both pointers are merged.
Any advise.
Thanks in advance.
|
5caf8f81-71fa-03e6-a2ea-7ddf8e744da3 | ['5d6f7a9e-1173-cbbf-40a4-5e8816dff860'] | Hello All
A thought on printers....
We generally have two types of print jobs
1. light jobs :- are usually needed "right now" typically 1-2 pages economy is not the criteria
2. Heavy jobs :- Its OK if we get the jobs next day, done from a Print Shop in Andheri or so.. SHOULD BE ECONOMICAL and preferably managable via e-mails..
This seems to be a good business proposal and very good for us who need heavy printing jobs
If anybody knows such printer peopl pl let us know
ONE MORE India could become a PRINT AND POST HUB FOR THE WORLD !!!!!!!!!
Any Takers ;)
MAHESH DESHPANDE
Trainer for
Linux Installations,
| 77633daf-5281-b76c-d5a7-ebd600f58ed0 | ['5d6f7a9e-1173-cbbf-40a4-5e8816dff860'] | CISSP
A cissp study book is available at computer Book Shop DN Road Fort Mumbai 400 001
For CISA visit their site and enroll yourself
You will start getting their regilar mails and snail-mails
Also at Matunga the CISA conducts regular courses for their June exams
Bookings usually starts at Nov-Dec and closes by Feb
<PERSON>
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.