query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
e9b1f92a-f300-8a83-4e3c-f78b591e2b16
['772f70ef-b17e-6dc6-b37e-dd271f62b451']
Dear All, I have installed dspace 1.5 with PostgreSQL 8.3, JDK 1.6, Ant 1.7 and Tomcat 6.0 successfully. My DSpace was running successfully and I have added some communities and collections but all of sudden there is an error Internal server error and nothing . Also if I click on any other link then there is same problem, even I run the ant command again and there is same error. Can anyone help me? Thanks <PERSON>
bcf097f3-55de-2733-605c-05d57da8e76b
['772f70ef-b17e-6dc6-b37e-dd271f62b451']
Dear All, I have installled DSpace 1.4, 1.5 successfully. Now I want to change its desgin and some fields of second step of describing any item. I want to add some fields. In dspace 1.5 I am confused that jspui files are multiple locations and I do not know to which I should edit. On the other hand in dspace 1.4, this may be easy, I think so. But I am confused which jsp file corresponds to second step of SUBMIT that is to decribe an item. Also I want to change desgin of all the pages. Can you help me please. Thanks <PERSON>
8c7ec397-c635-1e7a-e57a-85b1794648bf
['773c443b-f108-30c6-cd3c-adde6a666d45']
My question is, has anyone else done much performance testing of their code to compare v12 to v13 performance? We are running on a Windows 7 system using 4D Standalone. Users access our site via their browser. Our experience with v13 has been a bit disappointing and we are trying to discern whether it is something unique to us. If it is we are hoping that someone might have some magical suggestion of some global setting change in v13 which we could tweak and regain our v12 performance rates. If not that, perhaps someone knows of some changes in v13 which might affect performance and perhaps specific commands which should be avoided in v13. Below are some additional details. We converted over to v13 to do testing with our application last week. All went well and was straight forward until we moved into what we thought would be a quick performance comparison. What we quickly noticed was that v 13 appeared to be significantly slower than v 12 with our application. In testing we stumbled on a bug with the millisecond command under Windows 7. To ensure that wasn't the cause for our performance problem we amplified the execution of our code for a common request in our application by running the entire code cycle through a 500 loop FOR statement. This allowed us to confirm, using a stop watch, whether there was a performance difference rather than relying on the millisecond command. We saw two things. First was that an uncompiled copy of our application executed the loop in v 12 in 85 seconds. In v 13 the same execution took 115 seconds. That meant that it took 35% longer to execute in v13. We could have accepted that result but even more concerning were the results when running a compiled copy of our code. In v12 the same loop took 14.6 seconds while in v13 it took 34.9 seconds. That means it took over twice as long to execute in v13 than it did in v12. We found similar results in other examples of our code. A simple custom trace of the performance which looked at time stamps for each of several hundred steps along the way saw a steady degradation across the entire code cycle rather than, as we hoped, perhaps one offending method. Again, my question is, has anyone else done any performance testing and / or does anyone have any suggestions of settings which may have changed in v13 which we could adjust to improve performance or are you aware of any commands or other changes in v13 which might negatively impact performance. Thanks for any suggestions and feedback. <PERSON>
390965bf-51ab-0824-b0a0-00bc61f62c60
['773c443b-f108-30c6-cd3c-adde6a666d45']
I've encountered a mild annoyance since using v11 - from 11.0 to 11.5 hf1 - and was curious whether anyone else has encountered the same thing. I've used 4D Developer Professional on Windows Vista and 7. Over the past two years, often, but not always, when I quite 4D with methods still open in the design mode, upon opening 4D the next time all my methods are closed. Since it doesn't always happen I wonder if there is some subtle variation about how I close 4D that makes the difference. Has anyone else noticed this and if so is there some little trick to keep the methods from closing upon quitting 4D? As I said, it is merely a mild annoyance but I thought I'd ask. Thanks, Mike <PERSON>
0115eb31-b4f9-2da6-84a1-828009897a9b
['77582eb7-1da3-3704-bc9b-fc8d32d1345c']
Good day I'm new to this list (though I've done some searching, I'm sure I could have missed something), so please bear with me. I filed a bug report recently against Ubuntu with respect to the problem I'm experiencing (as per the subject: jacking in my headphones doesn't mute the speaker output) and was informed that the decision had been made upstream as a design intention. I was linked off to: http://mailman.alsa-project.org/pipermail/alsa-devel/2010-August/<PHONE_NUMBER>.html with respect to the thread on this topic. I'd like to raise it again though, perhaps as a configurable feature, for the following reasons: 1) headphones muting the speakers used to work some time ago (before Ubuntu 9.04, iirc), so we have an established user expectation even under Linux distros. 2) headphones mute the built-in speakers of my laptop (windows and ubuntu) -- why the inconsistent behaviour for a desktop? Just because the speakers in the laptop are bolted into the frame of the machine? 3) Windows "gets it right": plugging in headphones mutes external speakers -- it's convenient (Though I think you can disable that feature if you want). I'm told by a person at work that his Mac does the same. Again, the user expectation is unmet under a newish Linux distro. 4) The user expectation theory is held up by the number of bug reports against this design choice on Ubuntu Launchpad alone (I haven't looked at other bug lists). The average user is not expecting the current behaviour, obviously 5) For people without hardware volume controls on their speakers, the problem is exacerbated: instead of being able to plug in headphones to disturb others around them less (ie: contain their music/noises), they have to unplug their speakers too -- which may not be trivial since speakers are normally plugged in at the back of the machine, not at the front where the headphone jack would be found. On a machine in an enclosure, this is especially troublesome. So I'd really like to know if it would be possible to allow the behaviour most users are expecting. Sometimes it's useful to be able to output to both speakers and headphones, but I would say that the average user doesn't want both at the same time. -d
9d9c86df-2daa-e122-6207-70505cbbc980
['77582eb7-1da3-3704-bc9b-fc8d32d1345c']
I'd just like to add probably-not-even-two-cents-worth: Whilst I personally can't see any immediately viable (read: in the next 10 years, if ever) work to use HURD (*shudder*) or MINIX, the OP might get some satisfaction from Nexenta (http://www.nexenta.org). From what I've read (project maintainers and users), it's in a really good shape. If it's so important to the OP to use something other than Linux (personally, I don't really see the point, but that's my overly-biased opinion), then get something with a kernel which has actually been shown to work on a fairly wide array of devices and which has constant, active development. Oh yeah, and get ZFS... Man, I wish the whole licensing debacle for ZFS was over and I could mkfs.zfs... But I digress. -d
7ccde542-4776-6ec0-a800-cd3ee9ce3d5f
['7764c0be-ecf8-e1aa-a812-02532629210e']
As most of you guys may already know, I maintain the FreeBSD - the unknown Giant (www.freebsdnews.net) blog. This blog tracks news and development of FreeBSD and FreeBSD based OSses, e.g. PC-BSD, desktopBSD, m0n0wall, FreeNAS, FreeSBIE pfSense, etc. is there anything you guys think I should add/change to make the website even better for newbies, for people moving on from Linux and those who just want to read what's happening in "FreeBSD land". Many thanks in advance <PERSON>
7c55bb1e-227a-32e1-84b9-411c34dde52a
['7764c0be-ecf8-e1aa-a812-02532629210e']
* *Good day, I wish you a prosperous and Happy 2011. Hope you're well and hopefully we'll see some new updates for Debian GNU/kFreeBSD this year. I'm planning to do a little series of FreeBSD based operating systems on freebsdnews.net: a bit of background, history and the plans for 2011 etc. Would you mind and share your plans and roadmap for 2011 (if any)? Hopefully we can generate some excitement and a looking forward to releases this year. Many thanks in advance <PERSON>
ba118f49-53f1-19f6-42c8-8cdea61885ca
['7777b4ae-1bcc-1da2-4c92-64d27124dda9']
Folks, We have a faculty that has been importing imscc files generated by Blackboard <https://www.coursesites.com/>. It worked fine under SAKAI 2.9.x. In SAKAI 10.7 the resulting questions pools in Samigo are corrupt. When you look in the pool you don't see a question type. When you attempt to enter a question in the pool it exits you back to the listing of pools page. Anyone encounter this (and hopefully, solve it)? Thanks, <PERSON>
df38750a-1dce-1c26-fd13-1d97d9516d38
['7777b4ae-1bcc-1da2-4c92-64d27124dda9']
Folks, If a student submits an assignment and within the CKEditor embeds an image by uploading … where is that supposed to save? By default it's pointing to resources and of course students don't by default have write access to that folder … resulting in an error 203. Is this something due to configuration? We're also having a similar issue with Mneme where faculty are doing the same thing embedding an image via uploading resulting in a error 203. Now if we take that tool out and add it back in via the Mneme admin tool it works correctly. We have mneme.manage and mneme.grade checked off in !site.helper for Instructor, but it's not there for the realm of the site. After deleting the tool and adding it in via Mneme admin mneme.manage and mneme.grade is there for the realm of the site. Shouldn't !site.helper for Instructor work in this case? Thanks, <PERSON>
6b86d73f-e0e2-a982-8d73-f72adaa129b7
['77830c3c-aae7-2280-bf2d-86f3863da0bb']
What version of unix is the backup client? I can tell you that Solaris 2.7 (my admin host) *will* "ufsdump" an nfs mounted FS. It used to be back in the SunOS days that "dump" would not. I remember trying. Can you use a Solaris 2.7+ client to try it on? <PERSON>
445df28b-2156-6430-c0ff-0ca052cf9c2f
['77830c3c-aae7-2280-bf2d-86f3863da0bb']
I may be asking the obvious here but who dedicates their Oracle->Filer traffic to just *one* Filer? IE: That Filer *ONLY* handles Oracle->NFS->Filer traffic. Our CIFS traffic has grown significantly since we bought our FAS940 and I'm now dubious about mixing CIFS/Desktop services with Enterprise critical Oracle services. Currently we have a few small DB's (read-only) on the Filer but are moving to step things up over the next year with Regards to Oracle. (we are looking at RAC). So: Would ALAWYS or SOMETIMES or NEVER mix Oracle and Desktop Filer services on one filer? :-) <PERSON>
c166bce7-e533-7308-0dc6-592a569355cc
['77868b9f-1a3d-6215-fa96-0890632e2053']
Hi, The usual argument that not all tricategories can be strictified appeals to the existence of nontrivially braided monoidal categories. But this leaves open the possibility that a tricategory can be strictified completely aside from interchangors for endomorphisms of unit 1-morphisms, or something similar. Are results along these lines known? There are of course the results of <PERSON> and <PERSON> that it is possible to put all the weak structure in the unitors. One could see this as mildly in support of the above proposal, where also the weak structure is closely related to unit morphisms. In the absence of hard facts, reasoned speculation would also be good to hear. Best wishes, <PERSON>
b4cbf3d6-b391-8909-8188-8d0e6d288762
['77868b9f-1a3d-6215-fa96-0890632e2053']
Hi, In a category with a zero object and biproducts we obtain a unique enrichment in commutative monoids, which we will write as +. If the category is also monoidal with left and right duals for objects, then the tensor product distributes over +, in the sense that f (x) (g+h) = (f (x) g) + (f (x) h) for all morphisms f,g,h with g and h in the same hom-set. I have a proof of this but it is a bit clunky, and rather long. Can anyone give a beautiful one? Best wishes, <PERSON>.
db15421c-7713-4696-9427-eddace5c71a7
['7792e402-e764-c828-1389-94e8a33a7294']
I am using igraph "0.7.1" from R interface. I noticed that when I read a graph in GraphML, an additional id attribute is created for nodes in the igraph graph, even if there exists no GraphML id attribute for nodes. This might create problems if there exists already an id attribute for nodes in the igraph graph. An example follows. Best, <PERSON>
0ab7251f-dbc8-c879-cc33-701d67aaed2f
['7792e402-e764-c828-1389-94e8a33a7294']
Hi! I have prepared a full training in igraph (for students of a data science course I am teaching). It comprises an analysis of a social network of dolphins, featuring static and interactive visualization, centrality, similarity, communities and network structure. You can find it at: http://users.dimi.uniud.it/~massimo.franceschet/teaching/datascience/network/R/dolphin/dolphin-revealed.html Well, text is in Italian, but commands are self-explaining (I can translate it if useful). I also delivered some "interactive igraphics" Shiny apps here (requires RStudio): http://users.dimi.uniud.it/~massimo.franceschet/R/index.html Thanks for your careful work. Love igraph! Massimo
1c1cd8b8-d00e-c5dd-49ab-072e9c32bbe9
['77a2399d-9f57-d8f2-cbf3-1b51e3954b9c']
Hi <PERSON>, My T|E is over 1 year old and going strong. I mainly use it for Docs to Go - the reason I bought it - , MP3 player and the WW program that was banned (shame on them). Prices differ in the US, but in the UK, this is a bargain price machine, even a year after it came out. I have only had 1 problem with it - Find crashed all the time - but an update of DtG fixed that. I managed to get the "bulletproof" OSs on the Psion 3a and 5 machines to crash more often than my Palm T|E. Personally, I think they're great. Note I am not a rabid Palm fan - just bought a J720 handheld PC for text input - but consider the T|E an excellent tool. thenikjones
b1658ce8-76ba-4dc9-a44f-51076e414c70
['77a2399d-9f57-d8f2-cbf3-1b51e3954b9c']
I have used both and currently use a T|E as I think it is best value for money. DTG is better than Pocket Word/Excel, but Textmaker/Planmaker (for PPC) are better still, but you have to pay. I now have a C500 Smartphone and like that Tasks, etc can have multiple categories that sync with Outlook. At the ned of the day, both platforms are pretty similar, certainly for GTD stuff. If you're happy with Palm, you aren't missing much by sticking with it, but a change to PPC shouldn't be dismissed out of hand. I you want wireless, you are probably better going for a PPC. <PERSON>
e2bc4964-249f-5098-d3d1-73717b51c4f9
['77b61a82-9486-4b61-5409-9949a0cea9d7']
Hi all So sorry to be a total newbie, but i'm just finding my feet with R. I heard on the grapevine i could recreate a scanned microarray image, or at least get a good graphic of it from a just a data file. I have .txt files for illumina beadarrays but no images cos the service we used didn't send them. The 'beadarray' package seems to require TIFFs to get the graphic output. Does anone know of a way/a package i can use to rebuild the image? Many thanks!
13f9b13d-23b0-1a37-81f1-428576617e31
['77b61a82-9486-4b61-5409-9949a0cea9d7']
Hi Does anyone know how to convert the binary data in an affymetrix .CEL file via data.frame or any other way? The way i normally convert to a data frame doesn't seem to work here. I was to see the raw data from my arrays, but not the grouped probeset data. I trawled the net, tried numerous commands and i dont think the Affymetirix Expression Console Software does it for the raw data. Alternatively, how can i log2 the .CEL file data. I want to present it as a heatmap. Cheers!
bd975691-9025-de43-bb24-f616d7c7df6c
['77ca76a7-7bf8-d8c7-a19a-60f19b3ebbe1']
Hi, We decided to use common IRC channel #lf-releng for Releng Support issues so we can get help from our friends who work with other LF Networking Projects and increase the collaboration. Please get in touch with us on that channel for faster response. Few of us are already there. Automatic redirection from #opnfv-octopus to #lf-releng will soon be enabled. /Fatih
5a690ac0-37d3-e1e1-6efa-a6155266bd78
['77ca76a7-7bf8-d8c7-a19a-60f19b3ebbe1']
Hi, I think there is a misunderstanding, especially with this: "The first hurdle is that all test containers run inside a VM on the jump host today." I would like to clarify this and add some more details. Test containers do not run inside a VM on the jump host today. Test containers directly run on the jumphost. What I was referring to you when we had chat is the initiative Infra WG has been driving to move installers into VMs on jumphosts so the installation process can be isolated from the jumphost and driven from a VM in order to ensure installers can use any PODs we have. The reason for this is that we currently use dedicated PODs per installer and these installers directly run on jumphosts, depending on OS and some other stuff on jumphosts (I have suspicions that they even expect certain usernames for the login user which is really strange). This really makes things difficult for us in Infra, for the developers, and for the end users since noone knows what we have on those jumphosts and the state of CI POD jumphosts is a big question. (some jumphosts haven't been cleaned/reinstalled for 2 about years.) This is not really a good practice since we basically do not have any change control on these machines and whatever is done there has possibility to cause headaches. (for example, strange failures which you can't reproduce elsewhere.) And finally, our releases come out from these PODs so it is another big question; reproducibility of entire release... We haven't been able to get the above idea implemented by installers and this results in other issues such as resource shortage and so on which is nothing compared to issues listed above. As part of XCI initiative, we aim to apply all the ideas brought up by Infra WG over time and demostrate the benefits of what we have been proposing. (dogfooding) This is not limited to having a VM on jumphost where the installation process is driven. All the PDF, SDF, VM, dynamic CI, CI evolution etc. activities actually pursue to ensure we can have full traceability, reproducibility (CM101), and we use our resources wisely, shortening the time it takes for us to do things rather than things waiting in the queue for days. (Long list shows that none of the ideas we brought up was actually implemented which should be another concern for the community and a topic to brought up to TSC.) One of the first things XCI Team started working on is putting the components used by XCI into a VM on jumphost so we always get the same behavior by the use of a clean machine and we can use any and every POD OPNFV has. We also support 3 distros; Ubuntu, Centos, OpenSUSE. So we are basically independent from whatever POD we might be run on. PDF is the other and important part of what we are doing so we can have total independence from PODs. SDF will put the missing piece to get the BOM for any and every deployment and testing that is done in XCI. Putting test containers into VM and running testing from there have been part of our overall strategy. However, after talking with <PERSON> on IRC, it was clear that putting test containers to VM is not applicable to all the testing projects. As summarized above, the reason for us to use VMs is to isolate things from the physical jumphost and ensure everything starts clean for reproducibility and so on. Using containers for test projects serve the same purpose; isolation and reproducibility so we do not see any issue here and will do our best to support you, ensuring you get what you need as long as it fits into the overall strategy set by Infra WG. If we find something that doesn't fit, we can bring that topic to Infra WG and can do necessary adjustments if possible. /Fatih
3494b086-1a4c-d90f-d06c-14ae6ac7c44b
['77cea5e2-c86a-de98-9ce4-f87beb2aac8e']
Hei there I have some questions regarding checkpoint/restart: 1. Until recently I thought that ompi-restart and ompi-restart are used to checkpoint a process inside an MPI application. Now I reread this<http://www.osl.iu.edu/research/ft/ompi-cr/tools.php>and I realized that actually what it does is to checkpoint the mpirun process. Does this mean that if I run my application with multiple processes and on multiple nodes in my network the checkpoint file will contain the states of all the processes of my MPI application? 2. Can I restart the application on a different node? Thanks a lot, <PERSON>
ac2a1594-a068-86cb-a824-eae1fa67138e
['77cea5e2-c86a-de98-9ce4-f87beb2aac8e']
Hi, i am quite new to Xen, I want to do some experiments and I would like to know the followings regarding some of the Xen Hypervisor features: *Context A: *2 hosts in the same LAN, 2 VMs on top of each host and a NFS storage with the disk images between the hosts. Host 1 contains VM1 and VM2 and Host 2 contains VM3 and VM4. All VMs have private IPs. 1. Save/restore: I run an application that was communicating with all 4 VMs in the same time. I suspended the application and I saved the state of VM1 to the shared storage and than restart VM1 from that file on top of Host2. Than I resumed the application and it continued it's execution successfully. So this means that the connections were preserved. 2. Live migration: same thing. Application communicating with all 4 VMs and I migrated a VM to another host and all the connections were preserved. * Context B: *2 hosts in the same LAN, 2 VMs on top of each host and *no* NFS storage. All VMs have private IPs. 3. Save/restore: Can I save the state of a VM on one host and then restore it to a new host without having a shared device between the 2 hosts (maybe if I copy the disk image to the new host)? If an application is running on it are the connections going to be preserved? *Context C*: 2 hosts in different subnet, no NFS storage, VMs having public IPs from different subnet 4. Is there any way that I can save the state of a VM on Host1 and than restart it on Host2? And more than that, is it possible to preserve the network connections? I would really appreciate an answer for question 3 and 4, Thanks a lot, <PERSON>
890b1b80-ef3a-af7b-63d1-40ced2a3bd24
['77d38839-ea80-a111-65b2-b5ee9daa115f']
Hello, I have recently upgraded from Evolution 1.08 to 1.44 on Redhat Linux V9.0A,2.4.21-1.1931.2.399.entsmp with the KDE desktop 3.1.3-3 Red Hat. I often send out the same messages again and again ( i.e. like this message ;-)), so need to be able to save these messages somewhere in Evolution. If I put them in my Drafts folder, when I send out the message, the program now deletes the copy of the message in the drafts folder, and transfers the message to the sent folder. What I need is for the program NOT to cancel the messages I have in my drafts folder when I send them out, so as my "message templates" stay in my Drafts folder (as occured in previous versions of Ximian I used i.e 0.9, 1.0, 1.08). I am using a "work-around" for this problem by storing these messages in another folder called "templates". However, as these messages no longer appear as "editable", it is difficult to cut and paste from them ,plus when I want to work with them, I have to manually paste the message contents into the other message that I am sending out. Is there any way I can change the relevant properties of the folders to achieve my objective ?? Thanks for any input ( even if it it to say no this is not possible in 1.44), Mr e-mail
f3cf8c9c-9a7c-3cdc-a9a4-82d95f78d56e
['77d38839-ea80-a111-65b2-b5ee9daa115f']
Dear <PERSON>, thanks for your kind input. where do I switch "hide deleted" on and off ? Also I thought of another way to skin the cat. You could save your template messages as signatures and then add the signature instead of editing the draft. or the option to turn delete off would be really preferable. BTW going off on a complete tangent here, why does the address book sort e-mail fields in a case sensitive way if e-mail addresses are not case sensitive ? that's really dumb.Every time i save an e-mail address that starts in an upper case letter, i have to go in and edit the first letter of the e-mail so as it sorts it in lower case.aaarg Mr. e-mail
4f417d8a-657b-1825-9c39-e6ab0f347642
['77d3f2c3-a67c-9019-c9e3-971e645da1bc']
Hi, I am newbie in osg based on example's I can do off-screen rendering and auto-screen-capture then make a movie from my scene graph with single view and camera when I use composeitview ,these are my views and cameras: topview->topcamera leftview->leftcamera rightview->rightcamera captureview->capturecamera I can capturing and then make a movie too but no one should can see my captureview ,its just for capturing! how can I off-screen rendering on my capture camera in captureview ? Thank you! Cheers, <PERSON>
815cef06-1ebd-f08c-2a23-fd0f68d0fdeb
['77d3f2c3-a67c-9019-c9e3-971e645da1bc']
Hi, on/off method does not make any change! I have a question... I use osg autocapture example,so is that example unpractical? how can i write a better example for autocapturing? and about that red flag,... I found "The start/stop of threading and releaseContext() " in this forum as a solution for multi-thread model in off-rendering. Thank you! Cheers, <PERSON>
62ed04ce-a67d-ac55-ad74-3f1585004fff
['77df8bff-4596-1687-25d3-2ebfc1eec8a6']
Would it be unacceptable to just say that, when using #rrggbbaa notation, the aa part maps 0x00-0xFF directly onto 0.0 - 1.0 (that is, take the aa value, convert it to decimal, and divide by 255), and if an author needs to specify a color with /exactly/ 0.5 opacity (or another float that is not represented exactly by one of the 256 values 0x00-0xFF) for some reason, he or she just has to use rgba() (or hsla()) notation? I think that most authors using (or rather, who would use) #rrggbbaa notation, and I am among this company, would be just fine with 0x80 representing 0.50196... <PERSON>
abb5cc5e-916b-29e7-e6c8-b180743cdaa5
['77df8bff-4596-1687-25d3-2ebfc1eec8a6']
Speaking as an author who will almost certainly be using flexes... While the asterisk notation does make a fair amount of sense, especially to authors who are familiar with it from HTML, the fact that * is already an operator in calc() is a really strong argument against it, I think. As for the min-width/max-width/starting-width flex notations[1], I'd prefer something that's straightforward. We already have min-width and max-width, and they seem to be pretty well-suited to the needs at hand. And while there's something that nags at me about using calc(10px + 2fl) to set a starting ('preferred'?) width, I haven't seen a notation that makes more sense yet and can't think of one myself. I just think of it as a kind of algebraic expression - the width is 10 pixels plus "however much". "However much" might be negative, which is totally fine in my mind. Anyone who has the hang of negative margins (...no offense, <PERSON> ^_^;;) should probably be able to handle the concept of the flex part of a width being negative. - Eli Morris-Heft
588a2516-5a20-81b1-1a32-805a7dc0a98e
['77df9ce3-3f22-1236-0898-07e2463a3fdf']
I've been using darkice with a soundcard source but am moving to a setup where I'll be recording video while broadcasting live audio to shoutcast. Ideally I'd take the audio from the video to save additional connections. My plan is to use dvswitch for the video, and use dvsink-command to take a stream out. The example talks of streaming to icecast using ffmpeg2theora|oggforward. With darkice I've two questions: 1) how do I set up darkice to take the piped input (no need for named pipes of course). 2) What audio format should I get out of ffmpeg? Any help much appreciated. <PERSON>
1ba1bdb0-0f6a-0eee-ec47-29072e499212
['77df9ce3-3f22-1236-0898-07e2463a3fdf']
I have a several SD cards that boot fine on my xM Rev C board but not on my xM Rev A2 board. These include cards I've built with narcissus, and also the xM test beta 4-25 card supplied with the xM Rev C. These all 'hang' (although the sd card light keeps flashing) on the Rev A2 immediately after Uncompressing Linux.. .. done, booting the kernel. However I believe the board is OK as I have other cards (incidentally Ubuntu) running fine. I've drawn a blank locating information on this from search's: the closes I could see was http://processors.wiki.ti.com/index.php/Kernel_-_Common_Problems_Booting_Linux#Problem_.232_-_No_more_output_is_seen_on_the_console_after_.22booting_the_kernel.22Where its essentially Problem #2, but these seam to focus on issues with bootargs being passed which do not tie with the Kernel being loaded, hence are not applicable here. Any advice on where to look much appreciated. Regards <PERSON>
4ebe5912-9a07-f013-3578-cc132ab6eee8
['77ed5be8-c35b-1eb7-2961-79af047ef0eb']
i find i have got to page 199 and i have no files in usr/src. I came by this as i was trying to compile the keyboard into the linux kernel. I would have missed it otherwise. Have i missed a step? on this my 3rd attempt i have ticked every page as i have completed it. I can find no pages unticked nor can i find any switch to point to the usr/src directory. My original thought was that the kernel needed to be compiled before this directory got populated which is why i tried to compile the keyboard after page 199 and not on page 187. i don't want to reboot until i understand if this is right or not thanks <PERSON>
f556e8e1-6c29-aee7-3fae-1f246985c353
['77ed5be8-c35b-1eb7-2961-79af047ef0eb']
new lfs user, looking to get an nds login, i have downloaded the following package ncpfs-2.3.2. Has anyone had any experience installing this software? I am using LFS 5.0 as a base sytem, this will in effect be my firt install since the completion of the book. i shall have to recompile the kernel for IPX support. but other than that the package looks from the readme compatible with gcc andlinux 2.2.*. the NLS readme suggests using the included gettext. however, i am trying to find out if the lfs gettext is more current? <PERSON>
2390f7a8-5fd9-2332-bad1-85fed28cd20d
['77f24479-cad5-f287-cf19-fa556da2c939']
Is it possible to have svnserve run in daemon mode for read-only mode, and use svnserve/ssh for actions that require authentication? Is there support for this in the client? Please correct me if I'm wrong, but if you use SSH, you must authenticate, even for repository ls or checkout/update. I think it would be great to not have to authenticate when you don't need to, but still have a secure login mechanism for svnserve when you do need it. Thanks! <PERSON>
d2eec908-ffbf-f57e-7be5-cc4ad342cb4b
['77f24479-cad5-f287-cf19-fa556da2c939']
Hello, all! I believe I've come across a problem with Schema validation. Here's my scenario: I have a custom type (in this example, "pricetype". My Element is nillable, but the validator always bombs on it, because the "" item isn't in my xs:pattern list. Nillable should hit before we check patterns, no? Below my example is a small patch. Thank you. <PERSON> @AT@ hackswell.DOT.com
d082c6db-4ec5-924d-fb00-cf4f58506932
['77f6a000-6fb5-5266-8085-8369d1a5d85e']
Hello, If opensips receives a request with a R-URI the same of the proxy it will make any changes on the R-URI as i want, right? but what about if the domain of the R-URI is not the one of the opensips proxy? (its behavior should be as stated on the section 16.12 of the rfc..) could i make changes on the R-URI before sending it to the destination? like for example changing the userid or just the domain, or maybe the hole r-uri (i think this should not be possible because of the rfc 3261 16.12 section) BEst Regards
ccec0c08-9e66-6037-073a-4c85deb06ba1
['77f6a000-6fb5-5266-8085-8369d1a5d85e']
Hello, Is it necessary to have a rtpproxy installed into opensips server to be able to change/modify SDP headers in messages through the proxy? i'm telling this because what i need is to send the media through a sip recording server (and external server) and not peer to peer, but i do not need the rtpproxy (at least, not yet..), how can i do that? Thanks in advance!
43c35249-1cc4-474a-232c-e95983fd848a
['77fcf633-c7af-cb31-90a7-dbf142b998d7']
Hi, i need some clarifying about how the defines ALSA_PCM_OLD_HW_PARAMS_API etc. behaves. for alsa-1.x its clear that if the app uses the new api i have to define ALSA_PCM_NEW_HW_PARAMS_API, but what is for the case if i have an app with the new api and the user uses only alsa9? is alsa1.x then backwards compatible? regards <PERSON>
3dc8e042-34c7-3fc6-f54e-b757329afa0b
['77fcf633-c7af-cb31-90a7-dbf142b998d7']
[Automatic answer: RTFM (read DOCS, FAQ), also read DOCS/bugreports.html] hi there, i didn't used mencoder so much and i currently have no possibility to use it (man pages, no mplayer/mencoder installation) so it may be a stupid question. is it possible to use mencoder for encoding parralel into mpeg1, quicktime, realmedia and windowsmedia? or what codecs (codec familys) are supported with mencoder? are they the same as with mplayer? the reason i ask for is that i want to make an all in one solution for grabbing some clips from an analog in and encode it to the mentioned formats. best regards / mit freundlichen gruessen <PERSON>
f2ebeaa6-0c62-42f2-790d-019d4da73155
['78027e28-792c-6928-bec2-8b7ebd616896']
Hi I have a camera which uses the PAL_Nc norm this way I can see it with xawtv using the bttv driver. I recompiled my kernel to use the Video4Linux2 API. With the bttv2 driver I can see the picture in xawtv but I see 3 of them mangled.So I think it is because I can select the norm well. The norms that I see is: PAL NTSC SECAM PAL-60 PAL-M PAL-N NTSC-N And this is what I see with my small program too. I have RedHat 7.2 with Kernel 2.4.7-10. I would like to have a frequency value for the color subcarrier so I could add this to the videodev.h file.Should this work? Anybody can help me or with a new videodev.h file? <PERSON>
4969a7e2-eaa1-bf9b-7503-e90ef5902bfd
['78027e28-792c-6928-bec2-8b7ebd616896']
Hi all I have the following problem: Using bttv with xawtv I can see fine picture setting the norm PAL_NC, and it synchronizes OK. But using bttv2 with xawtv I can not sincronize the picture because I can not select the PAL_NC norm(It is not available i think)I made also a program which uses bttv2 and could not set this norm.(this is my PROBLEM).I don't want to rewrite my whole program to v4l because only for this if it is an easier way. I installed the videodevx interface by <PERSON> and <PERSON>. The videodev.h file is dated Mar 31 2002. Please help me how can I set this norm using bttv2 and this interface. I've used Dome and Micro-Compact B/W cameras and the same problem applies for both. Best Regards <PERSON>
7594175d-7f38-6e25-4464-bf7cfedf172d
['7804b03e-794d-89f4-66e9-56a5b6138f73']
Hi, EVMS list people! I tried a couple of times now to communicate with the list, and with the maintainers, about the SMI-S. I am not sure about why I don´t get any respons. For the reason that some people don´t know what SMI-S is, I will explain it briefly: SMI-S stands for Storage Management Initiative Specifications, and is it an open, vendor independent standard. SMI-S is of outmost relevance for all vendors, developers and end users of Storage Management software. All future software, of ANY significance, will have to comply with SMI-S, because this will be the sole standard in the future. No software that doesn´t comply with SMI-S will have any future, because it will not be compliant with all the other developers and vendors software, it will be an isolated island that not many people would like to use. I write this, because I care about EVMS, I would like to see a future for such a nice application. SMI-S is a major break through in the storage world for open standards. Everyone that has been around in this world for a while, knows what I mean. And I would hate to see the open source world be left behind. I send a link for those who want to know more about SMI-S: http://www.snia.org/smi/about/ What I originally wanted to know, was if EVMS developement team had a strategy for SMI-S? No one has answered me about that, and that makes me wonder? Why? Because there is no interest? I can´t believe that, this is of too much importance for that! Is there for any other reason that I can´t imagine? Is this sensitive in any way for IBM? I don´t know, I can not guess! Best regards from/Med vänliga hälsningar från <PERSON>
96452eaa-e0c5-ae02-571c-8cd97dbf029d
['7804b03e-794d-89f4-66e9-56a5b6138f73']
Hi, <PERSON>! What do you mean with this? "ps, since u have emc box, just curious, have u ever try snapshot on it? i found linux lVM is too slow to use with IET." Do you mean that LVM is too slow to use for snapshotting, or what? If you mean that, I would say that that depends on the underlaying subsystem, more than of LVM. If you snapshot sw raid5 set´sand put it on another sw raid5 set, or if you snapshot LUN´s from an external fc raid controller, and put that on another external fc raid controller, you will have all the differencies in the world. Best regards from/Med vänliga hälsningar från <PERSON>
9a44635e-884e-4148-fa68-48248503dcf5
['780f06c0-efd5-4eb7-aad1-15af7427de53']
Howdy NM, I have a bcm4318. Ya .. Thanks for the work it took for this chip. My problem now is how do you delete you next door neighbour's wifi. I seem to be picking it up across the street.....thank much. Were is a good link to help with the camannds for log chexs and script on different device's. Thanks very much in advance Bubba 8Fg/5CfuogqXt5jL@example.com
154d6a66-b248-91d7-8b5e-adda5d772178
['780f06c0-efd5-4eb7-aad1-15af7427de53']
Hi <PERSON>, Hope that I can be a + but Evolution is new for me. I am a MS moron.. Is there any way that I should set my account for List. I have also other ubuntu list. could you give and tip what works best for a newbie. What are some step's or tip to set up more account with Evolution. Thank you in Advance <PERSON>
2757f672-13fb-88f9-da00-1d74711604d3
['78106f39-9ffb-5a0c-254f-1681c92cd700']
Hello! My name is <PERSON>. I'm from China. Now I'm a master student in Computer Engineering major at University of Louisiana at Lafayette. I have a new idea for GSoC 2011, using Google code and kml language to develop a web-based, web-accessible tool with Google Earth embedded into it. I worked with a research team to develop an educational tool called Hydroviz<http://hydroviz.cilat.org/hydro/index.html> (http://hydroviz.cilat.org/hydro/index.html). I have those experience and want to work on it in GSoC 2011. I hope to find a mentor who would be interested in this idea. Thanks, <PERSON>.
3b4da319-2109-23be-036d-5310226a795b
['78106f39-9ffb-5a0c-254f-1681c92cd700']
Hello! My name is <PERSON>. I'm from China. Now I'm a master student in Computer Engineering major at University of Louisiana at Lafayette, USA. I'm very interested in this idea "Social Analytics Plugin for Status.net". I have some development experience with Java Script, HTML, and Google Earth/Map Plugin API. I worked with a research team to develop an visual educational tool called Hydroviz( http://hydroviz.cilat.org/hydro/index.html). I have those experience and hope to work on StatusNet for GSoC 2011. Thanks, <PERSON>
5a61ae55-b615-017f-ad8e-dfcf6bd193e3
['781193ca-5cdf-80fa-4bf3-43fedb1ec02c']
It is my privilege to announce the release of Varnish 1.1.2, the latest bugfix release of the Varnish high-performance HTTP accelerator. A summary of changes since 1.1.1 can be found in the ChangeLog file in the release tarball, or online: http://sourceforge.net/project/shownotes.php?release_id=563022 As usual, source code for Varnish 1.1.2 can be downloaded from SourceForge: http://sourceforge.net/project/showfiles.php?group_id=155816 The FreeBSD port has been updated, and packages for Debian, Fedora and a number of related distributions will be made available on SourceForge within a few days. Merry Christmas! DES
b1ab61d0-ad68-7cc6-0bfe-04f83c373764
['781193ca-5cdf-80fa-4bf3-43fedb1ec02c']
I thought I'd let you know what's currently going on at our top-secret underground Varnish R&D facility: - Poul-Henning has been working hard to resolve the stability issues and assertion failures some of you have reported (#136, #137, #138, #139, #140, #141 and #143 should be fixed, while #132, #142 and #144 are still being worked on) as well as working on 2.0 features. - I have been working on other bugs and build issues (#128, #130, #131 and #135 should be fixed, while #129 is still being worked on) as well as improving our test framework (and writing additional test cases) and trying to get Varnish to build and run reliably on Mac OS X and Solaris 10. - <PERSON> will be back on Monday and resume work on 2.0 features. - I just merged a ton of bug fixes from trunk to branches/1.1. Please please *please*, if you are running 1.0.4 or 1.1 today, take the time to test branches/1.1 and let me know about any remaining bugs or regressions. I would love to be able to close all outstanding tickets on 1.0.4 and 1.1 and release 1.1.1 on August 20th. Pretty please? With sugar on top? Thanks,
b9cccf7b-8f4a-90ff-20a7-876309f7fd53
['7819d3b5-753a-346e-401a-aaceb4dcf450']
Hi! I would like to create my own branching. It should assign values to the Var's in the order i choose. I was always using branch(..size_max, split_min) and now I can not figure out how to code a different branching. if anyone have an example on it, this would be perfect to grasp the way to implement a new branching. Thank you all, <PERSON>
f4ecd607-f901-81b8-3a07-3af01b2e42d4
['7819d3b5-753a-346e-401a-aaceb4dcf450']
Hi all, I am using gecodeJ latest version. - I use Options.java class for benchmarking. The problem is runtimes depends on System.nanoTime() method. So, what I get is basically elapsed time for program of that time. Is there a way to get the actual user time which does not depend on load on my. - For symettry breaking I use rel () to lex my var arrays. However I do lex several arrays in a sequence. Is there way (I checked the documentation, failed so far) to have something like lex-chaining all these arrays to get better propagation rather than one by one. <PERSON>, <PERSON>
30a780e9-fda6-b5cf-19e2-2afa2521ff3f
['78355aff-7492-3eb4-6f65-d6459c7cc311']
Hi, I am new to Python and coming from the EDA/VLSI Design background. I wanted to know if there are some active projects going on EDA modules written in Python. Usually, use of scripting tools in VLSI Design is on a per-project basis and rarely reaches the level of a general framework, mostly due to extensibility and speed performance limitations. That Python overcomes atleat one of them...may be a good reason to look at Python based EDA. I was interested in knowing if such efforts are on. Apart from that also any input on EDA softwares using Python will be useful. regards, Vishal
42024d77-3aeb-bc1e-5cb8-4364c9383e7f
['78355aff-7492-3eb4-6f65-d6459c7cc311']
Hi, Has anyone worked here with Clearsilver (using its python wrapper) and CherryPy? I would like to make a simple application using ClearSilver as the templating engine. The examples on ClearSilver website seem to be using CGI verymuch...can we use ClearSilver otherwise also? Is it necessary that the data to be presented should be available as an HDF file? I would like to do this at runtime...which is why we are using the templating system in the first place. I am confused about how to hand-off the data from the database using Python, assuming we get the data into some local variables...and then make an HDF out of it(would that be a Python dictionary)...and then pass this to the neo_util.CS()? any help would be great!! Thanks and best regards, <PERSON>
1b2aa0d2-acb9-37ab-74cb-8fce1cb8f81f
['783ac0df-e5b8-7690-14d1-4c8470e8c575']
Hi Are there any better error handling options for Kafka streams in java. Any errors in the serdes will break the stream. The suggested implementation is to use the byte[] serde and do the deserialisation in a map operation. However this isn't ideal either as there's no great way to handle exceptions. My current tactics are to use flatMap in place of map everywhere and return empySet on error. Unfortunately this means the error has to be handled directly in the function where it happened and can only be handled as a side effect. It seems to me that this could be done better. Maybe the *Mapper interfaces could allow specific checked exceptions. These could be handled by specific downstream KStream.mapException() steps which might e.g. Put an error response on another stream branch. Alternatively could it be made easier to return something like an Either from the Mappers with a the addition of few extra mapError or mapLeft mapRight methods on KStream? Unless there's a better error handling pattern which I've entirely missed? Thanks MIkeG
62c2e320-ebb3-3325-1779-6c9035589c35
['783ac0df-e5b8-7690-14d1-4c8470e8c575']
Hi Are there any better error handling options for Kafka streams in java. Any errors in the serdes will break the stream. The suggested implementation is to use the byte[] serde and do the deserialisation in a map operation. However this isn't ideal either as there's no great way to handle exceptions. My current tactics are to use flatMap in place of map everywhere and return empySet on error. Unfortunately this means the error has to be handled directly in the function where it happened and can only be handled as a side effect. It seems to me that this could be done better. Maybe the *Mapper interfaces could allow specific checked exceptions. These could be handled by specific downstream KStream.mapException() steps which might e.g. Put an error response on another stream branch. Alternatively could it be made easier to return something like an Either from the Mappers with a the addition of few extra methods on KStream. Unless there's a better error handling pattern which I've entirely missed? Thanks <PERSON>
7e2045c1-707c-c27e-d001-041424e62936
['7849e5bc-0670-d6bd-9093-612e2920808b']
Hi <PERSON>, If you are telling us that ICOM doesn't update their DSP code libraries... that is very sad! That said -- the ICOM radios are not true SDRs -- but simple DDC/DUC boxes locked into old and stale DSP code. Please tell me that ICOM is not that "dumb". SDRs are modular -- add new and better AM modules as the state of the art progresses. If not -- then, be left behind. <PERSON>
6b173c2e-44d5-4f92-af19-561807ee6e1b
['7849e5bc-0670-d6bd-9093-612e2920808b']
Just an observation (not a problem) but, I see that audio features like Squelch and Speech work just fine using the built-in speaker/phone jack/rear speaker jack -- not on the USB Audio CODEC output. There may be other differences. Is this by design -- or just a minor bug? <PERSON>
89bfcfe1-734f-ee14-e945-6a8011bc8538
['784b0445-dd4e-918c-a25b-6d162262e94f']
hello has anybody already read the book "python programming patterns" published by prentice hall. I thought this would be about design patterns in python but the review at amazon says different things about this book. but there is only one review there yet. I do not need another introduction to python, but if this is really an advanced book then I would like it. so, if anybody has already read this book it would be nice if she or he could give a short review of the book. thanks <PERSON>
fde42e27-313d-bda2-f67c-1f0a011176b7
['784b0445-dd4e-918c-a25b-6d162262e94f']
hello I am currently trying to switch from Xemacs to jedit for my Python and Ruby programming and I have the following questions 1) is there a shortcut for deleting an entire word like Xemacs "Alt-D" or vi's "dw" ?? 2) I have some trouble with my keyboard. I am using RedHat 7.1 and KDE 2.2 and java 1.4 I have a german keyboard and this are my problems when typing the "^" sign, it is not printed, even not if I press a space after typing it. This is serious problem because without "^" I can not write many useful regular expressions in Ruby and Python when I type C-[ for marking a whole block nothing happens. maybe this has the same reason like the problem with "^". does anybody know what to do here?? question 1 is not so important but question 2 is crucial if I really want to switch to jedit. markus
e55bd632-2186-aeed-dc33-6a4b9cde60e4
['784fdca9-bdf5-9bc1-34b3-a40fb2119543']
Hi There, I am trying to use TCP checksum offload under Linux. I found that if I install the netfilter, I cannot use TCP checksum offload feature. Because "nf_hook_slow" is called if there is any HOOK registered in the output path, and "skb->ip_summed" is set to "CHECKSUM_NONE" eventually. Am I right about that I cannot use TCP checksum offload feature if there is any HOOK registered in the output path? If it's right, how can I use TCP checksum offload feature while having the netfilter installed. Can I do it but disabling some netfilter options? Thanks, <PERSON>
b3f894d3-9da8-e760-e84b-788203769a40
['784fdca9-bdf5-9bc1-34b3-a40fb2119543']
Hi All, I am trying to add some partial TCP/IP Offload features to the protocol stack, which means I have to modify the source codes. Could anyone tell me what to begin with? Can I just take out the network module from Linux and install my own one? Where can I find the information of the interfaces between socket and TCP/IP, TCP/IP and NIC driver? Thank you, <PERSON>
9803689e-c04d-9185-cb15-382bfe4676b4
['78778a86-0391-7dbe-8348-9578d4e0f828']
Hi everybody, sorry, noob here. So we've just started using postgre and right at the start we had a hardware problem with a hard drive failing. We can still start postgre but a tablespace for a db that we don't even need anymore is lost. Now we would think that shouldn't be a problem, we could simply drop that db. postgre, however, doesn't allow us to, but tells us to vacuume the db first. Starting postgre singleuser and trying to run vacuum still gives us the same trouble. Now a naive question: is there any (sane) way to simply get postgre to entirely forget about the broken db - and that's it (dumping the other db and loading it back might not be the best solution, since it's rather large..)? TIA, best, <PERSON>
8248d49f-87fd-c8b5-709e-f64d85868a3c
['78778a86-0391-7dbe-8348-9578d4e0f828']
Hi everybody, considering the improvements and ubiquity of machine learning applications, and the impact they (may) have on the privacy of individuals, we want to bring together people interested in either of the two sides, here in Dresden on Sept 21/22 2015 ( http://www.prinf.eu ). To ignite discussions, we see two (rather obvious) topics, that we think are important and interesting: 1) considering all the public personal data by all sorts of people, we wonder how good */inference attacks/* can actually get, even on people who don't publish (a lot of) information about themselves - directly related of course is the question, if we can find new */privacy metrics,/* and - if there are ways to encounter such attacks, without diminishing the utility for the users (too much). And 2) on a related note we want to further explore ideas towards */privacy-preserving recommenders/*. We will organize this event as a "traditional" scientific workshop soliciting submissions, which will be reviewed and subsequently published, to make it easier for the academic audiences to convince their funding entity of its importance ;-) - but we really mainly want to engage in discussions, may be fostering some future collaborations, as well. I'm attaching the usual Call for Papers - and I hope, of course, that we will receive some interesting submissions (to make the official part interesting - so help in advertising is appreciated). We will accept both novel scientific contributions, but also datasets and replication studies. But most importantly, we hope to attract a broad audience of interested participants from different professional backgrounds, to facilitate great discussions! Thanks a lot! <PERSON> & <PERSON>
9b6f01bc-ead8-e862-3570-09864dcd7534
['787f41eb-820c-b0de-788b-2f533338d098']
Hi <PERSON> and Ansible realease team The letsencrypt maintaniers would like to request consideration of https://github.com/ansible/ansible/pull/37572 to be in 2.5 final. This rather small patch, fixes an issue with the wildcard certificate support, which was released by Letsencrypt just a few couple of days ago. The implementation for wildcard certificate support in the letsencrypt module for 2.5 was only possible with an enormous effort by <PERSON>), who worked closely with Letsencrypt, I see the value for Ansible to have an excellent support for the latest feature of letsecrypt is considerable and that is why I vote for this patch to be included in 2.5 final. Thanks
7639e988-24fc-139e-368d-b0b39e220066
['787f41eb-820c-b0de-788b-2f533338d098']
Hi I wanted to bring awareness of a little patch I submitted a while ago (3 lines change to the basic module_utils helper. (I even would call it a bug fix). https://github.com/ansible/ansible/pull/16492 Currently, for an undefined argument in a task, ansible uses the default value of this argument. That is absolutely fine. When a user passes the _same_ value as the default value for the argument, should it make any difference? It should not but it does: the "required_one_of, mutually_exclusive, required_together" will handle it as "user defined", even it has the same value. As a consequence for a required argument: If you use a task with an arg that is required, the default value of this arg is "None/null". The uses uses "null", what happens? Ansible won't complain. The patch fixes this. As a consequence in role defaults: In a role you specified role default value for a task with two mutually exclusive arguments with their default values. But Ansible complains, it is currently "not" possible. The patch fixes that. Even if there may be some situations, where "omit" may help. Why should it make any difference if a user passes the same value myarg=None, or the module uses its default value "None"? Thanks and take care <PERSON><PHONE_NUMBER> Currently, for an undefined argument in a task, ansible uses the default value of this argument. That is absolutely fine. When a user passes the _same_ value as the default value for the argument, should it make any difference? It should not but it does: the "required_one_of, mutually_exclusive, required_together" will handle it as "user defined", even it has the same value. As a consequence for a required argument: If you use a task with an arg that is required, the default value of this arg is "None/null". The uses uses "null", what happens? Ansible won't complain. The patch fixes this. As a consequence in role defaults: In a role you specified role default value for a task with two mutually exclusive arguments with their default values. But Ansible complains, it is currently "not" possible. The patch fixes that. Even if there may be some situations, where "omit" may help. Why should it make any difference if a user passes the same value myarg=None, or the module uses its default value "None"? Thanks and take care René
cb5547fd-0555-903a-1cba-ba77fd27fccb
['788949d0-257a-a386-e819-e5d4da54ba54']
Hi. In order for restoration to work fine, I have to create a folder using the interface (e.g RounCube) which also creates the required structures. Then I copied the emailfile e.g xyz:s to /myrestoredemail/cur/ How do i properly create a folder with its structures(index,etc) without using the RoundCube web inteface ? Thanks CoolAtt
a576c61e-a669-c6b4-71db-bbce24f45030
['788949d0-257a-a386-e819-e5d4da54ba54']
Hi All. I want to know how filenames are changed by dovecot e,g more details on how it renames the files when a mail is read. For read mails i found that it appends an S. <PHONE_NUMBER>.P22376Q0M765148.ecs,S=1161:2 <PHONE_NUMBER>.P22376Q0M765148.ecs,S=1161:2,S Want to know if other modifications done in file names in INBOX & in SENT (maildir/.Sent/cur) Thanks CoolAtt
18ba83e8-620a-1650-8361-b8eda4039640
['789094ee-efca-2a34-69f3-99f9b291134a']
Call for Papers “Ethics and Ontology. The Moral Phenomenology of Charles Taylor” June 10-11, 2016, University of Antwerp This conference centers on <PERSON> paper “Ethics and Ontology” (2003) and its central theme of the relationship between ethical beliefs and ontological views. One of the most striking features of <PERSON>’s philosophy is his consistent phenomenological approach to ethics in avoiding and criticizing reductive scientistic accounts of moral experience. This approach can be found in many of his writings, but it is most explicit in “Ethics and Ontology”. This paper raises the issue of scientific naturalism by arguing that it sets us the following challenge: either we correct our (implicit) naturalist ontology or we must revise the most salient features of our moral experience. <PERSON>’s moral phenomenology defends our commonsense moral reactions against reductionist views that attempt to dismiss these reactions altogether as mere projection on a neutral physical world. His criticism is that this naturalist ontology annihilates our very sense of morality, that is, the sense that moral values are in some way different from, higher than, or incommensurable with natural desires. Against this background, the central question of the conference is: what do our ethical views commit us to ontologically? Assuming a familiarity with <PERSON>’s work, people are invited to deliver a paper about both his article “Ethics and Ontology” and related topics of his moral phenomenology, such as: the critique of reductionist naturalism, the relation between moral values and natural desires, etc. In this way, this conference aims to discuss <PERSON>’s moral phenomenology in order to open up the question of the implicit ontological commitments behind our ethical beliefs. Keynote speakers <PERSON> (University of Notre Dame) <PERSON> (Macquarie University, Sydney)
835917f4-cfef-db88-1176-f55b2b6dc07a
['789094ee-efca-2a34-69f3-99f9b291134a']
Final Call for Papers “Ethics and Ontology. The Moral Phenomenology of Charles Taylor” June 10-11, 2016, University of Antwerp This conference centers on <PERSON> paper “Ethics and Ontology” (2003) and its central theme of the relationship between ethical beliefs and ontological views. One of the most striking features of <PERSON>’s philosophy is his consistent phenomenological approach to ethics in avoiding and criticizing reductive scientistic accounts of moral experience. This approach can be found in many of his writings, but it is most explicit in “Ethics and Ontology”. This paper raises the issue of scientific naturalism by arguing that it sets us the following challenge: either we correct our (implicit) naturalist ontology or we must revise the most salient features of our moral experience. <PERSON>’s moral phenomenology defends our commonsense moral reactions against reductionist views that attempt to dismiss these reactions altogether as mere projection on a neutral physical world. His criticism is that this naturalist ontology annihilates our very sense of morality, that is, the sense that moral values are in some way different from, higher than, or incommensurable with natural desires. Against this background, the central question of the conference is: what do our ethical views commit us to ontologically? Assuming a familiarity with <PERSON>’s work, people are invited to deliver a paper about both his article “Ethics and Ontology” and related topics of his moral phenomenology, such as: the critique of reductionist naturalism, the relation between moral values and natural desires, etc. In this way, this conference aims to discuss <PERSON>’s moral phenomenology in order to open up the question of the implicit ontological commitments behind our ethical beliefs. Keynote speakers <PERSON> (University of Notre Dame) <PERSON> (Macquarie University) <PERSON> (University of Tampere)
5ae4d733-f312-083d-ba91-2cb56ada55cf
['789da657-7ba6-b4e0-78e0-9f091edcc5dc']
Hi, I'm currently working on a wireless network load-balancing algorithm. The scenario i createad is made up by a network controller node, linked by a full duplex wired connection to some base stations. The controller keeps track of each mobile node and decide whether to change the base station of a mobile node in order to obtain a fair distribution of the load among the base stations. For the sake of simplicity, i suppose all mobile stations not to move and that a mobile station can "see" all the base stations avaiable in the scenario. I want to force a mobile station to roam from a base station to another one. I thought it was a good idea to exploit the mobile ip mechanism, the only thing i had to do was to force it to start every time the controller decide to change the mobile nodes distribution among the base stations. I did several attempts, but did not succeeded in. Any suggestion? Thanks in advance. <PERSON>
a4381120-3c90-006b-8b17-fa76227b9368
['789da657-7ba6-b4e0-78e0-9f091edcc5dc']
Hello everybody, Say I want to write a function that takes as input a timestamp (in the form {{y,m,d}, {h,m,s}} ) and returns another timestamp, obtained by adding a certain amount of {hours, minutes, seconds} to the input. It wouldn't be difficult but, since I'm dealing with tuples, I wouldn't have any guarantee on the validity of the obtained timestamp, e.g. I could have obtained something like {{2011,02,30},{...}}. In such cases, the calendar:valid_date/1 and calendar:last_day_of_the_month/2 would be useful. I was just wondering if there already exists a function that does such a work. :) Thanks in advance.
6c551f6d-2025-5f10-1b75-fcc01b84b42d
['78b10495-c782-643b-5f65-c3cbca45619c']
Hi, I have a Scheme function that returns a music expression. Unfortunately, the expression returned depends on the properties of the context it's evaluated in. I.e., based on the current setting of Staff.dynamicAbsoluteVolumeFunction it will return a different number of events in the music expression. Is this possible? The background of this is that I'm writing a Scheme function that generates single-note crescendos by emitting a series of \set Staff.midiExpression = ... events. Currently I've written a Scheme function that's able to generate these events; however, it requires me to specify numerical starting / ending values for midiExpression. I'd like to be able to specify dynamics instead of numerical values, and have the Scheme function automatically look up the numerical values based on the current volume function defined for the Staff. However, so far I haven't been able to figure out how to read the Staff properties except using \applyContext with a lambda, and AFAICT it's not possible to return a music expression from this lambda? Or is there another (hopefully simpler?) way to achieve what I want? --T
22074482-80e1-8232-5faf-5ae2add7a649
['78b10495-c782-643b-5f65-c3cbca45619c']
I couldn't remember when I started using D, so I had to go dig in my personal diary entries until I found it... In late 2011, I had already heard of D during my search for something better than C++, but I didn't really start seriously using D until by chance I came across Andrei's TDPL at a local bookstore. That got me *really* started. Apparently my first D program was a brute-force search to discover all augmentations of a certain class of 4D shapes called duoprisms. I had a very pleasant experience with it... and as they say, the rest is history. :P --T
7596fea5-a300-8b1c-e3dd-a1b5da770b5b
['78b7c6dd-3d8e-76b7-7947-f1a3b0760602']
Hi <PERSON>, Thanks for the information and suggestions. Robinhood v3 sounds quite exciting - I will give some thought as to how we might be able to participate in an early testing phase. I think our timeline probably points toward moving into production with v2, though. I am by no means a mysql expert. I'd very much appreciate if you can point me at tuned my.cnf examples. If I may ask for comment on high-availability using the current Robinhood releases and our lhsm+tmpfs setup - my intuition is that dual controller, direct-attached storage and two hosts each running a mysqld and an rbh manager - one for lhsm and one for tmpfs - would work well; the storage can be two separate LUNs (probably made from SSDs), one for each database. Then with a pretty conventional corosync/pacemaker setup, we can have one node ready to take over the services of the other as required but be running in an active-active mode in normal operation. Currently, our file system has 220M inodes in use (growing), which suggests each host having at least 256GB of memory. Having each LUN be a couple TB or so should allow for future growth (which we expect). Is this the kind of setup which is generally recommended if one needs a highly available Robinhood setup? I suspect that alternatives like mysql master-slave replication or DRBD will impact performance and/or not work quite as well - is my concern unfounded? FWIW, our cluster currently has ~22k cores, and we'll be growing to ~50k next year. Our combined create/open/close/rename/unlink activity is recently averaging in the few thousand ops/sec, though being in the 10s of thousands ops/sec isn't uncommon. Thanks again, <PERSON>
44fc1eda-cec9-b9e3-2318-13485fe4b752
['78b7c6dd-3d8e-76b7-7947-f1a3b0760602']
Thanks, <PERSON>. We are still running 2.5.3.90, which doesn't have that fix. That patch looks like it would solve our slow-to-mount MDT. FWIW, I don't think we have many (any?) empty plain llogs, but the removal of the llog_process_or_fork() call in llog_cat_init_and_process() looks like it addresses our issue - I see that in the stack of the osp-syn-* threads when the MDT is being read like crazy during mounts. As a followup - is there any reason *not* to unmount the MDT, mount it as ldiskfs, and simply delete the plain llogs in our MDT's O/1/d* folders that contain only CHANGELOG_REC records? Or even every file under the MDT's O/1/d* folders? I'm a little unsure. It seems that most of (if not all of) the files there now are just taking up space, and nothing else is going to remove them. FWIW, our intent is to start using changelogs and robinhood again after we upgrade to a later version of Lustre than what we are currently running, at which time we'll just start over - register new changelog users and rescan the whole filesystem. We won't care about any prior history. Thanks again, <PERSON><IP_ADDRESS>, which doesn't have that fix. That patch looks like it would solve our slow-to-mount MDT. FWIW, I don't think we have many (any?) empty plain llogs, but the removal of the llog_process_or_fork() call in llog_cat_init_and_process() looks like it addresses our issue - I see that in the stack of the osp-syn-* threads when the MDT is being read like crazy during mounts. As a followup - is there any reason *not* to unmount the MDT, mount it as ldiskfs, and simply delete the plain llogs in our MDT's O/1/d* folders that contain only CHANGELOG_REC records? Or even every file under the MDT's O/1/d* folders? I'm a little unsure. It seems that most of (if not all of) the files there now are just taking up space, and nothing else is going to remove them. FWIW, our intent is to start using changelogs and robinhood again after we upgrade to a later version of Lustre than what we are currently running, at which time we'll just start over - register new changelog users and rescan the whole filesystem. We won't care about any prior history. Thanks again, Craig
c7996736-0eaf-2869-a777-e8629ad37177
['78b8b77f-a8c1-7e2e-177e-f1a2cb2f9230']
Hi, A few more sci-fi books I'd recomment are: <PERSON> Mars Trilogy; while he might sometimes get bogged down in scientific details, his books are very well researched, very realistic and quite well-written. <PERSON> Hyperoin Series is one of the best ever written IMHO - sure, the first book isn't that good, but it gets better and better as you progress in the series, and he is both a great storyteller and has VERY creative ideas. I'd also recommend <PERSON> "Songs of a Distant Earth", which I see as one of his best books. Yours,
3f65fe94-21fc-4ee9-b8f3-c05631c92e21
['78b8b77f-a8c1-7e2e-177e-f1a2cb2f9230']
Hi, I've cobbled togather a modified and consolidated version of the Classic Traveller character generation tables, based on CT's LBB1 and S4: Citizens of the Imperium, with a few elements of MT ("Special Duty"). It is a "basic" system modified to contain all the skills in the LBBs and S4, and yeald characters with skill amounts similar to those of "extended" system characters, but using the simpler and faster "basic" way. A CoTI forum user has converted my original MS-WORD-2000 document into a PDF, which could be found here: http://neil.smudgypixels.org/temp/chargen.pdf Yours,
ac808af6-237c-7f4d-5d99-f9d8c55b1261
['78ba0ee8-ce81-dcbc-18b5-4a75f0a422a0']
Hi, This is most likely trivial but I cannot find it anywhere... I have a text_field and a button. I would like to use the input in the text_field as a value to compare against a column in a mySQL database that's not the id (call it 'name'). How can I do this? Thanks for any help.
e94f742c-1569-16e7-264f-7161021acd5b
['78ba0ee8-ce81-dcbc-18b5-4a75f0a422a0']
Hello, I am brand new to RoR trying to use the framework to build a simple web application (for an imaginary online shop) just to learn the ropes. Once every so often I come upon some problem that seems to have a really simple solution but I just cannot see it because of my inexperience. I was hoping you could help me out... So here's the issue. I have a simple mySQL database and a scaffold model for it. The database will contain products with a name and two different prices (eg. per unit and per kg). However the products themselves have 5 different price type combinations (each, unit, kg, 100g, 100ml). Each product has to have at least 2 out of the 5 prices, but I will be importing them directly and I I have no control of which 2. I am trying to set up the validation for this, but I have no idea how. It needs to check that at least 2 of the 5 prices are present, but it does not matter which. The railes validates_ methods seem a bit simple for this, but maybe I'm just not seeing it. When the controller invokes the create method is there any way to use the params passed in the validation? If not then any idea how I can solve this problem? Thanks for any and all help, W
76c4759d-87ad-dcd2-a9e8-e06bc025e423
['78ba4915-543e-99c1-ad2e-4acb6d2006ef']
Hello, i am trying to use erp5 with erp5 livecd.I am upgraded with all the bt5 necessary to utilize erp5_accounting. i am installed erp5_accounting_ui_test too. After having created predicates, in portal_rules/default_invoice_transaction_rule, according to erp5 handbook ,when i confirm sale invoice no transaction lines are created, and in simulation_movement appear only the applied rule "default_invoinice_transaction_rule without movement. I need help please no error log appear. erp5 handbook tell "Do not forget to fill Source and Destination views with proper accounts" how ? in the proper organisations i am insert the proper region. thank's for answer <PERSON>
e681ce74-d974-fe26-f47d-3749fff39897
['78ba4915-543e-99c1-ad2e-4acb6d2006ef']
Hello, i am trying to use erp5 with erp5 livecd.I am upgraded with all the bt5 necessary to utilize erp5_accounting. i am installed erp5_accounting_ui_test too. After having created predicates, in portal_rules/default_invoice_transaction_rule, according to erp5 handbook ,when i confirm sale invoice no transaction lines are created, and in simulation_movement appear only the applied rule "default_invoinice_transaction_rule without movement. I need help please no error log appear. erp5 handbook tell "Do not forget to fill Source and Destination views with proper accounts" how ? in the proper organisations i am insert the proper region. thank's for answer <PERSON>
c1d24793-53ba-2ee7-e762-e3b2eec5b001
['78bdea0d-890f-c187-726a-c9a67fc161d8']
Hello, I use a CMakeLists.txt script to build an application which works pretty well so far, except for the tar.gz package build. In the install procedure, I copy the application executable, the config files and, SysV init and SystemD service files, and an entire directory. In the make install and in the .deb file generated, all those files are correctly installed. But the .tar.gz package only has the executable and the config file, not the rest. Can anyone help me point out where my mistake is? The script is available here: https://github.com/babelouest/glewlwyd/blob/cmake/CMakeLists.txt The install commands that are missing in the .tar.gz package are the following: https://github.com/babelouest/glewlwyd/blob/cmake/CMakeLists.txt#L211 https://github.com/babelouest/glewlwyd/blob/cmake/CMakeLists.txt#L212 https://github.com/babelouest/glewlwyd/blob/cmake/CMakeLists.txt#L213 And the .tar.gz package generator commands are here: https://github.com/babelouest/glewlwyd/blob/cmake/CMakeLists.txt#L252 Thanks in advance
5ac0c3fe-d67a-dbc7-0026-6ed38c3e7806
['78bdea0d-890f-c187-726a-c9a67fc161d8']
Hello, I'm currently working on a ldap backend for the addressbook of owncloud. The goal is to use a ldap server to store the contacts (obviously :p ). The work in progress can be seen there for now : https://github.com/babelouest/apps/blob/contacts_ldap/contacts/lib/backend/ldap.php https://github.com/babelouest/apps/blob/contacts_ldap/contacts/lib/ldap/connector.php I have a couple of questions about ldap and user_ldap. In my implementation, the connection between a ldap entry and a vcard is based on an xml configuration : https://github.com/babelouest/apps/blob/contacts_ldap/contacts/appinfo/ldap_vcard_connector.xml I'm currently using user_ldap but only for myself, without any groups, so I don't know how it works with groups and user list. My first question is about groups. In general, do you know if there is an ldap attribute that could be connected to the VCard 'CATEGORIES' attribute ? There are the groups attributes (CN) in the DN, but these are part of the DN, so they can't be modified easily. I thought of 'o' or 'ou' attributes, but these may be used for different purposes, depending on the application behind. In you opinion, what should be the best way to connect the 'CATEGORIES' property into an ldap entry ? The second question is more technical and is about the ldap sizelimit attribute on the server. When an addressbook has more entries than the sizelimit, how do you manage this with user_ldap ? Do you perform multiple ldap_search as a workaround ? I have several ideas how to manage sizelimit parameters, but I still haven't found one yet. Thanks in advance for your help. /Nicolas
b6077ff2-d499-8c99-8684-21fd9310b7fc
['78cda4b4-bc60-d852-3343-97e532ced296']
<PERSON>. But what i want to know is where exactly is this call(CreateBitmap will eventually result in calls into probably X11 code) done. I have looked into the documentation in winehq.com, i get a fair idea of how it is done, however when i look into the source code, im unable to understand much. What i want to know is, this program that loads the windows exe and which runs them, where is this program located. Also, the code where CreateBitmap is mapped into the Xlib, where is this code. I am looking at some kind of low level documentation, or atleast a place where i could start. Thanks, <PERSON>
a6468243-cb07-296b-43cb-520883990087
['78cda4b4-bc60-d852-3343-97e532ced296']
Hi All: I was looking thru the source code of the XFree86 server. I was wondering if it was possible to change the background color/image of the Client Display. Also, who creates the display buffer for the display. i.e. if there are 3 windows open on a desktop lets say 4, where is the Background added to the display buffer. Is the Dislpay beffer same as the Frame buffer Thanks, <PERSON>.
30f8de9b-13b1-0cbb-cdc3-fcb1988d20c8
['78d7fd20-9f1e-9a9c-04dd-e8b9d3ecc959']
We are a group of students in telecommunications engineering from Uruguay. We are studying some network tools for our final project and we would like to know if someone could tell us which of this tools have IPv6 support: · Bprobe · Cprobe · Pathload · Pathrate · Pathchar · Clink · Nettimer · Spruce
c57953f0-f2ed-ed81-8d38-2aee41fc22d2
['78d7fd20-9f1e-9a9c-04dd-e8b9d3ecc959']
Hi all, I’m using Smokeping (http://oss.oetiker.ch/smokeping/) and I have some doubs that maybe you can answer: - Why Smokeping needs to be connected to internet even when I have only one Target to a host in my LAN ? - What is exactly the Standard deviation calculated at the bottom right of the graphs? Can it be interpreted as Jitter? Or is there another way to know the jitter? - Why sometimes when I add a new host, for the first minute I get a list of errors saying that the RDD file doesn’t exist? This happens commonly? Is there any way to solve this problem? Many thanks, <PERSON>
63bbeaa0-56a7-061c-8394-9c3d7ace9268
['78ff792c-9a11-2bb3-1805-83bbd49ffdf3']
Hi all, Somewhat of a basic CEDET question here. I have a root folder called "mware", and within mware, there are several projects and codes which are related. for example, ATL, gen_thread, evpath.. etc. Some of them have dependencies (like evpath requires the previous two). How do I "build a database" in semantic, or something, that will have information for all of the associated projects, so that when I come across a struct or function, I can do semantic-ia-fast-jump and it will bring me to the appropriate definition. Maybe I have to do this via csopes? but it seems that cscopes just shows you where things are defined, whereas semantic will let you jump. Any ideas on how I can do this, preferably something wtih my .emacs file so I don't have to do it each time? Thanks so much! <PERSON>
dd8f13a6-1376-5bf2-291e-aba6ce9dd0b4
['78ff792c-9a11-2bb3-1805-83bbd49ffdf3']
Hi, I use CEDET quite extensively and it works great with code completion and jumping to symbol definitions. I have some questions, though. 1) When I create a new struct somewhere, it often won't become "visible" to semantic in terms of code completion and semantic-ia-fast-jump until I close everything and re-open. How do I control when semantic parses my files? Is there a command I can enter in emacs to force semantic to analyze the code, instead of waiting on close? 2) I generally like find-tag to find global definitions instead of the semantic jump features, because I like to see the definition in another window, and also, I like the pop-tags ability. Because of this, I generate tags for my project manually myself with ctags -e -R . and then M-x visit-tags-table. I see that CEDET/Semantic has ctags support (semantic-load-enable-all-exuberent-ctags-support) Does this mean I don't have to do the tag management myself somehow? I've tried not generating the tags manually and then use the "find-tag" command, but Emacs always prompts me for a tags table. Where does CEDET/Semantic store the tags table, and how do I visit it? When is the tags table generated? 3) Similarly for Cscopes, which I like to use to "find all functions calling this function", etc. I have cscope support enabled via CEDET, but where is this database kept, and how do I force it to index it? Right now, i use the xscope interface and load and generate the cscope files myself OUTSIDE of CEDET. I think this is a bad way to do it because CEDET generates them internally? Sorry for the long email and thanks for the AWESOME tool that is *really* helping me with my development for larger projects. <PERSON>
ae19ea74-02f5-dde5-d057-3a41e86b7586
['79038d41-9a38-019d-dbbd-224c5c262148']
Hi all, Thought I'd send a RFC out to you all to get a little feedback before I make any changes to this since I didn't originally start the project. If no one responds with a good reason why NOT to make the changes I purpose, I will proceed in a somewhat timely manner. It's been over a year since I manually moved the Presentation Project from wwwold to the wiki. No one who was previously involved in the project has presumed membership (or leadership) with the project. There's only been one meaningful edit by <PERSON> to one of the project's two subpages. He kindly added links to a Gentoo presentation he recently gave. I currently see no reason why we need a separate project for Presentations. Maybe once when Gentoo PR was busier and we had more people giving talks it was necessary. If we're just looking for a place to host links or files I think merging this project with Public Relations will work nicely. I propose hosting the content of this project under the Public Relations project, since it has members who are actually overseeing it as Project specification (GLEP:39) demands. Empty projects should be removed. They are dead. It's not me who says it, it's GLEP 39. Be mad at it, not me. :P I already have a repo for PR setup at projects.g.o. I will make a presentations/ directory for storing the PDF files that <PERSON> was recently looking for...I'll have to dig them out of the Wayback Machine, but that shouldn't take me too long. Your wiki warrior, Maffblaster
adad85b0-2c2f-b435-c130-24b795a3f652
['79038d41-9a38-019d-dbbd-224c5c262148']
All, After _many_ essential and necessary years of service the Gentoo Documentation Project (GDP) has fulfilled its purpose[1][2] and is in the process of being phased out. As a platform, the wiki has enabled every Gentoo developer (not just those in the Documentation team) and many in the community to maintain Gentoo documentation. All the relevant "HOWTO" articles and other articles produced by former GDP members (and other contributors) have long since moved to publicly editable wiki namespace. These days the community and every Gentoo developer should step up to the plate from time-to-time and maintain the docs. There is very little that is currently "locked down" on the wiki. The Handbook (and everything else in the Handbook namespace) is one exception to what is publicly editable. Since it contains our official and crucial installation instructions; it is a bit risky to put it publicly editable namespace. This creates an inherent 'who's responsible to keep it updated?' So, in order to clarify who will continue to assume responsibility for the Gentoo Handbook, the Handbook project has been newly formed as a sub-project of the Wiki project. Currently just SwifT and me are in this new project, but anyone who's interested in updating/maintaining the Handbook is welcome to join. Here are our current goals, short and sweet: 1. Actively modernize and maintain the official Gentoo Handbook documentation in the Handbook: namespace. 2. Handle 'marking' the Handbook for translation. 3. Maintain the Handbook Development guide. Comments on these are welcome. Suggestions for Handbook updates or fixes can be left on each page's discussion page. Among other areas of contribution, those who are a part of this project will make it a priority to respond to these queries. I suspect next we'll need an e-mail address separate from the wiki's mail address. Other than that we'll still all hang out in #gentoo-wiki on Freenode and continue advancing Gentoo to a better place. Again, comments are welcome. Kind regards, <PERSON>
cd99c884-0645-1730-b2be-e1b8b16d49e9
['790a5864-927d-1f84-0377-dcd77c8a4f20']
Hi folks, Just for letting you know, my talk "Accelerating Random Forests in Scikit-Learn" was approved for EuroScipy'14. Details can be found at https://www.euroscipy.org/2014/schedule/presentation/9/. My slides are far from being ready, but my intention is to present our team efforts on the tree and ensemble modules, including along the way some of the lessons we have learned. In particular, I would like to thank <PERSON>, <PERSON>, <PERSON>, <PERSON> and <PERSON> who have contributed a lot these last months to improve these modules! Thanks guys! Cheers, <PERSON>
44207ef4-8241-322a-7b0c-10e777ac8c65
['790a5864-927d-1f84-0377-dcd77c8a4f20']
Hi Team, Given the increasing maturity of the project, we have decided (or, more precisely, I convinced my advisor :-)) to use Scikit-Learn in the machine learning course given at my university. Our objective is to make our students use Scikit-Learn for three assignments. We were previously using Matlab with some home-made modules. Has any of you already tried that? As the teaching assistant for this course, I plan to give my students a tutorial to Python+Scikit-Learn. I was wondering if besides the tutorials in our user guide, any of you had made (or knew) teaching materials targeted for students? I am aware of <PERSON> and <PERSON> tutorials. Also, if you have anything that comes to mind regarding elements I should include in such a tutorial, please feel free to suggest! Best, Gilles
0a00f1a6-3f96-235f-aab7-3443a7399a32
['790ae423-335a-af80-c641-b6cefcccdb54']
Hello I'm trying to migrate our software from 2.3 to 2.4 and I'm struggling with the new timebase concept. Here is how it works in 2.3: - I have a small kernel module that calls xnpod_stop_timer and then xnpod_start_timer with a handling function of my own. - This handling function does some housekeeping and then calls xnpod_announce_tick to trigger the nucleus (and eventually the pSOS skin). This is basically what <PERSON> suggested about a year ago. Now, all these functions are gone in 2.4, how's the best way to do something similar now? I couldn't figure out how the new timebase concept can be used for that. Thanks in advance. Regards, <PERSON>
e01bbffb-2d9a-e509-2852-483c6116ca6a
['790ae423-335a-af80-c641-b6cefcccdb54']
I'm trying to get oprofile working on a MPC8270 with xenomai. However, hardware perofromance counters don't seem to work due to a bug in the 603 processor. I tried it with oprofile's timer mode without success. Before digging deeper: Has anyone experience with this? Since oprofile's timer mode requires interrupts not to be disabled, I suspect this interferes somehow with the adeos ipipe ... <PERSON>
8243979d-4a90-f12b-bd24-94ebc40d4ebe
['791a81d5-c3ec-d290-12ba-610f8731625e']
I am running 1.2.8 and I have made an /etc/exports file to share and I have started /etc/init.d/nfsd manually and everything works great. I ran the setup program and checked NFSD under services and did a save-config. The problem is that nfs does not even try to start at boot. I poked around in /etc/init.d/rc3.d and noticed there are no K/S links to nfsd, maybe that is the problem? Should this work? -thanks
e1c69bc8-23d5-1b35-8d4e-cfcb676c51a4
['791a81d5-c3ec-d290-12ba-610f8731625e']
I have an app that uses a "details" buttons to expand/collapse a window to reveal more info. When expanded I would like to fix the upper portion to its current size, and let the bottom (expanded) part be resizeable. I tried setLayoutHints(LAYOUT_FILL_X) on the top frame and that seems to work, except it appears the top will always be resized to the default height?? I tried the logical setHeight, resize, etc and they don't have any influence. I even tried all the strange recalc, layout, etc and still nothing. Is there a way to set the size of something after construction? I did not see a setDefaultHeight function. Thanks
1b48c058-329b-f649-2037-c19276a25576
['7924f447-6d20-d981-7833-7635e00f5904']
Hello fellow hackers, I know nothing about 3d printing but I want to print a servo mount, I have seen some interesting designs in thingiverse but openSCAD is terse to say the least, could someone recommend some resources (app and tutorials) for mac? I would be on the hook to provide something of equal value, like the head of a rival or an antarctic flower. Thank you.
a61edbad-a891-67ec-6f7b-0669ff6fd4e0
['7924f447-6d20-d981-7833-7635e00f5904']
Hello <PERSON>, Just wanted to introduce myself as I recently joined this mailing list. My name is <PERSON> and hail from the sunny Dominican Republic, my background is software development and have just recently started to learn the hardware aspects of computing. I have been to the space twice and the energy is great, something about a room full of smart people working on projects they feel passionate about puts a smile on my face. Thank you.
70b90848-2251-1d41-6108-e87578ff9dce
['792bdb77-b8e4-26aa-7b3a-191b1d43dd0d']
It seems like the votes have stagnated! I'd like to see this go to the front page, yet we only have 39 diggs so far, I can't believe there are only 39 people that read these mailing lists who are interested in GNUstep. I'm not sure how many diggs we need to get to the front page, but I think it's 40, which means we only need 1 more vote, come on people! Just do it! <PERSON>
00fffed7-78e2-5108-a02a-68a65d8aa6c9
['792bdb77-b8e4-26aa-7b3a-191b1d43dd0d']
I just added the missing bits to the toll-free bridging code in corebase but I'm running into some problems when comparing NSCFStrings to other types, like NSConstantString. The PASS_CFEQ() macro in CFTesting.h calls CFEqual, which in turn calls -isEqual: if either object is an objective-c object. The CFSTR() macro just turns into @"" when called from within a .m file, and since all test files are .m files all the test strings are an instance of NSConstantString. That should be enough background... The problem I'm running into is that [NSConstantString -isEqual:] is returning NO on instances of NSCFString. I tried following the GSEQ_STRCOMP() code in GSeq.h but I can't make heads or tails of it. I setup a breakpoint in CFStringGetCharacters() and the range passed onto it is {0, 0}. So it doesn't return anything useful. Can someone more familiar with the code there shed some light on what might be going on? I've already had to fix a few bugs in CFString that this exposed, so I've got a mistake somewhere else, but I can't follow the code well enough to figure out where. Thanks <PERSON>
e6d4f3aa-43a4-5e72-6e4e-bb5dc5c86a4a
['792c1aaa-3044-a3b5-e582-44e5b13eb192']
Hi <PERSON>, Hi <PERSON>, The disadvantages in using something at the filesystem/block is specially regarding load balancing. We have two servers, with different IP addresses inside our network, where courier is running. At the front, we have the firewalls and LVS doing a load balancing. They will redirect to the active courier server. So, if we replicate at the filesystem/block layers, we have to take out several files from the replication proccess, because the different IP addresses, different server names, and so on. As a solution for that situation, I think that a "replication point" feature in courier will be a nice solution. We have a software develop team, so we can do that contribution to the courier project. Maybe we could improve this idea... Best regards, <PERSON>
27d8a628-8636-b7e2-c85b-b03cff054bad
['792c1aaa-3044-a3b5-e582-44e5b13eb192']
Is it any feature in courier that allows me to inform backup storage paths? Today I use unison to maintain a replica of our email files, configurations, ..., in a second server to use as a fast recovery backup solution. The problem is that unison uses a lot of resources and IO_WAIT is always very high. So, I was wondering if it is any feature to inform backup storage paths for courier to save all its files always into two different paths. Does something like this exist? If it does not exist, don't you think it is a interesting feature to be implemented? Thanks and regards, <PERSON>
812d1124-4cf5-4f1f-8479-04c70e67585f
['793d3533-2ee3-a301-df70-aeed2fac7b61']
Hi all Just a heads up that entry to this years Sheffield Hat will open at 7pm BST on Tuesday 2nd June. The event itself takes place at Hillsborough Arena in Sheffield on the weekend of 5/6 September 2015. All the details can be found at the following website, with more information being posted as and when it is confirmed: www.ninety2ultimate.com/sheffield-hat/2015 There is also now a Facebook event you can sign up to - just search for Sheffield Hat 2015 and you should find it. If you have any questions, you can drop me an email. <PERSON>
e5fece69-49b2-aa36-e983-4699db4c3dad
['793d3533-2ee3-a301-df70-aeed2fac7b61']
Evening all Mix & Mingle 2013 is now officially open to entry, but be quick as places will go fast!! If you're a regular then you will know exactly what to expect, but with a few changes for the better (see below). If you've never played M&M before where have you been??!! Big tournament, big party, great costumes and a whole weekend of fun with an 80s theme!! This years' event is the weekend of 9-10 February 2013 which happens to coincide with my 30th birthday so the party is going to be huge!! It does also clash with Uni Mixed Nationals (I think?) so please bear that in mind when you enter. Head over to http://www.ninety2ultimate.com/mix-mingle/2013/entry and fill in the little form. Payment details will be shown when you submit the form but you likely won't hear from me until your payment has been received. You place will only be confirmed on receipt of your payment. Some things stay the same: - great venue - great fun - great competition - great party There are a couple of changes to this years event: - the cost per team is down to £110 - the party ticket price is up to £4 - the party is at a new venue and will run until 2am - unfortunately we are unable to offer accommodation this year* ** you can currently get rooms at the Travelodge for £19 per night, other options on the website - if you really struggle to get somewhere to stay let me know and we may be able to sort something out, but it's not at all guaranteed I'm afraid (even I will be in a hotel this year!)* It's in Sheffield, at Ponds Forge which is a fab venue. Loads of details on the website. 'Like' www.facebook.com/MixAndMingleTournament or 'Follow' @ninety2ultimate Any questions, get in touch. Si
2bf59792-5772-d4fa-da2b-d1f33cababb1
['793fe0e8-bcc2-b889-5f6b-402c69f62d61']
I _think_ that I have fixed the problem with m0n0wall rebooting itself on a Compaq 1850R. I removed the ethernet card for my LAN and replaced it with a different card (different manufacturer). Now it has stayed up for more than 24 hours, which it never had before. So, I'm hoping that fixed the problem. <PERSON>
f8c41793-3ba3-02a7-6fd6-49324adec4e5
['793fe0e8-bcc2-b889-5f6b-402c69f62d61']
Hello - I recently decided to switch from an IPtables firewall to m0n0wall. I downloaded the current version - pb26r614 and installed it on a Compaq 1850R with 256 meg of RAM and 2 network cards in addition to the built in interface. (3 interfaces total) I tested the new installation for about a week without it being connected to the Internet. All testing worked out fine, and m0n0wall ran like a charm. It stayed up the whole week, with no problems. So, yesterday, I decided to put it into production. Within about 3 hours, it rebooted itself (about 12:30 pm yesterday). Then, this morning at about 8:30, it rebooted itself again. Does anyone know what might be causing this? It stayed up just fine in the testing phase, which was, of course, a very light load. However, even now, in production the load is pretty light. I should add that this is a relatively new machine for me. However, before I installed m0nowall, it had stayed up just fine for weeks running linux, but not acting as a firewall. Do I have some kind of NIC problem? If I start to replace my ethernet cards, is there any way to reload the old ruleset, or do I have to manually enter them again? Thanks, <PERSON>
67319672-3172-0367-196a-e5f2d83fb8a5
['7950743d-e8cb-b83c-6c37-4850809335aa']
Why the heck not... <PERSON> here, 22, Male. Started with a bit of javascript, then bash, then awk, then I changed my major to computer science =). Learnt a lot of Java, some C++, but Ruby was the first language I felt extremely comfortable with. Ruby taught me what it really meant to be comfortable with a language. I had been trying to learn Haskell (a very good thing to study, but less than practical to use, for most people), but I fell in love with Ruby while the Haskell books were in the mail... I'm writing C++ nowadays (for the summer, at least), and it seems that C++ with the Boost libraries and the STL certainly isn't as beautiful as Ruby, but it gets the job done with its own equivalents for ary.each { } and ary.map { }, delete_if, Proc.new, etc. Sometimes I even find strict typing to save me effort in catching mistakes. Work: student / intern, for now. <PERSON>
c565b1b9-aa99-a7f4-0447-ddee692b7786
['7950743d-e8cb-b83c-6c37-4850809335aa']
Hello, everyone. I have just written a tutorial (sort of) on Haskell concurrency, specifically on the derivation of a function mapP, a parallel version of map. I wrote this because I am fairly new to Haskell, and I didn't realize how easy concurrent code is until I wrote this, and because I could not find a tutorial describing a parallel implementation of map. If anybody is curious to see this, the page is http://zwell.net/content/haskell.html#mapP . I would appreciate any criticism, too. I hope somebody finds this helpful. <PERSON>
b29ac77f-1c82-a596-1009-84959f3d32bb
['7962f4e8-2e06-67ff-0438-3b4e16ad7b48']
Hi all! I want to print a hyphen to a pdf() or postscript() device. As the documentaion of postscript says ASCII Character 45("-") is mapped to a minus sign (ASCII Character 95) by default. The advice given is to use "\173" for a hyphen. But, the following code produces a curly brace instead of a hyphen. Thanks for any advice? <PERSON>
4b249623-bb12-88b0-9b18-0e8d74f93d02
['7962f4e8-2e06-67ff-0438-3b4e16ad7b48']
Hi all! To add to my previous posting I want to give some more deatils give a more precise I want to print a hyphen to a pdf() or postscript() device. As the documentaion of postscript says ASCII Character 45("-") is mapped to a minus sign (ASCII Character 95) by default. The advice given is to use "\173" for a hyphen. But, the following code produces a curly brace instead of a hyphen. Thanks for any advice? <PERSON>
472887bd-c4e8-ad61-69fc-7b58df8aeb44
['79690eb2-f75c-e1ad-b78e-195943c13c14']
Hi, ADCON0bits struct def seems to miss CHS bit field which is available for 18F4550 in pic18f4550.h Also CHS0..CHS3 are missing. Further pic18f4550.h defines GO_NOT_DONE, GO, NOT_DONE and DONE fields for ADCON0, last of which seems wrong because the bit meaning is "go/not done" so DONE does not make sense. Where do these definitions come from, do they have Microchip origins or is this just a regular typo/bug in SDCC? Should I file a bug report? I'm using sdcc 3.4.0 wbr Kusti
6e9adacd-3e8d-9a34-5f21-90572f3a3539
['79690eb2-f75c-e1ad-b78e-195943c13c14']
I tried and succeeded in using the jpackage from jdk13 (not really) to package 32 bit java8 for Windows. Everything goes nicely at build time and when I double click the .exe an install is performed, however the resulting installed .exe fails with "Failed to find library C:\Program Files\jDraft\runtime\bin\jli.dll" But that file is there, though it is a bit suspicious that it is not in the Program Files(x86). Should it be possible to embed 32 bit JRE or am I doomed to fail? <PERSON>
b7b1f5ad-01cb-4aed-d698-318158bd5a2a
['7969498f-9a4f-55c2-b9eb-a5b8c08e70ba']
I've written a PHP5 extension in C++ that builds great on Linux with GCC, whether I compile it into PHP or build it as a shared library. On Solaris with Sun CC, however, I've had to manually edit the Makefile or the libtool script in order to successfully build my extension. When I compile the extension into PHP, I have to change $(CC) to $(CXX) in certain places in the Makefile. When I build it as a dynamic library, I have to edit libtool so that it uses $(CXX) and not ld to create the shared library. Am I overlooking some configuration step that eliminates the need to manually edit these files? Thanks, - Mark
4fabe1fc-b991-7346-ca36-4c1660f60545
['7969498f-9a4f-55c2-b9eb-a5b8c08e70ba']
The docs for this class state: "Future instances are created by Executor.submit() and should not be created directly except for testing." https://docs.python.org/3/library/concurrent.futures.html#future-objects We have a need for a thread-safe future type in our extension but this statement makes us hesitate to use it. We don't need the executor functionality. We can write our own future class easily enough, we're just wondering what the justification was for the limitations mentioned in the docs. Thanks, <PERSON>
c3ade9d3-e089-a21f-9495-4a1879a7a849
['79728401-4147-9403-9843-b266b3999556']
Hi <PERSON>, I've been playing around with leftsubnet and leftsubnets to see if either leftsubnet can be used for multiple subnets (it can't) or if leftsubnets can be used for a single subnet (it can with or without the braces). Is there any disadvantage of using leftsubnets for a single subnet apart form it appending an instantiation marker to the conn name? While doing this checking I was using the "ipsec auto --replace" command and I think I have a problem. If you have leftsubnets={subnetA subnetB}, xfrm policies are put in place for both subnets. If you change your file and remove subnetB from leftsubnets and do an "ipsec auto --replace" it leaves the xfrm policy for subnetB in place rather than remove it. Is this the expected behaviour? It is like it reads the updated file and uses this to change the conn but it only changes the bits it sees from the file and not the bit that was removed. Regards, <PERSON>
a6e8a6e4-c794-39ae-0b3e-a6734a9be724
['79728401-4147-9403-9843-b266b3999556']
Hi <PERSON>, In the conn you can use left=%defaultroute which automatically picks up your left IP. There does not seem to be an equivalent in the secrets file or am I missing something? I can use an FQDN or I can set %any to get round it but %any has other side effects like limiting you to one secret across all conns. I found an old thread between us 9 years ago asking the same question and I am wondering if there has been any progress? In that thread it pushed me to %any which I'd rather not do. To me if would be nice if you could also use %defaultroute or something like %myip to automatically pick up the WAN IP. I can also work round it using IKEv2 and a leftid. <PERSON>
beb0a7b6-9823-b4aa-3688-d88637796df2
['797d1d68-ab56-5d00-308a-4a24d5e49e9e']
Hi, I have a few (more) question about various Gecode constraints that I am interfacing to ECLiPSe: 1) the constraint for lexical ordering rel(Home, IntVarArgs, IntRelType, IntVarArgs) requires the two arrays to be the same size. It seems the same restriction applies in the global constraint catalog for the lexical ordering constraints as well. However, the existing ECLiPSe lexical ordering constraints does not have this restriction -- i.e. the two sequences of numbers being compared can be different lengths. What is the reason for the same length restriction? The lexical ordering for sequences of different length used in ECLiPSe seems to be the natural one, where missing numbers in the shorter sequence are replaced by 'bottom', i.e. a value smaller than any number. So (using . to separate numbers in a sequence) 1.3 > 1.2.5 1.2.5 > 1.2 etc. Could such a definition of ordering be used in the Gecode version as well? 2. What IntConLevel is supported in the binpacking constraint? I can't find a specification for this in the reference documentation (in doc-latest) [beyond the default of ICL_DEF] 3. For various versions of the same constraint (e.g. cumulative), some arguments can be IntArgs or IntVarArgs. In general, does it make a difference if I always use the version with IntVarArgs, even if all the IntVar has only one element in their domain (i.e. they are integers)? I guess what I am asking is if (in general) Gecode does a check of its IntVarArgs argument to see if they are all integers and calls a more efficient implementation for the constraint if there is one? [I was specifically looking at the cumulative constraint, but there are obviously others]. The reaosn I ask is that in interfacing such constraints to ECLiPSe, I currently don't check if such arguments are all integers or if they contain some domain variables (IntVar). I can do this check, but if it is done in Gecode already, there is no reason for me to do so. Thanks in advance for any help and information! Cheers, <PERSON>
5e6056b4-68fd-41bc-2660-73194e940ee5
['797d1d68-ab56-5d00-308a-4a24d5e49e9e']
Hi, I am looking at using activity (IntActivity) when the search is done outside of Gecode to ECLiPSe's Gecode interface. I have added support for IntActivity when the search is done in Gecode, i.e. using a search engine, but I had thought I can't support IntActivity outside because IntActivity is declared with a fixed set of variables, while ECLiPSe can dynamically add variables to the problem. However, the ECLiPSe interface does provide a predicate (procedure) that perform variable selection on a fixed set of variables, so I think IntActivity can be added as a selection method here -- an IntActivity can be declared for the set of variables being selected when the predicate is first called. Does this IntActiviity have to be added to my space class? The activity applies globally, and is not specific to a space, but the IntActivity is declared with space as an argument along with variables for that space. If it needs to be added to the space (with a IntActivity* in the space, set to NULL if IntActivity is not used, I assume?) How should it be copied when the space is cloned? Does this look like the right way to do it? Thanks in advance! <PERSON>
e99b142d-f112-bf9c-d12e-c7bebf078dd0
['7997c89a-74fa-ff70-f3e8-3e90258f11ed']
I have the desire to utilize my slave mongod's for some long-running queries (read-only). I have a sharded environment, so the only way I see to do this would be to have a config + mongos setup dedicated to the slave mongod instances. Is that kosher? Is there another way to query replica sets globally? thanks, <PERSON>
a583ab30-9cbb-d4c4-1a03-6c60a596db9a
['7997c89a-74fa-ff70-f3e8-3e90258f11ed']
Hi, The sharding docs at http://www.mongodb.org/display/DOCS/Configuring+Sharding#ConfiguringSharding-3.ShardingaCollection say this: "One note: a sharded collection can have only one unique index, which must exist on the shard key. No other unique indexes can exist on the collection." The _id field is a unique index. Does the statement above exclude the _id field? If it doesn't exclude the _id field, the doc seems to imply that you can only shard on the _id field but from practice, I know this not to be the case. Am I reading this wrong? thanks, <PERSON>
7afdc810-cc94-3dec-7d40-c221ea99419a
['799b0acf-2e53-3175-b1b1-e6eef0d5d72f']
Hi <PERSON>, Does re-installation of your product involve service re-registration? If yes, and the service restarts correctly immediately post-install, I'd say you are looking at a problem with the service entry. DEP hasn't been an issue for 4D since v2004, so I'd be surprised if that is the answer (although we're still deploying v11, so maybe it's back with v12....). Are you using local system account to log on to Windows, or have you specified a different account? Are you allowing the service to interact with desktop? Will 4D Server open your database correctly as an application via file > open recent? If you've never dealt with 4D Server running as a service in Windows 2008 you won't be familiar with the interactive services dialog detection service. This allows you to see the 4D Server UI when running as a service, as long as the 4D Server service is using the local system account to log on, and the option to allow it to interact with desktop is checked. It is a standard Windows service, make sure it is started before you try to start your 4D Server service. It might let you see what is going wrong with 4D Server. <PERSON>
4c9d8d5e-a8e1-ac79-3017-4577d1c4200b
['799b0acf-2e53-3175-b1b1-e6eef0d5d72f']
Hello all, Has anyone else encountered problems with the backup in v11.8? I'm seeing several sites' backups fail. On investigation the backup settings seem correct, but trying to take an immediate backup fails with the message that a backup cannot be started, because a backup is already in progress. We have started using journal files as part of our standard deployment, whereas we never have done so in the past. Our standard deployment methodology includes creation of a new backup.xml - we do not continue with the previous settings from 4D 2004. This is proving to be a big time-waster for the support team, so any advice will be gratefully accepted. <PERSON>
75701168-6394-cf2d-4cad-02f2873d0c63
['799d4808-4527-7c39-c88b-4e84645dba61']
Hi, My setup includes: * One Openstack Controller with ODL (of course with SFC) which is deployed through OPNFV Colorado 3.0 * One aarch64 Compute Node (OpenStack Mitaka), which is attached to the above OS Controller I downloaded OVS 2.6.1 from OVS git hub and applied NSH patches from below URL: https://github.com/yyang13/ovs_nsh_patches/tree/master/v2.6.1 I am able to create the SFC attributes like VNFD, VNF, Chain, and Classifier through Tim Rozet SFC walkthrough (https://github.com/trozet/sfc-random/blob/master/tacker_sfc_apex_walkthrough.txt). And also, I launched VNF, http_server and http_client VM's on the same compute node. All the related NSH flows are added to the bridge 'br-int' in the compute node without fail. But, when I execute the command 'curl --local-port 2000 123.123.123.4' from the http_client VM, I am getting the below error message: curl: (7) Failed to connect to 123.123.123.4 port 80: Connection timed out When I execute the command 'wget 123.123.123.4' from the http_client VM, I am getting the 200 OK response. But, in both the cases, no VxLAN packets are coming at the tacker VNF VM (using vxlan_tool.py). OVS bridges and flows details of Compute Node are given at the below link: http://pastebin.com/MEN7dk8n Can anyone give me some clue, to debug the issue. I also want to know, whether anyone one are succeeded while executing SFC across compute node through ODL and OVS 2.6.1 with NSH patches from <PERSON>. Thanks for the help. Regards, <PERSON><IP_ADDRESS>' from the http_client VM, I am getting the below error message: curl: (7) Failed to connect to <IP_ADDRESS> port 80: Connection timed out When I execute the command 'wget <IP_ADDRESS><PHONE_NUMBER>' from the http_client VM, I am getting the below error message: curl: (7) Failed to connect to <PHONE_NUMBER> port 80: Connection timed out When I execute the command 'wget <PHONE_NUMBER>' from the http_client VM, I am getting the 200 OK response. But, in both the cases, no VxLAN packets are coming at the tacker VNF VM (using vxlan_tool.py). OVS bridges and flows details of Compute Node are given at the below link: http://pastebin.com/MEN7dk8n Can anyone give me some clue, to debug the issue. I also want to know, whether anyone one are succeeded while executing SFC across compute node through ODL and OVS 2.6.1 with NSH patches from Yang. Thanks for the help. Regards, Srikanth.
4c153434-9d72-4547-6e23-986656837314
['799d4808-4527-7c39-c88b-4e84645dba61']
Hi, I am working on ODL SFC. I am able to execute the basic SFC usecase. But, I just want to understand, what vxlan_tool.py does in SFC vNF. Is it forward packets to next vNF based on NSH header? Can anyone please let me know, what are the specific tasks, which are handled by vxlan_tool.py? Regards, <PERSON>.
f6e25a99-c84a-5e9f-8bc4-fbddeb434fff
['79ad8d51-9bcc-53d5-e04f-bb017139f1dc']
*TLDR: There is a new chromium.android waterfall.* Hi chromium-dev! We're creating a new chromium.android waterfall! This waterfall will eventually contain all of the public android devices and configurations. Right now, we're moving over the bots from chromium.fyi; we've already moved over some, and more will be moved over soon. These bots will also eventually have CQ coverage, and will be sheriffed by the android sheriff. As of now, no action is needed. They will potentially be tree-closers and/or on the CQ, depending on how useful this would be, how important the configuration is, how much it breaks, etc... For more information, please feel free to shoot me an email. Thanks,
a12b58db-0948-587f-54ad-9990a34ccd03
['79ad8d51-9bcc-53d5-e04f-bb017139f1dc']
If you don’t care about layout test results, you can stop reading now. Hi, As part of my work to improve CQ speed and reliability, I noticed that archiving layout test results on Windows takes ~10 minutes, when it should take less than a minute (tracking bug here <https://crbug.com/792737>). The easiest solution to this problem is to remove the recursive upload of the entire results directory. This would break the existing test results “Show results” functionality (screenshot of this attached). This functionality can be fixed, but it’s unclear how widely used this is; if you have used the “Show Results” functionality recently (past few months), please let me know. You can reply all to this email, to me directly, or post on the bug. I’m also planning on removing the “latest” result upload for tryservers. This is used on the waterfall to compare a particular run of a test to the “latest” result for the bot. This doesn’t make sense for tryservers, since the “latest” result is roughly meaningless. If the above paragraph’s plan of removing recursive upload is implemented, this probably won’t be done. If you have any comments or issues with this, feel free to contact me. Thanks, <PERSON>
0100c481-5028-9653-b505-510fd7073bc0
['79afc9f6-c5d2-1aa8-b4e8-0cd7dea00d5f']
Now that the X100P is no longer being offered by Digium, what is the best solution? I seem to have run into a few posts where people talk about problems they've had with their X100P clone cards (dropping calls, echos, etc) other people seem to not have any problems. Of the three chipsets that will work: Intel 537EP, Ambient MD3200, and Motorola 62802 (as seen here http://www.voip-info.org/tiki-index.php?page=X100P+clone ) Is one of them more stable than another? What modem cards have you had the best luck with? Are there any ones to stay away from? I'll be using this for home use, so absolute reliability is not necissary. As a result I'd like to stay ~20 or less, and get the best quality I can for this price range. Thanks! <PERSON>
d6c7cdf0-51dd-e34d-9cf8-426b9e40249c
['79afc9f6-c5d2-1aa8-b4e8-0cd7dea00d5f']
I ran into this on the net today, and thought it could have some interesting implications for Asterisks/Vonage integration. It seems integrating this with Vonage/Asterisks server could allow you to have Vonage call you, give you a dial-tone, and call out a different line. (Or other such uses). I thought it might be useful to post here, so people could use their imaginiations and perhaps come up with better possible uses of this. Feature Described Here http://www.vonage.com/features.php?feature=click_2_call Open Source Firefox Extension That Uses Click-2-Call http://mcgrof.com/firefox/c2c/ Doug Logan
aae65868-e366-d694-7314-4669ae3b4ee6
['79b885ba-f336-3ee4-6673-4bcf28572406']
Hi I'm trying to get a tree control to expand when I left click on the label and not just on the +/- button. The expansion using the button is handled in the default right down event of the tree control and I am toggling the state in the left up. This leads to the desired behaviour when the label is clicked but not so when the button is clicked. To avoid this I was going to use the flags return from HitTest but the flags returned on the button is 4104, an integer, and wxTREE_HITTEST_ONITEMBUTTON is 8. I'm using wxPython 2.4.0.2 on Redhat 8.0. <PERSON><IP_ADDRESS> on Redhat 8.0. Nigel
b9f61237-0be8-7dee-98b9-92e2a2f6956e
['79b885ba-f336-3ee4-6673-4bcf28572406']
Folks I'm using bitmaps with masks so I can draw bitmaps on other bitmaps and see the colour through the overwritten bitmaps. My problem arises when I try to change the top bitmap. If I draw a line, the mask stops it being shown. What I would like is to get the mask bitmap and draw the same things on it in monochrome as the bitmap so it will show. Can I do this? wxMask doesn't have a GetBitmap function documentated. Should I do this some other way? The working with images wiki was helpful but didn't address with issue. <PERSON>
f332b6cf-5d13-5889-0f56-66fd9f940b18
['79bad54a-90e7-fae5-1bf1-34857b1afe95']
I am confused as to why backports.org exists. If I want to use the latest software, should I use backports.org or I should upgrade to etch (I am currently using sarge). What happens if I use backports for some time and then decide to upgrade to etch. Will I end up with a broken system? What is the philosophy of backports? Thanks, <PERSON>
6a9c03e0-c17e-d6d2-3f5c-72bd062f2202
['79bad54a-90e7-fae5-1bf1-34857b1afe95']
I'm using Debian Sarge (stable) and searched for a mono package to install: apt-cache search mono. I didn't get any packages related to the mono project back. How can I install mono and run .NET applications on my debian box? I searched the internet and found the backports.org website. However I don't quite understand how it functions. There are lots of packages under the mono directory ( http://www.backports.org/debian/pool/main/m/mono/), but I don't know which of them to install and, moreover, how to install them. They are not visible with the apt-get tool. -Martin
c4ca4b5c-b6ed-1022-ca54-bdb3ca4c791e
['79be3164-cd89-7f82-a7e0-8997d391e02f']
After lurking here for a while, I finally took the plunge this weekend and unslung my Slug. Great directions, went through them step by step and everything worked -- thanks everybody for your hard work! Once I got SSH up and running, I installed first appWeb (couldn't get to load), php-tthppd (got working) and Apache with PHP configured as a module (php-apache). It was exciting, but when I tried to load two of my favorite applications, the RSS reader Gregarius (http://gregarius.net/) and the streaming music server (http://streamline.sourceforge.net/). I couldn't get either of them working with php-tthppd, seemingly due to permission problems. I'm a little more familiar with Apache, and it turns out that to get Gregarius working requires that the expat xml parser be compiled in as part of the PHP module, and Streamline requires that iconv be compiled in, and it seems that neither of them is in the php-apache package. Is there any chance of getting a version of PHP with iconv and expat in them? Or does someone have a suggestion for a PHP based non-itunes streaming music server they have gotten working on their Slug? Thanks in advance, <PERSON>
d6eceedc-bb3b-f859-1606-5b4714e8699a
['79be3164-cd89-7f82-a7e0-8997d391e02f']
I recently found ECCO and discovered that unlike the other programs in this category that I have tried, ECCO works the way I do. I love it. I have run into one problem that I am hoping to get help on. When I first setup synchronization with my Palm, I set it up to synchronize Calendars and Contacts only. Now I would like to add To Do's, but when I change the settings in Ecco (Tools->Pilot Synchronization Options->, check To do, To-Do's tab, tried both normal and full synch, it doesn't seem to change what happens. I've rummaged around the documentation, but haven't found anything. If someone knows the answer, or could point me to appropriate documentation, I would be grateful. Thanks in advance, <PERSON>
233bc61b-2257-2dae-a717-31fd54193e60
['79bec30d-a1f3-aedc-f61e-aa4869062262']
Hi, I've created a cucumber test to test a metal I've created that connects to a third party. I'm using rspec mocks to create a stub for the 3 party. It seems that stubbing/mocking isn't extending all the way to the metal. It calls the read object and not the stub. Any idea of how I can make stubbing/mocking extend to the rails metal? I'm stubbing in a cuc step. Thanks.
0b936535-0e76-2201-50ef-e9cc104a00cd
['79bec30d-a1f3-aedc-f61e-aa4869062262']
Hi, I've created a cucumber test to test a metal I've created that connects to a third party. I'm using rspec mocks to create a stub for the 3 party. It seems that stubbing/mocking isn't extending all the way to the metal. It calls the read object and not the stub. Any idea of how I can make stubbing/mocking extend to the rails metal? Thanks.
80b99386-54b3-8392-cd15-e386c7fc4110
['79c957fa-6c72-05cf-b2b9-c4577491a073']
Dear all, For the continuation of our websites and (plone) multi-site-CMS we want to make an analyses of our system. Who knows an independent professional with experience in Plone, Django and Drupal, who can make an evaluation of our system and advice us in our decision for continuing in plone and upgrading the system or switching to drupal or django, taking in consideration our wishes? Thanks, <PERSON>
3e5b1866-e7ef-b78a-9786-a03db046fbc4
['79c957fa-6c72-05cf-b2b9-c4577491a073']
Dear people, I would like to introduce myself. My name is <PERSON>, I work for SOMO. We have several websites (five) with a different CMS and at the moment I am starting a project to convert all this sites to Plone. I am asking some proposals to companies who work with Plone in the Netherlands. Also <PERSON> is giving me advise. Because we have several websites with a lot of shared information, like news and agenda, I am interested in using Multiplone for this project. Now I am very interested in your experience with this. My main questions are: 1. One company said it is a bit tricky to use Multiplone, because it is solely designed for Oxfam, so to use it for another project would ask quit some adoptions. What is your opinion/experience on this? 2. The conclusion of one of the companies is that it would be so much easier and less work to make five separated websites then to make this five sites connected by Multiplone, therefore using Multiplone for this project would cost a lot more money. What is your opinion on this? Furthermore I would like to hear more about your experiences and maybe some tips & tricks on this project. <PERSON>, maybe we can meet? kind regards, <PERSON>
f6a2a756-991b-061e-8834-4318fb15654a
['79c9627a-a377-4f94-a90d-55aa9c430581']
Hi, I am using the community version of mysql workbench, tried both 5.2 an 6.0 version. When I connect to a database, trying to get ER diagram through the Reverse Engineering module, I am not getting the connections (or relationships).  The foreign key relationships were not shown in the diagram.   I am not sure this is due to that I am not paying for the product or it is a bug or feature not implemented yet. I am connecting to a mysql database, using mysql syntax (SQL).  When I use a commercial product, I was able to generate a full ER diagram. Anyone has similar experience? <PERSON>
7952b4eb-a3d8-a1fe-beb1-be538741ac02
['79c9627a-a377-4f94-a90d-55aa9c430581']
Hi All, Not sure I am reaching the right group or not. Here I am having a tough problem with gobject-introspection.  I have at least build > 2000 packages in my career as a programmer.  This is the first time, it failed at the 'make install' stage.  Here is the background: My machine (brand new) has RedHat 6.4, I realized this is very old, so I have installed a few hundres packages in the /usr/local/ including mostly system related (the gnu 4.8.1 compiler included).  So far so good. My pkgconfig path points to /usr/local/lib/pkgconfig I have tried with gobject-introspection-1.37.4, and 10 versions down.  They all have exactly the same problem: setting the include directory (glib-2.0) directory to the old /usr/ directory instead of the /usr/local I saw several other people had the same problem, but have not seen any solutions.  If anyone has a hit please let me know. Thanks. <PERSON>
5acebf1a-f3cb-c169-f302-99f49d68a8e5
['79cf3b55-26b5-d3f3-02af-61c3cc40de08']
Is there a way to control an array resync process? In particular, is it possible to skip read errors? My setup: LVM2 Phisical Volume over a two disks MD RAID1 array Logical Volumes didn't span whole PV, some PE free at the end of disks What happened: disk1 broke I installed new disk1 and started sync from disk2 to disk1 but at 99.9% disk2 gave some read errors and the sync process started again, over and over I didn't notice errors on disk2 because they were in unallocated PEs at the end of the disk. The MD device spans the whole disk, while the LVs don't. I'd like to complete sync ignoring read errors, then replace disk2. I think this is a not-so-uncommon situation, leaving some PEs free for future expansion is a good idea and errors go undetected until you "use" those free areas. Thanks.
6c31d0c0-5242-c115-2ef6-796090f82059
['79cf3b55-26b5-d3f3-02af-61c3cc40de08']
Hi, I'm using shorewall (4.5.18) and lsm (0.163) with a two ISP setup. I followed documentation and the linuxfest presentation (all provider balance), but choose to ping remote ip instead of the local gateway. lsm is started as a service, not by shorewall. The setup is working, but I'm not sure on what to do when lsm detects a link down event. I tried shorewall disable ispX, but it deletes the routing rules, so the link cannot come back. I could adjust the mangle file and restart shorewall: would it be a good idea? Any other suggested option? I choose not to ping the connection gateway because both gateway are local and never go down, while especially one connection (wimax) goes down once in a while and I can detect status pinging a remote ip. Thanks in advance.
18e85af9-6ff4-d596-038c-67897383a5ac
['79d15c45-4f01-fe92-1769-845560a32510']
Dear all, Thanks to my co-authors and all the people involved throughout the years of aerial monitoring of the seals in the Wadden Sea, I am happy to announce our recent publication in the open-access journal PLoS ONE Echoes from the past: Regional variations in recovery within a harbour seal population <PERSON> , <PERSON>, <PERSON>, <PERSON>, <PERSON>, <PERSON>, <PERSON>, <PERSON>, <PERSON>, <PERSON> Published: January 3, 2018https://doi.org/10.1371/journal.pone.0189674 Read the full publication via: http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0189674 Abstract Terrestrial and marine wildlife populations have been severely reduced by hunting, fishing and habitat destruction, especially in the last centuries. Although management regulations have led to the recovery of some populations, the underlying processes are not always well understood. This study uses a 40-year time series of counts of harbour seals (Phoca vitulina) in the Wadden Sea to study these processes, and demonstrates the influence of historical regional differences in management regimes on the recovery of this population. While the Wadden Sea is considered one ecologically coupled zone, with a distinct harbour seal population, the area is divided into four geo-political regions i.e. the Netherlands, Lower Saxony including Hamburg, Schleswig-Holstein and Denmark. Gradually, seal hunting was banned between 1962 and 1977 in the different regions. Counts of moulting harbour seals and pup counts, obtained during aerial surveys between 1974 and 2014, show a population growth from approximately 4500 to 39,000 individuals. Population growth models were developed to assess if population growth differed between regions, taking into account two Phocine Distemper Virus (PDV) epizootics, in 1988 and 2002 which seriously affected the population. After a slow start prior to the first epizootic, the overall population grew exponentially at rates close to assumed maximum rates of increase in a harbour seal population. Recently, growth slowed down, potentially indicative of approaching carrying capacity. Regional differences in growth rates were demonstrated, with the highest recovery in Netherlands after the first PDV epizootic (i.e. 17.9%), suggesting that growth was fuelled by migration from the other regions, where growth remained at or below the intrinsic growth rate (13%). The seals' distribution changed, and although the proportion of seals counted in the German regions declined, they remained by far the most important pupping region, with approximately 70% of all pups being born there. It is hypothesised that differences in hunting regime, preceding the protection in the 1960's and 1970's, created unbalance in the distribution of breeding females throughout the Wadden Sea, which prevailed for decades. Breeding site fidelity promoted the growth in pup numbers at less affected breeding sites, while recolonisation of new breeding areas would be suppressed by the philopatry displayed by the animals born there. This study shows that for long-lived species, variable management regimes in this case hunting regulations, across a species' range can drive population dynamics for several generations. Best regards, <PERSON>
715d9b5f-39a4-8a0a-1834-8a608f1a0ff9
['79d15c45-4f01-fe92-1769-845560a32510']
We are pleased to inform you that the following paper has been published: <PERSON>, J. P. A., <PERSON>, J., <PERSON>, K., <PERSON>, J., <PERSON>, S., <PERSON>, <PERSON> and <PERSON>, G. (2015), Estimating the spatial position of marine mammals based on digital camera recordings. Ecology and Evolution. doi: 10.1002/ece3.1353
b5549fdd-5331-b043-8a33-c731dd01e90f
['79d47130-90ec-8763-7ca8-46e599abd0d2']
Hi, <PERSON> and Kmail, walking hand in hand, are giving me itches yet again. So I dug around and came across some tutorials and hints for PostgreSQL. Unfortunately I can't find a package for PgSQL as a backend to Akonadi while other distributions have it. I already asked on /r/openSUSE but without any luck. Any hints how and where to start with Tumbleweed? Regards, vinz.
715006e1-4e4a-0ed6-c1d8-0221aa0cd3f6
['79d47130-90ec-8763-7ca8-46e599abd0d2']
Hi there, I recently went through many parts of Xfce's i18n infrastructure and documentation. To make it short: lot's of old stuff, many things unmaintained but truly a lot of useful gems grown throughout the years. So following up is my proposal for reorganizing i18n in Xfce, split into two phases. I put it up on this list to reach more people. A. Short term 1. Dump mailing lists. One list (xfce-i18n) is enough and for language-specific questions, discussions a short [LANG] tag added to the email topic should do the distinction job. Bonus: questions might be answered quicker as more people read them. 2. Before dumping email all lists subscribers directly and (of course) telling them about. Plus asking them if they are on this list alread and, which is even more important, if they would like to help out on Transifex. 3. Transifex needs coordinators for every language. If there is no one coordinating a language there should be some "super-coordinators" as backup. These persons shouldn't be devs or at least not core devs. Stacking work leads to problems. B. Long term (mostly RFC) 1. Ease up the introduction process. Someone dropping by at Transifex sees the need of some translation work in his language. If he decides to register at TX, to join team *and* do work: yeah, cool! But forcing him to additionally join the mailing list and introduce himself is a major blocking thing. I guess it's something from the pre-Transifex days where quality assurance had to be done manually. But with the split of translators & reviewers over TX this should be obsolete. 2. Wiki needs clean up. There are lots of really, really useful things. But as well there is lots of ancient, outdated stuff. Unfortunately most stuff is mixed so there needs to be manual work done. TBD: a process of cleaning up, either dumping everything and rebuilding or the other way around. 3. Revise the use of Transifex. It's not self hosted and it's proprietary but it does the job. There are up- and downsides of it so in the light of the Git hosting discussion TX should also be reconsidered *openly*. Generally speaking I'd step up for both doing "dirty jobs" as well as coordination work. Thanks for your time reading this! Looking forward to your comments, <PERSON>.
29c83af6-3224-abeb-8455-72681100f395
['79d92539-9b4e-25c5-ea7b-400501cdffa9']
Hi All, I have some problem when signing app, i'm very new to things like public key, private key, etc.... I try to build my project using ant, but always failed in signing part (debug.keystore), when i try to build using 'ant release' command, it was successfull but we still need jarsigner to sign our app. Can anyone give some simple example from start to finish, the signing process. I really appreciate that. Thanks.
d3f7c00e-44f8-2ac7-177f-16746f6acd26
['79d92539-9b4e-25c5-ea7b-400501cdffa9']
Hi all, This is may be very stupid question, but how do I close an running android app e.g: browser. When I press the back button it seems application only minimized not closed. So I try to list running app by holding Home button and browser app still running. I'm running android on emulator. So, how do I close an App? I didn't find any close button :( Thanks.
4ea2adec-0fef-51ab-4e40-d526b9af3249
['79dc62b8-ef4a-4881-78ed-6fadd6938680']
Hi, I am using the install guide from the confluence site, and have created the makefile. When running make, I get the following error: error: The requested URL returned error: 400 Bad Request while accessing https://stash.freeswitch.org/scm/fs/freeswitch.git/info/refs Any idea why this would be happening? I have installed using this method before and it worked perfectly thanks <PERSON>
41603629-ed3c-4baa-abea-55fb3204a495
['79dc62b8-ef4a-4881-78ed-6fadd6938680']
Hi, is it possible to change the location of the scripts directory? I want to have one central location for multiple Freeswitch instances to read the javascript files from but have hit a snag. I have added the central location to the dialplan file when calling the javascript application e.g <action application="javascript" data="/central/freeswtich/scripts/test.js" /> This loads the initial file just fine, but any included files (include("includes/file.js")) are still trying to resolve back to /usr/local/freeswitch/scripts/includes/ directory. I have checked the java.conf.xml and spidermonkey.conf.xml files but these do not seem to hold the answer. I would prefer not to have to "hardcode" the scripts and include locations if it can be avoided. Any help would be appreciated. thanks <PERSON>
ea04cd0f-deba-629e-e1cf-b41153298d6e
['79e03378-4661-6993-e073-f982675c36d5']
Hello, I have just discovered jQuery and it looks cool. I used the jQuery UI builder at ui.jquery.com to build the UI and it looks and works great :) Unfortunately I do not know enough about javascript to answer my own question, so I am deferring to you :) I would like to have a tab with a dialog button on it, and when that dialog comes up there will be a form. This part I can do easily enough and have already implemented, however, what I need to do after that is where I am stuck :( I want the user to fill out the form, say there are fields like "Customer Name" "Telephone Number" and "D.O.B." and once that form is filled out, and the user clicks OK, I want it to do the proper inserts, etc into my mysql database. This is probably a very easy question, but I have not found out anything useful so far through various google searches. I would like to do the bulk of my mysql stuff via perl/mason. Is there a tutorial anywhere for this or perhaps I should be using a different technique? I am trying to avoid PHP simply because I would prefer to not have to learn PHP while I am in the process of learning the things I need for jQuery, but I can if I need to. Any advice/help/examples would be greatly appreciated., Greg Evans
00f1ba06-4f49-6524-d693-658ef2ec9eab
['79e03378-4661-6993-e073-f982675c36d5']
Hello, I have just discovered jQuery and it looks cool. I used the jQuery UI builder at ui.jquery.com to build the UI and it looks and works great :) Unfortunately I do not know enough about javascript to answer my own question, so I am deferring to you :) I would like to have a tab with a dialog button on it, and when that dialog comes up there will be a form. This part I can do easily enough and have already implemented, however, what I need to do after that is where I am stuck :( I want the user to fill out the form, say there are fields like "Customer Name" "Telephone Number" and "D.O.B." and once that form is filled out, and the user clicks OK, I want it to do the proper inserts, etc into my mysql database. This is probably a very easy question, but I have not found out anything useful so far through various google searches. I would like to do the bulk of my mysql stuff via perl/mason. Is there a tutorial anywhere for this or perhaps I should be using a different technique? I am trying to avoid PHP simply because I would prefer to not have to learn PHP while I am in the process of learning the things I need for jQuery, but I can if I need to. Any advice/help/examples would be greatly appreciated., <PERSON>
f15d5327-cd80-dcc7-0ce5-926d7beb0687
['79e6dd15-9ae7-c09a-5d9d-5a25d18f609f']
Hi, The dynamic library(say x.so) I'm loading has a function( foo() ) defined as a weak symbol __attribute__((weak)). When x.so is used from a C program, foo() will be defined in the C code and when linked with x.so, foo() will override the weak foo() from x.so. How can I achieve the same when I dynamically load x.so using ctypes. I want to be able to define foo() in python, and after loading x.so through ctypes, somehow add the new foo() to the symbol table, so that it can override the weak foo() that was brought in after loading x.so. Is this possible with current ctypes? Any pointers on what it would take to implement this ehancement? thanks <PERSON>
e4c998b1-0e78-fde2-85e3-f9748bd7cade
['79e6dd15-9ae7-c09a-5d9d-5a25d18f609f']
Hi, To maintain backward compatibility with released .proto definitions, there are certain rules that needs to be followed while extending them(they are well documented). I am wondering if there are existing tools that enforce these rules by comparing the released .proto files. May be the protoc compiler itself can be enhanced for this, by taking input the previous version(s) of the .proto file. Does anybody know of such tools? thanks <PERSON>
a3dffb71-f8a8-85cf-229f-44ff0e6c8e17
['79e6f783-0e36-dda5-3207-cbc97e0d7e93']
If there is any uncertainty regarding if having a fair use quotes is in conflict with GFDL for us, or for others, one safe way, it seems, is to declare something like the following: All Wikipedia contents are available under the terms of the GNU Free Documentation License. This includes all the texts and images except for the following. - Images hosted in other servers than wikipedia's own. They may appear as if part of Wikipedia contents, but they are not. Therefore, you should contact the owner of the image regarding its copyright status. - Images, texts, and other contents hosted in Wikipedia server(s) under fair use or similar justifications. They are not part of Wikipedia's contents that are released under GFDL. Your use of these materials based on the terms of GFDL could result in copyright violation. I think this kind of message could be placed on Wikipedia:Copyright , at the bottom of each page, and the download page for Wikipedia's database dumps. Any opinions? Regards, <PERSON>
8c8cef07-6c86-88a0-75e3-da18976ade97
['79e6f783-0e36-dda5-3207-cbc97e0d7e93']
If my 2cents count here, <PERSON>'s position sounds quite reasonable to me, too. If <PERSON> is the one taking the legal risk, and if he honestly thinks that it falls under fair use, then I think we should let him work on the list. At Japanese wikipedia, we sometimes talk about taking legal risks. Current understanding is that if administrators are too cautious taking risks (or too bold) and other users do not like it, the users recall the admins, and find substitute who can take reasonable risk. This idea is again, because admins are far more likely to be sued for knowingly keeping suspected copyvio, obscene, defamatory, or other allegedly illegal material available - more so than other users. We take admins' opinions a bit more seriously when it comes to risk. But admins have to serve the users, so if they think their idea of reasonable risk is too far away from other users', then they should resign. I think there is little dispute regarding Jimbo's good leadership and contribution, and there is no need for recall or resignation. So we can support his move. If Columbia comes to notify us, then we would discuss again. Regards, Tomos
8661439a-41ab-d4c2-bdd2-96fe55ecb2dc
['79f5c299-9873-e135-d5fb-58db2531708c']
I have recently downloaded a vbscript "plugin" called "simple COM-wrapper for Terminal Services Manager". You can find it at http://kentldyer.com/index.htm under "KIX scripts", look for wtsmanager.zip. After registering the supplied DLL on a terminal server, it allows me to list any and all information about all sessions on that server with a few lines of VBScript, such as the session status and, what I actually wanted to use it for, the IP-address and netbios name of the client computer. On Win98/2k/XP workstations, this works like a charm. I can now retrieve the current workstation name and ip-address, instead of just the workstation that was used during the initial login. However, when I run the script when logged in using a linux client (specs are below), the script reports <IP_ADDRESS> as the IP-address. The netbios-name is detected correctly (samba is installed for cdrom+floppy access). Although it's not a problem for me (I can easily find the IP-address using the DNS-entry for the netbios-name), I think that rdesktop is not reporting the correct IP-address. If I connect to the same session using a Windows client, then the script reports the correct address again (the name and ip-address of the windows client). Two servers are running using NLB. Connecting to the shared ip-address or the private ip-address of one of the servers gives, as far as I have tested, the same result. The linux client is a member of the same domain the Terminal Server is in. I have tested this on multiple (identical) linux clients with the same results. Did I forget or misconfigure something? <PERSON>
611b507e-8fd3-fb7b-25ac-3c69ff19c1e3
['79f5c299-9873-e135-d5fb-58db2531708c']
Hello ha-users, I am currently trying to configure two servers with: mdadm <-- drbd <-- lvm2 <-- iscsi-target (ietd). The final goal would be to provide disk space via iscsi to several servers running vmware where each of those vmware servers would store its VM images on the iscsi mounted partition. This way, if a vmware server dies I can easily bring it back online on another server... a poor-man's substitute for VMWare ESX. I will configure both nodes to share an ip-address using LVS to avoid having to reconfigure things when the primary node goes down. I am uncertain about the following: Each node will sit in a different server room next to a vmware server. If the link between the server rooms goes down, both HA-nodes will think that they are the primary server and will take over servicing the vmware server next to them. No problems here because all servers will simply stay operational. However, I dont know what will happen when the link goes down, both nodes start servicing requests on the shared IP, and then the link comes back online a few minutes later. I certainly don't want vmware servers on either side to connect to the HA-node with the "old" copy, but I can't see how I can prevent that as both nodes must operate using the same ip-address. I would like to shoot down one of the nodes, but in this situation I can't, as both HA-nodes have a chunk of "live" data which is being used by one or more of the vmware servers. I can probably prevent this from happening by using crossover UTP or serial connection between both servers for heartbeat, but in case it does happen I would like to know how disastrous things are. I have read a whitepaper (http://www.drbd.org/fileadmin/drbd/publications/drbd8_wpnr.pdf) which (if I understood correctly) states on page 9 that DRBD can automatically resync after a split-brain scenario. However, it seems like DRBD can only sync from node A to node B or from B to A and not both ways (which is what I want). After all, since each partition is used by only one vmware server, I know for sure that if a chunk of data on node A is modified during the split-brain, it cannot be that it was also modified on the other node. Theoretically :). So all in all, I have the issue of two nodes using the same IP address, and I have two nodes both of which have a bucket of blocks that need to by synchronized with the other node (these are the blocks that have been written to during the split brain). Judging from the whitepaper and the various docs on the web, it looks like DRBD cannot handle this situation (using v0.7, v0.8 does not seem to have these features either judging from http://svn.drbd.org/drbd/trunk/ROADMAP). Hence my question is... is this a viable setup? The alternative would be something like mdadm <-- lvm2 <-- drbd <-- iscsi-target (ietd) where I would have to create a drbd device on top of every lvm2 partition so I can tell drbd to resync a partition based on which node the vmware server has connected to during the split-brain. I suppose this would solve the issue, but it seems this would make the whole setup alot more complex and I have no idea what to expect performance-wise when running multiple DRBD instances on the same set of disks... I hope this is the right place to post this question, many thanks in advance for your time. Regards, <PERSON>
d8b7afc9-f1d8-035b-b39f-f297b8b6add2
['7a0d7a4e-cded-980d-6759-e7ea1c6fccd8']
Hi, I'm trying to install snare from source code ( a requirement). When I unpackage and compile the snare-core package ( 0.9.7) i do not get source code for auditmodule.c as is called out in the HOW-TO documentation link. I do get an auditd but when I try to run it it states thate there is no proc/snare running. The documentation at thehome of the project is, in my opinion, sorely lacking, so I am contacting the group to see is anyone has successfully installed the package from scratch. Thanks
2425baf6-c175-e59e-b111-49afbddf71c3
['7a0d7a4e-cded-980d-6759-e7ea1c6fccd8']
Can someone tell me how I can generate a chronological listing between 2 dates of all the issues I have designated to watch in the audit capabilities of the system? I need to get a system with audit capability sufficiently palatable to the Defense Department for classified use. I currently use Sun workstations and generate the audit logs via praudit and was looking for some way to set up simaller archived weekly audit logs. Thanks,
5afb276f-6481-1d55-e057-cedfcf613cf3
['7a23630e-b0b3-c59b-c7d4-2c1744a0a267']
Hi, 1. xft support I would like to use the new release to build tclkit. What I´m missing is xft-support in the tkConfig.sh file (XFT_LIBS, XFT_FLAGS) and/or the related parts in the TEA config. 2. Win builds Using the installed "tclPort.h" file failed at line 25: # include "../win/tclWinPort.h" After removing the "../win" part it works. <PERSON>
0c64f45f-404e-923a-4573-022dd44d3c62
['7a23630e-b0b3-c59b-c7d4-2c1744a0a267']
Hi, currently we have a public structure Tcl_Time and related Tcl_*Time functions. But it is not possible to get date information from the values in this structure without using external functions (localtime() etc.) or calling the tcl clock function. The tcl C-function TclpGetDate() is currently only available with the tclInt.h header file. I would like to have a public function doing these work. I do not insist of a particular interface. A very simple one, not requiring the tm struct, could be: Tcl_GetDate(long time, int useGMT, int *sec, int *min, int *hour, ..) Would it be possible to add such a function? Thank you very much, <PERSON>
3d3d98da-8f0c-bac1-0878-7bc43d4f3e1a
['7a3dba88-87c7-4dcc-47bd-3683f75a1ed5']
Hi, I have spent countless hours trying to get the Artist name from a DiscID. The CRelease has a CArtist in it but it's always NULL, yet when I cout << *FullRelease in the sample code the Artist Name is there. The documentation is not at all helpful. ANY help would be greatly appreciated. Thanks, <PERSON>
69613b25-a59f-8c97-cf40-fffedde4edf6
['7a3dba88-87c7-4dcc-47bd-3683f75a1ed5']
I am new to Music Brainz. Downloaded the 3 libraries. Am trying to get the CD Title, The Artist Name and the track titles of the cd from the C++ class library. I ran the libdiscid classes fine to obtain the discid. tried the short demo program passing the discid to the CQuery example. It dumps the info to cout fine but I can't figure out how to obtain the Artist Name and the track titles of the songs. The Title of the CD was easy with FullRelease->Title() but I can't get the above out of the data! Also when stepping through the code in debug the for loop drops out after one iteration when the NumItems() count is 2. Not sure what to make of that either. Any help would be appraciated. I'm finding the Doxygen help to be not enough and the example doesn't go far enough to show the relationships of the classes to one another. <PERSON>
aec7d06b-8829-8e36-dbd3-c28ad76510b7
['7a6773b7-0815-0ca7-4a16-443cfed8c6ee']
Hi folks, Is anyone aware of any projects that enable recording the interaction between a client and a server so that the conversation can be replayed as a unit test? I am attempting to develop something myself but have come up a cropper when the interaction continues via a client clicking on a href that submits the page via some javascript. Calling javascript functions is fine it's just automatically recording the javascript function that was called is proving difficult. Thanks, <PERSON>.
a63af09d-3a33-014f-5e45-bf3f36801824
['7a6773b7-0815-0ca7-4a16-443cfed8c6ee']
Hi folks, I've two different proxies sitting between my client and server. One is a sockets proxy, simply passes the info from the client to the server much like apache's tcpmon. The second strips out the http request from the client's request and makes the call to the server using httpUnit. My problem arises where the heading TRANSFER-ENCODING is used. With the first proxy, at the beginning of each response's entity-body there's a hex value indicating how long the first chunk is and subsequent hex values for following chunks. This is fine and this gets transferred back to and is interpretted by the client correctly. However, when I attempt to get the second proxy to make the same call, in the response all the hex values indicating the chunk sizes have been removed. Has httpUnit been implemented to strip this purposely so as to make HTML parsing easier? Is there any way to prevent it from being removed or should I manipulate the response headers so that the client no longer expects chunked data? Thanks in advance, <PERSON>. - Hi Spoon.
52285e64-ab56-1b18-8c8c-031292bd86aa
['7a6eacd7-b889-9675-a280-2a06056891e7']
what embedded (or embeddable) http servers with cgi support would you suggest for use in applications. Reliability, small memory footprint, so the server is aware about its secondary status unlike Apache, support for Windows, MacOSX and Linux as well as cgi are all essential features. Open source or proprietary - I would appreciate any suggestions, good or bad experiences... I wish to add p2p-like collaborative working and distributed resource sharing option into my GIS application. Plus the ability to show distributed resources provided by different users on a dedicated web-site. Once I used Apache as an embedded server in a CMS application created for one of my customers... Now looking for other alternatives... all the best! <PERSON>
40e9e14f-29cf-dbf4-89f2-fd5246935660
['7a6eacd7-b889-9675-a280-2a06056891e7']
hi friends ;-), just wanted to ask for any practical hints (algorithms) on downsampling very large raw binary (32 bit integer) raster file from disk. Looking for an algorithm or library that could be implemented or used in Ada, C/C++, Fortran, Assembly, Perl, Python or whatever else. The simpler the better :-). My aim is to subsample 2-4 gigabyte raster data file without loading it into RAM within 10 seconds on an "ordinary" PC with Intel processor (Windows, Linux, MacOSX on Intel). Wishing to explore as many options as possible... If you know any paper or website dealing with this kind of algorithms, I would appreciate the url. For example I have 50000x50000 raster and I need a portion of it resized to 700x700... Thanks in advance All the best! <PERSON>
16323ad1-b6b9-4887-2ea8-f392a830e8db
['7a742623-f512-2e6d-1d7f-d58afaa5b861']
Hi all. I'm new to the list, so first off I'd like to say hello. The reason I've joined the list is that I've tried to use the updatable ResultSet feature with the JDBC drivers for Postgres and found that they won't work if you have columns in your database that have capital letters in it. After a lot of fumbling (which included not downloading the source, but decompiling the driver (I'm not too proud of that :))) I've found the AbstractJdbc2ResultSet.insertRow method, which prepares the query for execution, but ignores the fact that columnames must be included between "- s if they are capitalized. I'd be happy to post a patch, but I'm really a newcomer when it comes to open-source shared development, and haven't got the slightest idea on how to do it. I've tried to look this up in the archives, but haven't found any threads dealing with this issue. Thanks <PERSON>
d6e0c1e2-1d20-d692-b5f8-91e439dfdd1c
['7a742623-f512-2e6d-1d7f-d58afaa5b861']
I was surprised to see that the getColumnLabel method in the ResultSetMetaData object returns the name of the column instead of the comment (when available). The documentation says that the method: "Gets the designated column's suggested title for use in printouts and displays." Any thoughts on that? I'd do it if you guys agree. Regards <PERSON>
6c75ba3c-3fa0-309a-1b63-b0dc3893b527
['7a7b5aeb-2660-5bb4-2b66-bce64b1f9b3e']
How secure is old school HTML forms? I regularly buy from the Internet using personal data via HTML. But only use a "secure" connection. However, I would NOT want a remote XForms renderer to do things locally unless I know what it is doing. I would NOT want a local XForms Browser plug-in to send stuff out over the wire unless I knew what it was doing (gave it permission?). I would want a browser plug-in XForms renderer to be verifiable (digitally signed?). I would want a statement of assurance (usual legal stuff) from the XForms product vendor to make sure it has no "hidden" features. Would they need to be insured for likely losses? Yep the tool vendors will need to provide comprehensive answers. DAN
1ce3da5e-bef4-fe8e-76de-ff65ca6ed5ad
['7a7b5aeb-2660-5bb4-2b66-bce64b1f9b3e']
Hi All, I'd like some thoughts on the following issue, I sent a mail to the editors on 27/6 so it was probably to late for the Proposed Recommendation. The Proposed Recommendation refers to the use of "the terms may, must and should in accord with [RFC 2119]". The last paragraph in section 3.2.1 states "A host language must include an attribute of type xsd:ID on each XForms element." However, the example "G.1 XForms in XHTML" does not include an attribute of type xsd:ID on all the XForms elements. Thus in the last paragraph in section 3.2.1 I believe that the word "must" would be better replaced with the word "should". This is supported by the wording in the section 2.4 which explains how 'scoped resolution' works for when the binding element does not have a model attribute. And in such circumstances the id attribute on a model is not neccessarily used ("the first XForms Model in document order is used"). This and the fact that XForms can be authored and work in current implementations without an id attribute on XForms model elements suggests to me that under 3.2.1 it should state "A host language should include an attribute of type xsd:ID on each XForms element". (My understanding of the difference between "must" and "should" comes from RFC 2119). Or am I missing something, should EVERY XForms element have an attribute of type xsd:ID? Furthermore can someone please clarifying exactly how the attribute of type xsd:ID should be declared. For example could a host language which defines an attribute of type xsd:ID called "key" use this "key" attribute on the XForms elements as suggested in the last paragraph in section 3.2.1. There is also some inconsistency within the Proposed Recommendation with regards to the id attribute so if anyone could also clarify it for me I would be grateful. In several places the Proposed Recommendation uses ID, id or xsd:ID. It is not always clear whether or not these occurrances are refering to an attribute of type xsd:ID or an attribute called id of type xsd:ID. Here are the occurances of id, ID or xsd:ID that I have found that are not consistent: a) Section 2.4 1st paragraph - id b) Section 2.4 paragraph beginning "The main difference..." - ID c) Section 3.2.1 last paragraph - xsd:ID d) Section 3.2.3 in paragraph under "model" - ID e) Section 3.2.3 in paragraph starting "It is an exception..." - ID f) Section 3.2.4 in paragraph under "model" - ID g) Section 3.2.4 in paragraph starting "It is an exception..." - id h) Section 3.3.4 1st paragraph - xsd:ID i) Section 4.5.1 1st paragraph - ID j) Section 7.3.1 - ID k) Section 7.5.1 paragraph starting "Another dynamic..." - xsd:ID l) Section 9.3.2 last paragraph - id m) Section 10.1.9 in the "setvalue with Expression" example - id This should prove useful to those polishing up the Proposed Recommendation before it reaches full Recommendation. Cheers, DAN <PERSON>
4968a31d-108c-4add-08b6-25fbf2c621fa
['7a9b9b1e-87a8-ea06-08df-9137694b1ac3']
I finally got the first shot at my "User Story Traps" post done. I'm posting the top 3 traps here, and the link to the full blog posting. (Incidentally, I'm not doing that to redirect traffic to my blog, but I didn't think you folks would enjoy an even more ginormous post on here) Any comments/complaints/controversies/corrections appreciated. :-) <PERSON>
786bec19-2b1d-04c0-7826-a493930d46a1
['7a9b9b1e-87a8-ea06-08df-9137694b1ac3']
<PERSON>, The only comment I'll add to what others have said is that, if your team is new to Scrum, it is perfectly ok to experiment with sprint lengths. However, rather than plan in advance that the sprints will be short at the project start, then bigger, then short again at the end, your team should simply pick a length for the first sprint, then retrospect on it and adapt(possibly changing in future sprints). Typically, when teams are experimenting with sprint length, I encourage them to pick a length and stick with it for 2-3 sprints, then decide whether to adjust it. I expect a team to experiment with Sprint length for the first 3-4 months they're doing Scrum. Over time, teams eventually prefer shorter sprints, but every company and every team is different. I also strongly agree with others that you should NOT plan "bug fixing" sprints into the project. You should get your team to have running, tested, bug free features, EVERY sprint. Planing bug fixing sprints ahead of time just gives team members an incentive to leave undone work. Be sure and read the section closely on "undone work" in the Scrum Guide. Actually , read it 3-4 times. It takes a while to absorb. It might be ok to "secretly" leave room for unknown project risks later in the project, but my advice is: don't call it "bug fixing" publicly to the team. Charles
7f5dc0b8-5cf5-5655-806d-1efda922d153
['7aa080a3-4e19-46d6-5562-2c2e4e2bbf99']
Hello, I've thought a bit about diplomats which are rather super-powerful attacking units with default options (diplchance 80). Used as attackers against cities, they can't (or very hardly) be countered. Used as defensors they are rather useless: if a defending dip in a city makes the chances of success for bribing the city be lowered to 64% instead of 80% this is still too much imho. I don't know how diplomats work on civ2 but in the game i'd like to play such easy ways to take a city shouldn't exist or just be possible in certain limited cases. Here what i think are super-powers: super-power 1: unlimited moves One thing which seems to be the bigger super-power is the ability of engaging an unlimited number of combats against a city regardless of the moves left for the attacking dip and without suffering any damage until the end of the contest. A lucky dip can bribe a city defended with several other ones in one round, engaging as much combat as needed against the defensors. For example 3 defending dips (which is rather unthinkable in term of affected prod) leave a 41% chance of taking the city for the attacker. Here is an example of an attacking dip far of two tiles from a targeted city (this city is defended with 3 dips): 1/ first the attacker uses one move to go to a adjacent tile of the target (on move left) 2/ then a first combat is engaged against one of the defending dips -> 100% chance to reach this stage, 80% chance to win 3/ a second combat against the 2nd defensor -> 80% chances to reach this stage, 80% chances to win 4/ a third combat against the 3rd defensor -> 64% chances to reach this stage, 80% chances to win 5/ a fourth combat against the city itself -> 51% chances to reach this stage, 80% chances to win -> 41% chances globally to win the contest In such a case, this behaviour makes the dip to have kind of 3 extra moves. Also, the dip isn't damaged by the several contests since it keeps its 80% winchance for each of them. In comparison with that, a horse attacking a 3 warriors defended city has no chance : it needs one move to approach the city, then one move to attack and (possibly) kill the 1st warrior. After that no moves are left and no more combats can be engaged in the same round. The horse is by the way probably damaged which can make it easily killable by one of the 2 warriors left in the city or at least lowers its chances to win another combat against the city and reduces its ability of 2 moves for the next rounds. super-power 2: ability of having 100% chances in aggressive actions regarless to diplchance These actions are: - bribe a unit - investigate city I don't remember exactly how "poison city water" works for spys so i cant speak about that. What could be done to counterbalance those super-powers is to : - limit the cases where a city is bribable - raise the efficiency of defending dips in a city (for example by really taking a move to the attacker in a dip contest). - increase the risk in bribing a city (for example the cost is lost for the attacker even if the city doesn't revolt) G. Gracian
63a21a29-e1ed-0798-fa6f-4ce342336c02
['7aa080a3-4e19-46d6-5562-2c2e4e2bbf99']
Hello, When one of you ally dies (ai or human), you're able to see the whole map. In the savegame, connect as h2 to kill the ai and as h3 to see the map. It's possible also to /ai simon to make him human and connect on his spot as h1, the map is also visible. If they are teams, the problem also occurs. <PERSON>
a6445932-6962-205d-bf41-fef0e122b320
['7aa8c31f-2364-430b-5eff-686cbc131aa2']
Hello everyone, I've recently posted a problem to SO (http://stackoverflow.com/questions/1751408/boost-python-opengl-segmentation-faults), where I haven't had much luck with answer. My problem is that if I import OpenGL python libraries before importing my B.P bindings, I get strange segmentation faults for no apparent reason. I isolated a small test case which I'm attaching below. The current boost library version I'm using is 1.37; you'll have to make a change to Makefile if you have any other version installed. I'd greatly appreciate if anyone can give some insight on what's happening. <PERSON>.
118d0478-75d0-a6d3-0650-0bdcc0c98e68
['7aa8c31f-2364-430b-5eff-686cbc131aa2']
Hello everyone, I'm trying to tag certain load instructions with specialized metadata and make use of it in an AA pass. However, it seems that metadata is getting lost when instructions are merged together. Here's a toy problem with one intentionally duplicated load to illustrate the issue. http://susepaste.org/3af3a676 The above paste has two before/after IR pairs (the `before` code gets passed through `opt -O3 test.ll | llvm-dis`). Regular `tbaa` tags are still present when loads are merged, however custom ones seem to perish. Is this expected? If yes, is there any way I can make them persist from within a custom AA pass at all? I tried overwriting copyValue() in the AA pass but it doesn't seem to get called at all. The main idea is to recognize certain types of loads and have other information in the my.tbaa metadata that regular tbaa doesn't provide. The only other solution I can think of right now is to encode that extra information in tbaa node strings, but that seems rather gross. Thanks, <PERSON>.
e057fd0f-8ac3-f3be-29b8-2837b4414453
['7abe4812-2dda-4597-40d4-a4e582721d35']
Hi Folks, Over the years I've written four drum/percussion tutor books using lilypond-book. I've just put these up on the web, with the added feature that you can play back the small exercises that make up the majority of the books with noteheads highlighted and sound. Volume 1 is free, the other three volumes require a small monthly subscription. http://percussion360.com/ This works by taking SVG generated from lilypond, munging it a bit (including matching it up with the .ly source), then playing it back using javascript and html5 audio. Playback doesn't work in IE but does work in other recent (proper) browsers. Feedback and users appreciated! I'm happy to consider adding functionality; I'm particularly thinking of adding some pupil management tools for teachers. Enjoy, <PERSON>.
e4d74075-5cd7-752b-262a-9fed954b7d88
['7abe4812-2dda-4597-40d4-a4e582721d35']
Folks, In4/4 time, \times 4/6 { sn16 sn sn sn sn sn } creates a six, with six notes, semiquaver lines joining them together. Perfect. However, \times 4/6 { sn8 sn16 sn sn sn } in lilypond 2.6, produced a quaver joined to a group of four semiquavers, whilst in 2.10 I get a quaver, a semiquaver, and then a group of three semiquavers. The odd semiquaver's lower beam line is pointing left rather than right. Can someone point me towards a fix for this please? Can I tell the time signature how I want it to beam things like this? Thanks, <PERSON>.
a27ce8a1-1420-b836-4036-63effa3cc54d
['7ac3bca2-c6d7-79c0-cfac-8d2163101215']
First let me apologize for not doing a direct reply to your email. I don't seem to be getting the mail forwarded from the list, even though I re-enabled it in my profile. So I'm hacking this to your reply. NOW, let me express my sincere appreciation to you for solving my problem. I don't know why Xine backend no longer works, but gstreamer fixed it. I added this problem to BZ #551496. I'll update that bug, with your solution. I have checked the other machine and it has Xine as the backend. Maybe after another set of updates, either to KDE or pulse, I will try switching it back. Even after 20 years of *NIX, and 10 of Linux, I keep learning. I don't mind the occasional problems, as long as there is a solution. Gotta love Linux! Thank you again, <PERSON>
579c00a8-97fc-a451-7f8a-9285c0de400c
['7ac3bca2-c6d7-79c0-cfac-8d2163101215']
This is my first post to this list, so I hope I get this email formated correctly. If not I will endeavor to correct my mistakes in future postings. I tried sending this previously but did not see it show up in the mail, so I'm trying again. If they both show up, my apologies. I've seen several other postings on hangs and slow system time updates during hangs, but no indication if anything has been found to resolve the issues. Here is my system info: Soyo Dragon Plus (Via KT266A chipset) AthlonXP 1700+ 512MB DDR2100 using on-board lan and sound Geforce3 ti200 WD 120GB primary master Maxtor 120GB master on Promise Ultra I have the hangs both with and without the livna drivers. The hang occurs at varying times. Anywhere from a few minutes to many hours of running. I do not leave this machine up 24/7, so it always happens after a cold boot for the day. Some days I may be on for 4-5 hours and not have a hang. It appears that if I wait long enough during a hang, I can usually get a virtual terminal and do a shutdown. 2111 kernel appears to be OK, although I have had it hang after a reboot from a lockup in another kernel. Never had a hang with cold boot into 2111. One thing I have noticed is that sometimes after a hang, the system time is off by a significant amount (10s of seconds to minutes). I have two other computers running FC5 and they do not exhibit this problem. Both of those systems are different mixes of hardware vs the problem system. All three systems are upgrades from FC3, and one has been upgraded from FC1 => FC3> FC5. That one runs fine. I hope someone can point me at the solution to the problem. B.T.W., I'm using the problem system right now, and it has been fine for the past week on 2111 kernel. I've run smartctl and checked my drives and they are OK. No errors other than a few reallocated sectors. The sectors are not new and have been slowly growing for a couple years. I find that all my WD boot drives have the tendancy to do this, while my Maxtor and IBM do not. Thank you in Advance, <PERSON>
a62cf973-be37-ccf6-833d-0542b17e3741
['7acdd230-4c94-d64f-7f08-591fd97c898d']
So ECMAScript does allow a regex to start with /{ or /+ or even /} and only forbidding // /*. The javascript language definition was not allowing this which caused highlighting problems with some code (for example the zotero source). Attached is a super simple patch for Bundles/JavaScript.tmbundle/ Syntaxes/JavaScript.plist to fix the problem. I'm attaching it as both as gzipped file and a straight diff <PERSON>
ee282830-e87e-e60e-83b8-2915aefbad86
['7acdd230-4c94-d64f-7f08-591fd97c898d']
As both the linked_file and linked_url templates don't include a path field but linked_file happily makes use of such a field I wanted to check how I should treat a linked_url in a zotero client using the web_api. In particular, can I assume that a linked_url is always simply a url with no filesystem content? Or should I assume that a linked_url can have whatever filesystem content that would be uploaded to the server for a imported_url? Thanks, <PERSON>
b91ef52c-e3ab-efbf-7ce3-e6f7ce4aeaaf
['7ad113ab-413f-46fe-41d7-9be2c18933a0']
Hello <PERSON>, <PERSON>, Sorry that I bother you but I'm under some hurry 'cause a school server crashed and I have to set it up over the weekend, again. It had double disk fault in RAID5. The second disk show the damage during RAID5 reconstruction ;-( So I changed all 4 disks with another brand and put a fifth one as spare in. hda (the spare) is on the on board VIA 686b controller and show some "bad" logical numbers so that I get a different partition layout. hde and hdg are on the on board HPT 370 hdi and hdk are on an additional HPT 370A IDE card hdb is for installation only Any chance to change the logical disk layout of hda? Second to you and <PERSON>: Is it possible to boot from the mirrored RAID1 partitions (hdX10) with the current lilo-22.x (SuSE 8.0)? With ReiserFS? Thank you much! -Dieter
b330a127-dad1-a34c-707e-31ed398d1865
['7ad113ab-413f-46fe-41d7-9be2c18933a0']
Page coloring for 2.4.18+ isn't preempt save? It gave ~10% speedup for memory intensive apps on my single 1 GHz Athlon II SlotA (0,18µm, L2 512K) but look the system hard from time to time. Nothing in the logs. I've changed the patch for 2.4.19-pre7 + vm3 + latest rml-O(1) + preempt. Thanks, <PERSON>
2b84bb06-ddfc-574a-f053-5571a7919394
['7ae0d4ea-030a-3d51-c3b5-facb05c01d8e']
I have made a KIX script that uses two external utilities to check the round trip time for e-mail. First it sends the mail, then it waits for it to run and finaly it puts the result in a log file. I do this with two different mail server and have two different log-files, but can easily make a third log-file in MRTG format. So the question is: Is it possible to tell MRTG to skip the gathering and just make graphs from a log-file? <PERSON>
7e4d7aca-8881-5af7-09d9-ce3ff6db6434
['7ae0d4ea-030a-3d51-c3b5-facb05c01d8e']
Shouldn't really be a problem. You have a firewall in between that regulates whats allowed and whats not. That's why I moved the controlling server from the DMZ to the inside, since I didn't want to open up for all the ports it needed to the inside. But theres people out there that knows much more than me about security stuff, so maybe they can give a better answer. <PERSON>
340c07ac-aeb1-2d8d-d56d-3c84ef6885bb
['7ae1e027-6c2f-32d9-b93a-a4716cc77a4d']
File>Export Multiple > Options give window that is on top of all world -- it remains on top even when I switch to a different application. Also it is annoying that double Alt-Tab does not return me to Audacity, contrary to what happens normally. I think it might be connected to wxSTAY_ON_TOP parameter to wxDialog in files src/export/Export<format>.cpp Windows XP (tested with Audacity 1.3.6a1 (no newer one aviable to me yet)) Jan
acab5b7b-eaa2-060e-40ec-6686aac37d1a
['7ae1e027-6c2f-32d9-b93a-a4716cc77a4d']
Hi all. By the way, I *much prefer* single effect menu (letter keys jump nicely where I need, and the menu remembers last position ! Much faster the old way!) Where can I change it in Preferences? This following makes me crazy: Can not generate 0,8 amplitude Sine I think Generate>Tone had default 0.8 (or, 0,8). I changed it to 1 once. Since then it often changes 0,8 or 0.8 (both!) to zero and generates silence. (Now reproducing with the latest alpha from Gale) Sometimes it works, which might (and might not) depend to switching the versions. Note: English Win XP, a lot of Czech in localization, Audacity's lang changed many times. Jan
c9fc5a62-c0c0-a1f8-e66e-6d45386de389
['7ae51cd8-9fdb-aa69-0087-fdba12250a1d']
i now have a requirement to look at both grinding and cam turning on my small lathe looking around i notice shirline have done some coding in this respect to emc for cam grinding has anyone else looked or done the same . i'm looking at how to approach this or alternatives . my own fault i think wanting to make a 18 cyl radial ! <PERSON>
1183a1f8-24a2-e49b-a02b-8fdb8ef01ba2
['7ae51cd8-9fdb-aa69-0087-fdba12250a1d']
Hi All As everyone knows people do have difficulty with these Chinese TB6560 combined stepper driver boards , over the last year , I have tried to collate some method to support these . Their appears now to be some resemblance of normality of a few pin outs , rather than the earlier , total mash up of even incorrect paperwork with the boards . attached is a patch to Master, to add additional examples for the TB6560 3 Axis and 4 Axis boards , please note that the pin outs do change dependant on the number of axis on the PCB . Could I respectfully ask that these be given some consideration to be included in master . I'll continue to monitor these boards should any others pop up , and add an example if needed or amend as required . I Hope that this helps new users And thanks to the many forum and Mail replies I have had over the months , with configuration testing and supplying working versions to my database . <PERSON>
87cf0bda-d79d-a25f-5174-d296cf30d48d
['7ae76cec-cf6c-08c9-8b45-c555c04bc3bc']
Dear <PERSON>: Being half a world away I am just reading the news of your Dad's passing.  May his soul rest in eternal peace.The passing of a loved one, and especially a parent, is very poignant, as one tend torecall all the days of one's life, the fun times and the sad times.  It takes a while to overcome the grief of the passing.  But time heals, and we move on.We will include him in our prayers for all our near and dear ones, and friends who havedeparted this world.Take care.  Best wishes to your wife and children.   <PERSON>
d6e2a3ac-9a64-f1a9-bb48-f052ba85f1fd
['7ae76cec-cf6c-08c9-8b45-c555c04bc3bc']
Lets all go on hunger strike to change things in Goa.  Lots of things we can go on hunger strike about............ 1. Indiscriminate mining 2. Corruption 3. Filth and garbage everywhere 4. Haphazard traffic 5. English and Marathi in our schools instead of only     Konkani,   and on and on and on....................... Forget about the democratic process, where power comes from the ballot box, and the party or persons who get the majority get to govern.  Now we have a short cut to get our way.....call a press conference and declare a hunger strike till death !   (fat chance of any one dying of hunger it is all theatrics....doctors will proclaim that the faster is on the verge of death, prominent persons will appeal to the faster to give up his/her fast, and on some pretext the "fast until death" will come to an end) No I am not belittling recent events in India.  But havent things gone too far ??  Persons with some name recognition, petulantly playing on the emotions of those who look up to them, go on hunger strike giving un-achievable deadlines or else !!!   Sounds like blackmail to me. Now, I am all for removal of corruption and corrupt officials and bringing back ill-begotten gains to the country,  if they are adjudicated to be such.  But this is a democracy and collectively the people's voice is supreme, Not the voice of one person or a handful of persons claiming to be civil society.  And yes I am just as frustrated as anybody else with the slow pace of change and the slow grinding wheels of justice.  Unlike some of the countries in the Arab world where tyrants ruled, India is a democracy.  Not a perfect democracy, but as good a democracy as any place else. So we cant mimic what the citizens did in those Arab countries.  All political parties should respect the rule of law and remember that they too will have their turn to rule.  What they forget is that the un-democratic means that they support are like a two edged sword that will come back to haunt them when they are back in power. <PERSON>
3c2539bd-0f2b-87a2-8f00-3763ec315997
['7ae87aed-48ce-3e2d-3ca7-12ed34b8420d']
Hello All I am currently looking for a security role in the Akron/Cleveland, Ohio area. I am a CISSP with over 8 years experience and a current DoD Secret Clearance. I am a quick learner and passionate about my work. I have experience with firewalls, routers, vulnerability assessments, audits, policy, awareness, etc. I am available immediately. Please e-mail me if you have a position open and you are interested in seeing my resume. Thanks for your time. <PERSON>
fef6cc54-39e7-c972-dbf6-1aa04130e371
['7ae87aed-48ce-3e2d-3ca7-12ed34b8420d']
Hello I am looking to keep up my skills of technical auditing while I take a paperwork position for the next three years. I am based out of Ohio, but could audit anywhere from here. I am also open to traveling for weekend audits. I can complete firewall reviews, router configuration audits, vulnerability assessments, war dialing, etc. I am looking to do this on an ongoing basis for either a few clients or working with an agency to set up one off's. Thanks! <PERSON>
e95657af-ccee-ecfb-a0df-a1eb482480ac
['7afff8ce-4ebe-4bdb-a4e4-896e56cb66bb']
Dear Sir, Dear Sir, We renamed our project file from tourism.pprj to tourism_db.pprj. And while converting our project file to database , we were not asked to rename any tables. We were just asked to give a table name which we named as tourism. The snapshot of the dialogue box is attached. So now we don't know how to use the tables created. Using the data master, we are able to view the tables created but not access it to either enter the data or retrieve data. Please help us out. Regards, <PERSON>
d964cef1-fe7e-be24-a9a6-8e77d167803b
['7afff8ce-4ebe-4bdb-a4e4-896e56cb66bb']
Dear Sir, We renamed our project file from tourism.pprj to tourism_db.pprj. And while converting our project file to database , we were not asked to rename any tables. We were just asked to give a table name which we named as tourism. The snapshot of the dialogue box is attached. So now we don't know how to use the tables created. Using the data master, we are able to view the tables created but not access it to either enter the data or retrieve data. Please help us out. Regards, <PERSON>