query_id
stringlengths
4
64
query_authorID
stringlengths
6
40
query_text
stringlengths
66
72.1k
candidate_id
stringlengths
5
64
candidate_authorID
stringlengths
6
40
candidate_text
stringlengths
9
101k
3ab082fb-f71f-2d9a-556f-d3597f11dddc
['1cea16e6-bcb0-1af5-ad8c-954adbaa33d0']
Hi all, I have just a couple of hours trying to install the Server Administrator 1.4.0a on a Debian/Woody box. Searching the maillingist has revealed nothing. Attached is short writeup on how i finally managed to get it all running. I hope this will save some time for others. Let me know i you find any errors or mistakes. Have fun <PERSON>
463029b8-4cd8-0e9a-ffeb-036b73696ba1
['1cea16e6-bcb0-1af5-ad8c-954adbaa33d0']
Hi, Since upgrading to 2.6.28-rc5 and also with -rc6, the syslog gets spammed with errors like "NFS: v4 server returned a bad sequence-id error on an unconfirmed sequence ffff8800779fb828!" This is not only a cosmetic problem, but breaks applications like evolution if the home directory is on NFS4. Client is 64bit, server is 32bit with 2.6.24 (Ubuntu Hardy). <PERSON>
c959e86c-35a7-3037-62ac-a4d78875839a
['1cf5af50-e14f-13d0-7d15-626717d0edea']
I need to generate a DEM on a rather large dataset: 130x100km at 1:20000. The output resolution will be 20m. The area is very mountainous, and the input is contour lines. There appear to be a number of surface generation tools available. I am running v.surf.rst now, but it is very slow. Is there another, faster module that will generate acceptable results? <PERSON>
2fcaf3d5-15d9-886b-177f-d01aec21fc98
['1cf5af50-e14f-13d0-7d15-626717d0edea']
I have recently started to get commands terminating with a `bus error'. Some examples area d.what.vect (only when I use the postgresql query), ps.map, but only when processing a integer colortable (float colortables are fine) and when importing shapefiles. These used to work, and I think the problem might be my upgrade from Mac OS 10.2.5 to 10.2.6. This problem affects GRASS 5.0.0, 5.0.1 and 5.0.2. Has anyone else experienced this? <PERSON>
bcc158fd-b6bd-faea-4cf4-4abe308d9091
['1cf9a65a-7375-b37e-4333-50f6a1661d8d']
I have a self referential many to many association I am unsure on how I should handle in my entities. We have the concept of a Team. A team can be associated with one or more other teams. Its not like the friendsWith and friendsWithMe example in the docs because you cannot ever have a team affiliation that is one way only (where team A is affiliated with team B but team B is not affiliated with team A) Currently we have a Team table that looks like team
4e9a04a5-7712-8a74-d4ec-ccada71f6819
['1cf9a65a-7375-b37e-4333-50f6a1661d8d']
Hi All, I am trying to inject a dependency on @security.context into one of my services and and am getting a circular reference error [Symfony\Component\DependencyInjection\Exception\ServiceCircularReferenceException] Circular reference detected for service "security.context", path: "security.context -> security.access.decision_manager -> security.access.decision_manager.delegate -> swim_security.authorization.company_per mission_voter -> swim_security.permission_manager -> security.context". I tried changing the injection to use setter injection instead of constructor but that didn't help. Can anyone recommend a way to achieve this? Regards,
9cb4153b-e936-5edf-5164-6d1d98856961
['1d091694-c283-b8cf-35e3-b9355eab6d5f']
has anyone rebuilt red5 on windows from the src files using ant. I need to rebuild it in a different directory than what is in the installer. I managed to use ant to rebuild it, but it doesnt work, so i'm curious if there is anything else I need to do besides just running ant in the red5 directory. like any pre-configuration before I run ant? Maybe I just didn't start it correctly? I clicked on the red5.jar file after i ran <PERSON>. Shouldn't that start it? if anyone has done this and could give me a rundown of what they did. that would rock!. thanks
55633c2c-e591-ca21-6931-4572c46a699b
['1d091694-c283-b8cf-35e3-b9355eab6d5f']
Sorry i'm asking this here. I put it in some forums online and am not getting any responses. So I figured ya'll could help me out. Does anyone know of any webcams that work with MAC OS X other than an iSight?? If not, does anyone know if AOL AIM Video chat will work with iChat on OSX? ( AIM to iChat ).. I'm having to do this video conference soon, the other guy is for sure gonna be on a mac. I need a solution for the above... thanks in advance to anyone. sorry i put this here.. not getting much from other forums.. <PERSON>
c5a5ea9a-c69a-165c-2b8c-aa20dd3320c5
['1d0b058f-eebe-3bbd-c07a-afb98c098303']
I´ve been looking for this function but I can´t find the Align to Baseline. It´s very usefull to make perfect aligned text and images. If it is not implemented, I would like to request the introduction of this function. As it can be used in any professional DTP aplication as QuarkXpress and InDesign. Thnak you. <PERSON>
23ccd31b-24d5-c8f8-b357-de0243c82d9e
['1d0b058f-eebe-3bbd-c07a-afb98c098303']
Hi People, When I started This thread I did not had the intention of starting somekind of war from what I´ve been seeing in the mailing list. I Just thought that It would be great to have some kind of "Open Source Creative Suite", just like Adobe Creative Suite. With more than CMYK and spot color in Gimp and Inkscape. Creating a full integrated design environment combining the 3 software. I know it´s not an easy task. And to be honest I know that it would take years to this became true. But this is my dream, maybe not the dream of the developers of each individual software listed above. I´m only a graphic designer searching a replacement for adobe dominjation of the digital press world. I wich I could help more with the development team working as a programer, but my skills in doing this are less than zero. Thnak you for reading my thoughts, <PERSON>
df4d9747-307e-dcf1-2e8c-70d4e31b8162
['1d1f32ee-b9d7-bb68-b173-b55e8b89717a']
Hi, I am using Grinder for load testing a Java application and Grinder is really fine for that purpose. I try to set up a scenario in which Grinder agents run on multiple hosts and I ran into a distrubution problem. - Distributing the .py scripts via the console works really fine. - Distribution of the grinder.properties file seemingly has no effect as the Grinder distributes everything into the HOSTNAME-file-store directory grinder.properties and tge Agent reads its grinder.properties from somewhere else. It would be very handy if I could distribute grinder.properties so changes in grinder.sleepTimeFactor or grinder.threads could be distributed to all agents easily. Is there any way to distribute grinder.properties vis the console so that it is effective for all remote agents or do I simply have to resort to a shared filesystem for all agents? Kind regards <PERSON>
82224561-7e04-99dc-fa77-a15ab246a33b
['1d1f32ee-b9d7-bb68-b173-b55e8b89717a']
Hi <PERSON>, Using Subversion (or im my old-fashioned case cvs :) seems to be a good idea. So I take it for granted that Grinder can distribute almost everything except grinder.properties. Distributing grinder.properties seems to be a chicken-and-egg problem anyway. Still it would be nice to modify grinder.sleepTimeFactor or grinder.threads remotely via the console to have an easy way to increase the load on a test system, but for now I can live with the current set of solutions. many thanks <PERSON>
169ef05e-e4a7-7b85-1ada-1f1bbd084f6a
['1d32ceb8-f528-b154-39c6-7e725a04823e']
Hello, I am planning to use the nano sim based mobile phone. Could you please tell me if its possible to use the SIM TRACER for the same. For example samsung galaxy S9. By looking at the wiki I see the sim tracer would fit in for standard sim slot, will the end point of the flex PCB fit the nano sim holder of S9 ?. Could any one please guide me on the same. Best regards, <PERSON>
d03d4f11-0b83-948b-afd3-4db4e0516492
['1d32ceb8-f528-b154-39c6-7e725a04823e']
Hello all, I have newly ordered the Simtracer 2. I have received the same today. I would like to have some guidence from you, how to proceed now. Should I load a firmware on the module?, which software should I use? how to track the message exchanges from the sim to mobile. There are several manuals, could u please tell which one to read? Could you please guide me on the same. Best regards, M.Sc. <PERSON>
e5a9e498-fc99-91c3-6bb6-df17717a1f57
['1d41cc8c-32cd-500e-109f-c334aa3e7510']
Folks: I'm new here, but I've taken a look at the archives and don't think I've seen anything about this sort of problem in the last year and a half. Maybe no-one uses Nuppelvideo any more... I did however, and as the Subject: line indicates, I've got a problem. I'm on linux 2.6.10 kernel with transcode-1.0.2 If I play my .nuv file with 'mplayer 1.0-pre8' the A/V sync is fine. If I transcode to .mpeg (MPEG-1 or MPEG2) and play with mplayer, the sound is late by about a second at the start of the file and gets progressively worse as time goes on (sound seems to be getting later by 1 or 2 seconds per minute). I notice that 'transcode' announces that its nuppelvideo import filter is dated sometime in 2002. You may also care to note that I get this same A/V sync problem if I transcode "by hand" using 'nuvplay' or 'mplayer' to get the sound as a .wav file and 'lame' to encode it, along with using 'exportvideo' to get the video and use 'mpeg2enc' to encode it. This is the route suggested in the nuppelvideo package. Transcode may well be tripping over the same problem (presumably shared in the 2002 import filter). Note though that 'mplayer' isn't having this problem. I've put a very short extract of both my .nuv file and .mpg (MPEG-2 as it happens) on ftp://tallyho.bc.nu/pub/steve/nuvsyncproblem.tar.gz just in case anyone wants to look at it. You should easily be able to see my initial 1sec sync problem, though my clip is only about 12 seconds long so you won't see the progressive sync slippage. The guy who blows a whistle about 10 seconds into the clip illustrates the effect nicely. The clip is 12 seconds long so as to keep file length down (currently about 21Mb). Any assistance appreciated. Steve Hosgood
f0422d82-4b52-4b24-6a44-c10c7553fc9a
['1d41cc8c-32cd-500e-109f-c334aa3e7510']
Folks: As most of you know, I've recently been working on a FDM for the Colditz Glider. I was surprised and encouraged by the amount of comment that the original thread generated. I've not been sitting still, and have now got a second version that you may like to play with: ftp://tallyho.bc.nu/pub/steve/flightgear/colditz_20050525.tgz Changes are as follows: After *much* grovelling the net, I eventually discovered a reputable claim (from <PERSON> at UIUC) that the Colditz Glider used the classic 1930's "Clark YH" wing profile. This ties up with a comment from <PERSON>'s "The Latter Days at Colditz" to the effect that the bottom surface of the wing was flat (most of it is indeed flat in the Clark YH). So I went looking for lift and drag coefficients for the Clark YH, and found them after a long search in a tutorial document on the web originating from Strathclyde University in Scotland. The new Colditz Glider FDM now uses these stated figures for the Clark YH, and though I don't have proper stall hysteresis figures for that wing, it seems almost impossible to fly the Colditz Glider model so that the airfoil actually does stall anyway. As airspeed decreases, the glider just loses lift to the point of mushing through the air below about 32 knots. With the machine flying normally, its best-case rate of descent is about 4 or 5 ft/sec, agreeing fairly well with the estimates of the original designers in Colditz. Likewise, its glide ratio is about 18:1 as estimated by the pilot who flew the replica in 1999 or 2000. I've adjusted my estimate of the locations of the CG and locations of the pilot and passenger after measuring around the reproduction of the original plans. I've played with (but commented out) an attempt to model the launch catapult with a very short-lived rocket engine. Basically, I need a rocket with a burn-time of 2.2 seconds and a thrust of 1866N (that's about 420 pounds in Flintstones units). However, my attempts have failed so far. Suggestions welcome. For instance, what are the units of fuel capacity for the tanks and fuel usage for the engine? [ Presumably tank capacity is in American Gallons or maybe "Barrels", and fuel usage is in Bushels per Nanofortnight, eh? :-) No chance of litres per second or cubic metres per second around here I suppose? ). The next version might even include a 3D model. <PERSON> is working on one right now. Thanks, <PERSON>. Whatever - enjoy escaping from Colditz. You should be able to make the intended landing site on the far side of the Mulde from the castle roof with height to spare if the prisoners' estimated distance to that landing site was right. Does anyone here actually live in or near Colditz? <PERSON>
e315eea5-4010-7a2f-d784-a707f9908996
['1d499800-f6d1-de88-5df8-1327fff422b9']
Folks, Would anyone out there who has working config for Frost & Sullivan please share, and also any advice for this one? OCLC's config (http://www.oclc.org/us/en/support/documentation/ezproxy/db/frost.htm) doesn't make things behave as expected. The vendor just got through walking me through their two "IP" authentication options, one of which seems like referring URL to me, and the other which still requires individual user sign in. At least I'd like to be able to test the latter via proxy, even if we have our objections. <PERSON>
a6db9c80-5a09-fe9b-5c94-12d022d36cd6
['1d499800-f6d1-de88-5df8-1327fff422b9']
Folks, We are on EZproxy 5.1c and are employing the user limits functions to detect excessive usage and to shut users off, our ~expires function is not specified. I have an instance that I am trying to make sense of and could use some expertise. Looking at the usage limits it appears that a user 's access was shutoff 2009-08-31. However looking at the audit events, and the logs, it appears that this user has been active before and since that limit was enforced. As we are using the ~expires function and the user still appears to be clocked via the global usage limit, I don't get why or how access is still being made available. An additional issue is that "real" user professes not only not to have accessed any of the materials all summer, but is not in the part(s) of the world that this access continues to take place. Ca anyone talk to why and how this is happening, and what steps we should take next, in terms of proxy? <PERSON>
92b9c73f-7a65-9fd3-0b19-a5f215738be5
['1d56cb5b-a0ff-7350-da67-d9f8ed1f2692']
Hi. I have an r50p with a 2.6.10 vanilla kernel and had problems with suspend-to-ram. I had to add a kernel option (acpi_sleep=s3_bios) because without it my notebook does not wake up again. For the IBM special keys you should also check out the tpb package. If you can speak (and read ;-)) German, visit my homepage: http://www.kai-hildebrandt.de/linux/thinkpad.html#ACPI I try to document all stuff about my r50p (or Thinkpad in general). Regards Kai
dd669804-c783-127b-0a62-c9f70b7a73ab
['1d56cb5b-a0ff-7350-da67-d9f8ed1f2692']
Hi all. I just compiled a 2.6.9 kernel on my debian sarge notebook with ACPI support. I also configured tpb but some keys do not work. The notebook is a IBM thinkpad r50p. What can I do to get the key to change the video output (Fn + F7) working? Is there any chance to get my notebook to work after an suspend-to-ram with a working screen? It stays blank after resuming not depending on framebuffer (vesafb) or the X-server (I deactivated both to test it). Thanks. <PERSON>
822539d9-5c81-e41c-afaa-fd5b2cfe5448
['1d5ff30b-e885-e9b1-ebba-fbde8380dc64']
I have an educational website designed for technical writers creating API documentation. I recently added a tutorial on how to create an OpenAPI 3.0 specification document. This tutorial uses a sample weather API for context, shows how the various fields appear in Swagger UI, and tries to guide the user step-by-step through each object in the spec. If anyone is interested in reviewing the tutorial and providing any feedback, I'd be grateful. Here's the link: http://idratherbewriting.com/learnapidoc/pubapis_openapi_tutorial_overview.html. There are approximately 9 pages in the tutorial. I link abundantly to the relevant places in the OpenAPI 3.0 spec on GitHub and also to other documentation. Thanks, Tom
411954d8-1776-29ec-d790-28332f67fa8b
['1d5ff30b-e885-e9b1-ebba-fbde8380dc64']
My group is thinking of adopting the OpenAPI spec as a standard for documenting REST APIs, but there's some disagreement about whether it's better to generate the specification document from code annotations (through Swagger codegen libraries), or to create the specification document manually (separate from the code). I'm compiling a list of pros and cons and wanted to get a quick pulse check from the community. Can you tell me if I'm off-base about my pros and cons? *Cons of code annotation method* - the annotation syntax will differ by programming language - developers might not want to give others (tech writers, PMs, testers) access to their code, limiting collaboration - there's more of a learning curve to the syntax of code annotations than there is with the standalone YAML file approach - if engineers are unfamiliar with OpenAPI / Swagger and are uninterested in learning it, managing the spec within the code becomes like inviting yourself to someone's house without invitation - the specification file can only be generated *after* the API is coded - annotating the code puts the whole documentation process directly in the engineer's domain, prompting the engineer to be the primary person responsible for the documentation - not all codegen libraries support the latest version of the OpenAPI 3.0 spec - the annotations clutter up the code with a lot of documentation *Pros of code annotations method* - reduces potential for documentation drift - consolidates your doc and code in the same location - enables engineers to write documentation using their existing IDE - automatically creates the model definitions for requests and responses Am I correct in assuming that the prevailing trend is to create the specification file manually outside of the code repository, primarily so that the specification document can be used as a contract or blueprint for first testing and then building the API? <PERSON>
fe613bb9-08cc-3645-c5c5-5d02a0726613
['1d750e6d-e5ee-0af7-c6a8-dc7bc069c58c']
Hi all, I want to use SLIDE to store files. By the Architecture of SLIDE I think I only need to invoke Server API of SLIDE, and don't need to use WebDAV Servlet. who can show me a sample of using Server API of SLIDE. Thanks a lot. <PERSON>
583a2090-2889-528e-5bf8-cfa48da5b726
['1d750e6d-e5ee-0af7-c6a8-dc7bc069c58c']
Hi all, I want to use SLIDE to store files. By the Architecture of SLIDE I think I only need to invoke Server API of SLIDE, and don't need to use WebDAV Servlet. who can show me a sample of using Server API of SLIDE. Thanks a lot. chengj
987d9da9-f3a3-b57e-f312-f9fcf78874a1
['1d8702c1-7d88-67b0-c721-6e0465a61772']
Hi All, I would like to contribute an option to be added to the Virtual host section, so, could someone explain to me what sections of the source code need to be modified, where they need to be modified(to see the new code and access it), and a copy of the source code in un-compiled version soas i can study it and work into it. what i am looking at dooing ios something like the <directory ....> </directory> option Thanks <PERSON>
6464a03f-9c53-ff6e-066b-c4a9d74144e6
['1d8702c1-7d88-67b0-c721-6e0465a61772']
I am in support of a system to moderate, even ban members if need be, but there must always be a course of redemption left open to those that may have acted improperly, to cut the avenue of redemption to those that may have done something wrong, can push people to worse actions, there are many examples in todays society ranging from friviouls to serious!. remember, don't judge a person by how many times they fall over, but by how many times they pick themselves up, learn by their mistakes, and keep going! outright ban is inexcusable unless all other avenues have failed!, and is a failure onf moderation! <PERSON>
259627d5-ed47-5230-3a3e-269c23ebab90
['1d9a4f00-0d6c-58be-8e53-b45b4544d8c0']
By anonymous, I mean a source that traces only to an email address with no phone number, no address, no anything. Not even a significant history of email from that account. This source has no accountability, because they are anonymous. By 'System software' I mean software whose integrity a company relies on to perform its functions. If the software were remotely exploitable, it could potentially result in remote access being obtained, and/or confidential information being exposed, firewall being circumvented, etc Here are the specific questions: 1. Would you trust (meaning use) system software from an anonymous source? 2. Would the fact that the software is a derivative of well known software, but with apparently gratuitous "security fixes": would that increase or decrease your willingness to trust the software? 3. Would the fact that source implements a variation of discredited changes advocated by gray-hat or black-hat hackers increase or decrease your willingness to trust the software? 3. Would you consider it a bad judgment to use such software knowing (1) for sure, and perhaps (2) and (3)? How serious is the bad judgment? 4. Would it be reasonable to fire the admin responsible if they knew of (1), and perhaps (2),and (3), but used it anyway? 5. Does the reasonableness of termination depend on actually knowing (1)? That is, supposing the admin didn't know (1), should the admin have made an effort to find out if the software was from a dependable (or at least accountable) source? Please reply off list. Thanks, <PERSON>
2cb45437-084e-bb66-392b-938868a92c7b
['1d9a4f00-0d6c-58be-8e53-b45b4544d8c0']
Hello, I have recently discovered the hercules project, and I have to say bravo folks! I did a sysgen of turnkey mvs over the break while stuck at airports, with a strange sense of glee that I can't really explain. Some background on me: I worked on mainframes at college, then at Draper Lab as a pl/1-db2-ispf programmer in the late 80's. I ran a set of hitachi G4 (370/390 clones) (VM and MLPF systems) for porting OSF/1 to the mainframe for hitachi in the mid 1990s. I also worked in network engineering and IT at Kendall Square Research (which build KSR supercomputer in early 1990s), so I have come to appreciate mainframe architecture. I remember <PERSON> and <PERSON> from Draper (greetings, if you are here). But I also did a lot of work on unix systems at Draper (to the consternation of the computing facility dept, I think ;0), went to OSF in '89, worked on Motif, OSF/1, DCE, and other stuff since. Recently I've been collecting old unix systems, simh, and that's how I found hercules. So, anyway, now I am looking for the IMS DLI/360 tapes, which a couple of people appear to have (judging from email messages anyway) running in some state. Emails I found suggest it should be downloadable from <PERSON> site, but the shedlock.org site is down or not responding for a couple days at least. Thanks, --Dean
4eab582c-828f-75f2-4b68-d00e5af494ac
['1da59760-b49f-3d5c-fdfc-266647235099']
Dear Weka Folks, I am an amateur Weka/Java user. I wrote some java codes in Windows Using Weka API's and it works good. now I need to train and test the program with a larger data set so I want to use a Unix/Linux system, but i don't have any idea how to add Weka Libraries to the code to compile in the unix system. Any hint or suggestion would be appreciated. Hope to hear back fro you guys. <PERSON>
9c56b871-314d-edfc-a166-1a9da45bb6ee
['1da59760-b49f-3d5c-fdfc-266647235099']
Dear Weka gangs, I am trying to train and test random Forest on my data set. My train set has a class variable but since my test set is a real case which we want to predict we don't have any class variable for them . in fact we want to predict the class for them based on training set. When I am trying to use the weka on that i will receive this error : Weka exception: Train and test file not compatible! Does any body have a suggestion for me? Sincerely <PERSON>
13dc4880-24e5-7827-fdef-c159ce1d22af
['1db62ca6-188d-86f0-7c8a-9813703e9b37']
Hi All, Can someone clarify the usage of the 'sched_hangup' application in mod_dptools. In the wiki it suggests the timer starts when the call is answered, however testing proves this not to be the case and the call is released while in the ringing state if the timer is short enough. Or should it be used in conjuction with channel variable 'execute_on_answer'. Thus forcing setting of the timer at this point of the call/ If you need the debug or any other info please let me know. Thanks, <PERSON>
8a36ccb8-0f5b-19c0-0632-ae4c8005b44e
['1db62ca6-188d-86f0-7c8a-9813703e9b37']
Hi All, I am using call-timer app and found that it does not forward any CANCEL messages to the 'B' leg of the call.  Based on the debug and wireshark, SEMS replied to the CANCEL on the 'A' leg with a 200 Ok. It then sets the timer to 0 and waits for the 200 OK with SDP from the 'B' leg and then will send a BYE to the 'B' leg. I would have expected SEMS while working in B2B mode to forward the CANCEL to the 'B' leg of the call and stop the terminating phone from ringing.   Does anyone else have some thoughts on how SEMS should handle this scenario? Any help in debugging would also be helpfull. Regards, <PERSON>
4dcd5d82-f514-b8fc-529c-7fedb3cc01c3
['1dc7b569-2963-ed5a-c2cf-7bea67110c78']
RESEARCH ASSISTANT, Right Whale Habitat Studies The Provincetown Center for Coastal Studies (PCCS), a private, non-profit research, conservation, and education organization based in Provincetown, MA, is seeking a full-time Research Assistant for the 2009 winter/spring season (Jan. 2 – May 15) of the Right Whale Habitat Studies Program. The successful candidate will play an integral role in the rapid assessment and reporting of habitat characteristics that influence right whale distribution in the Cape Cod Bay Critical Habitat. Primary responsibilities include extensive time at sea conducting zooplankton sampling, microscopic identification and enumeration of zooplankton samples, right whale behavioral studies and photo-ID, and database management/data entry. Field work will likely occur in harsh winter weather conditions. Additional duties may involve assisting with field and laboratory operation of a Conductivity-Temperature-Depth profiler (CTD) and Optical Plankton Counter (OPC), and with the downloading and analysis of cruise data. Familiarity with the identification of Gulf of Maine zooplankton species is highly preferred, but not required. Experience with Microsoft Access and Excel, GIS software and biological oceanography software/data management techniques are preferred. Candidates must have a relevant 4-year college/university degree (B.A. or B.S.), should be highly motivated and independent, and should have strong attention to detail; completion of a Masters degree or equivalent experience is preferred. This position provides housing and a monthly stipend of approximately $2450. To learn more about PCCS and the Right Whale Habitat Studies Program, and to read field notes from last year's season, visit www.coastalstudies.org. Interested applicants should submit a cover letter and resume with references to: <PERSON>
9242fbf7-bb40-a3ee-641b-71338b3f0020
['1dc7b569-2963-ed5a-c2cf-7bea67110c78']
Dear <PERSON>, Please note the CHANGE IN APPLICATION DEADLINE for the following position. RESEARCH ASSISTANT- Right Whale Habitat Studies The Provincetown Center for Coastal Studies (PCCS), a private, non-profit research, conservation, and education organization based in Provincetown, MA, is seeking a full-time Research Assistant for the 2009 winter/spring season (Jan. 2 – May 15) of the Right Whale Habitat Studies Program. The successful candidate will play an integral role in the rapid assessment and reporting of habitat characteristics that influence right whale distribution in the Cape Cod Bay Critical Habitat. Primary responsibilities include extensive time at sea conducting zooplankton sampling, microscopic identification and enumeration of zooplankton samples, right whale behavioral studies and photo-ID, and database management/data entry. Field work will likely occur in harsh winter weather conditions. Additional duties may involve assisting with field and laboratory operation of a Conductivity-Temperature-Depth profiler (CTD) and Optical Plankton Counter (OPC), and with the downloading and analysis of cruise data. Familiarity with the identification of Gulf of Maine zooplankton species is highly preferred, but not required. Experience with Microsoft Access and Excel, GIS software and biological oceanography software/data management techniques are preferred. Candidates must have a relevant 4-year college/university degree (B.A. or B.S.), should be highly motivated and independent, and should have strong attention to detail; completion of a Masters degree or equivalent experience is preferred. This position provides housing and a monthly stipend of approximately $2450. To learn more about PCCS and the Right Whale Habitat Studies Program, and to read field notes from last year's season, visit www.coastalstudies.org. Interested applicants should submit a cover letter and resume with references to: <PERSON>
759d1291-dfbd-2454-f356-95840852973a
['1dd934b1-91a2-4a49-e34d-2538d5ec65c9']
I am having problems inserting pictures. It seems if the picture is too large for the frame, their is no way for me to "shrink" the picture down to fit the frame. I have not had this problem in Publisher or other programs, so it is probably something that I am doing or not doing. I am on the final layout, so if anyone has a quick answer, sent directly to me. thanks <PERSON>
ceedad39-cae1-6b20-79f4-8eda9a71782a
['1dd934b1-91a2-4a49-e34d-2538d5ec65c9']
Hi, Last week I downloaded 1.3.3.9. I have gone over the tutorial, however, my needs at this point are really basic. For instance, I need to have my text flow from one page to the next. I think I have done what the tutorial tells me to do, however, I just get the origianl text pasted in teh new page. I am 72, sort of computer literate, but I need things explained in basic everyday language, as I don't always understand the short cut intial-type instruction. For instance I have not the vagues idea what SVG or eps mean. Also, I don't understand exactly how to set up the style for headings. I assume I can set a a style for the headings, and each time i want a heading I should be able to call on that style -- at least once I figure it out. Please cut an old lady some slack, and help me with some very basic problems. Also I too have had the problem with the making changes in the story editor only to find something different on the document. I don't know if this is something I am doing, not doing, or perhaps it is a program problems. thank you for any generous person who is willing to help me. <PERSON><IP_ADDRESS>. I have gone over the tutorial, however, my needs at this point are really basic. For instance, I need to have my text flow from one page to the next. I think I have done what the tutorial tells me to do, however, I just get the origianl text pasted in teh new page. I am 72, sort of computer literate, but I need things explained in basic everyday language, as I don't always understand the short cut intial-type instruction. For instance I have not the vagues idea what SVG or eps mean. Also, I don't understand exactly how to set up the style for headings. I assume I can set a a style for the headings, and each time i want a heading I should be able to call on that style -- at least once I figure it out. Please cut an old lady some slack, and help me with some very basic problems. Also I too have had the problem with the making changes in the story editor only to find something different on the document. I don't know if this is something I am doing, not doing, or perhaps it is a program problems. thank you for any generous person who is willing to help me. Joan
3d6992a5-9015-8252-b6f2-51df668aaef4
['1dee1643-6120-0846-412e-cacc5556f382']
Hi, For my portal page I need a layout where left side is smaller than the right side. Jetspeed-2 have "VelocityTwoColumns2575" and "VelocityTwoColumnsSmallLeft " layout. No matter what layout I apply to my page, it is always divided into two columns of equal length. Do I need to specify some properties for layout fragment for my page? Can someone point me what wrong I am doing? Regards, Raj
8eb4d255-49ce-46f7-0564-6127f4371f4f
['1dee1643-6120-0846-412e-cacc5556f382']
Hi, I am developing portlets based on jsf bridge. I have a form with input text field and it's required attirbute set to true. When I submit this form with empty input value, validation error message should be printed at <h:messge> tag location. This seems to be not working in jsf bridge based portlets. It looks to me that fields are being validated as I don't see the action handler's process action method called however, validation messages is not printed. Have someone experienced similar problem or I am missing something? Regards, Raj
239f5e0e-2698-3a27-5947-5820c2bd6fbf
['1df0bc2d-1da8-fc1a-3b8e-198259a0de59']
Using a quote for another beginner blender dev :) "I am a beginner in Blender development and this is my first patch. Expecting comments and suggestions from you." (<PERSON>) This patch adds the two keys that don't work in my abnt2 keyboard (with 107 keys). For now only for win32. An advantage of this patch is that don't modifies the normal functionality of original source, just adds two lines in one switch/case, and respective ifdefs on begin of file. Patch link: http://www.pasteall.org/11733/diff P.S: sorry my bad English. I hope it's helps. Thanks!
b558336f-cb96-7d99-da3c-b1afd7e44335
['1df0bc2d-1da8-fc1a-3b8e-198259a0de59']
Hello, I'm creating a binding for one library that have one class with mutable field, which the best way to bind this field? .def_readwrite work only for non const class. With the error message for const instance: "No matching overload found, candidates: void <unknown>(CMod1&,int const&)", where the mutable field is an integer. .property work, but is this the only solution for this case? Thanks,
c00cf569-3c9d-259c-9159-4d66dc2cb182
['1e1c9582-51c6-e959-6b38-1ee6b040843e']
Hi, I have mpeg4 (h.264) video streams encapsulated via mpeg2 transport streams. I'm using the live555 media server as an rtsp vod server. I was hoping to use the 'MPEG2TransportStreamIndexer' tool. But when I run it I get a .tsx file with zero length so it looks like it only operates on mpegts payloads that are mpeg2. Is there any other way/tool I can use so I can use trick play features of the live555 media server? BTW: The live555 has worked flawlessly as an rtsp server, not once ever crashed. Thanks, <PERSON>
63cd08b8-04b1-0300-0203-2cddee986e4e
['1e1c9582-51c6-e959-6b38-1ee6b040843e']
I found your project on the ClamAV website. What I'd like to know is what are the general limitations using this type of scheme for anit-virus protection for LAN workstations behind the firewall? For example, when using a web proxy anti-virus solution the software must completely buffer long web file downloads before it can do virus scanning. I don't see how this could be done using iptables were you have to "vote" on a packet-by-packet basis. Secondly, is the current inline snort version suitable for production use? Thx, <PERSON>
0f6e1f0f-1710-25be-06fc-aca492b42bd8
['1e2652ee-d12e-419e-ad85-ab805b297dd7']
John Vosmek questioned (in a footnote) the validity of Mr. Eden's statement that "only 10% of our customers are using" the catalog. I'd like to question Mr. Eden's statement that "a library spends roughly 60-70% of its budget on personnel salaries and benefits in technical services, OCLC fees, and vendor OPAC fees." Basic question: is Mr. Eden referring to a library's total budget or its operational funds? It's an important distinction that no one wanting to invoke a business model should ignore. Assuming that the statement* refers to operational expenses and is correct (which I doubt but have yet to investigate at my institution), I have to wonder what percentage of that sum applies to *cataloging*? "Technical Services salaries" certainly includes people in Acquisitions, and depending on the institution may also include folks in serials, physical processing, government documents and/or others. OCLC fees -- unless Mr. Eden has eliminated them from his calculations -- include money for resource sharing (ILL), access & support (connectivity), regional network fees, and public use of WorldCat, various subscription databases, and other online resources. The expensive "vendor OPAC" is (I hope) a lot more than an OPAC at Mr. Eden's institution. A robust ILS supports acquisitions, billing, circulation, inventory, serials receipt & claiming, and statistical analysis of operations. Only part of the ILS cost supports the OPAC, the rest is an essential *cost of doing business*. Mr. Eden appears to favor a "business-model" for running libraries. (Although in discussing changes at LC, Mr. Eden states "Politically, LC must put its resources into mass digitization." Apparently, business models and cost-benefit analysis only apply to some operations.) Fine -- for those select operations, a business has to have control of inventory, efficient ways of billing and paying bills, ways of acquiring the things customers want, etc. -- all the things that are part of technical services costs. I've just spent days in "business-model mode" writing my department's annual report (rather than cataloging) and find that over the last 9 year, an average of 52% of the nearly 10,000 books we've bought for our circulating collections have been borrowed within their first year in the library. I think that's pretty good value for those technical services costs. Could Mr. Eden's library run at all if technical services salaries & benefits, OCLC costs, and his ILS were eliminated? Why does he put all that baggage on the backs of catalogers and blame it on our "infatuation with order"? Perhaps the reason there was deafening silence on Autocat when the Mar. 2008 American Libraries hit our desks was because we're just so very tired of this type of undocumented, unsupported article? Ann Kebabian, Head of Cataloging Case Library and Geyer Center for Information Technology 13 Oak Drive, Hamilton, NY 13346 Phone: 315-228-7309 Fax: 315-228-7943 Email: <EMAIL_ADDRESS><PERSON> questioned (in a footnote) the validity of Mr. <PERSON>'s statement that "only 10% of our customers are using" the catalog. I'd like to question Mr. <PERSON>'s statement that "a library spends roughly 60-70% of its budget on personnel salaries and benefits in technical services, OCLC fees, and vendor OPAC fees." Basic question: is Mr. <PERSON> referring to a library's total budget or its operational funds? It's an important distinction that no one wanting to invoke a business model should ignore. Assuming that the statement* refers to operational expenses and is correct (which I doubt but have yet to investigate at my institution), I have to wonder what percentage of that sum applies to *cataloging*? "Technical Services salaries" certainly includes people in Acquisitions, and depending on the institution may also include folks in serials, physical processing, government documents and/or others. OCLC fees -- unless Mr. <PERSON> has eliminated them from his calculations -- include money for resource sharing (ILL), access & support (connectivity), regional network fees, and public use of WorldCat, various subscription databases, and other online resources. The expensive "vendor OPAC" is (I hope) a lot more than an OPAC at Mr. <PERSON>'s institution. A robust ILS supports acquisitions, billing, circulation, inventory, serials receipt & claiming, and statistical analysis of operations. Only part of the ILS cost supports the OPAC, the rest is an essential *cost of doing business*. Mr. <PERSON> appears to favor a "business-model" for running libraries. (Although in discussing changes at LC, Mr. <PERSON> states "Politically, LC must put its resources into mass digitization." Apparently, business models and cost-benefit analysis only apply to some operations.) Fine -- for those select operations, a business has to have control of inventory, efficient ways of billing and paying bills, ways of acquiring the things customers want, etc. -- all the things that are part of technical services costs. I've just spent days in "business-model mode" writing my department's annual report (rather than cataloging) and find that over the last 9 year, an average of 52% of the nearly 10,000 books we've bought for our circulating collections have been borrowed within their first year in the library. I think that's pretty good value for those technical services costs. Could Mr. <PERSON>'s library run at all if technical services salaries & benefits, OCLC costs, and his ILS were eliminated? Why does he put all that baggage on the backs of catalogers and blame it on our "infatuation with order"? Perhaps the reason there was deafening silence on Autocat when the Mar. 2008 American Libraries hit our desks was because we're just so very tired of this type of undocumented, unsupported article? <PERSON>, Head of Cataloging Case Library and Geyer Center for Information Technology 13 Oak Drive, Hamilton, NY 13346 Phone: <PHONE_NUMBER> Fax: <PHONE_NUMBER> Email: yAj5vIHJDN8mDXvp@example.com
e72b014e-5ed8-431b-1e1e-22f07f790a10
['1e2652ee-d12e-419e-ad85-ab805b297dd7']
Classification of films is a free-for-all, since LC does not classify their films, and therefore the LC classification has no good place to put them. The choices appear to be: 1. PN1997 & PN1997.2, the LCCN for screenplays 2. PN1995.9, special topics, with a cutter for the genre (frequently) or special topic (less often) 3. Either of the above, but adaptations of pre-existing literary works classed with the original. At Colgate we started classifying all fictional films in PN1997 - PN1997.2. The advantage is that it keeps the screenplay, critical works about the film, and videos of the film all together. The disadvantages for us were l-o-n-g cutters, and a less useful arrangement for faculty & students in literary studies who are looking at the film & text together. We started classing everything with the original literary work if there is one, adding a '2' (abridgement or condensed version) to the work cutter. We've also recently found that using the publication date of the video in the call number is problematic, as when we have several different versions of Wuthering Heights or remakes of earlier films. We're starting to use the production date of the original movie with work letters, which seems to make more sense for our users as well. ("I'm looking for a 1930s musical...") It's a mess, and I sometimes wish we had gone the PN1995.9 route, which would have helped our casual users who want an arrangement more like a video store... With films, it often seems like we can please none of the people any of the time! Ann Kebabian, Head of Cataloging Colgate University Libraries 92 Broad St., Hamilton, NY 13346 Phone: 315-228-7309 Fax: 315-228-7943 Email: <EMAIL_ADDRESS><PERSON>, Head of Cataloging Colgate University Libraries 92 Broad St., Hamilton, NY 13346 Phone: <PHONE_NUMBER> Fax: <PHONE_NUMBER> Email: yAj5vIHJDN8mDXvp@example.com
5e432851-3856-bdb2-ec08-9a01d10a45a0
['1e33a121-c2ab-858b-ee2d-715db5d68ded']
Hi, I was able to get my new Mifos dev server running, however I am getting 19 JUnit Test Suite failures attempting to build my Mifos test server. I did a fresh Mifos checkout night (Thursday) prior to the build. Attached is the test xml file and the Mifos log in case it helps. There are several exceptions in the latter. Does someone know what would be causing these failures? As far as I can tell both my dev and test servers are set up the same, both with Mifos checkouts done tonight. Thanks, <PERSON>
070544cd-d14c-5647-ec07-8519d169162b
['1e33a121-c2ab-858b-ee2d-715db5d68ded']
Hello, I am still getting build errors. Attached is the test suite xml file. My version of mifos is 12171. My environment is: O/S XP, SP2 JBoss 4.0.5 Sun JDK 1.5.0_08 My SQL 5.0.24 Apache Ant 1.0.7 Does someone know what could be causing out problems? Help would be greatly appreciated. Thanks, <PERSON>
481335df-4ce1-1743-f89e-e14dea8b4e04
['1e4f3fcb-fad9-1cb5-8308-3565f1b3b337']
I can't figure this out. I get a NullPointerException in the following circumstances: - using a JXTable to display an EventList - pipeline chain is EventList -> SortedList - JXTable has TableComparatorChooser.install() on the table - when the following user actions are executed: -- sort table by one of the columns -- delete two of the rows -- page down to the end of the table -- hit page up, and BOOM, I get the exception:
241eecb5-521e-584c-5d31-4af83f074eb1
['1e4f3fcb-fad9-1cb5-8308-3565f1b3b337']
I have a virtual server running Windows Server 2008 R2 Standard SP1. I have successfully installed TortoiseHg, Java, Python, and several other programs. I have tried installing both the Windows Python installers for Mercurial (both the "Mercurial 2.5.1 Python 2.7 package - x64 Windows" and "Mercurial 2.5.1 Python 2.7 package - x86 Windows"), and both hang and crash when the install goes to start. (when the install wizard says "Ready to install" and I hit the "Next" button) The installer finds the correct version and location of Python (2.7.3 in my case; I have both x86 and x64 installed). Any idea what I can look for to figure out what's going on? <PERSON>
78ea275a-ab28-a911-8d2e-71d90bf662df
['1e754630-1da8-64cb-8f84-5c8d370d8441']
Has anyone had any luck getting the addon's to compile under OS X? I have been able to get 1.06 to build and run great, but I really want to get the cdr_addon_mysql.so to enable mysql writing. When I try to compile I get a bunch of Undefined symbols and I have tried adding everything to my path. I've got mysql_devel, etc. I have posted this to the Mac Asterisk mailing list about a week ago with no response, so I thought I would just check in here. Thanks,
c0a1bd5d-8dd0-85ce-d8da-6dc119342ac1
['1e754630-1da8-64cb-8f84-5c8d370d8441']
This is driving me crazy, when making an outgoing call the first 30 seconds is always perfect, then the party on the receiving end can always hear me perfectly but after 30-60 seconds the audio coming back to me from them starts to get choppy and drops out. I've tried this with multiple devices, from multiple locations some behind NAT, others not. This is using the ulaw codec, although i've tried it with alaw as well. Problem happens via IAX as well as a SIP channel, both calling PSTN numbers. Network performance to the Asterisk server is good, 15-20ms, performed ping tests for 1-2 hours with almost no packet loss. I'm willing to check or post anything needed here, but I need some fresh ideas since i've checked everything I can. Regards, -J
be538071-f5fa-e6eb-a5ec-082d61125f5b
['1e824b9c-b931-66ee-bb35-77111c82457b']
Dear Modellers, ScrumPy is a package of Python modules that allows you to do metabolic modelling. You don't need to know about Python, Programming, or Modelling or Metabolism to use it (2 out the 4 should suffice, any more and you're already a power user). Current status is 0.9.0, it is IMHO in a state capable of being useful, but I would like to hear your opininions before bumping up to 1.0. Full details here http://mudshark.brookes.ac.uk/ScrumPy Regards, Mark
f0b20adf-3ce5-e36c-70b6-0c6a922fa259
['1e824b9c-b931-66ee-bb35-77111c82457b']
OK, very sorry, it's not a driver bug, I'd accidently moved a small slider on the front of the case which duplicates Fn/F5. Although I feel a bit silly about it, it's a truly awful piece of ergonomics: a black switch on a background located in a position that you can't see, but can easily brush against, that duplicates and over-rides another more visible control with equivalent function, and is not mentioned anywhere in the paper documentation (at least that I received). Is there any possiblity of enhancing the driver to ignore or overide it (or at least put a message in the logs) ? I can't believe that I'm the only person whose going to get caught out by this - or maybe I am ... :-( Thanks for the attention,
be6fbb56-8b5f-ff57-5b73-9e60598bf7a9
['1e9868b1-527e-2421-dcbf-f3372e3a24fe']
Good day, Further to previous mail, pls note that SOGO log shows the following errors: - "[ERROR] |SOGo| did not find Locale for language: Russian" - [WEStringTableManager] > missing translations directory - [WOComponentDefinition] > Note: did not find component class with name 'SOGoFolderRussianAdditionAdvisory' May the error 'did not find Locale for language' be the reason for improper Calendar behaviour? Other matters related to Calendar: I can create new calendar via web UI although I am unable to delete it afterwards. Kind regards, <PERSON>
3ad2fd86-d6f3-017e-241e-0b664e559843
['1e9868b1-527e-2421-dcbf-f3372e3a24fe']
Good day, The SOGo 1.0.1 is installed with CentOs5.2. The web interface does not show days in ''weekview" mode, i.e. the numbers 2, 3, 4 of the days are not visible in column headers. Looking at Calendar options in Preferences I may see that the setting "Week begins on" contains the list of <NIL> instead of week days names. Any suggestions how to resolve this? Best regards, <PERSON>
7f673d79-3a26-ef79-504b-3fc20308348c
['1e9f69d4-ce36-0db6-c058-6043294fb28a']
Hi, I have a customer who has a NAND device with two MTD partitions and each of the partitions contains one UBI volume with a UBIFS filesystem. Now U-Boot can mount the UBIFS from the first partition just fine, but if the UBIFS from the second partition is mounted afterwards this fails in some cases. I can reproduce the error and tracked it down to uboot_ubifs_mount() in fs/ubifs/super.c. If this function is run for the second mount, the struct ubifs_fs_type is reused and it contains a list fs_supers, that still holds one entry for the first mount. I guess, that if the second mount would happen on a volume that is on the same MTD partition as the first volume, than this will work. The second entry is added to ubifs_fs_type.fs_supers. In my case however, the second entry being added to ubifs_fs_type.fs_supers is invalid and causes the mount error. Reinitializing the list in uboot_ubifs_mount() before each mount, solves the problem, but I guess that it will cause failures in other setups, where there are actually multiple volumes on one MTD device. So how can I solve this properly? Do we need one instance of struct ubifs_fs_type for each MTD device? I tested this on an old version (2017.03), but looking at the current code, it looks like the same problem applies to current mainline. Thanks for any help and suggestions! <PERSON>
d81412c4-84a2-f851-80b4-50a957ff24dd
['1e9f69d4-ce36-0db6-c058-6043294fb28a']
Hi, I have a board with i.MX6 and a WM8510 codec and I can play audio using an old vendor kernel (3.14) and my own asoc machine driver [1]. Now I want to run this with something more recent (currently trying on 4.14) and upstream drivers like simple-card or fsl-asoc-card. The problem is I'm only getting a little bit of noise from the speakers but nothing more. I would really appreciate it if someone more knowledgeable could have a quick look and maybe come up with some ideas how I can proceed debugging. Thanks, <PERSON>
fc27a6e2-6e0f-0f9a-972d-298ac818e31b
['1ea45290-debb-96de-c7c9-67d65793f5a0']
Hello all: I am going through and optimizing one of our applications, and I have found something a bit interesting. In some of our scenes, we have thousands of models that could be displayed and each of them have very high resolution textures. To help with this, we are using a large number of screen size LODs in an attempt to reduce the graphics resources needed at any one time. What surprised me though (although it probably shouldn't have) is that once a texture or display lists (or VBOS/FBOs) are allocated for a geometry and material, they remain allocated even though they may be in a sub-tree of the LOD node that is not currently active. This leads to the unfortunate side-effect that we run through 1.5 GB of texture memory (and system memory) *very* quickly when flying over the models in our scene because each one loads the highest resolution texture at some point and leaves it loaded in the system. So, what I propose doing is modifying the screen size lod to include a flag that tells the system to unload all graphics resources when a sub-branch is deactivated. Unfortunately, I don't know how to do this. Can anyone point me in the right direction? Thanks, <PERSON>
db7f49e5-f6fa-f2d0-67b6-ce1cb441b05f
['1ea45290-debb-96de-c7c9-67d65793f5a0']
I have been writing a GUI to connect to trac using XMLRPC. It is all up and running now but I seem to have problems with notification. When I use the XMLRPC interface to make a change to a ticket, it changes the ticket correctly, but it does not trigger the associated notification code to be called. Can anyone point me in the right direction to fix this? I am using the head of the 0.10-stable branch right now. <PERSON>
996df2b8-d2d7-b7c6-505d-e5820f915b1f
['1eadc6a3-3b38-0f99-7349-f94faadbb83a']
Wow!!! <PERSON>, that is working very nicely! The time that must have taken is greatly appreciated. Thank you. The typing in of values and selection works well for me and I am pleased. If you are able to incorporate reference information for the attributes like that I think that is more than enough to supply the detail to help the user choose the right attribute value. One Issue I have found: If you enter any unknown value into an attribute the context does not close correctly. In coldfusion (and other programming languages) we are often entering variable names into attributes: <cfset direction_type="rtl" /><cfoutput><h6 dir="#direction_type#" style="background-color: blueviolet;">Thank You!</h6></cfoutput> In my testing, everything after dir="#direction_type#" looses it's syntax highlighting. Maybe the string used to end context could be a regular expression allowing unknown attribute values to end the context. I'll keep working with the code you sent and see what I can do with it. Thank you!! <PERSON>
58c6448d-d747-fa26-ebe7-417d49725925
['1eadc6a3-3b38-0f99-7349-f94faadbb83a']
<PERSON>, thank you for continuing to give your valuable input. I have written the JSON parser that outputs bflang2 using the "<tag>" options currently available. At this time I have no way to utilize the attribute autocomplete information I have. I believe in making reasonable goals and achieving them. The first is a working cfml.bflang2 file using what we have. See attached files to sample that. I would be glad to modify my JSON parser to format according to a more advanced set of syntax. ..
80ee6a21-0cd2-bb46-da29-89f4492bc1e6
['1eb43ce0-f55b-63f2-ef55-f1570e34957a']
I've been trying to get a simple little scp implementation going that tries to download a specific file. As I understand it, I can provide an interactive authenticator that will be used to prompt the user for login information, if required. However, when I use this, my Challenge implementation is simply not called. I'm on Windows, if that makes a difference. Here's the code. Please forgive the quality - this was something I threw together in about 15 minutes. I would appreciate any pointers in the right direction. Cheers, <PERSON>
7e77fc86-9a50-eaf3-01d9-d71cc0c1af31
['1eb43ce0-f55b-63f2-ef55-f1570e34957a']
Hi, I'd like to understand why a conversion from a channel of typeA to a channel of InterfaceB isn't automatically handled, when typeA satisfies InterfaceB. A function that takes InterfaceB as an argument is perfectly happy to take TypeA in as a parameter. However, if the function accepts a channel of InterfaceB, it cannot be given a channel of TypeA. Here's an example: http://play.golang.org/p/CNjVmMkXeK Given that the channel conversion function isn't really doing anything smart, is there a good reason that channel type conversions aren't automatically handled? Cheers, <PERSON>
b9bbbde1-75ee-7c05-de8f-45817d3758d8
['1ec4e475-0e45-dcd8-b0da-60be820d3de4']
Well, I solved it even though I have no idea what was going on. After disabling UAC, I happened to glance at the httpd.conf again and it was back to the default. I edited it AGAIN to set the correct configuration and lo and behold, now I can run httpd as a service. Did UAC do that? If it did, it's just plain evil... <PERSON>
89cba799-d480-d8cf-898e-fb9eff54238a
['1ec4e475-0e45-dcd8-b0da-60be820d3de4']
I installed Apache as a service as normal, then tried to run it. It ran...but it won't read the httpd.conf file. Even changing the service to call httpd.exe -k runservice -f (path to conf file) *does not work*. I can only run httpd manually, from the command prompt, with the -f parameter, and get it to read the conf file. Why??? FYI, nothing shows up in the error log to explain this. Also, apparently the registry key pointing to the proper path IS set (HKLM\SOFTWARE\Apache Software Foundation\Apache\2.2.11 ServerRoot) but isn't being used. <PERSON>
1dd3d322-80f1-952f-d88a-eed711070029
['1eca52cd-e66a-179f-fe55-c1b840ab8943']
Hello, Here are 3 patches to the mouse issues in the Help Viewer. 1) mouse-rightbutton.patch Returning to a previous node by pressing right mouse button doesn't work in xterm. It works only in Linux console with GPM. The problem is that help_event() catches only GPM_UP event, and it seems that xterm doesn't report which mouse button was released. Handling GPM_DOWN instead of GPM_UP will fix it. 2) mouse-offbytwo.patch Last two lines of the help window (the bottom line and a frame) are not mouse-clickable. So the user will not be able to follow a link with a mouse, if it is situated at the bottom line. And the frame needs to be clickable because it serves a special purpose - page scrolling. 3) mouse-linkfollow.patch When following a link with a mouse, an extra '\n' is insterted at the top of the window (just follow any link with a keyboard, then return back, and follow the same link with a mouse, and you should see the difference). P.S.
254ceb8b-e0bb-078f-682f-8424f82bd9c6
['1eca52cd-e66a-179f-fe55-c1b840ab8943']
Hi all, src/man2hlp.c utility sometimes replaces spaces in the link target with newlines. There is a workaround code in src/help.c:141, which handles that case. But it doesn't always help. If the link is used inside the indented block (made by .TP tag), man2hlp will insert additional spaces inside the link. For example, the link "Directory tree" will become "Directory\n Tree". This will result to a broken link. Currently, in Russian help file there are two such broken links (one in "Listing Mode..." section, and the other one in "Options/Virtual FS" section). The result of trying to follow such broken link is displayed at this screenshot: http://img227.imageshack.us/img227/179/mcbrokenlinkmo3.png It is lucky that there are no such broken links in English help, but they can appear in the future. So I made a patch for the man2hlp utility, to make man2hlp never insert newline and spaces inside the link target. Only the link target (the invisible part of the link) is affected by the patch. The link name itself (that is visible in the Help Viewer) can always be wrapped to another line, and it never causes problems. Regards, <PERSON>.
ec3cb15e-685d-95e3-27be-3ee35298e8e3
['1ecd8df0-4c81-012e-6aef-0a46f8d1383e']
Hi everyone, if a dGrid has focus, the default behaviour if one hits -enter- is that the cell cursor goes one row down. Right now i writing a 'navigation dialog' (to navigate to a dataset) with just a grid and two buttons (OK, Cancel), and I want the OK button to be pushed when the user hits -enter-. Setting the 'DefaultButton' property of the button to True doesn't work (because the Grid catches the event before the dialog ?). Which event of dGrid covers the hit of -enter-? I tried the KeyEvent and KeyChar of the Grid, neither works. Is there a possibility to tell the Grid 'ignore that event and put it through to the form'? As you might guess, I'm doing my first steps in gui programming, especially with dabo. I'd appreciate it if someone could point me to the right direction. Best Regards <PERSON>
9aed8912-ea7e-c5c5-fea8-b7ac523d9f10
['1ecd8df0-4c81-012e-6aef-0a46f8d1383e']
Hi, I recently installed x2go, and the feature that I really miss is the possibility for session shadowing. It would be great for me to help my users from a remote location if they have a problem. Is something like this planned for a future release? One other question: is it possible to connect a usb device other than memory sticks to the client and "tunnel" it to the server? I'd like to use a card reader for the home banking software (money plex) with x2go. Greetings <PERSON>
deed9b79-aa04-5eb5-d967-6e9f7142a842
['1ed5bb85-0c1c-6fae-8316-4d8a8edd133c']
Hi. CTRL+Space and I found the "Editor: build file or project", but it doesn't seen to be what I'm looking for. I'm doing some Appcelerator Titanium project and I would like to instruct Light Table in how to compile and launch my project/workspace in the Android emulator (a couple of shell commands), to speed up the coding/testing routine, but can't find how to accomplish that. There is some magic command or plugin that offer that possibility? Thx.
674e6e9b-efc1-7aaf-352d-38e7274300f3
['1ed5bb85-0c1c-6fae-8316-4d8a8edd133c']
Hi. On Eclipse, when I click over a word, all the occurrences in the file get highlated, plus a mark is showed besides the scroll showing the lines in the file that are the occurrences. On LT the ocurrences get underlined in blue, but I can't see how to find out if there is other occurrences in other parts of the file besides the little piece I'm looking to ATM. There is some plugin or config that show me that info? Thx.
21770d66-327f-03e4-a422-ac8a76a1ed49
['1edaaabb-5652-1915-3146-2b0bcde79deb']
Hi all, I generally like to keep a little manifest file with my rootfs images, containing a list of installed packages and their associated versions. On images where I keep package data around, I usually generate this via 'opkg list'. In my current situation, I have a small read-only image where I can't afford to ship pkg data, and wanted to gather this info on the host side of things. I assume this information is tracked somewhere in the poky build directory, but I'm having some trouble finding it. Could someone kindly point me towards this information or the documentation section(s) that I may have overlooked. Thank you! <PERSON>
93c822cb-61e5-8e9f-e185-2b838258c536
['1edaaabb-5652-1915-3146-2b0bcde79deb']
Is there a simple way to disable the use of PREMIRRORS and MIRRORS within a recipe? (Perhaps the answer here might be worth mentioning in Section 12.23 of the Poky Reference Manual?) My use case for this is the situation where the code hasn't been released yet, so there's no point in checking mirrors. Thank you, <PERSON>
fa94e398-61b1-b86f-8347-df82f4e7e979
['1ee1efb1-9b32-f178-7b4a-69314a717f2f']
Dear, I suddenly think one question on CMS, why need to use CMS or Midgard to build website?? I am Midgard beginner and need to change existing website using CMS; however, I think midgard is so difficult and complex to use. Anyone can tell me what is the advantage of using CMS or Midgard?? Use CMS or Midgard no need to write code when build website?? Before I use CMS, I guess no need to write so much code when building website; however, when I learn how to use it recently, I think the structure is so complicated. <PERSON>.
3efbdd3e-3c64-7173-99fe-4764e994a416
['1ee1efb1-9b32-f178-7b4a-69314a717f2f']
Dear, If I want to add company's upcoming event into database and need to get the most latest in the website. Also, I need to provide interface to our staff to update events(not the schedule event in mdigard, is company event and need post it in web).Then, need to create new table to store it?? Also, need to build the interface for our staff by myself?? or midgard provides this function. Thanks. <PERSON>
42fc9cf0-d551-26bc-d7f8-71f87eb90e7e
['1ee57fb5-145a-b3f5-2d79-c70a511a00e9']
IVR question: Users dial my DID numbers and get connected to macros and other vectors that guide them to the appropriate context. Once connected to a specific context I would like to send a text message to their phone. Do I need a PERL script or is there something native in Asterisk 1.6 that can trigger a text to the endpoint? Thank you
5751b49d-52e3-26e2-0043-cb8a52b9b6a1
['1ee57fb5-145a-b3f5-2d79-c70a511a00e9']
This may be an obvious reflection of my Asterisk/Linux/Windows weaknesses but I want to know in any case! Can a vb script run somehow on a Linux machine or does it only work on Windows? If I were to build a call file script (described in this link http://www.voip-info.org/wiki/index.php?page=Asterisk+auto-dial+out ) then how does it work if my Asterisk machine is running on Centos 5.5? I simply want to execute a script that helps me automate the voice broadcasting/IVR of up to 10000 phone numbers. Thank you <PERSON>
104c41e3-1b35-92c6-59cc-875a3b99975b
['1f1a61cc-dac0-8c48-4240-dc2ec036fc20']
Hello, I am new to UEFI development, so any help or advice you could provide would be appreciated. I have loaded the BeagleBoard build onto a BeagleBoard, and I am trying to access the UEFI shell. I am able to boot into the EBL environment, but I cannot figure out how to start the shell from EBL. On the wiki, it mentions that this is possible, but I cannot find any instructions for doing so. Can anyone provide me instructions on how to launch the EFI shell from EBL, or point me to a resource that would help me? Thanks so much, <PERSON>
5ca41272-0279-3c59-2faf-f2167943ba9b
['1f1a61cc-dac0-8c48-4240-dc2ec036fc20']
Hello, I am working on a UEFI project on the BeagleBoard, and I am trying to flash it to NAND so that the BeagleBoard can boot directly into UEFI. I have successfully built the BeagleBoard package, and I have been successful in running the BeagleBoard_EFI.fd file from RAM using the instructions here: http://sourceforge.net/apps/mediawiki/tianocore/index.php?title=BeagleBo ardPkg#Booting_EDK2_in_the_Beagle_Board_DRAM_using_U-Boot However, I have not been able to load the BeagleBoard_EFI_flashboot.fd file into RAM and boot successfully. I first tried the u-boot instructions, found here: http://sourceforge.net/apps/mediawiki/tianocore/index.php?title=Beagle_B oard_Wiki#Put_edk2_code_in_the_Beagle_Board_NAND_using_U-Boot This bricked the board, and forced me to do a recovery. I then tried the EBL instructions, located here, running EBL from RAM as noted above: http://sourceforge.net/apps/mediawiki/tianocore/index.php?title=Beagle_B oard_Wiki#Put_edk2_code_in_the_Beagle_Board_NAND_using_U-Boot This seemed to have no effect. My BeagleBoard is revision C5, which may be newer than the BeagleBoardPkg was targeted for, but I cannot find this documented anywhere. If anyone can provide advice, instructions, or point out potential errors in my process, any help would be much appreciated. Thanks, <PERSON>
bde707f5-b3d0-e516-7cf0-48fd24863f58
['1f207961-7988-e69f-a475-9730300511dc']
We have been using the Colibri covers for about two years. It's a huge time-saver for us, the covers haven't yellowed, and they are easy to replace (they're not glued or taped down.) It's true that the covers are a constant cost, but I believe that is offset by the time saved by our staff. We only use it to cover paperbacks (the ones that cost $4.99 and up) and donated hardcovers as our book jobber covers our purchased hardcovers with Mylar jackets as part of our contract with them. <PERSON>
45c1b917-b982-031c-4826-d8580ffa2422
['1f207961-7988-e69f-a475-9730300511dc']
<PERSON>, We (I) implemented the Acquisitions module of our Horizon (SirsiDynix) ILS in 2003. We never had a dedicated acquisitions department before I was hired to create it so we never used the Acquisitions module. Now we wonder how we ever got along without it! We are a medium-sized public with a couple of small branches, but centralized ordering/processing/cataloging/invoicing. Our department consists of three full-time clerks (not paraprofessionals), one part-time cataloger (MLS - 10 hours/week) and one full-time MLS (me!) One clerk receives all the adult materials orders and one receives all the youth/ya orders. They submit them to the vendor, download the MARCs, and enter the line-by-line information into Acquisitions, including the budget line for each item. I set up the acquisitions module, creating the vendor accounts and budget lines for the clerks to use. The selectors (public service librarians) now have a way to see online the following: Each budget line's total allocated amount, amount on order (encumbered), and amount spent. They also have the option of looking at the budget reports online by budget line or vendor account to see what has been ordered, received, and the amount paid for each title. When the items are received, the clerks "receive" them in Acquisitions and pass the packing slips and invoices to me. As I invoice them (line-by-line) I can easily see (and correct) any errors they made in assigning budget codes. When the clerks go to process (barcode/label) each item they can see if the MARC looks complete enough. (I showed them what is acceptable and what isn't.) Those items with incomplete MARCs are passed on to the cataloging shelf which the part-time cataloger and I work on. For your information, I also supervise a clerk receiving all our serials/periodicals and our Interloan system which I also set up. <PERSON>
834633d1-1dfc-a608-8d0e-54c52e5ddac5
['1f21e7d4-76f4-5166-ea66-ff4a3802a31d']
Hi, Actually, I use the wsdl2java tool only to get the Data transfer objects from the wsdl. The services.xml is actually handcoded by me. The steps that I follow are as follows: 1. Run the WSDL2Java first to get the Data Transfer objects.(It's is a one time activity) 2. We take the created objects and code the web services based on them (DTOs created). 3. The services.xml is also written as the one I attached in my previous mail. 4. We create the .aar file and then deploy it. As I'm new to Axis2 , I'm not sure as to where its going wrong. Thanks, Regards, <PERSON>
f519d610-d646-d13f-20bf-180972b3643b
['1f21e7d4-76f4-5166-ea66-ff4a3802a31d']
Hi, I've a web service that uses a wsdl of doc-literal style. I use Axis2 1.3, Spring and hibernate for this purpose.(Earlier, I had been using Axis 1.4). The method that I follow is to first create the data Transfer objects from the wsdl using the WSDL2Java tool. After that, I code the rest of the service (business, DAO...) and finally, at the time of building, I add the wsdl also into the .aar file. The service comes up on deployment, but on invocation, neither am I able to pass data to it, nor am I able to get anything from it. I use ADB kind of data binding. I'm attaching the services.xml with this mail for reference. Please help me. I got stuck here and am unable to proceed further. Thanks in advance, Regards, <PERSON>
279a655b-6d48-cbba-68fb-c33caa83a9a3
['1f2a072e-103a-6fd2-3147-f867a6316ec2']
Hello, I am trying to call wxListCtrl::GetSubItemRect in wxMAC but I get a compiler error that tells me that GetSubItemRect is not a member of wxListrCtrl. I have taken a look to the sources and GetSubItemRect is not really implemented in /include/wx/mac/listctrl.h. I am working with 2.8.12 version of wxMAC. How can I use wxListCtrl::GetSubItemRect in wxMAC? Thanks in advance. Best regards, <PERSON><IP_ADDRESS>GetSubItemRect in wxMAC but I get a compiler error that tells me that GetSubItemRect is not a member of wxListrCtrl. I have taken a look to the sources and GetSubItemRect is not really implemented in /include/wx/mac/listctrl.h. I am working with 2.8.12 version of wxMAC. How can I use wxListCtrl<IP_ADDRESS>GetSubItemRect in wxMAC? Thanks in advance. Best regards, Ferran
bec0da2c-3712-e315-ecbc-61655815414a
['1f2a072e-103a-6fd2-3147-f867a6316ec2']
Hello, I use wxFTP::SendCommand method and the command contains non-ascii characters (a file name). When I do this my application crashes. The crash is in the following line of the wxFTP::SendCommand method: if ( Write(wxMBSTRINGCAST tmp_buf, strlen(tmp_buf)).Error()) precisely, in the strlen call. I think that the crash is due to the use of strlen function (no unicode) with a unicode string (tmp_buf). I am compiling with UNICODE define and I'm using unicode version of wxWidgets libraries. Is it a known bug or i'm missing something? My wxWidgets version is 2.8.10 and I'm under Windows XP. Thanks! <PERSON><IP_ADDRESS>SendCommand method and the command contains non-ascii characters (a file name). When I do this my application crashes. The crash is in the following line of the wxFTP<IP_ADDRESS>SendCommand method: if ( Write(wxMBSTRINGCAST tmp_buf, strlen(tmp_buf)).Error()) precisely, in the strlen call. I think that the crash is due to the use of strlen function (no unicode) with a unicode string (tmp_buf). I am compiling with UNICODE define and I'm using unicode version of wxWidgets libraries. Is it a known bug or i'm missing something? My wxWidgets version is 2.8.10 and I'm under Windows XP. Thanks! Ferran
c73f7c91-4758-fbbc-7cd8-e818dd4acf3a
['1f34ec69-d27d-96dd-8f9e-31bc7e1afd9e']
Meeting started at exactly 3 pm. We had some new members. <PERSON>, a young student who has been using Linux for last four years. He shared his experience of installing Linux for the first time and also using Linux in School. <PERSON>, another newcomer, who is an experienced sysadmin and programmer, said that he has been pushing FOSS in his company. We were happy to see an old ILUG member Prof <PERSON> from IIT-B who is also a Debian developer. After the initial introduction, <PERSON> started with the basic introduction of gpg. he explained how symmetric and asymmetric encryption works, the concept of secret keys. he explained how symmetric encryption is coupled with asymmetric encryption in the gpg to optimise performance. <PERSON> showed us how to set-up gpg key pairs and also how to share the public key with other users, he then gave a demonstration of how to encrypt, decrypt and sign a document with gpg. He explained the concept of the web of trust and how to import and sign gpg keys of others. After the demo, there was a small question and answer session. ( Prepared by <PERSON> and <PERSON>) Thanks and Regards -- <PERSON>
40a76529-9fed-502e-5868-39c3ce115c9b
['1f34ec69-d27d-96dd-8f9e-31bc7e1afd9e']
Hi, 47 Supporters so far. Thanks <PERSON> for your comments and support. I met <PERSON> today. He appreciated the petition drafted by <PERSON>, JT and others. He has assured his active participation in this activity along with his gnukhata team - <PERSON>, <PERSON>, <PERSON>, <PERSON> and <PERSON>. We have decided to take small informal sessions in schools and colleges for students/teachers to boost the activity. At the same time we will try to get supporting letters from some well known schools and colleges stating that these institutions themselves do not want proprietary softwares and want to use FOSS. Letters will be sent to MHRD along with petition & signatures. (I know few such institutions, <PERSON> also knows a few. He is going prepare a draft in couple of days. So we can get it printed quickly on letter head after convincing them). We know it is difficult to do all this in such a short amount of time but let us try. (Orators are welcome !) Regards, Milind
7bf103c1-f38b-3d48-333a-fa47aab7b0c9
['1f55634d-6baf-af6a-9512-a90d89283b4f']
Hi gang, Searching for a part for my 'modern' Saab (...which is consistently more finicky than my '62 96 stroker!), and wondering if anyone can point me to any parts houses they've had good luck with. Mine is a 2003 Saab 9-3 (Linear), and the right (passenger) side, upper front motor mount is shot. Apparently it's a hydraulic mount, and the fluid has emptied. I'm told the part is #12785090. Rock Auto doesn't carry it, and eeuroparts lists it as "NLA"...but can special order an aftermarket part for $200 (which is about 30% higher than the original price). Should I just nab this one, or does anyone have any ideas where else I might search? Many thanks, <PERSON>
f0158092-106d-c456-135c-7d3b67ebcaa3
['1f55634d-6baf-af6a-9512-a90d89283b4f']
Hi all, Hope that I'm not stepping out of the bounds of proper etiquette here, but I wanted to let the group know that I'm selling a bunch of nice 96 interior parts on ebay right now. Front and rear seats (burgundy) in excellent shape, door panels and rear side panels (gray on gray) in good shape, hat shelf in good shape, carpet set in okay shape; more to come in the coming weeks. These were salvaged from my nice '63 Bullnose driver, which I purchased from Chip back in '02. Long story short, my cool car lost a battle w/ a minivan, and while the car was a write off, I was able to strip it bare of all its good stuff. I'm back freewheelin', smokin' and buzzing down the road again, in a terrific '62, thanks again to leads from <PERSON> (and <PERSON>). But these extra parts have been gathering dust, so it's time to let 'em go. Feel free to have a look, any of you working on your own rolling restoration. And, again, sorry if I wasn't supposed to use the list for this plug... <PERSON>
9e238d5f-619b-8345-bf2c-556c826196c6
['1f5e542c-d13b-25db-c8d5-c7b7e266ccab']
Forwarded to me by a former student: From a <PERSON>, <PERSON> book: "A gowned <PERSON> .... was particularly pleased with himself that day; and with the Classical Quarterly, whose review of Greek Moods and Tenses (<PERSON>, 204 pp., £45.50, Classical Press) contained the wonderful lines that <PERSON> had known by heart: "A small volume, but one which plumbs the unfathomed mysteries of the aorist subjunctive with imaginative insights into the very origins of language." Cheers, <PERSON>
1ab25ab1-2060-b60e-81a8-57479c022b7d
['1f5e542c-d13b-25db-c8d5-c7b7e266ccab']
Does anyone of my generation remember the old Classics Comics? Like most grammar-school kids I wasn't that happy about the violin lessons I had to take, but I appreciated the stock of Classics Comics Mr. <PERSON> kept in the waiting room. It was the first time I met <PERSON>, <PERSON> and many others - the first time I "read" the Iliad, including the last half-page panel showing the Trojan horse being dragged inside the walls. Cheers, <PERSON>
6a7b4920-d978-2f34-9ea4-68576debbf14
['1f6b0b9a-b501-491e-6bd4-fcf1facd7625']
Hi guys, I'm building a new cocoon app (cocoon2.2) and performance really comes into play so after reading http://developer.yahoo.com/performance/rules.html I'm trying to get an A-grade for all rules ;-) Now I'm focusing on "Add an expires or cache-control header" so I started googling around. I was reading this thread http:<EMAIL_ADDRESS>/msg16951.html about adding support for settings expires headers. But reading the official cocoon documentation I did not found anything back. Does cocoon have built-in support to do so and are there examples/documentation of how to accomplish this? And by the way... anyone who has experience with configuring ETags?? Thx in advance,
cc718165-2f02-2761-2a8e-1ec834c8ccdc
['1f6b0b9a-b501-491e-6bd4-fcf1facd7625']
Hi all, I'm a bit puzzled about how to reference an external javascript file. Let's say I've put it in src/main/resources/external/js/test.js I thought about referencing it the same way as my css file but somehow this does not work. Can someone explain how I can accomplish this? Now adding the <script> line in my xslt gives me an empty html page. Thx in advance, <PERSON>
9f73d89e-c5c1-193a-f0c2-0bcf2889e406
['1f704d52-ae55-27ad-7269-d0bd9f4736ef']
Hi, This patch is built on top of Part 1/2, and is replacing constant heap logic. Oprofile shows that quite a lot of wined3d CPU is used in constant heap maintenance and walks. As DX initially sends constant ranges to be updated, splitting ranges into individual constant updates (through heap versionning) should be avoided if possible. Especially if we intend to send GL ranges at the end. This patch maintains (and somewhat merges) ranges sent to DX, so that walk_constant_heap() (to be renamed ?) can directly issue range updates. Thanks,
738baae9-0401-d51f-6e0d-4b9db591da59
['1f704d52-ae55-27ad-7269-d0bd9f4736ef']
Hi, I am using wine to play World of warcraft (this game is using opengl, and is threaded - 2/3 threads IIRC). I did recently find a reproductable performance regression on versions 0.9.24+. Under the same conditions (always difficult to get in an MMORPG game, but this is reproductible), version 0.9.23 gives me 20/25% more FPS than versions 0.9.24->0.9.26 (have not tried .27 yet). - Is this performance regression known/expected on opengl/threaded applications (new codebase) ? - Is someone using WOW here (and has found a similar preformance regression) ? - It is worth for me to start looking at it (more measures, throw oprofile at it, bissect the change...) Thanks, <PERSON>
8dbf25cb-f49b-7018-972c-1c5c0700dc67
['1f7a71fc-02f1-89e6-5a94-344369d566fe']
Hi *, Is there a way to use a Native query and fetch the data as an array without providing a result set mapping. The problem is that I'm pivoting a table in SQL and I don't know the names of the columns before hand as the column names will be the data from a specific row. Regards, M
7e5019ae-51c6-d4df-10b0-8758950bcc3e
['1f7a71fc-02f1-89e6-5a94-344369d566fe']
Hi *, am I the only one who finds the "There is no extension able to load the configuration for XYZ" annoying? * the sf is preventing me to have a future proof config file * you can't simply just comment out the bundles when trying to pinpoint which bundle is causing a problem without removing a bundle configurations * sometimes one would like to run a command in dev enviroment with no-debug switch to prevent any logging which would cause the OOM errors What's others opinion on removal of this? Regards, M
ca7de2b8-386b-6ac5-d2da-a2d773683858
['1f848535-99f0-943d-a0e1-dd441a6c574e']
Hi! I wonder, if someone knows the rational behind the standardized block comments, which seem overly cumbersome to me. I'd expect commenting out code to work by just inserting a macro that eliminates itself: So instead of: (+ 1 #;(* 2 3) 4) -reads-> (+ 1 4) I'd expect something like: (+ 1 (#; * 2 3) 4) -reads-> (+ 1 4) Where #; is the comment macro. Any insights? Just curious. Greetings, <PERSON>
69103a5a-1c5d-cca0-61b0-6f116058a438
['1f848535-99f0-943d-a0e1-dd441a6c574e']
Dear <PERSON>! I was wondering: Did anybody compile a set of example programs for Chibi-Scheme to learn from? From my experience, running "generic" scheme programs is not always easy on each dialect, esp. for a beginner learning scheme (like me). So I thought it would be nice to have an example library with small programs that have been tested on chibi scheme. Is there such a thing? Does anyone like to start one? Anything we could adopt? Cheers, <PERSON>
90079882-f9c5-cc32-8ab3-787f44b361de
['1f93f1d9-a494-ed30-f6b6-ae58aa9a897f']
Hi, some time ago I upgraded my machines from kfs to fossil/venti using the homegrown kernel (as described on the wiki page, this was before the fossil/conf times) that uses a fossil without srv -AWP support. In order to do a pull from sources, I created a special admin user, who is a member of the sys group. So, obviously, lots of new files that I have pulled subsequently are now owned by my admin user. And this leads to those annoying "cannot change attributes" messages during pulls. I have now rebuilt my kernel using the latest, greatest fossil. What is the most straightforward way of reverting all files owned by my admin user to sys? I have so far attempted doing just a pull, this has no effect, bar giving me the usual warning. I have attempted a pull -s, which has absolutely no effect whatsoever, doesn't pull anything as far as I can tell. I have attempted removing the offending files, but a pull only reports that it won't update "locally removed files", and an explicit pull -s of an explicit deleted file has no effect at all (again). Any ideas? Do I have to build a special purpose fossil that allows srv -AWPp? Thank you very much in advance, <PERSON>
ee697fc9-c67d-b317-1ab2-de8b9532e180
['1f93f1d9-a494-ed30-f6b6-ae58aa9a897f']
When writing text in left-to-right languages like english, you are more likely to be focussing your attention on the left side of your window. Cursor positioning is therefore also more likely to be nearer to the left edge of the window. So, your mouse pointer is more likely to be closer to the left edge of your window. <PERSON>
93888944-c8ab-eaec-76e7-6477dbc4e925
['1f9b22f4-bac3-c500-8889-14caeeedfeb5']
Anyone have a clever method to have a client failover between udp and tcp servers? <PERSON>: if I can't connect to a UDP 1194 server, I want to try a TCP 443 server I am thinking of just writing a little perl script to check if the client is connected every 5 minutes or so and if not, change the config file and restart openvpn. That seems like it would work okay, but is kludgy. <PERSON>
236e942f-5fea-94b4-dd7e-8d69376b41bf
['1f9b22f4-bac3-c500-8889-14caeeedfeb5']
Update to OpenVPN 2.1_rc1 and use the "topology subnet" option. This will let you use every IP address in a subnet instead of a bunch of virtual /30 subnets. That or you can make your subnet bigger, for example: Server <IP_ADDRESS> <IP_ADDRESS><PHONE_NUMBER> That will get you about 200 clients and is probably a smaller/easier change to make. -Darren
98309609-5ff7-194b-487c-d3082139b4b7
['1fa1f63c-e695-be5d-53e4-76f8800b24ad']
Hi all, I'm running RTEMS on a LEON3 (sparc) platform. It's downloaded from Gaisler and claims to be rtems-4.9.99.0(SPARC/w/FPU/leon3) It seems to me that the function rtems_clock_get_uptime() fails when it is called during a timer tick. In this case a single time value is wrong, the time is skipping backwards by 1 millisecond (TICK). The next call returns a correct time. I'm cross-posting this problem to the rtems-users and leon-sparc groups, because I don't know who is to blame: generally RTEMS or the Leon-BSP. Details below. Thanks & regards, <PERSON><IP_ADDRESS>(SPARC/w/FPU/leon3) It seems to me that the function rtems_clock_get_uptime() fails when it is called during a timer tick. In this case a single time value is wrong, the time is skipping backwards by 1 millisecond (TICK). The next call returns a correct time. I'm cross-posting this problem to the rtems-users and leon-sparc groups, because I don't know who is to blame: generally RTEMS or the Leon-BSP. Details below. Thanks & regards, Rolf Schroedter.
21bed3e9-6aeb-3350-787f-be57de9f0021
['1fa1f63c-e695-be5d-53e4-76f8800b24ad']
I need to add, that the longjump test fails with optimization: gcc -qsvt -O0 and -O2 It works with gcc -qsvt -O1 and -O3 (SVT) gcc -O1 ... -O3 (MVT) !!! Between testing SVT and MVT modes, the board needs a RESET, because the MVT startup code doesn't reset the SVT bit in %asr17. Any ideas ? <PERSON>.
b018cac3-7028-94dd-614a-ca57bb95729f
['1fc51492-e308-dae8-6ec2-b3341cfa672b']
Hello, i'm using postgis to store 3D multipolygons and i want to get the the maximum z-corrdinate of a set of multipolygons. Right now i first create a set of polygons which contains the polygon from all multipolygons, then i create a set of points from the polygons and at the end i loop through the points to calc the maximum, but i think this way is slow, so i want to know if there is an other way, f.e. is there a way in postgis to extract all points from the multipolygons with one function. Hope someone can help me. thank you in advance best regards max
0d06dc1c-bc78-9761-30dd-3000b050163b
['1fc51492-e308-dae8-6ec2-b3341cfa672b']
Hello, because i'm very new to the postgis topic, i want to know where to find informations about handling 3D geometry in Postgis. I found out that i can store 3D points, polygons and so on without any problems. What i want to know is are operations like intersect working for these 3D objects and second is there an good API to view these objects, because i used openjump, which, of course, only handles 2D objects. thank you max
8bbb67f0-9ab3-9eea-0c93-83158892be33
['1fcd6c50-be84-904e-91fe-eacfd1e94b28']
Hi, I would like to create character bitmap from script-fu logos (using a single character) and the zapf dingbats font. But I need to use character codes out my keyborad range and this for several characters. Please, could someone tell me how to do or give me a small script (perl ??). I don't really understand how all this script stuff works. Any idea ? Thanks <PERSON>
ffa62de3-1e68-0830-881f-7a1a682e609b
['1fcd6c50-be84-904e-91fe-eacfd1e94b28']
Hi, First, thanks for balsa. I would like to know if it's possible to have smallest icons on the icon bar. Or, at least, to have a scroll bar on the icon bar because when they are to big, the left ones disapear. Indeed, they are quite big (it seems they get the size of the text). Or is it possible to set the icon text size ? By the way, I'm using the french version. Some comments are very long, such as "Selectionner un identite", how is it possible to shorten it to "Identite" ? Thanks
a1f1f321-7bc9-0901-8825-49ff97d893db
['1fce933e-d514-d351-7115-e31915dce966']
Updated patch available at <http://lcs.mit.edu/~wollman/openafs-fbsd52.patch>. Changes in this version: - Doesn't crash when doing heavy recycling (e.g., listing /afs/sipb.mit.edu/users with a small cache). - I've rebroken shutdown (will investigate when I am sufficiently awake again). - Redundant vop_lock family routines eliminated (the standard ones are just fine for our purposes). - Includes correct fix for vop_symlink. - Includes preliminary autoconf glue to test for POSIX regex(3), eliminating need for libcompat. (Same change should be made for OpenBSD.) <PERSON>, feel free to commit these parts. -GAWollman
edeff415-4b21-622f-b36c-6c9a96e34938
['1fce933e-d514-d351-7115-e31915dce966']
Those who follow the Austin Group lists will be aware of an issue raised (and not resolved there) about the unsuitability of the return type of dlsym() for its most common use. Specifically, Standard C leaves undefined the result of converting a data pointer to a function pointer. Some compilers and lint programs may warn about doing this conversion, even though it is perfectly well-defined on all of the architectures FreeBSD supports. An alternate interface, dlfunc(), has been proposed to remedy this problem, by providing an alternate interface to dlsym() which returns a function pointer type. However, the POSIX committee is loath to add a new interface without some prior art. The following patch implements dlfunc() in order to provide that prior art. -GAWollman
c30a5e5f-cfca-149b-ade3-0c3769c12d47
['1fd04109-e1a6-2d0b-4481-a3b8c0bf0af4']
So I finally got back 'round to finishing off the work I was doing with salt-cloud a few weeks back. Since stuff's changed, I'm now integrating my changes into the current develop branch for testing. I use the Asia Pacific region, which means I need to create a connection based on my location. This means changing the API of get_conn() to supply a location (and for all other cloud types), and then changing the methods in libcloudfuncs which are shared. What do you think? I'll throw together a pull request if you prefer. <PERSON>
04314602-6d5d-bd43-51d9-852e1d807f15
['1fd04109-e1a6-2d0b-4481-a3b8c0bf0af4']
Hey y'all, What's the general feeling on having support for elastic IPs in salt-cloud? I think it would be useful for AWS users. Do other cloud providers have any conceptually similar features? Is this outside the remit of salt-cloud? I'll be doing work to support these for some of our server deployments. Makes sense for me to add this salt-cloud if it's desirable.. Cheers
e38d121d-7e4a-0d46-0608-8ec56e5b0ebe
['1ff1b154-1378-574d-81de-8a87f11dc0f0']
Now that the kitchen is off limits and has been cordoned off (electrified barbed wire and machine gun emplacements like at the Berlin Wall) and as a reminder of <PERSON> and <PERSON>, I've been evicted to eat at a restaurant for 2 days :-) I had more time to check mareh mekomot about corn and kitniyot. Both the MAHARIL (Hilchot Afiyat haMatzot) and the Leket Yosher (I Orach Chayim) translate SHIFON (rye) [one of the 5 species] as "korn". By the way, the Maharil translates "orez" [a.k.a. "rice"] as "hirzen", with "dochen" [millet] translated as RICE (resh yud yud zayin). So what we think of rice (Uncle Ben's) may be the wrong item. The gemara in Pesachim may have been talking about millet!! An expert in German I consulted couldn't give any explanation for the meaning of "hirzen' [except for "irzen" = confused]. After almost 58 years, I finally figured out why the American breakfast cereal from corn is called KIX. The Chatam Sofer III Even HaEzer 52 translates (American) "corn" [tiras] as "kokritz"; the Yeshuot Malko I 17 "kakaritze" (likewise the Aruch haShulchan who uses the term "kikas') whereas <PERSON> (Hea'orot #453) and <PERSON> (American) corn as 'Turkish wheat' [tirkishen veitz]. The very fact that these poskim thought it looked like wheat (as a matter of fact, corn growing in the field takeh does look like wheat) may have brought our American corn into the category of kitniyot. Last but not least: the language of the Chelkat Yaakov Orach Chayim 195 #1) regarding "tiras" is strange. <PERSON>, he differentiates between corn and kitniyot! "rak pat shel tiras u'kitniyot u'k'doma". CHAG KASHER V'SAMEACH <PERSON>
38e4f7cc-0a9b-080c-1467-9f16d70ea68e
['1ff1b154-1378-574d-81de-8a87f11dc0f0']
One would think that with the explicit passuk KI SHESHET YAMIM ASAH .... one would be required to believe in a literal 144 (human) hour length of MB. But look in <PERSON> 32 on "shorshei ha'mitzvah". [she'ha'olam nivra b'shiva yamim chalukim"]. Note, there is no mention that a YOM is equal to 24 hours. It could easily refer to a TEKUFA. Had <PERSON> insisted that we must believe that a day = 24 hours, I would concede. But he doesn't and <PERSON> was very exact and precise in his wording. The only HALACHIC matters that deal with zeicher" are: yetziat mitzraim (with regard to tzizit, tefillin, mezuza) [Minchat Chinuch # 21] and zeicher amalek. No zeicher of 144 hours of MB. When we make a bracha OSEH MAASEH BRESHIS on hearing thunder (gemara in Brachot 59a) [Orach Chayim 227:5] there's no mention anywhere that we have to believe in a literal 144 human hour period. Nor is there any mention of this in the Aruch haShulchan. Think about it: had there been a requirement to believe in 144 human hours, the bracha we say would be a bracha L'VATALA if we didn't believe in an exact 144 hours. Do you have a special girsa "Oseh maaseh breshis b'me'ah arbaim v'arba sha'oht" ? :-) MAASEH BRESHIS is supposed to be a SOD (that's why we learn in Chagiga 11b EIN DORSHIN ..). If it's literal (144 human hours), where's the SOD ??? All we are required to believe is that <PERSON> created the world (Rambam Hilchot Yesodei haTorah 1:1, 2:9). Nowhere on Shemos 20:11 or 31:17 is there any chiyuv to believe in a literal 144 human hours. If there is, show me a source. Incidentally, open up a Shas (Chagiga) to the peyrush of the YAAVETZ (it's way at the end) of Daf 16a. He specifically mentions PHYSICAL worlds being created and destroyed by the RBSO before being satisfied with ours (as per the Midrash) and this was not allegorical. KT <PERSON>
40283c0e-a406-4bc2-9873-6da61efb0ca1
['1ffa8dc0-45af-dd9d-e47b-e7c470d9640d']
Can I 'seed' my 5.5 repo with the previous version's files and perhaps save some rsync time/traffic; or is this is bad idea? A bit of 'Friday finking' by a non-SysAdmin: I have local repositories for CentOS 5.3 and 5.4. rsync is utilised to pull down a copy from my closest CentOS 'mirror' site. Thereafter this is fed into my LAN's Cobbler server. All local installs and updates run from there; saving multiple update downloads, my systems don't have to run yum-updatesd, ... Now that CentOS 5.5 has been released, the repo/Cobbler server needs to be expanded. Given that some files will not have changed between 5.4 and 5.5, if I were to rename my 5.4 directories to 5.5 (and update the rsync command accordingly) might this save me some download traffic? (before feeding the combination into Cobbler as "5.5") Conversely, would the older version be a faulty foundation for the new release/repo? Anyone thought this one through, been-there-done-that? Regards,
ab116e04-e6ab-4544-a01d-6748a03cd64c
['1ffa8dc0-45af-dd9d-e47b-e7c470d9640d']
Is 'now' the time for rpm-based distro users to upgrade from Cobbler to Spacewalk? I'm (still) working (in my ?spare? time!) on refreshing my CentOS systems. Currently I have 5.3 sitting happily in a Cobbler server, and have been downloading 5.5 when I remember/there is unused allowance in my monthly allocation... 'My' repo is primarily used to proxy multiple machines on my SOHO LAN, and is sometimes taken to client locations or used remotely to support client activities. When I start using VMs 'installation' will take on yet another flavor... The Spacewalk project has been making strides* recently. Is it really 'ready' for live firing? Is there an 'official' policy that Spacewalk has/is/will absorb Cobbler (and Puppet, and ...)? Rather than update my Cobbler repos, should I bite-the-bullet and learn to implement Spacewalk? Is there a 'tipping-point' of 'scale' whereby 'little guys' should continue with Cobbler whereas Spacewalk better suits larger operations? Any and all advice gratefully received - and I'll be v.happy to be pointed to authoritative web references, etc. Regards,
bb97d939-bc34-8ef6-e2b2-3e9da4bcefb2
['20081fae-10d0-923d-850a-8df6795e714d']
Hello all, This is to let you know that the IANA has allocated 222/8 and 223/8 to APNIC for further distribution. A notation of the allocation has been made at <http://www.iana.org/assignments/ipv4-address-space>. If you are filtering these ranges you need to be aware that they will be bought into use and you may wish to adjust those filters. <PERSON>
c321e8a5-88c7-4899-7474-d01954936667
['20081fae-10d0-923d-850a-8df6795e714d']
Hi folks, I hope you consider this operational and on Topic:) We have been receiving some reports of rejection of TLD strings at the ISP level. Some of this may be due to length limitations set in peoples software, there may also be other causes. We want to make sure that there is a place where people can go and pick up a current list of the strings that are in the root zone. (Without having to parse the zone file itself). To this end the IANA is maintaining a list that can be downloaded from the following information page http://icann.org/topics/TLD-acceptance/ If you have questions please feel free to use the e-mail listed on the above URL. <PERSON> CTO ICANN
3279aabe-9f3f-ebd5-76d2-a7c3150e6142
['200d4897-246a-a503-5bd6-20206af81c72']
Hi, we have a 2 MBit internet connection. The firewall and the router are connected to a 3COM 10 MBit-hub. On this I want to connect a snort box too. Does a Pentium-1 200 MMX with a SCSI-Harddrive and a LevelOne 100 MBit Network-Card achieve the requirements of a snort box for this Network-Traffic? Does anyone know if such a box is enough for iptraf or ntop? I have a dual Pentium 200 MMX with SCSI-Disks. Does this old system fit the requirements for network traffic analysis or protocol? Greetings <PERSON> <PERSON>
8c195286-b5a2-f141-1263-a099993c3b29
['200d4897-246a-a503-5bd6-20206af81c72']
Hi, I just changed from Bind 8.3.1 to Bind 9.2.1 I read the file /usr/doc/bind-9.2.1/misc/migration and changed my named.conf. The nameserver starts, but after all the information about the sended notifies it tries to get the sones from the master for the zones, this server in configured to beslave. then the following is put in the /var/log/messages: "lame server resolving 'bla.com' (in 'bla.com'?): 1.1.1.1#53" i changed the domain an the master-ip ;-) What does this mean? Greetings <PERSON><IP_ADDRESS>#53" i changed the domain an the master-ip ;-) What does this mean? Greetings Alex Alexander Bruns
f7374b7a-20ca-226e-caa9-75dd26558f43
['201632e8-a59c-91fc-c94a-4f700a1c46d0']
Dear all, I have torque (2.3.6) and maui (3.2.6p21) running on my cluster. Maui is running under my personal account (not root) and I am ADMIN1. Other users can execute qalter on their own jobs only for queued ones. Once the jobs are started and a user tries to increase the walltime for example, he gets an "Unauthorized request". Is this normal behaviour? Is it possible to change the configuration such a way that a user can qalter his running jobs? Thanks in advance! <PERSON>
3d90c8d3-dd2e-1dbb-b9e1-0944455434eb
['201632e8-a59c-91fc-c94a-4f700a1c46d0']
Hi all, is there a chance to give a normal user permissions to hold, release, delete jobs of other users of a single queue only? Maybe some kind of queue configuration? The problem is that defining a user as "ADMIN1" would give him the permissions to administer all jobs. (We are using Torque 2.4.6 and Maui 3.2.6p21.) Kind regards, <PERSON>
11df2a22-1efb-0e44-6f31-5313a32a7242
['2017be9f-13ba-5d0b-7656-94614ed704d0']
Hi <PERSON>, I have a big, big...big problem with Horde. I simply exported/imported my bookmarks from a test install (CVS) to a prod install (last stables modules) of Horde (same web server with two diffents mysql databases on a Kolab server) and, suddently...a totally blank screen after a successfull connection, in each of theese environments ! Nothing in log files, no messages (except after import : something like "error, mailbox does not exit"). It clearly has to do with Trean import, but I absolutely don't know how to deal with this. I cleared Trean install and disabled it in "horde/config/registry.php", but got always this totally blank screen. I wonder where are Trean's data stored ? Other users can connect with no problem at all... just me, the administrator !! Can somebody point me in a right direction ? (it's not critical, but, I'd like not to loose all my mails, events, notes and so on...if I have to create a new account) Many thanks in advance, <PERSON>
28469c9f-8774-673a-fdd3-fffdfdc6533b
['2017be9f-13ba-5d0b-7656-94614ed704d0']
Hi list, I'm about to make my Horde/Kolab server in production after quiet a year of tests. Important thing : my preferences driver is Kolab My configuration isn't absolutely perfect but...most of fundamentals are there. One exception : I have no Horde users ! All of them are created via Kolab...and don't appear in Horde's users list. Very difficult to set permissions, groups and other stuff in this case. When I add a new user within Horde, its mailbox is created and appears in Horde's users list. But I can't see users created with Kolab's web interface in Horde. And, of course, I can't create an existing Kolab's user with Horde's interface. So I'm wondering if users created with Kolab should appear in Horde's users list ? If not, how can I deal with it since I need Horde's users to set up permissions and groups ? Just a point in the right way will be extremly appreciated. Thanks in advance, <PERSON>
547384fc-cdd6-bf93-e7e9-119ad88e3929
['202cb357-61e1-4aa4-9ed4-a1e46bb8e922']
<PERSON>, Thanks for your answer. It covers the scenario very well where there is only one MG, and the MG doesn't have a redundant MG to protect it. But there are still to scenarios in the case that the failed MG has an redundant MG protecting it, 1. The rendundant MG detect the failure of its pair before the MGC does. In this case, should the redundant MG initiate a failover? How ? 2. The MGC detect the MG failure first. In this case, shoud the MGC initiate a failover to the redundancy MG? How? These two scenarios are not clearly addressed in the Megaco spec. There is another issue regarding the second paragraph of section 11.2 It says, " Allowing the MGC to send duplicate message to both MGs accomodates pairs of MGs that are capable of redundant failover of one of the MGs. Only the working MG shall accept or reject transacctions." It also says, " Upon failover, the primary MG sends a ServiceChange command with a Failover method and the MG Impending Failure reason. The MGC then use the secondary MG as the acitve MG". The above text addresses the scenario of primary MG failure, while still be able to talk with MGC. But It is not clear to me how does the MGC then use the secondary MG as the active. If it just simply directs the new transaction requests to the secondary MG, the MG may still think itself as Non-active, and thus ignore all the requests. It seems to me either the MGC should send a kind of ServiceChange command to the secondary MG to indicate the change of role, or the secondary MG get the role change information directly from the primary MG. Which one does the Megaco spec imply? Thanks,
90f687d0-cbb6-45d2-91ea-cde11d403469
['202cb357-61e1-4aa4-9ed4-a1e46bb8e922']
All, In section 11.4 of the Megaco protocol spec, it decribes the scenario of a MG failure, while THE MG is still capable of sending a message to the MGC to indicate the failure. It seems to me that we are missing the scenario that a MG is totally dead. In that case the MGC won't get a indication of failure from the dead MG. How does the protocol handle this scenario? Thanks,
102b68e7-34db-ab4d-e7b9-2ef74c2783b2
['2035ae00-edf4-fd48-8b42-6bc94cd342fc']
<PERSON> i was going to reply to just one section of this yesterday but did not get time to find my code. You mentioned problems with getting best object size to work with listbox columns..i cant remember if i had a problem or just never really tried using it on the list box column. When i was trying to to this i used a ‘dummy’ object on my form to put the data into(the object font etc set to match the LB column) and tested the data in there.(distinct values..find widest data(loop)) I cached the results-because you dont really want to do that every time-too slow on million plus records-but faster then you might expect…. ..and only did it on text (and string then) fields-i use better techniques on pictures and a different strategy for sizing numbers(although here is it good to contextual to the users-if you have customers that do invoices that are all less then 10,000 dollars and others that do invoices of 1,000,000,000 you dont really want to size for the 1,000,000,000 on the customers that do 100,000 max. I used this to optimize column widths overridding defaults that had been set there were not valid. <PERSON>
de2a93e1-11bc-bc9f-6a17-551cc960070e
['2035ae00-edf4-fd48-8b42-6bc94cd342fc']
A small Yosemite Gotcha...probably wont affect most people I noticed a problem in my 4D system the other day with window sizing on Yosemite. In my code have an long standing function that checks the status and the location of the Dock on screen on OSX. In that code I test the OS version number(because the plist changed after 10.3). The thing to look out for is that if you have any code where you are testing if the OS version number on OSX is greater than 10.x you need to remember that 10.10 is less than 10.9. <PERSON>
4d30f43b-8d23-b95f-528d-e6b5903c1ed0
['2035f157-b910-d51b-a083-0382a201f360']
Hello everyone, After playing with patch files for a while, now it seems that the doxygen port is completely corrupted. Is there a way to return everything to what it was before I started messing around with the Portfile? I tried clean, clean --all, uninstall, everything, but nothing seems to work. Thank you all,
8cbfa14d-1b5d-efd9-919b-4d2d154b8737
['2035f157-b910-d51b-a083-0382a201f360']
Hi everyone, I'm new to MayaVi. I've been searching the internet for a solution to a problem I have but I didn't find the answer, therefore this email. I have 2D cell vector data that I would like to display in the cell centers. I used Paraview before MayaVi, so I used to extract cell centers first and then with a glyph filter show the arrows, but it was impossible to replicate this in MayaVi. Can someone point out if this is possible? and if it is, what are the steps to replicate this? Thanks a lot, aa
1ac2a6e0-78cb-dbe4-ebe0-2d0caa8d0264
['2053cd9a-1240-905c-b347-774e96e5aaf6']
Hello guys, I have been running a server (running RH 7.3) with WINS support for quite some time and it works well. I am seeing that once Windows clients (Windows XP) are connected to the server, the wins.dat in the server gets updated where the clients IP etc being added. Even the server's IP has been changed (for example, the machine is moved and connected to a different network with a different WINS server), the clients still automatically find and talk to that WINS server and have themselves registered. However, the linux clients do not seem to be so smart(?), my finding is that "wins server" has to be specified so the client's samba will then talk to the WINS server and therefore get registered. This becomes very inconvenient if the linux machnes (e.g. laptops) are used/connected to different networks where WINS servers' IPs are different or even unknown. I am wondering, how does Windows (act as a WINS client) do that? Can linux running Samba dynamically find WINS server without having to specify it in "wins server"? Thanks
1982ea7d-f267-3474-42e2-e5f14c2ab030
['2053cd9a-1240-905c-b347-774e96e5aaf6']
Hello guys, Since I upgrade OpenVPN from 1.5 to 1.6, I'm having some problems about maintaining connection from my Windows XP clients to OpenVPN server (Red Hat linux, using tap/bridge). There was no problem to make a connection initially ("Peer Connection Initiated with xxx" shows) and it'd work from there. It'd then stop working after a while (say 10 mins) and then if I close the connection (from the client) and try to reconnect, it no longer makes the connection. I can make it to connect again if I disable and re-enable the TAP adapter from Network Connections (in Windows XP) and that will then stop working again after a while...same thing can go over and over. I'm wondering if it could be the way I upgrade the client, which corrupted the virtual adaptor? I install OpenVPN 1.6 without first uninstalling 1.5 Thanks, Best Regards, <PERSON>
6042c44a-c1ba-1ad8-d97a-eacfdccc61ae
['20617634-f9f1-3e61-1a0b-1359ab952856']
Dear All, Is there a way to only allow a certain user group to logon from the local network, but to allow another user group to logon from external IP addresses as well as local? I'm using OpenLDAP for authentication, so I'm wondering if there's a way to pass the IP address of the connecting client through to the LDAP authentication backend and then use an LDAP filter, but I'm not entirely sure how that filter would work yet, so I was wondering if there's another easier way through Horde.. Many thanks, <PERSON>
d5d5da09-861f-0ddf-d37a-7a27c03b0db3
['20617634-f9f1-3e61-1a0b-1359ab952856']
Hi All, I'm running linphone 2.2.1.1 on Android 4.4.2 on my Galaxy Note 3. Linphone registers perfectly with my Asterisk PBX. When I'm in a call, I cannot use the phone's touchscreen - it stays off. If I press the power button twice quickly, the screen flashes up for a split second then goes off again. If I'm quick and use two hands I can hit the end-call icon, otherwise I need to wait for the other person to hang up. I've tried disabling the lock screen and password with no effect. Anyone got any other ideas I can try? <PERSON><IP_ADDRESS> on Android 4.4.2 on my Galaxy Note 3. Linphone registers perfectly with my Asterisk PBX. When I'm in a call, I cannot use the phone's touchscreen - it stays off. If I press the power button twice quickly, the screen flashes up for a split second then goes off again. If I'm quick and use two hands I can hit the end-call icon, otherwise I need to wait for the other person to hang up. I've tried disabling the lock screen and password with no effect. Anyone got any other ideas I can try? Tim
8fa36eb4-cee4-c2d8-4a7c-0422ddd012e6
['206cf63a-ef67-7373-1f6f-18b8b310287c']
Hi all, I'm syncing two local trees that have identical files but different permissions. I thought I could simply ignore the permissions by putting 'perms = 0' in the prf file, but this doesn't seem to be doing the trick. Unison still detects permission differences on about 3500 files. Is there a way to tell unison to sync files but ignore the permissions, even if they're different on both sides? TIA Joost
a217dc6e-6449-35f9-b144-269687f33da3
['206cf63a-ef67-7373-1f6f-18b8b310287c']
hi all, my situation is this: i have an iBook running linux/ppc that i'd like to install sbcl on. i also have access to an iMac running OS X 10.3.4 that has sbcl installen through DarwinPorts. now i'm wondering whether i can cross-compile sbcl on OS X for linux/ppc. my first attempt wasn't very successful. sbcl building on OS X assumed it was building for darwin, not linux. so my questions: does sbcl always assume it is compiling for the OS it's running on, (it would seem so, because make-config.sh uses uname to set the target OS), and if so, is there a way to tell it to cross-compile for a different OS, not just a different architecture? TIA Joost
dd44140a-7501-69ed-d5e7-c2d2d83f11bf
['20866414-d9d5-5bc7-362f-c94e0d21c8ef']
I have the Engage security EagleX package running on a windows2000 box. It is a flavour of snort, msql and ACID. I think the default config is that of and IDS but i would like to configure it as a sniffer that would allow me collect any analysis data on a continual basis. I have ethereal but it cannot continuously collect data. Are there any open source solutions that do that (something similar to NIA's Sniffer Portable).? Regards
245ff088-142e-6082-5de9-d5411c0a9116
['20866414-d9d5-5bc7-362f-c94e0d21c8ef']
I have a software for web filtering ( websense ). All URL requests are sent to the server which then determines if the site is to be allowed or blocked. When this happens i get spp_portscan2 alerts on my snort console. How can i filter out this alert? I am running snort on windows 2000. Thanks in advance. Regards
dd29615d-ec2f-6478-56d0-5e41453b6464
['20ab8509-3921-1129-fadf-75d562f9129d']
IRAN: 14,635 ARRESTED IN MORALISATION CAMPAIGN
ae8a3940-74a9-be7d-b783-f51c9a38ebb3
['20ab8509-3921-1129-fadf-75d562f9129d']
Eight Years Requested For <PERSON> At Mafia Trial Regional authority president accused of aiding and abetting Cosa Nostra says, "I've never given favours". At trial of Palermo anti-Mafia directorate moles, public prosecutor requests nine years for Carabinieri special operations group officer. PALERMO - Eight years' imprisonment is the sentence requested by assistant public prosecutor <PERSON> for the president of the Sicilian regional authority, Christian Democrat (UDC) <PERSON>. Mr <PERSON> is charged with aiding and abetting Cosa Nostra and passing confidential information about the trial to the so-called moles in the Palermo anti-Mafia directorate. The Palermo public prosecutor's office also requested eighteen year's imprisonment for businessman <PERSON>, who is accused of Mafia-style criminal association. For former Carabinieri special operations group officer <PERSON>, charged with external complicity in Mafia-style criminal association, the prosecutor's request was nine years, reduced because the accused opted for a summary trial. OTHER REQUESTS - Substantial sentences were also requested for the other ten defendants at the trial. Five years were requested for Dr <PERSON>, charged with illegally accessing the prosecutor's office's computer system; four and a half years for <PERSON>, the former assistant of public prosecutor <PERSON>, who is accused of illegally accessing the computer system of the prosecutor's office and revealing confidential information; one year and four months with mitigating circumstances for former UDC member <PERSON>, one of <PERSON> closest collaborators, charged with aiding and abetting; three years and six months for police officer <PERSON>, charged with abuse of power; five years and a 1,000 euro fine for <PERSON>, charged with criminal association and fraud; four and a half years and a 1,000 euro fine for <PERSON>, charged with criminal association and fraud; nine months' imprisonment for <PERSON>, charged with corruption; two years' imprisonment for <PERSON>, charged with corruption; two years' imprisonment for <PERSON>, answering charges of corruption; and five years and a 1,000 euro fine for <PERSON>, charged with criminal association and fraud. Finally, a fine of 1,549,000 euros was requested for Mr <PERSON>'s Diagnostica per Immagini company as well as one million euros for another <PERSON>-linked compay, ATM, Alte tecnologie medicali. NO FAVOURS - Mr <PERSON>'s terse response came later yesterday evening: "I have learned of the requested formulated by the public prosecutors with regret, a sentiment exacerbated by the inner awareness that has never abandoned me in these years of never having behaved in such a way as to benefit the Mafia". SOLIDARITY FROM <PERSON> - The president of the region received several messages of solidarity, the most significant of which was from the leader of the House of Freedoms, <PERSON>. "I vehemently confirm my solidarity for the president of the Sicilian regional authority, <PERSON>", was Mr <PERSON>'s comment. "My esteem for his intelligence leads me to rule out absolutely that he could be involved in the episodes in which attempts are being made to involve him". English translation by <PERSON> www.watson.it
44fb7e30-ecd6-d137-272d-5981fd5b1733
['20bf649a-0d09-5cab-8a80-261d7c516732']
Hello everybody, I'm new here on this list, so I'll introduce myself shortly. My name is <PERSON> and I'm a Linux consultant from the Netherlands. I'm using Quanta mainly to do PHP development, and I've been using Quanta for about 1 year now. I compiled the Bleeding Edge Beta today, and I was having some performance problems when using and editing large files. I had loaded some small files (<3k bytes) and one large (105k bytes) file. This last file is a PHP function library that consists of about 3500 lines of PHP code. I ran top to see how much memory Quanta was using and I was a bit shocked to see that it consumed 154 MB of memory! So I closed Quanta, opened it without files and it got to 30 MB. Added the 105k function library and voila, 141 MB! For some more testing I did the same using the Quanta version that comes with SuSE 9.0 (Quanta 3.1.4 on KDE 3.1.4), same results. Doing the same trick with <PERSON> was a little less bad, but still some 125 MB. Can anyone explain here why those editors consume more than 1MB of memory for every 1kB of code? Isn't that a bit excessive? Ok, I've got code highlighting turned on, maybe that costs a lot, but still... Regards, <PERSON>
cd26963f-b55b-078f-cacd-28d92a27db94
['20bf649a-0d09-5cab-8a80-261d7c516732']
Hello everybody, I've been running SuSE since 6.0 on a lot of different desktops and laptops, but lately I noticed something very strange. When using SuSE 8.2 with an IBM A31p loading the kernel takes about 19 seconds (the part where it says Loading linux.............). On all other systems this takes just a few seconds at most. This is a P4 1.8 Ghz laptop, so it should load very fast. I'm using the stock 2.4.20-4GB kernel from SuSE with the following append line: acpi=off hdc=ide-scsi hdclun=0 showopts. Does anyone have an explanation for this? Could it be something in the kernelparameters? Regards, <PERSON>
35442932-05c6-4776-ce36-d4554d8fea89
['20cbd166-aadc-5c10-a607-8026821dc854']
Correct...I have OVERLAY on my control record format AND on the subfile "footer" record... I suspect you have OVERLAY on all your record formats, and none of them overlap the subfile, so the old subfile is not being erased. If this is the case, you can do one of the following when there are no records: 1. Make sure that you write at least one record format WITHOUT using the OVERLAY keyword (condition the OVERLAY with the same indicator that you use to condition SFLDSP). This will erase the entire screen. This option may be the easiest, depending on your screen layout. 2. Write a dummy format that occupies the same space as the subfile. This is actually a nice option, because you can output a nice "no records found" message. <PERSON>
56a5a900-bc4a-e5c3-88ef-f8bd0b93cea3
['20cbd166-aadc-5c10-a607-8026821dc854']
I'm working on a file maintenance program (add, update, delete). The records are being displayed in a subfile. After each add, update, or delete, the subfile is reloaded and re-displayed to the user. Everything works fine except when the file contains only 1 record. The record is displayed in the subfile and then I delete it using the program. After the record is deleted from the file, the program clears the subfile and reads the file to reload the subfile. The program doesn't find any records and correctly sets the indicators for SFLDSP and SFLDSPCTL. BUT when the screen is displayed, the record that was deleted is still being displayed on the screen. Any thoughts as to why this may be happening.... Thanks, <PERSON>
ce1ef854-94eb-52a1-17c0-b204e816cf2f
['20e5005b-e584-c80b-eac8-340fb7aca59e']
Just bought a Bitx40 SSB xcvr kit for less than $60. I am not an EE nor an IT guy. Just a ham. I am wondering what it would take to make this a digital xcvr as stand alone (eg., for EmComm purposes?) Linux microcomputer, sound card, TFT for waterfall display? Your thoughts? <PERSON>
070fa4d8-495c-cbd4-72f4-47a378e240ef
['20e5005b-e584-c80b-eac8-340fb7aca59e']
OK 'Check the use xmlrpc, initialize save and close. Do not check the xmlrpc ... that is for use with earlier versions of flrig and other 3rd party xcvr control software.' so, check use xml-rpc program or do NOT? I left use xml-rpc program unchecked and initialized. followed all other directions and saved config "restart to continue with configuring" continue how? I have flrig installed but don't know how to configure it. I'll send screenshots of flrig to file <PERSON>
851c9453-ceea-af69-8e75-739d945009df
['20f330bc-33bd-6759-73ba-27e2afdc49d7']
Hi <PERSON>: We have a tool that helps bridge the gap between what <PERSON>'s mentioning here (using the right thicknesses and Dk's), HyperLynx, and your PCB tool. A thread on the SI list just a couple of weeks ago discussed how difficult it can be to find and track all of the material properties from different vendors, and we've tried to make that easier by pulling that data into our libraries. When you're done planning your stackup (in our tool), you can shoot the results to HyperLynx, if you like, for further analysis, and/or to your PCB environment. To your earlier question, it would give you two sets of field solver results to compare, as well. E-mail me offline if you want to give it a try. <PERSON>
8874a180-dbb6-71f0-81c5-1e98064f7e55
['20f330bc-33bd-6759-73ba-27e2afdc49d7']
Hi <PERSON>: You can go from Gerbers to PADS to HyperLynx using software from DownStream Technologies. http://www.downstreamtech.com/reverse-engineering-products.php The problem with conversions from Gerber, though, and the reason that HyperLynx doesn't support Gerber input natively, is that Gerber files are pretty "dumb," relative to the data you're going to characterize the board for signal integrity simulation. You'd really be swimming upstream. <PERSON>
91e1111f-80c8-9119-75b0-b43ed99fc251
['20f90f2a-6c5c-e8e9-d029-a8c3af6c9978']
In keeping with JEP-4 <https://github.com/jenkinsci/jep/tree/master/jep/4>, I'm posting to the developer mailing list my plan to create and lead the Jenkins Docs special interest group. I'll submit a pull request to jenkins.io with the structure described in JEP-4 <https://github.com/jenkinsci/jep/tree/master/jep/4>. The first Docs special interest group meeting is scheduled for Friday, May 10, 2019 at 13:00 UTC. Proposed meeting agenda is at https://docs.google.com/document/d/1uNNo0QJKPHnNp8PGr_jLI8p3K_94ZYD-M0evZOEZ93c/edit# . We'll be using Zoom for the meeting, URL https://zoom.us/j/557525312 . We're using Zoom rather than Google Hangouts on Air so that we can test the Zoom account and so that we participants from China can attend the meeting. Draft mission statement: The Docs SIG focuses on improving Jenkins documentation. Improvements are targeted at both existing and new Jenkins users. Scope of work includes reviewing improvement project ideas and working with potentials mentors and contributors in order to improve Jenkins documentation. Thanks, <PERSON>
b2f89e7f-7091-aa71-a8f7-7b1f9be6e10d
['20f90f2a-6c5c-e8e9-d029-a8c3af6c9978']
The structs plugin (structs.hpi) seems to be missing from the current/latest directory at https://updates.jenkins-ci.org/current/latest/ . It is still available in https://updates.jenkins-ci.org/1.651/latest/ It is still available in https://updates.jenkins-ci.org/stable/latest/ Is the difference between those directories intentional? Without the structs plugin, the pipeline plugins fail to load. I'm using a separate script to download plugins from the update center to my Docker image definition, and that separate script expected to find all the plugins in the https://updates.jenkins-ci.org/current/latest/ directory. Should I refer to a different location instead? Thanks, <PERSON>
0930b10c-a721-52a5-a96b-54659c0632ee
['20fa64d0-7778-0c6d-69de-567154fed9fc']
I have recently updated my computer to FC5 and am in the process of getting everything running. Everything, of course, includes Komodo 3.5.2 I hope. I brought the libstdc++.so.5 from the FC4 system and the install went ok. However, Komodo exits with the error "Cannot find mozilla runtime directory. Exiting." probably because FC5 uses Firefox rather than Mozilla. Anyone have any suggestions? Thanks, <PERSON>
848f1f36-6293-f378-05c7-80411783c947
['20fa64d0-7778-0c6d-69de-567154fed9fc']
This is one that has bugged me for a long time. If you go to a dojo, like the graphics dojo for example and extract a project as in svn checkout svn://labs.trolltech.com/svn/graphics/dojo/modelviewer the project will not import into Eclipse. So, what is the mormal procedure for setting these projects up to compile under Eclipse using the QT-Eclipse integration? (Yes I know I can qmake make to compile in a command window, but I may want to experiment with the project, hence, the need to Eclipse them.) Thanks, <PERSON>
43d447d4-423b-f3ad-9c7e-a1d72d5182d1
['20fe465d-4ccd-940e-b907-bf1902c0b4bd']
The Algorithms Group (with <PERSON> and <PERSON>) in the Department of Computer Science at Braunschweig University of Technology, Germany, has a vacancy for a computer scientist or mathematician, available immediately. We are looking for a scientific employee / postdoctoral research associate (TV-L13 = about 2800 Euros gross per month; DFG project "Kunst!"). The position will be for one year. Applications for a one-year research visit from excellent PhD students are welcome. The position is funded from DFG-Project "Kunst!", so applicants are expected to contribute to the project; however, ample time to work on independent research will be given. The project combines methods from computational geometry, mathematical programming, and algorithm engineering in an exciting way. The goal is to solve problems from the art gallery family (almost) exactly, and to design theoretically and practically efficient algorithms. See the paper <PERSON>, <PERSON>, <PERSON>, <PERSON>: Exact solutions and bounds for general art gallery problems. ACM Journal of Experimental Algorithmics 17(1)-2.3, 2012. http://dl.acm.org/citation.cfm?id=2184449 to get an idea on the subject. Interesting research directions can be in visibility computations, extended real-world versions of the problem, the analysis of notoriously hard cases, exploiting multi-core, and robust data structures. We are looking for applicants with an above-average Ph.D. in mathematics or computer science, who have a solid knowledge in computational geometry, an interest in and a capability for a wide range of scientific work (from developing ideas all the way to their practical realization), an open mind for the other activities of our department, and a healthy team spirit. Proficiency in English is a must. Familarity with C++ development and algorithm implementation is a plus. There are no teaching obligations, but researchers will have the opportunity to interact with students in the context of their projects. We are offering an opportunity to participate in exciting interdisciplinary projects that span all the way from basic research in algorithms to its practical and technical application. As a member of our group, you will be part of a motivated team that maintains various interdisciplinary and international contacts and activities. Among applicants with equal qualifications, handicapped persons will be given a preference. The Braunschweig University of Technology is an equal-opportunity employer and encourages women to apply. Applications are considered on a rolling basis until the position is filled. Please send your applications with the usual documents to: <PERSON>
84e73451-dee0-f434-9c58-9a5bf48529f1
['20fe465d-4ccd-940e-b907-bf1902c0b4bd']
The Algorithms Group (http://www.ibr.cs.tu-bs.de/alg) in the Department of Computer Science at TU Braunschweig (http://www.tu-bs.de) in Germany offers a wide range of interdisciplinary research activities. The group has two vacancies for computer scientists or mathematicians under supervision by Prof. Dr. <PERSON> and Prof. Dr. <PERSON>, starting in April 2013. We are looking for two scientific employees / PhD students or postdoctoral research associates (TV-L E13). The positions are in the context of the recently approved DFG project "Controlling Concurrent Change", see the announcement at http://presse.rz.tu-bs.de/presseinformationen/?p=4608 . CCC is an interdisciplinary project that spans all the way from basic research in algorithms to its practical and technical applications. The exact scientific directions will be shaped depending on the successful applicants' skills and interests. For position A1 (Kroeller), applicants should be familiar with some of these topics: - distributed systems, - programming for embedded devices, - distributed algorithms, - combinatorial optimization, - linear programming. For position B1 (Fekete), applicants should be familiar with some of these topics: - combinatorial optimization, - game theory and mechanism design, - online algorithms, - mathematical programming. There will be no teching obligation. Pay is according to the German bracket "TV-L 13" (between 3180 and 4090 Euros gross per month, depending on previous experience and employment). We are looking for applicants with an above-average degree in mathematics or computer science (Master's degree or equivalent, e.g. German Diploma, resp. PhD degree), who have an interest in and a capability for a wide range of scientific work (from developing ideas all the way to their practical realization), an open mind for the other activities of our department, and a healthy team spirit. We are offering an opportunity to pursue a doctoral degree (to Ph.D. candidates) or enhance and deepen the scientific record (to candidates with a doctoreate), as well as the participation in exciting interdisciplinary projects As a member of our group, you will be part of a motivated team that maintains various interdisciplinary and international contacts and activities. Among applicants with equal qualifications, handicapped persons will be given preference. TU Braunschweig is an equal-opportunity employer and encourages women to apply. Please send your detailed application by February 21, 2013, indicating your preference for A1 or B1 (or both) with the usual documents to: Technische Universitaet Braunschweig Institut fuer Betriebssysteme und Rechnerverbund, Abteilung Algorithmik Muehlenpfordtstrasse 23 38106 Braunschweig Germany
388f4ce7-7a09-03d8-853c-95ae505fc167
['21038c15-2ebd-5d84-7591-3e0eb400c84a']
Hi, I'm looking for some DeltaV scenarios. Meaning I'm interested in logs of request/response between a server and a client. Does anybody have such information ? The main problem isn't DeltaV itself. In RFC3253 there are many examples. But what about the basics like an OPTIONS reply ? RFC2518 and RFC2616 aren't that verbose. When I e.g. tried to connect as a Windows Webfolder my server got an OPTIONS request. But what should I give as a reply ? I tried to compile cadaver on my Mandrake 8.1 Linux but got an error :-( Cheers, <PERSON>
29880271-f9e5-3f3a-b4f3-3124c76f2ef9
['21038c15-2ebd-5d84-7591-3e0eb400c84a']
Hi, my repository naturally is giving me a time when a version of a resource was created. Did we define a name for this property ? I didn't find anything in 3253. Or should DAV:creationdate of a checked-in VCR be set to the creation time of the version ? Is this be possible for all operating systems ? Cheers, <PERSON>
324eed07-0c20-cf90-6062-fc7e9033df21
['210ebb65-de4e-7bff-82ba-1c46bcb1990a']
Anyone on list had any experience with PCI pass-through in KVM? We have had some partial success in getting an old PCI DAQ card "passed-through" to an ancient copy of Windows 2000 running as a KVM guest on a Debian Jessie host. There were issues with shared interrupts causing some problems which we were able to resolve, but we still have some issues with reading some memory areas from the DAQ card. Have also tried it with VirtualBox with less success. Next step is to try it with a Linux guest and write some code on the guest side to see what we can see. Any other tips or warnings? cheers, <PERSON>.
8db57a6e-dc89-a148-74c5-1f2d74240c9d
['210ebb65-de4e-7bff-82ba-1c46bcb1990a']
A sort of Linux'y question: anyone on this list played with interfacing to their car's Engine Management System? Care to share your experiences? Any tips on mailing lists/we sites I should look out for? I am particularly interested in Holden Commodore VS series 2 with 3.8l V6 engines, but any tips or pointers would be welcome. I am in the process of redesigning my in-car system and want to hook into the EMS this time, if possible. All running Linux, of course. Cheers, <PERSON>.
60faec29-bda1-4923-c3de-11047719f567
['2112e41c-2c09-e909-285d-8ecbf545e8e3']
Hi Guys, I've updated the program, and used the latest winavr distro, and its still see nothing in hyperterminal. Just to give more background, I have receiving working fine. Also the origional code seems to transmit ok, so I gradually removed parts of the code & re-tested, gradually it got more unreliable (only sending a message every 3 times it was turned on), then just sent one space instead of the message, until it did nothing, and this was just by removing code not relevant to transmitting! In one case removing an unsued global int stopped it working! What could cause something like this? Thanks for all the help, <PERSON>
609a6912-2a7f-70a0-2abd-65763957a200
['2112e41c-2c09-e909-285d-8ecbf545e8e3']
Hi, I noticed a bit of duplication between the zip & tar ioslaves, so I've replaced all instances of KTar to KArchive, so the tar ioslave is now generic for any archive type supported by the KArchive classes. Please see the attached patch, if its ok I'll add the relevent protocol files & remove the zip ioslave from the makefile.am. <PERSON>
9825e0f4-2f6b-7ffc-03cc-cf5a500d52ec
['21259823-37b6-e8bf-1804-759387193fb5']
I've just encountered a strange problem from one of our remote users. She called me to say that when she clicked on a proxied link she was directed to the appropriate database, but the database didn't recognize the IP as coming from our campus (she was not presented with an EZproxy login screen). She had been authenticated via EZproxy earlier in the day, but had spent a long time using non-proxied pages in her browser. When she clicked on a proxied link, she was not presented with an EZproxy authentication screen but was sent directly to the database, which of course didn't recognize the IP as belonging to our university and thus presented her with a login screen. Closing and reopening the browser fixed the problem. Any thoughts on what could be causing this? ERic
40a28e5b-b1a7-1a1e-15cc-ba62e0b78de8
['21259823-37b6-e8bf-1804-759387193fb5']
We would like to find a way to limit access to a specific resource to a subset of our students. We are using the flat file "loginid:password" approach. If I came up with a file listing only those students who have access to a particular resource, is there any way we can tell Ezproxy to only use this file for authentication of that specific resource? <PERSON>
9bf96e86-370a-6d41-6419-e8249ae8fe3a
['212b148f-1eb6-f81e-4538-d74fd4fdedac']
<PERSON>, In that case, I do not think there is exactly what you want (or not one I know of). I would write a simple subroutine to do this, and write it such that you can just plug it in where ever you need it, including (and maybe most importantly), into a Dict. Pass it several arguments (like the raw data, what the delimiter is (FM, VM, etc), and such) so that it can be reused anywhere. <PERSON>
9337cb91-cb94-aac2-c09b-a841ac595e34
['212b148f-1eb6-f81e-4538-d74fd4fdedac']
I have always preferred dynamic files. In the long run, they are easier to deal with. You should still determine the best block size. Also, if you are going to bulk load it, you should set the min mod so that it does not start splitting right off the bat. You still want the basic size to be accurate if you can determine that up front. That being said, there is the nasty "blink" error (backward link). The only time I have seen these were in a vendor's proprietary index files, but when we got them, the file was worthless. Fortunately, since they were not "data", we could rebuild these files. <PERSON>
d0ed9cf3-3288-e280-3efe-e5e667b610e2
['213223d0-6a61-4f84-c862-a23ab74874b0']
Hey All- My name is <PERSON>. I am a second year PhD student at the University of Illinois at Urbana-Champaign and am interested in working with the Nmap project for GSoC 2011. With this message I would like to get to know some of the devs, as well as outline my initial thoughts for the project. I would like to participate in the enhancement of Nmap's IPv6 capabilities. As I'm a researcher, I'm inclined to tackle more complex problems such as OS detection. I have previous experience in using statistical packet analysis to perform classification of encrypted traffic. At this point I still need to specify in greater detail my ideas and scope for the project, but figured it would be good to start here to make sure that I'm getting the right feedback throughout the process. I will say I don't know how current IPv4 host detection occurs, and assume that I should start there. The following lists an initial approach I would take in order to develop host detection: * Review IPv4 host recognition techniques and other literature on the subject * Review IPv6 RFC Specification * It seems as though host detection is very specific to the OS and other implementation specific issues, and as such profiling the different systems seems like a good first step. I would manually review packet traces from each OS in order to find any unique state produced by the system. * Review other state output visible to the network. This task would be focused on exposing any unexpected state that could be used for host detection. * After manually analyzing these traces and other output from the hosts I would start to develop some type of classification of different types of data we find valuable in performing host recognition. * Would need to analyze and define what type of pattern/classification technique we will use. * The next thing would be to build some type of initial prototype and see how it does. * Then make modifications and recurse over testing and modification until the application performs as desired. Like I said this is a very raw initial approach. Please provide any feedback to point the project in a direction that would better serve Nmap. I have a few questions: * What is the potential for publication coming from this work? Would Nmap be okay if attempted this, and would there be interest from Nmap to participate in this? * Is this too advantageous of a project, or would I also need to add in some other work? Additionally, the following link returns a 404 error: http://socghop.appspot.com/gsoc/org/home/google/gsoc2011/nmap As well as: http://www.google-melange.com/gsoc/org/home/google/gsoc2011/nmap These are the application and application template links. Thanks,
228da12f-58b8-1d01-9e01-74780f994dab
['213223d0-6a61-4f84-c862-a23ab74874b0']
Hi <PERSON> am a graduate student at UIUC and am currently working on a system that isolates the MMU from the rest of the FreeBSD kernel. For the purpose of enabling privilege separtion within the kernel. - This code is approximately 3k lines. - This base system also provides kernel code integrity (write protection in memory) as one of its base properties. - This is somewhat follow up work on previous publications VirtualGhost (ASPLOS '14) and KCoFI (IEEE SP '14). The new system uses similar MMU policies to create the isolation within the kernel. Controlling the MMU enables read/write protection policies (for memory pages) to be enforced within the kernel itself. I am interested in thoughts from the community on how to best use an intra-kernel isolation and mediation mechanism? One example policy or use of this mechanism would be to place critical function pointers into a write protected region of memory and apply a policy where function pointers (such as a system call function pointer) are only allowed to be set once (a write-once policy). Some initial ideas I have include: - Protecting against common root kit behaviors such as system call hooking, character device hooking (protect function pointers), DKOM (modifying process lists to hide data), kernel object hooking, etc. - Protecting capabilty enforcement - Mandatory Access Control Mechanism enforcement - Auditing mechansims (to ensure integrity of audit logs) Does anyone have insight into these or other important uses of this type of system? What would be a "killer app" type use in the kernel? I can be available on IRC if a real time discussion seems like a better place for discussion. Thanks, <IP_ADDRESS>nathan<IP_ADDRESS>
cacaaefb-7fe2-f65b-2539-9f33eb216dc9
['21491166-63b7-e63c-d521-9a98a365eeb0']
Thanks! That did exactly what I want for cos**2. I would also like to reduce the power on higher order terms. I tried TR7() on cos(x)**3 and cos(x)**4 and it did not make any changes. From quick glance here (https://github.com/sympy/sympy/blob/master/sympy/simplify/fu.py), TR7 seems like the right function for power-reducing. Do you know of a method to do the same thing with larger powers? A clean way to iteratively break off 2nd-order chunks, apply TR7, and distribute would be a viable, but kludgy, alternative. Thanks, <PERSON>
062c18ff-35a9-8efb-2a34-be87625904c0
['21491166-63b7-e63c-d521-9a98a365eeb0']
I just mean that if I'm trying to write a script and I want to reduce all of the powers of cos or sin, I'm not sure how I would code that, because I can't tell if there's a pattern of which calls need to be made in which order to reduce an arbitrary polynomial of trig functions. For my present purpose, I'm doing the analysis interactively, so a little trial-and-error is fine. For future projects I'm trying to get a clearer understanding of how these functions work. So for example, why does TR7(cos(x)**7) just return cos(x)**7? Thanks, <PERSON>
aa3f515e-b02f-2237-d6bc-b39d7e1eeec4
['2149cf36-46d2-29fb-aec8-5a18f4b8856d']
All, HI, I'm afraid I've fallen a little behind in keeping up to date with the latest version of Squeak. I think I'm using 1.2 or 1.23. I'm afraid I'm not quite sure where to FTP the latest verison of Squeak for Windows95. Could someone direct me to the location and needed files. Thank you, Take care, <PERSON>
91d6f60f-4f99-f4c7-9009-9d0d5dd76734
['2149cf36-46d2-29fb-aec8-5a18f4b8856d']
All, Hi. I'm using VisualWorks NC on Linux and have Squeak as well. I was going to be using VWNC for a semi-commercial usage but I may not be able to afford it commercially. I'm using it for a Web Server (using a Web Server I built in VW similar in concept to the Squeak PWS and an HTML framework I wrote). I'm curious if Squeak has general performance good enough to investigate using it as a web server. Thanks and take care,
3ded2b9b-e776-f37e-82dc-bd0536b6c269
['214ee8a4-4630-b43e-bc4d-b93c348d52f7']
I'm new to MySQL, and only have limited experience in modifying some legacy dBase IV code. One of the minor aggravations I occasionally encounter when using management systems is how best to handle names, particularly the many individual name segments of names. Typically the name field is of a fixed length which usually is way too long for most names, but occasionally too short for other (often important) names. Also, sometimes there are two name fields, both of the same length, which is great overkill for most needs, but still inadequate for some. Is there a good solution to the problem of handling a wide variety of name segments, that are part of a wide variety of names, that belong to a wide variety of entities? It would be ideal if it were possible to have an unlimited number of name segments that combine into one name AND also the ability to have an unlimited number of names. Any suggestions on how this might be done? Thanks, <PERSON>
5fe6010e-003b-ecd0-b8b0-14ae2cb910cf
['214ee8a4-4630-b43e-bc4d-b93c348d52f7']
From the manual: "If you don't have any variable-length columns (VARCHAR, TEXT or BLOB columns), a fixed-size record format is used. This is much faster but unfortunately may waste some space." Can anyone more precisely quantify how significant is the difference in performance if the table uses all CHAR fields versus all VARCHAR fields? Also, if VARCHAR() is used in a table, what would be the reason to use, say VARCHAR(25) rather than always simply using VARCHAR(255)? V Harris
a57dfbfd-5d64-22b0-02d3-8a1254d3abde
['2157d085-bd8b-1a28-5d4b-e76e9ebbe409']
Hi Everyone, I was looking over some pending tickets today (in preparation for a future 1.4 I hope)... and ran across an issue with MochiKit.Format.truncToFixed and twoDigitFloat: http://trac.mochikit.com/ticket/275 It's one of these boring floating-point rounding-errors. I think twoDigitFloat can be easily fixed by using roundToFixed. But the issue in truncToFixed seems a bit harder. Is there anybody lurking on this list with a good idea for solving this by any chance? Perhaps we should just use Number.toFixed() directly if the original problems in Safari have been resolved? Any ideas regarding this would be most welcome. Cheers, /Per
36c31627-e4d1-9173-35b9-4c9debbf46cc
['2157d085-bd8b-1a28-5d4b-e76e9ebbe409']
Hi, I just wrote up a new page on the Wiki with a few suggested extensions to MochiKit. None of it has been tested too much, but I use most of it in my own JavaScript applications and widget framework: http://trac.mochikit.com/wiki/MochiKitStyleExtensions I'd appreciate any comments or suggestions regarding this. Also, feel free to provide bug fixes or links to similar work (that I obviously missed). Cheers, /Per
3a811cbc-5cc2-81af-01c5-dbde9fd97c0c
['2169b468-5b7c-6381-d2cd-68fb1b7de02d']
All, Somebody pointed me recently to this ITSM Tool list<http://www.servicesphere.com/blog/2012/3/5/where-can-you-find-a-list-of-itsm-and-help-desk-tools-with-o.html> . - Not seeing BMC Remedy at the top makes me feel strange. - Seeing almost-unknown technologies have more "likes" than BMC Remedy makes me feel very bad. Should you feel the same, please thumb up BMC Remedy<http://www.servicesphere.com/blog/2012/3/5/where-can-you-find-a-list-of-itsm-and-help-desk-tools-with-o.html> . *Note*: to authenticate on this list.ly-powered platform, you can use your Facebook or Twitter account, quick and easy, no hassle, no spam :) You all have a great Friday and a wonderful week-end! ~ <PERSON>, BMC Software Senior Community Ambassador, BMC Communities cVsT03fgj/RnhoaY@example.com Follow me at @Matt_L <https://twitter.com/matt_L> Skype: matt.laurenceau
5426ba92-6345-36af-c851-3e9da8b14eaf
['2169b468-5b7c-6381-d2cd-68fb1b7de02d']
BMC is launching a monthly *BMC Atrium Ask the Experts* series to discuss Best Practices and to answer your architecture and implementation questions. The format is informal and simple. BMC product managers and architects will start the call with a short discussion on a select BMC Atrium topic, then open up the forum for YOUR questions. All Atrium-related topics are fair game. We will have *Architects, Product Managers & Product Marketing* on the phone to answer your questions. We ask
1d98bac1-18e1-165b-7370-f3ebe7409be2
['2185dbf4-8ced-4c30-98ab-ecdc1102bcfd']
Is there documentation that will show me how to modify or replace the portal_syndication tool during a migration? Is generic setup ready to tackle this in 3.0? Or is there an old-school way of doing it that I will have to master? The summer of code draws to a close and I face migration.... Thanks, <PERSON>
1e2c1f9c-634e-d3e6-0c10-32590d5254c6
['2185dbf4-8ced-4c30-98ab-ecdc1102bcfd']
I am interested in implementing PLIP #128 (Plone Pluggable Syndication) for Plone 3.5. I first expressed this interest on plone-user and <PERSON> suggested I post here and request input. First, let me say that I am an experienced software developer but only a novice when it comes to Zope + Plone. I've read parts of the Plone source code and patched a few small defects. Fortunately, I find that my knowledge of patterns enables me to get Zope3 concepts fairly easily. I expect the hardest part of implementing PLIP #128 will be understanding and porting the existing Zope2 infrastructure. However, I have experience porting from a platform that is being phased out. ;) I will be attending Camp5 and, thus, in mid-March should have proven a basic working grasp of Five development for Plone 3. Given that caveat, I am requesting input for desired features (with use cases) and architecture for PPS. I have read philikon's FiveFeeds tutorial and I am assuming that is a good starting point for what needs to be done. But, instead of being an add-on (as in the tutorial), the existing plone syndication tool should be rewritten for Five. Would updating the CMF syndication code be included in this PLIP? I assume so. I am also guessing that some DublinCore work may be required, as <PERSON> states that, as of the tutorial publishing date, we need "an adapter that does the mechanical translation of the CMF DublinCore API to the Zope 3 one," though perhaps this has already been done. Primary target formats are RSS 2.0, Atom, and RSS 1.0 (for backwards compatibility). I would also like this to be available as a product that runs on at least Plone 3.0, perhaps Plone 2.5.x, as well (how hard is this?). What am I missing? What is in your wishlist? I know I have a lot to learn here, but I assure you that I am industrious and a quick study. It occurs to me that it would be nice to have a mentor for this project. I can always post to this list, but the volume of requests for guidance I may have may look like noise here. ;) Anyone interested in taking on this job? If not, might this be a potential gsoc project, assuming my employer doesn't see a conflict of interest (I've already gotten approval to donate back)? I am an MS Computer Science student at the Georgia Institute of Technology. I will run the contributor agreement through my employer's legal department. Can the successful implementation of this PLIP serve as my proof of skill for svn access? I attended Plone Conference 2006 and was amazed by this community. I am eager to contribute back and this PLIP matches my employer's needs with a need of the community. But don't be reticent to push back if I'm biting off more than I should... ;) I am have time available to begin immediately. Thanks, <PERSON>
19655a12-7585-c1d9-a9b7-94045d71df9d
['2186cfbf-8ca9-9fe3-137a-5fc163fc155d']
This whole thread is striking me as akin to a child going "but why" every five seconds ad infinitem. Let's look at the issues here. 1. Google (and other companies, such as Microsoft and Yahoo) make available satellite / aeroplane images on the net, free at the point of use, for all to view. 2. Taking an image from a satellite or aeroplane image requires no copyrightable skill: Camera points down, takes images at fixed focus at regular time intervals. Images are published. 3. By using the geolocation of known reference points on each image block intermediate locations can be interpolated. 4. Geolocation data can also be obtained by hand-held devices and from paper maps and many other sources. 5. "mass database extraction" requires *mass* - ie large proportion or 100% extraction of that data. WP doesn't do that. Any extracts are made individually for each case required. There is no automation in place. Even if one were to presume to consolidate every WP editor into a supposed single entity (which is precluded by the very nature of the projects anyway) it still would not be "mass". 6. "mass database extraction" requires *database* - ie accessing that data in the form of data tuples. WP doesn't do that. If this source is used it is manually by eye by an individual reading and entering the data. 7. There is no database right applicable to this process for the same reason: there is no access by a WP editor to a *database( of tuples or otherwise. 8. Stating the source of a fact - any fact - within WP isn't required for anything which is self-evident ("The sky appears blue during the day and is dark at night") or easily provable ("gravity on earth pulls things down"). Geographic *co-ordinates* per se therefore do not need a source. 9. WP are not slavishly copying the source data, nor are we using the agglomeration of that data in the form of a map. As such we are not contravening the copyright nor contract law which may or may not apply. 10. What OSM does is not - directly - of concern to WP. And vice-versa. Obviously however one should hope that an equitable and valid way can be found to share information. <PERSON>
3c60f6ad-f129-f7cb-000e-e5eb5c177710
['2186cfbf-8ca9-9fe3-137a-5fc163fc155d']
I didn't want to ask this actually while voting was open in case anyone got worried, but not that voting has closed I'd like to ask something. How are our votes actually counted and, more importantly, how can we each be certain that the votes we made are actually the ones which are being counted? I ask this because of the issues raised in the USA about election fraud (http://en.wikipedia.org/wiki/Diebold#Security_Concerns etc.) and wondered whether the same could happen with us, After all, the voting isn't being carried out on independent servers it is on Wikimedia servers and, presumably, a lot of people have access to those who could do things without leaving a trace. I am *not* meaning imply that anything has been done, but I would very much like to know what security voters like myself have that our votes have been correctly recorded and tallied. <PERSON>
f75978ca-a8bc-9ff6-7a74-8c55f534a89b
['21993df9-3bd5-49b6-ef96-0a80258170d6']
Hi, I have a couple of questions about Quad-BRI solutions for Asterisk, and was hoping that I might get some feedback based on other people's experience. We currently use the Junghanns card, which is a pure Zaptel solution, which is fantastic, but they have no hardware EC solution, and their drivers are becoming increasingly un-stable with time (I back-port to a modified qozap driver from 0.2.0-RC8n which is the last one I can run without bad behaviour) I am aware of the Beronet BN4S0, which appears to be "exactly" the same card as the Junghanns card, but with mISDN drivers, still no h/w EC solution, and as a result, less effective (from what I read here) software echo cancellation. I thought I had struck gold when I saw the Digium B410P, which had a driver that builds as part of Zaptel, but then when I read on the list people describe it as an mISDN based card... Which is it? I prefer a ZAP based driver because I use that for the Sangoma and Single-BRI solutions that we build. I assume that if I switch to mISDN, I will need to install all of the Linux ISDN support, change my dialplan to use CAPI/ as a technology, use new and unfamiliar config files, and all sorts of other horribleness, probably losing one or two ZAP/ specific features such as the ZapEC() command in the process? Perhaps there is an alternative solution that I have missed entirely out there? The Single BRI (HFC) card has the "vzaphfc" alternative driver available, has anyone done the same for the Quad (HFC4S) card? Thanks for any pointers that can be provided. Kind regards, <PERSON>
f067e175-0e83-ab26-755e-beb335da0e01
['21993df9-3bd5-49b6-ef96-0a80258170d6']
Hi, I have a very simple setup with 2 nodes, using the basic resource manager. We are space constrained on the server so cannot easily install all of the many dependencies for the more complex resource managers, but our needs are simple :) - Start with 2 running servers in master/slave mode, and all happy. - Kill the master (A). - The slave (B) is coming up - Some transient issue prevents the RC scripts running on (B). - (B) backs down and requests to become slave again - (A) is down, so (B) never gets confirmation of its slave request. Nothing more happens. A is down and B is sulking! Can a node be persuaded to retry under these circumstances? Perhaps there is a way to identify this odd intermediate state so we can force a heartbeat restart or reinitialise? Thanks for any pointers. Regards, <PERSON>
ff520b7f-b0a9-8966-da5d-60a81241cbf6
['219bb199-2b13-dd9f-c3f4-72db90d2b6c5']
Exactly, displaying an 'ip' register which is not the instruction pointer is very unhelpful. The existing register names are currently different on posix-vs-Darwin anyway, so there shouldn't be any harm in making the change now. I think it's "fioff" on Darwin. It uses different display names vs. the internal variable names. I agree that it would be nice to make it consistent everywhere. Personally I'm in favor of it matching the Intel names as closely as possible. I'm slightly puzzled by how many places in the codebase (re)define the register sets, it seems like making changes like this may not be straightforward. Richard Mitton <EMAIL_ADDRESS>
c268a61e-79ba-8346-2d1f-937795619886
['219bb199-2b13-dd9f-c3f4-72db90d2b6c5']
Why does this replace a condition variable with a 10 millisecond spinlock? Spinlocks are bad. I don't know about the C++11 standard, but if you want the same effect on Windows you'd just use SetEvent. Also don't use volatile. In fact, never use volatile. m_output_flushed is a class variable, so the compiler knows to reload it each time anyway. Richard Mitton <EMAIL_ADDRESS><PERSON> pYV05PPkH2DJB7+p@example.com
d24c705f-de8d-f69e-8c88-4375d7d23ec1
['219c033e-76f4-24d8-6856-fdec095f04ed']
Hi All, I have use case to put data in avro format in Hbase , I have frequent read write operations but its not a problem. Problem is what if my avro schema get changed how shuld I deal with it? This should in mind what about older data which already inserted in Hbase and now we have new schema. can anyone suggest me solution for the same Thanks <PERSON>
5a8efec9-efbd-d50d-c07a-770662dc1e93
['219c033e-76f4-24d8-6856-fdec095f04ed']
Hi All, I have to find which is the best way to query on Hbase will give best result options are as below if any one can help 1. REST API 2. Using Thrift: 1. HBase and Thrift in .NET C# Tutorial http://pawelrychlicki.pl/Article/Details/52/hbase-and-thrift-in-net-c-sharp-tutorial-c-sharp-45-and-thrift-093 3. Commercial ODBC/ADO.NET Connectors 1. http://www.cdata.com/drivers/hbase/ado/ 2. http://www.simba.com/drivers/hbase-odbc-jdbc/ 4. Apache Drill 5. Apache Phoenix Thanks Manjeet
90c001c1-beee-e438-c5a4-53bea0943653
['21aa33ac-e6c1-aa05-5e14-d5c630e4a664']
Not sure it is specific to build 6111 but I see that the main window buttons are disabled if a composer is open. Ie. if I am composing a message I can not get back to the main browser and, say, delete from the message list using the button. The menu items and contextual menus are still active so it is not a major issue. A. Takacs
de28a1d8-7ba9-3087-e818-bb0f124e1ea6
['21aa33ac-e6c1-aa05-5e14-d5c630e4a664']
I'm seeing a fairly bad case of message corruption in the following scenario: 1. create a fairly substantial message (size wise - typically with some attachments ) 2. send it 3. while being transmitted cancel the send (via contextual menu, selecting the message in draft folder). This will - in my experience - yield an immediate "database corruption" error. If at that stage if you elect to "do nothing" and relaunch MM the offending message will have vanished (not in sent nor in drafts). Not good :( A. Takacs
b75dad59-0136-7ce6-76af-b71631517628
['21b99b88-89d8-3f4e-748f-d517af0f8f28']
I need help figuring out how to filter sysex messages. I noticed a lot of sysex messages (actually just the same message showing up over and over) in all of my MIDI tracks. After a lot of testing, I have discovered that the messages are actually being generated by one of my MIDI interfaces (I was sure it was some synth). I have a MOTU MIDI Express XT USB which is the offending unit. So, my question is: How do I filter all incoming Sysex messages? I am mostly a Digital Performer user, but I have been using Logic steadily on my laptop for the past year because of it's great strength with virtual instruments. Now that I have VSL, I need the additional horsepower my desktop machine can offer, but I am hung up because of this stupid Sysex problem! ARGHHHH! In case you are wondering, I read the manual, online help, and checked out infoweb. I tried using the transformer in my MIDI environment, but I can't figure out how to make the SysEx mapper work (maybe that's the wrong function?) I also tried the other filers in Transformer, but they don't seem to be able to filter out sysex messages. I would greatly appreciate any help with this! <PERSON>
02be90d6-0048-5e2f-1f87-edb6bd42a8dc
['21b99b88-89d8-3f4e-748f-d517af0f8f28']
I was once a die hard DP user before I switched to Logic. I came for the ability to use VSL, and stayed for the excellent soft synth handling. I'm frightened to admit that my 30+ hardware synths haven't been switched on in over three months now, tho the tracks keep rolling out. Ahyhoo, on to my question: As part of my transition to Logic-ness, I kept using MOTU's Mach 5 sampler plug because I had all of my samples already scanned into it. At some point over the last three months, through whatever updates I have done, Mach 5 doesn't work in Logic anymore. It always crashes the inspection. There's a note on MOTU's website on the downloads page which indicates "Attention Logic users: MachFive 1.2.3 requires an update to Logic 7.1 for compatibility under Tiger OSX 10.4.x." So... that doesn't help me. I have the most current version of Mach Five. Are they trying to tell me that I need a newer version of Logic? (I currently have 7.1.1) This will probably be the last time I ever use Mach 5 (Thank goodness) but I need to run it one last time to see what patches I used in three pieces I wrote some time ago. Thanks for any help you can provide! <PERSON>
682b06b0-7ea5-184e-b235-39fcfbd237dc
['21c2c8ad-6c0d-1b32-e950-6eab035080e8']
Hi, when I use the gradient filter (unstructured dataset), I observe edge effects which are IMO unphysical. For the nodes on the outer boundary of my domain, the gradient magnitude is smaller than I would expect. The behavior is like the filter tries to use the adjacent nodes to calculate the gradient, and since there are none outside the domain, it assumes zero and finds a lower result. If so, I would consider this a bug - is there a fix or workaround? Thanks, <PERSON>
e9161c25-c070-3478-3f53-a9c41722381f
['21c2c8ad-6c0d-1b32-e950-6eab035080e8']
Hi, I need to convert legacy vtk files to vtr files. The vtk files are in ASCII format and approx. 17 MB large. When I use binary or ASCII as data mode, I get a file size of approx. 28 MB. When I use appended data mode without encoding, it's 20 MB. When I use encoding with ZLib compressor (there is only this one), the shinks to approx. 800 kB! Also data access is a lot faster when I load that file in ParaView. My questions: 1. Why does the file grow for most of the modes I have tried? 2. Is the ZLib compression used in ParaView lossless, or do I lose information when I apply it? Thanks for some explanations... <PERSON>
a57fc21d-1eb6-a1bb-e342-cd886381f223
['21c542a3-dc81-3734-57a1-d0edc1fa88fe']
Marxism Today was published in the eighties by a group of young British CP dissidents. Quite a good magazine if a little shaky on "Marxism" - very good on queer and racial issues. Long gone however, though I still have a pretty good collection left over from college. <PERSON>
ccd7728f-b665-bded-14c7-12d0d029873e
['21c542a3-dc81-3734-57a1-d0edc1fa88fe']
<PERSON>; Like you, I would like to know more about the miners' day-to-day struggles. But, I jib at the idea that this was not a genuine demonstration against the westernizer <PERSON>, whom they have every reason to loathe. There are real resentments here against the type of affluent, English-speaking yuppie types who are everywhere on the opposition barricades, in Kiev and in the cities of western and central Ukraine. Too, I do not share your skepticism about the miners being wholly a captive audience. I work in the construction trades, and there are always around election time, rallies and demonstrations by the building trades that barely mention real grievances, and instead focus on the personalities of this or that candidate. Isn't that what was happening in the Ukraine? We must be careful I think not to give short shrift to movements or events simply because they cannot pass muster doctrinally by the lights of more socially conscious workers, especially here in the West. <PERSON>
b543fd64-fd9d-8a0d-bd05-bc34c73d4ee9
['21cda90b-c42d-c75e-d00f-fe1a76e045ca']
I'm attempting to map my domain object to a legacy database table. The legacy table is in SQL Server and has a primary key set to: int not null no incremental id generation I have tried unsuccessfully save domain objects to this table, but I always receive the error "Constraint violation, primary key cannot be null". How can I manually set the value of the primary key before saving? I've tried: employee.id = 1111 employee.save()
4aa0383f-f8fe-9ba1-27fc-45dabfbf3fea
['21cda90b-c42d-c75e-d00f-fe1a76e045ca']
I'm experimenting with the platform-ui plugin and was looking for some examples that show how to layout pages with navigation elements, like a sidebar page. Here's my first attempt at creating a page that uses a sidebar layout: https://github.com/ddelponte/furthercms/blob/master/grails-app/views/public/sidebar.gsp Is this correct and portable between themes? I'm concerned about the use of the specific classes in the list elements. Thanks everyone for your help.
7dccc8d2-cac0-2062-0a19-1339a65c6420
['21e74515-66da-5d7b-9028-83d24951a180']
Greetings I'm new at GRASS and I have been developing a few scripts toGRASS but I have afewquestions regarding Scripts wxpython frame (the one defined by %MODULE): - I want usersto input a value in an OPTION regarding a parameter (e.g. a string or a number). ==> This is EASY. But, I also want to compute the input, automatically, before I run my scritpt, in order to provide a list of possible values in other OPTION e.g. in option "Parameter" the users put a value (e.g. Vegetation). Automatically a small processa is run in order to provide a list of possible values for Option "value". And then, users will be able to choose "Values" options that is e.g. 100,10, 0etc". Only after this values were inserted is possible to run the script. This seems complex but my question is: is it possible to run a function, while I'm inserting values in scripts, and all this before I run the script? Best tregagrds <PERSON>
cdcfdf64-45a6-fe15-4f14-20ecf9b63b21
['21e74515-66da-5d7b-9028-83d24951a180']
Hi I'm a Environmental Engineer working in Remote Sensing and landcover classification. I have been using ITK for segmentation and not much. Now, I believe I should "increase" my uses by using for Supervised Classification. I saw that a couple of days ago <PERSON> sent a message regarding the use of Supervised Classigication methods. But, as far as I can see at the archives no one reply to him. Could anyone indicate me where can I retrieve some information regarding implemented Supervised Classification algorithms (except Software Guide ITK). Thank you for your help Best regards, <PERSON>
e1e81e10-1765-ad8b-304e-3ccda6902eba
['21ead8fa-c481-d0d3-e3e6-0c2d36388750']
I googled <PERSON> for Vietnam and Navy ROTC, figuring that as an officer he must have been in Naval ROTC at Yale. Several sites say that he enlisted. No further explanation is given. ROTC never showed up. Nor did I find out how he became an officer. Did he enlist and then go to a Naval Office Candidate School? <PERSON> also volunteered, but I've forgotten the details. Both were already planning their resume as liberal democrats, making sure that "old glory" would wave behind them. from <PERSON>
bc5dfa17-03b5-8080-378a-ed6db25470ef
['21ead8fa-c481-d0d3-e3e6-0c2d36388750']
I wrote "rules", when I should have written "rulers". And then in cutting and pasting I said "<PERSON>/Blair" instead of "<PERSON>/Bush"! How easy it is to be corrupted! I have also strengthened my second paragraph. I disagree with <PERSON> <PERSON> the nature of the "real world" and about the role of revolutionists in fighting against the patriotism (however hidden) for their own country. "<PERSON>/Bush" says: "Yes, we are equally morally responsible" for our rulers are conducting this war (including providing the lies about 'yellow cakes,' etc.) to the extent of our economic and political strength." Within Britain, adopting a slogan that puts <PERSON> before <PERSON>, i.e., the Blair-Bush War, may appeal to the revolutionary-pacifist sentiment that we are the enemy, i.e., that we are responsible for our own leaders and we must get rid of them. <PERSON> is there--in the United States; <PERSON> is here--in Britain. Of course, this particular slogan was meant to suggest a strategy. What would work best is up to the activists within any country. However, The Militant was basically correct. The overwhelming emphasis on <PERSON> during his March 2003 visit meant that an educational opportunity for revolutionists was lost. from <PERSON>