text
stringlengths
1
3.78M
meta
dict
1. Technical Field The present invention is related to integrated circuits. More particularly, the present invention provides inspection methods and structures for visualizing and/or detecting specific chip structures. 2. Related Art Many back-end-of-the-line (BEOL) semiconductor process layers and structures are optically transparent when planar. The optical transparency of such layers/structures makes undesired planar residual chip structures difficult to detect. For example, inadequate removal of a Cu BEOL TaN liner can result in low yields due to conductive shorts formed by the liner itself. An example of this problem is illustrated in FIG. 1. In this example, chemical-mechanical polishing (CMP) failed to completely remove the TaN liner 12 in the area 18 between Cu metal conductors 14 and 16. This can result in low yield due to shorts between the Cu metal conductors 14, 16 caused by the residual conductive TaN liner 12. It can be hard to detect such underpolishing of the TaN liner 12, since the TaN liner 12 is planar and therefore difficult to detect because of its optical transparency. A physical cross-section is often the only sure way to determine if such a structure is missing, not located where it should be, or present where it should not be. However, physical cross-sections are destructive, time and resource consuming, and must be provided off-line. Physical cross-sections also need to be directed to specific location(s) on the wafer or chip, unless the problem is very gross. There is a need, therefore, for inspection methods and structures for visualizing and/or detecting specific chip structures.
{ "pile_set_name": "USPTO Backgrounds" }
Q: How do I work with Multiple Recordsets in C++ ODBC I am trying to streamline a complex process of storing information in multiple tables and them linking them to a central table. The linking occurs using IDENTITY values generated in each table to provide the unique linking. I know I can use a combination of SET NOCOUNT ON and SELECT @@identity to get each identity, but that still requires me to call a separate SQLExecute() for each table. I have read dozens of articles saying ADO can handle multiple recordsets using an ODBC driver, so the question is how do I do it without ADO? I have already encapsulated all of the standard ODBC stuff for similar behavior to ADO. I basically just need to know what ODBC API calls will allow me to recreate ADO's NextRecorset(). I am working on a combination of MS SQL 7 and MS SQL 2005, using either the SQL Server ODBC, or SQL Native Client Drivers as appropriate. End Goal: SET NOCOUNT ON; INSERT INTO TableA (data) VALUES ('a'); SELECT @@identity AS NewID; INSERT INTO TableB (data) VALUES ('b'); SELECT @@identity AS NewID; INSERT INTO TableC (data) VALUES ('c'); SELECT @@identity AS NewID; ... RS = DB.OpenRecordset() RS.MoveFirst() A_ID = RS.GetValue("id") RS.NextRecordset() RS.MoveFirst() B_ID = RS.GetValue("id") RS.NextRecordset() RS.MoveFirst() C_ID = RS.GetValue("id") A: Use the SQLMoreResults() call as the analog to the NextRecordSet() function. However you probably don't need that if you are willing to make your executes consist of INSERT ...; SELECT @@IDENTITY Since the only result returned from this statement is the identity, you don't need to bother with the SQLMoreResults().
{ "pile_set_name": "StackExchange" }
Sydney Vincent Sippe Major Sydney Vincent Sippe (pronounced -ee) (24 April 1889 – 17 November 1968) was a British pioneer aviator. He designed, built, and tested early aeroplanes, being the first pilot to take off from the sea in Britain. He flew many missions in World War I, including some of the first ever bombing raids. He won honours from several countries, particularly for his part in the 1914 bombing of a German Zeppelin factory. Early life Sippe's parents were Charles Henry Sippe (1842–1924), a shipping export agent (whose firm, C H Sippe & Sons Ltd, still existed until c.2013), and Elizabeth Jane Thornton (born 1846). They had moved to Britain from Australia, both families having originally emigrated from Liverpool. The youngest of nine children, Sydney Sippe was born in 1889 in Brixton, London, where his parents lived at 17 Lambert Road. He was educated at Dulwich College from May 1903 to December 1905. Name and title Sippe was named after Sydney, Australia, where both his parents had lived. His first name is often misspelled 'Sidney', even on official documents, and his surname sometimes written with an accent, 'Sippé', as if of French origin. The form 'Sydney Vincent Sippe' is used both on his birth and death certificates. Following his service during World War I, Sippe continued to use the title Major – and was known as 'the Major' – even though the RAF rank was renamed Squadron Leader in 1919. He also went by the nickname Pi (pronounced like 'pie'), and was known to his family as Pipi. Aviation pioneer After leaving school, in February 1906 Sippe became an engineering apprentice with British Westinghouse in Manchester. Between late 1909 and early 1910, just a year after Wilbur Wright first demonstrated powered flight in Europe, Sippe (aged 20), his brother Arthur, and their friend James Jensen (or Jenson) designed and built a monoplane from steel tubing. Its attempted maiden flight on 24 April 1910 at Addington, Croydon, failed due to insufficient power: Sydney Sippe was thrown forward with some violence and his nose came into collision with one of the steel tubes. The nose came off worst, and a piece of flesh was removed from the inside of his thigh. That, with sundry bruises, was all. A well-meaning friend rushed up with a flask of whisky, which he thrust into the pilot's mouth, and so Sydney Sippe arrived home to his mother with a broken nose, a bleeding thigh—and slightly intoxicated. He learned to fly at the Avro school at Brooklands, gaining his licence in January 1912 "in a way which showed that he had thoroughly mastered the art". He immediately became a test and demonstration pilot. Three weeks after his flying test he survived a crash near Finchampstead caused by a frozen carburettor; the aircraft was wrecked, but Sippe escaped unhurt. In the spring of 1912 he test-flew the Avro hydro-aeroplane at Barrow-in-Furness, making the first ever flight from the sea in Britain on 2 April. He shortly afterwards tested a monoplane for Hanriot in France. Engine failure forced him to land in a cornfield; the plane ended up upside down, but was undamaged, and he flew it again in the first Aerial Derby. This was a race round London, competing against various European pilots including Thomas Sopwith. Engine trouble forced Sippe out of the race. Later that year he supervised the construction of aircraft in Milan for the Bristol Aeroplane Company, and took part in an Italian long-distance flying competition. On his return to the UK he took up flight testing and instruction at Salisbury. War service In 1914, at the outbreak of World War I, Sippe immediately joined up and was made a Flight Lieutenant in the Royal Navy Air Service. He took part in the Entente's very first bombing raids on Düsseldorf and Cologne; in the latter, after failing to find his designated target, he instead bombed Cologne railway station, causing serious damage. Friedrichshafen raid This was followed by a celebrated attack on the Zeppelin sheds and factories at Friedrichshafen, Germany on 21 November 1914, one of the first long-distance bombing missions. Sippe and two other pilots flew from Belfort, France, over mountainous terrain and in difficult weather—a risky flight near the limit of the aircraft's range. The distance was increased by the need to avoid flying over neutral Switzerland. Reaching the target area, Sippe crossed Lake Constance in mist while under heavy fire, descending to just ten feet above the water so as to use the mist as cover. Despite their aircraft taking damage, the three pilots succeeded in bombing their targets. Although substantial damage was claimed at the time and in some later histories, the damage inflicted was slight. One pilot was shot down and captured, but Sippe and the third pilot returned safely. The raid was announced by Winston Churchill, then First Lord of the Admiralty, who called it "a fine feat of arms". One historian concluded: "The pilots deserve all praise for their admirable navigation... this flight of 250 miles, into gunfire, across enemy country, in the frail little Avro with its humble horse-power, can compare as an achievement with the best of them". The bullet-damaged tail of Sippe's (or possibly one of the other pilots') plane was later mounted on a plaque as a souvenir. Honours Sippe and the other returning pilot received the French Legion of Honour (rarely given to foreigners) immediately after the Friedrichshafen raid, at the request of General Joffre himself. Sippe was awarded the Distinguished Service Order in the 1915 New Year Honours, and the OBE in the 1919 New Year's Honours. He was also awarded the Croix de Guerre, and made a Chevalier of the Belgian Order of Leopold. Sippe was mentioned in dispatches six times during the war. Cigarette cards Sippe was featured on at least two cigarette cards: one, with his photograph and short biography, no. 26 in a series of 50 Naval Portraits, was issued by Lambert and Butler in c.1917. Another, no. 2 in a series of 50 War Incidents, was issued by Wills. Both cards recounted the Friedrichshafen raid. Family In December 1915, Sippe married Mabel Frances D'Arcy, only child of Gerald d'Arcy of Dublin; they had two children, Angela and James, and four grandchildren. They subsequently separated, and were divorced in 1933 on the grounds of her adultery with a Greek man. Post-war life After World War I, Sippe went through a variety of jobs, mostly related to engineering or aviation. In the early 1920s he trained pilots in Japan. In 1922 he began a secret attempt to salvage what was thought to be £2 million worth of gold coins (£100 million in 2012 prices) from the SS Tubantia, a ship which sank off the Dutch coast in 1916. When a rival salvage operation started the following year, Sippe won a landmark court case which ruled that, as his divers had found the ship and started work, he had the sole right to salvage it. In 1925, after £100,000 (£5.2 million in 2012 prices) had been spent on the salvage attempt, he concluded that though the cargo had been in the ship, the difficulty of accessing it made it too dangerous for divers to recover, and so the project was abandoned. Sippe was a sales manager for ten years with Short Brothers (now a large aerospace business), then with Crossley Motors, followed by the Fairey Aviation Company from 1946 to 1955. He later founded his own business, Field and Forest Supplies, selling products of his own invention. In 1963 he became a consultant to the 20th Century Joinery and Packing Co Ltd, a company which specialized in packing aircraft parts for transport. He died of cancer in 1968 in Leatherhead Hospital. Despite Sippe's distinguished war record, it seems no obituary appeared in any major newspapers, as they were not aware of his death at the time. References Category:1889 births Category:1968 deaths Category:British World War I pilots Category:British inventors Category:Chevaliers of the Légion d'honneur Category:People educated at Dulwich College Category:Companions of the Distinguished Service Order Category:Recipients of the Croix de Guerre 1914–1918 (France) Category:Royal Naval Air Service aviators Category:Royal Air Force officers Category:Officers of the Order of the British Empire Category:20th-century inventors
{ "pile_set_name": "Wikipedia (en)" }
Various apparatus have been devised for dry-docking boats. Pleasure boats of the kind to which the present invention is particularly adapted are thought of, for purpose of the present description, as being generally within the range of about fifteen feet to twenty-four feet in length and within the range of about 800 pounds to 6000 pounds in weight. Apparatus for docking pleasure boats has been associated with both non-floating as well as floating docks. The use of such apparatus is intended to facilitate removal of a boat from the water, maintaining the boat in a stored position elevated above the water, and for launching of the boat into the water. Among the many methods known and used for lifting boats for dry docking, floating lifts (such as set forth in U.S. Pat. No. 7,021,861, issued Apr. 4, 2006 to the inventor of the present invention) are known to have particular advantages. Secured or anchored in some fashion, such lifts can provide dry docking for boats in areas which were formerly open water. Floating upon the body of water, such lifts provide convenient dry docking at a constant level above the surface, in contrast with fixed lifts whose use may at times be rendered difficult or impractical due to varying seasonal depth of the body of water. Floating lifts can provide dry docking in bodies of water in which securing a fixed boat lift is difficult, as in deep muddy or sandy bottoms. Floating lifts may also be relocated with considerably greater ease than fixed boat lifts. An improved form of floating boat lift apparatus is illustrated in U.S. Pat. No. 6,602,022 to Wilkins, incorporated herein by reference, in which a tilting apparatus, installed in a floating boat well, receives a boat driven by its operator. The apparatus is then tilted about a pivot point, thereby elevating the boat in dry dock. In some embodiments, the lift mechanism, responsive to the change in weight distribution on the apparatus caused by the receipt of the boat, pivots on its own and elevates the boat. In other embodiments, after the boat has been received by the apparatus, it is necessary to use a winch or other mechanical assist to tilt the mechanism to elevate the boat. Advantageously over the prior art, Wilkins provides an apparatus that incorporates a single pivotal frame onto which a boat can be driven and positioned ready for storage using only the motive power of the boat itself and which can be pivoted so as to either elevate the boat above the water for storage or discharge the boat into the water for further use. However, the operation of Wilkins boat lift is subject to several shortcomings. First, the elevation of the pivot point relative to the boat well (and therefore its elevation relative to the surface of the water) is fixed in Wilkins lift. It is desirable to allow adjustment of the elevation of the pivot point relative to the surface of the water in order both to optimize the tilting operation of the lift and to provide adequate clearance of the boat from the water for dry dock purposes when the lift is in the elevated position. Second, no additional mechanical advantage is employed by Wilkins to augment the leverage provided by the change in weight distribution caused by receipt of the boat. Indeed, in embodiments elevating the boat above a relatively low height above the water (approximately six inches) or when elevating heavier boats, the Wilkins lift does not tilt on its own, instead requiring the use of the aforesaid winch or other mechanical means to effect tilting of the mechanism to elevate the boat. It is desirable to provide augmentation of the leverage provided by the weight of the boat, in order to allow the lift to be pivoted to the elevated position with greater ease and to provide for embodiments in which boats may be elevated above relatively low heights by operation of the apparatus on its own without the need for additional mechanical means to effect tilting. Third, Wilkins does not provide a damping functionality for his tilting apparatus: in embodiments relying on change in weight distribution to effect tilting, as the apparatus is tilted to elevate the boat and the center of gravity of the mechanism passes over-center of the pivot point, the lift apparatus settles to the elevated position in an abrupt fashion which can be jarring for the boat operator and tends to cause wear and fatigue of the lift mechanism. What is needed is an improved floating boat lift, having a pivoting lift mechanism onto which a boat can be driven and positioned ready for storage using only the motive power of the boat itself, but which also further provides mechanical advantage to augment the leverage provided by the change in weight distribution caused by receipt of the boat while providing a damping functionality to mitigate abrupt pivoting of the apparatus. What is needed further is such a floating boat lift that affords a means of adjusting the elevation of the lifting pivot point relative to the body of water.
{ "pile_set_name": "USPTO Backgrounds" }
NTAP NTAP is an acronym for nonprofit technology assistance provider. The term generally refers to organizations and individuals that specialize in providing information and communication technology support to nonprofit organizations, without regard for whether the provider itself is formally incorporated as a nonprofit entity or a for-profit business. Nonprofit technology assistance provider is distinguished from a "nonprofit management assistance provider." The latter focuses on building organizational capacity in all areas of nonprofit management, some of which may include technology assistance. Readers should also understand that the term "technical assistance" has historically covered any form of capacity building assistance, technological or otherwise. See also Circuit rider (Technology) Nonprofit technology The Nonprofit Technology Enterprise Network (NTEN) TechFoundation Category:Non-profit technology Nonprofit Technology News
{ "pile_set_name": "Wikipedia (en)" }
871 N.E.2d 376 (2007) Daniel E. HOAGLAND, Karen Hoagland, and Hoagland Family Limited Partnership, Appellants-Plaintiffs, v. TOWN OF CLEAR LAKE BOARD OF ZONING APPEALS, Appellee-Defendant. No. 76A03-0609-CV-444. Court of Appeals of Indiana. August 8, 2007. *377 J. Michael Loomis, Fort Wayne, IN, Attorney for Appellants. W. Erik Weber, Neal R. Blythe, Mefford and Weber, Auburn, IN, Attorneys for Appellee. OPINION SHARPNACK, Judge. Daniel and Karen Hoagland and the Hoagland Family Limited Partnership (collectively, the "Hoaglands") appeal the trial court's dismissal of their petition for judicial review regarding a decision by the Town of Clear Lake Board of Zoning Appeals ("BZA") and the trial court's denial of their motion to correct error. The Hoaglands raise four issues, which we consolidate *378 and restate as whether the trial court erred by dismissing the Hoaglands' petition for judicial review of a decision by the BZA based upon the Hoaglands' failure to provide the statutorily required notice. We affirm.[1] In this appeal, the Hoaglands challenge the trial court's dismissal of their petition for judicial review regarding an Improvement Location Permit ("ILP") issued concerning property owned by Jim and Cathlene Nevin. Today, we also issue a decision in Hoagland v. Town of Clear Lake Bd. of Zoning Appeals, ___ N.E.2d ___, No. 76A03-0610-CV-495, 2007 WL 2257214 (Ind.Ct.App.2007). In that appeal, the Hoaglands challenge the trial court's dismissal of their petition for a writ of certiorari. That appeal relates to the BZA's denial of their appeal regarding a second ILP issued to Steven Tagtmeyer to rebuild an existing garage and add a shed to property owned by the Nevins. The relevant facts concerning the first ILP follow. On September 7, 2004, the zoning inspector for the Town of Clear Lake, Indiana, issued an ILP to Jim and Cathlene Nevin for the remodeling of their residence at 1120 Lake Dr. at Clear Lake. On August 26, 2005, the Hoaglands, who are the Nevins' neighbors, appealed the issuance of the ILP to the BZA. A public hearing was held on October 26, 2005, regarding the Hoaglands' appeal. The hearing was continued on December 20, 2005, and the BZA apparently denied the Hoaglands' appeal at that hearing but did not issue written findings. On January 19, 2006, the Hoaglands filed a "Verified Petition for Judicial Review" of the BZA's denial of their appeal, but the Hoaglands did not serve notice upon the Nevins. The BZA responded on February 6, 2006, by filing a motion to dismiss because the Hoaglands cited the wrong statutory authority in their petition and filed a petition for judicial review rather than the required petition for writ of certiorari. At a hearing on the matter, the BZA also argued that the Hoaglands' petition should be dismissed because the Hoaglands did not serve notice of the petition upon the Nevins. On February 8, 2006, the Hoaglands filed a motion to amend their petition for judicial review and filed an "Amended Verified Petition and Request for Writ of Certiorari." Appellants' Appendix at 114, 117. After a hearing on the motion to dismiss, the trial court entered an order dismissing the Hoaglands' petition as follows: 1. On December 20, 2005 the BZA entered its Order approving the grant of [an] Improvement Location Permit requested by Jim and Cathlene Nevins ("Nevins"). Subsequent thereto, Hoagland sought review of the action taken by the BZA by filing on January 19, 2006 a Verified Petition for Judicial Review pursuant to Ind.Code 4-21.5-5-1, et seq. 2. At the time Hoagland [sic] the Verified Petition for Judicial Review the Nevins were not notified of the filing. 3. On February 7, 2006 the BZA filed its Motion to Dismiss for Lack of Jurisdiction. 4. On February 8, 2006 Hoagland filed a Motion to Amend Verified Petition for Judicial Review. *379 5. Therein, Hoagland sought leave of Court to change the format of the Petition originally filed on January 19, 2006 to a Petition for Writ of Certiorari, and be granted leave to provide notice as required by statute to all adverse parties. 6. Hoagland seeks review of an adverse decision made by the BZA. 7. The statutory procedure to seek review of such a decision is set forth at Ind.Code 36-7-4-1005 not Ind.Code 4-21.5-5-1, et seq. 8. In the case of Bagnall v. Town of Beverly Shores, 726 N.E.2d 782 (Ind. 2000), the Indiana Supreme Court at page 785 observed: "Decisions by boards of zoning appeals are subject to court review by certiorari. Ind.Code Sec. 36-7-4-1003(a) (Supp.1995). A person aggrieved by a decision of a board of zoning appeals may present to the circuit or superior court in the county in which the premises are located a verified petition setting forth that the decision is illegal, in whole or in part, and specifying the grounds of the illegality. Id. Sec. 36-7-4-1003(b). The petition must be presented to the court within 30 days of the board's decision. Id. The court does not gain jurisdiction over the petition until the petitioner serves notice upon all adverse parties as required by Ind.Code Sec. 36-7-4-1005(a) which provides in pertinent part: On filing a petition for a writ of certiorari with the clerk of the court, the petitioner shall have a notice served by the sheriff of the county on each adverse party, as shown by the record of the case in the office of the board of zoning appeals. . . . No other summons or notice is necessary when filing a petition. . . . The Code defines an adverse party as `any property owner whose interests are opposed to the petitioner for the writ of certiorari and who appeared at the hearing before the board of zoning appeals either in person or by a written remonstrance or other document that is part of the hearing record.' Ind. Code Sec. 36-7-4-1005(b). We read the language of statutes pursuant to the codified rules of statutory construction, which provide that `[w]ords and phrases shall be taken in their plain, or ordinary and usual, sense.' Ind.Code Sec. 1-1-4-1(1) (1998). As the trial court noted, "[t]he plain and ordinary meaning of the word `on' in the statute's phrase `on filing the petition' is taken to mean `at the time of `filing the petition." . . . To comply with the statute, a petitioner must file, with the clerk, notices to adverse parties contemporaneously to the filing of the writ petition. Because "strict compliance with the requirements of the statute governing appeals from decisions of boards of zoning appeals is necessary for the trial court to obtain jurisdiction over such cases," . . . (Emphasis Added) (Case Citations Omitted) (Citations to Record Omitted) 9. The [Nevins] are adverse parties to this proceeding. 10. The [Nevins] were not timely provided with any form of notice that Hoagland was seeking review of the decision made by the BZA on December 20, 2005. 11. To grant Hoagland the relief he has requested would require the Court to nullify the clear statutory *380 language regarding notice to adverse parties as set forth at Ind.Code 36-7-4-1005. A trial court quite simply does not gain jurisdiction over a Petition Seeking Review of an action taken by the BZA until the petitioner serves notice of the Petition upon all adverse parties as required by Ind. Code 36-7-4-1005. See, Citizens v. Brazil Bd. Of Zoning Appeals, 565 N.E.2d 380 (Ind.App.1991). IT IS THEREFORE, ORDERED, ADJUDGED AND DECREED as follows: 1. Hoagland's Motion to Amend Verified Petition for Judicial Review is denied. 2. BZA's Motion to Dismiss for Lack of Jurisdiction is granted. Appellants' Appendix at 135-138. The Hoaglands then filed a motion to correct error. The Hoaglands contended "that a final reviewable decision ha[d] not yet been made by the BZA sufficient to trigger the running of the thirty (30) day rule within which time an aggrieved person must file his Petition for Writ of Certiorari with the trial court. The reason being the BZA has not performed its statutory duties pursuant to Ind.Code 36-7-4-915 and Ind.Code 36-7-4-919(f)." Id. at 5. Relying upon Biggs v. Board of Zoning Appeals, 448 N.E.2d 693 (Ind.Ct.App. 1983), the trial court entered findings of fact and conclusions thereon in which it found that the "failure of the BZA to make written Findings of Fact does not toll the running of the thirty (30) days within which it is necessary for an aggrieved person to file a Writ of Certiorari in order to properly invoke the jurisdiction of the trial court when seeking judicial review." Id. at 7. Consequently, the trial court denied the Hoaglands' motion to correct error. On appeal, the issue is whether trial court erred by dismissing the Hoaglands' petition for judicial review of a decision by the BZA based upon the Hoaglands' failure to provide the statutorily required notice. The standard of appellate review of rulings on motions to dismiss on jurisdictional grounds depends on whether the trial court resolved disputed facts, and if so, whether the trial court conducted an evidentiary hearing or ruled on a paper record.[2]Wayne County Property Tax Assessment Bd. of Appeals v. United Ancient Order of Druids-Grove # 29, 847 N.E.2d 924, 926 (Ind.2006). We review de novo a ruling on a motion to dismiss for lack of jurisdiction if the facts are not disputed or, as here, the court rules on a paper record. Id. The Hoaglands argue that the trial court erred by: (1) determining that it did not have jurisdiction; (2) applying Biggs rather than Holmes v. Bd. of Zoning Appeals, 634 N.E.2d 522 (Ind.Ct.App. 1994); and (3) finding that the 30-day requirement within which to file a petition for a writ of certiorari and serve the adverse parties was triggered.[3] *381 Decisions by boards of zoning appeals are subject to court review by certiorari. Ind.Code § 36-7-4-1003(a). A person "aggrieved by a decision of a board of zoning appeals . . . may present to the circuit or superior court in the county in which the premises are located a verified petition setting forth that the decision is illegal in whole or in part and specifying the grounds of the illegality." I.C. § 36-7-4-1003(a). The petition must be presented to the court within 30 days of the board's decision. I.C. § 36-7-4-1003(b) & (c). Moreover, the petitioner must give notice of the petition as follows: * * * * * (1) If the petitioner is the applicant or petitioner for the use, special exception, or variance, the petitioner shall have a notice served by the sheriff of the county on each adverse party as shown by the record of the case in the office of the board of zoning appeals. (2) If the petitioner is not the applicant for the use, special exception, or variance and is a person aggrieved by the decision of a board of zoning appeals as set forth in section 1003 of this chapter, the petitioner shall have a notice served by the sheriff of the county on: (A) each applicant or petitioner for the use, special exception, or variance; and (B) each owner of the property that is the subject of the application or petition for the use, special exception, or variance. The service of the notice by the sheriff on the chairman or secretary of the board of zoning appeals constitutes notice of the filing of the petition to the board of zoning appeals, to the municipality or county, and to any municipal or county official or board charged with the enforcement of the zoning ordinance. No other summons or notice is necessary when filing a petition. (b) An adverse party under this section is any property owner whose interests are opposed to the petitioner for the writ of certiorari and who appeared at the hearing before the board of zoning appeals either in person or by a written remonstrance or other document that is part of the hearing record. If the petitioner was an unsuccessful appellant in the administrative appeal, or an unsuccessful petitioner or applicant for a variance, special exception, or special or conditional use, and the record shows a written remonstrance or other document opposing the interest of the petitioner that contains more than three (3) names, the petitioner shall have notice served on the three (3) property owners whose names appear first on the remonstrance or document. Notice to the other persons named is not required. (c) Notice given under subsection (a) must state: (1) that a petition for a writ of certiorari, asking for a review of the decision of the board of zoning appeals, has been filed in the court; (2) the premises affected; and (3) the date of the decision. (d) An adverse party who is entitled to notice of a petition for writ of certiorari under subsection (a) is not required to be named as a party to the petition for writ of certiorari. Ind.Code § 36-7-4-1005. The Indiana Supreme Court has held that "[t]he court does not gain jurisdiction *382 over the petition until the petitioner serves notice upon all adverse parties" as required by Ind.Code § 36-7-4-1005. Bagnall v. Town of Beverly Shores, 726 N.E.2d 782, 785 (Ind.2000). In Bagnall, the court noted: "To comply with the statute, a petitioner must file, with the clerk, notices to adverse parties contemporaneously to the filing of the writ petition." Id. "[S]trict compliance with the requirements of the statute governing appeals from decisions of boards of zoning appeals is necessary for the trial court to obtain jurisdiction over such cases."[4]Id. Here, the BZA held a hearing on the Hoagland's appeal on December 20, 2005. The Hoaglands filed their "Verified Petition for Judicial Review" on January 19, 2006. The Hoaglands do not dispute that they did not serve the Nevins at that time. Further, the Hoaglands did not argue to the trial court and did not argue in their appellant's brief that notice to the Nevins was not required under Ind.Code § 36-7-4-1005.[5] After the BZA filed a motion to dismiss, on February 8, 2006, the Hoaglands filed a motion to amend their petition for judicial review and an "Amended Verified Petition and Request for Writ of Certiorari." Appellants' Appendix at 114, 117. The CCS indicates that the petition for writ of certiorari was served on the Nevins on February 13, 2006. If the BZA's decision was issued on December 20, 2005, pursuant to Bagnall, the Hoaglands were required to file a petition for writ of certiorari and serve notice of the petition upon the Nevins within thirty days. However, notice was not served upon the Nevins until February 13, 2006. The Hoaglands attempt to argue *383 that the BZA's decision was not issued on December 20, 2005, because the BZA failed to issue written findings of fact as required by Ind.Code § 36-7-4-1003. However, this same argument was addressed in Biggs v. Bd. of Zoning Appeals of City of Wabash, 448 N.E.2d 693 (Ind.Ct. App.1983). There, the petitioner's request for a variance was denied at a meeting on September 17, 1981. 448 N.E.2d at 694. The minutes of the September meeting were approved by the Board at their October 15, 1981, meeting. Id. The petitioner filed a petition for writ of certiorari on November 13, 1981. Id. The trial court then dismissed the petition because the petitioner failed to file the petition within thirty days of the Board's decision. Id. This court held that the date of the Board's decision was September 17, 1981, even though written findings of fact were not issued on that date. Id. Thus, the November 13th petition was not timely, and "[f]ailure to comply with the statute is fatal." Id. Even though the BZA in this case did not issue written findings of fact, the Hoaglands were aware of the BZA's December 20, 2005, decision. See Appellants' Appendix at 11-25 (Hoagland's Petition for Judicial Review, which alleges that the BZA rendered a "final decision" on December 20, 2005). As a result, under Biggs, the Hoaglands were required to file a petition for writ of certiorari and serve the required notices within thirty days of the BZA's December 20, 2005, decision, and they failed to do so.[6] Their failure to comply with the statute is fatal, and the trial court did not err by dismissing the Hoaglands' petition.[7]See, e.g., Bagnall, 726 N.E.2d at 785 (holding that the petitioners did not "secure jurisdiction" where they failed to serve notice of the petition upon adverse parties); Town of Cedar Lake Bd. of Zoning Appeals v. Vellegas, 853 N.E.2d 123, 126-127 (Ind.Ct.App.2006) (holding that the trial court did not have jurisdiction to consider a petition for writ of certiorari where the petitioner failed to serve notice upon an adverse party). *384 For the foregoing reasons, we affirm the trial court's dismissal of the Hoaglands' petition for judicial review. Affirmed. MAY, J., and BAILEY, J., concur. NOTES [1] The BZA filed a Request for Leave to Supplement Record to include the transcript of the October 26, 2005, BZA hearing, exhibits submitted at the hearing, and a document submitted by the Hoaglands at the December 20, 2005, hearing. The supplemental record is unnecessary given our resolution of the issue in this case. Consequently, we deny the BZA's motion. [2] In United Ancient Order of Druids-Grove, the Indiana Supreme Court held that such an issue "is properly raised by means of a motion under Rule 12(B)(1) for lack of jurisdiction or 12(B)(6) for failure to state a claim, depending on whether the claimed defect is apparent on the face of the petition." Wayne County Property Tax Assessment Bd. of Appeals v. United Ancient Order of Druids-Grove # 29, 847 N.E.2d 924, 926 (Ind.2006). [3] The Hoaglands also argue that the trial court should have treated the BZA's motion to dismiss as a motion for summary judgment. However, the Hoaglands did not raise this issue to the trial court. "A party generally waives appellate review of an issue or argument unless the party raised that issue or argument before the trial court." GKC Indiana Theatres, Inc. v. Elk Retail Investors, LLC., 764 N.E.2d 647, 652 (Ind.Ct.App.2002). Consequently, the Hoaglands waived this argument. [4] We note that the Indiana Supreme Court clarified jurisdiction concepts in K.S. v. State, 849 N.E.2d 538 (Ind.2006). The court held: Like the rest of the nation's courts, Indiana trial courts possess two kinds of "jurisdiction." Subject matter jurisdiction is the power to hear and determine cases of the general class to which any particular proceeding belongs. Personal jurisdiction requires that appropriate process be effected over the parties. Where these two exist, a court's decision may be set aside for legal error only through direct appeal and not through collateral attack. Other phrases recently common to Indiana practice, like "jurisdiction over a particular case," confuse actual jurisdiction with legal error, and we will be better off ceasing such characterizations. 849 N.E.2d at 540. The notice requirements in Ind.Code § 36-7-4-1005 implicate what courts have previously called "jurisdiction over a particular case" rather than subject matter or personal jurisdiction. See, e.g., Packard v. Shoopman, 852 N.E.2d 927 (Ind. 2006) (holding that statutory deadlines for filing a petition for judicial review in the Tax Court referred to what was previously known as "jurisdiction over a particular case"); United Ancient Order of Druids-Grove # 29, 847 N.E.2d at 926 ("The timing of filing the agency record implicates neither the subject matter jurisdiction of the Tax Court nor personal jurisdiction over the parties. Rather, it is jurisdictional only in the sense that it is a statutory prerequisite to the docketing of an appeal in the Tax Court."). Consequently, although the court in Bagnall referred to this concept as jurisdiction over the case, we will not do so. [5] The Hoaglands argue for the first time in their reply brief that notice to the Nevins was not required under Ind.Code § 36-7-4-1005. The Hoaglands did not make this argument to the trial court and did not raise the issue in their appellants' brief. "A party generally waives appellate review of an issue or argument unless the party raised that issue or argument before the trial court." GKC Indiana Theatres, Inc., 764 N.E.2d at 652. Moreover, "[t]he law is well settled that grounds for error may only be framed in an appellant's initial brief and if addressed for the first time in the reply brief, they are waived." Monroe Guar. Ins. Co. v. Magwerks Corp., 829 N.E.2d 968, 977 (Ind.2005). Consequently, the Hoaglands waived this issue. [6] The Hoaglands argue that the trial court should have relied upon Holmes v. Bd. of Zoning Appeals, 634 N.E.2d 522 (Ind.Ct.App. 1994), rather than Biggs. In Holmes, the BZA failed to make written findings, and this court held that "[t]he proper procedure by the trial court is to remand to the BZA for specific findings of fact." 634 N.E.2d at 525. However, in Holmes, there was no dispute regarding the timeliness of the petitioner's petition for writ of certiorari or notice to adverse parties. Here, the trial court could not remand to the BZA for findings because the Hoaglands failed to meet the statutory requirements regarding the notice to adverse parties. [7] The Hoaglands also argue that the trial court had "jurisdiction" because at least one of the issues presented fell within the trial court's "jurisdiction." Appellants' Brief at 15-16 (relying upon M.C. Welding & Machining Co., Inc. v. Kotwa, 845 N.E.2d 188 (Ind. Ct.App.2006), and Austin Lakes Joint Venture v. Avon Util., Inc., 648 N.E.2d 641, 644 (Ind. 1995)). Each of the allegations were presented as part of the Hoaglands' request for judicial review of the BZA's action. The Hoaglands presented no separate issue against the BZA for the trial court's review. In fact, the relief requested in the petition related only to the BZA's denial of the Hoaglands' appeal. We conclude that the Hoaglands presented no separate issues for the trial court's review, and the Hoaglands' failure to follow the statutory notice requirements was fatal to their claims. See, e.g., Johnson Oil Co., Inc. v. Area Plan Comm'n of Evansville and Vanderburgh County, 715 N.E.2d 1011, 1013-1014 (Ind.Ct. App.1999) (rejecting the petitioner's argument that the trial court had jurisdiction over its civil rights claims, inverse condemnation claim, and estoppel claim and that the trial court erred by dismissing its complaint "based upon [the court's] [apparent] belief that Johnson Oil had failed to exhaust its administrative remedies").
{ "pile_set_name": "FreeLaw" }
Q: Capitalizing a lower case screen name at the beginning of a sentence When starting a sentence with a lower case pseudonym, such as a screen name of a user account on a website, should it be capitalized? Or are there different cases where it would and would not be appropriate to do so? My curiosity was sparked by this meta EL&U page where a user comments that he does not mind if his name is lower-cased or not. Are there any established standards on this practice? A: When writing professionally, the first letter in the sentence is capitalized, sole exceptions being when the capitalization could result in a misunderstanding. In such cases, the word is usually typeset differently. However, when writing in an informal context, you might want to take into consideration the preference of the user. For example, Randall Munroe prefers his username xkcd to remain lowercase; however, as you've linked, some users like nohat do not mind it being capitalized. If you are unsure, I would suggest capitalizing it and adhering to professional writing style.
{ "pile_set_name": "StackExchange" }
0181385 to 3 dps. -0.102 Round -934.8294 to the nearest one hundred. -900 What is 0.0149856 rounded to 4 dps? 0.015 What is 984.443 rounded to the nearest integer? 984 Round -86981.23 to the nearest 1000. -87000 What is -7.218521 rounded to 1 dp? -7.2 Round 1353.656 to the nearest one hundred. 1400 What is 0.3993051 rounded to three decimal places? 0.399 Round -6.2066 to the nearest 10. -10 Round -9.87127 to 2 dps. -9.87 Round 10.259205 to zero decimal places. 10 What is 0.02543201 rounded to 3 decimal places? 0.025 Round -22.3042 to zero decimal places. -22 What is 0.0001569893 rounded to six dps? 0.000157 Round 2.4354309 to 2 dps. 2.44 What is -0.02023282 rounded to 5 dps? -0.02023 Round -0.00016758301 to 5 decimal places. -0.00017 Round -12600.8 to the nearest 1000. -13000 Round -0.007074456 to three decimal places. -0.007 Round -0.001253595 to five decimal places. -0.00125 What is 0.1193138 rounded to two decimal places? 0.12 What is -0.000000208553 rounded to 7 decimal places? -0.0000002 Round 723852 to the nearest 100000. 700000 What is 658682 rounded to the nearest one thousand? 659000 Round 0.0001410821 to 5 decimal places. 0.00014 What is 38.945384 rounded to two decimal places? 38.95 Round 0.000225105 to four dps. 0.0002 Round -51.6052 to the nearest ten. -50 What is 392114 rounded to the nearest 1000? 392000 What is -0.25156011 rounded to three decimal places? -0.252 What is -0.003965447 rounded to four dps? -0.004 Round 39800.27 to the nearest 100000. 0 What is -0.01928969 rounded to 2 dps? -0.02 Round -6010.639 to zero dps. -6011 What is 0.004593111 rounded to 3 decimal places? 0.005 Round 2191.176 to zero decimal places. 2191 Round -0.20395523 to 4 dps. -0.204 What is -0.1506579 rounded to 2 decimal places? -0.15 What is 0.8403108 rounded to 2 dps? 0.84 Round -32.3309 to the nearest 10. -30 What is -124.7171 rounded to 1 decimal place? -124.7 What is -7248491.2 rounded to the nearest one million? -7000000 Round -0.00129617 to six decimal places. -0.001296 What is 0.0000029199 rounded to 6 dps? 0.000003 What is -1429.319 rounded to the nearest 100? -1400 What is 1763.952 rounded to the nearest integer? 1764 What is 0.0000166298 rounded to 5 dps? 0.00002 Round 0.000086457588 to five decimal places. 0.00009 What is -1.835324 rounded to the nearest integer? -2 What is 0.0000127012 rounded to 5 decimal places? 0.00001 Round -0.004671615 to 6 dps. -0.004672 Round 1137.483 to the nearest 10. 1140 What is -0.058949 rounded to 1 decimal place? -0.1 What is -0.000185077 rounded to 5 dps? -0.00019 Round 56000.71 to the nearest one hundred. 56000 Round -4495300 to the nearest one million. -4000000 What is -0.03964646 rounded to 5 decimal places? -0.03965 Round -0.00004275612 to six dps. -0.000043 Round -0.004677052 to four decimal places. -0.0047 What is -581198 rounded to the nearest ten thousand? -580000 What is 0.00889614 rounded to five dps? 0.0089 Round 0.0000303083 to 6 decimal places. 0.00003 What is 15240976 rounded to the nearest ten thousand? 15240000 What is 225484.6 rounded to the nearest one thousand? 225000 Round 0.1710254 to four dps. 0.171 What is 888.116 rounded to the nearest one hundred? 900 Round 0.00005199651 to 5 dps. 0.00005 What is -0.052074 rounded to two decimal places? -0.05 What is -742.6 rounded to the nearest ten thousand? 0 What is 0.061037 rounded to three dps? 0.061 What is 743991.3 rounded to the nearest 100000? 700000 What is -0.009457 rounded to three decimal places? -0.009 What is 0.000000328306 rounded to 6 decimal places? 0 Round 0.006980503 to five dps. 0.00698 What is -0.0000010907014 rounded to seven decimal places? -0.0000011 Round -0.0000282862 to 5 dps. -0.00003 What is -0.000001054324 rounded to seven decimal places? -0.0000011 What is -134.0417 rounded to the nearest 10? -130 Round -0.00163814 to 5 dps. -0.00164 What is 0.02850544 rounded to five dps? 0.02851 Round -3352.216 to the nearest ten. -3350 Round 2.8294977 to two decimal places. 2.83 Round 2.1362 to 2 dps. 2.14 What is -384.19927 rounded to the nearest 10? -380 What is -3092271 rounded to the nearest one hundred thousand? -3100000 Round -0.0027223118 to five decimal places. -0.00272 What is -3578.32 rounded to the nearest ten? -3580 What is -0.00000924937 rounded to 6 dps? -0.000009 What is 0.0000862334 rounded to 6 dps? 0.000086 What is 0.00011538731 rounded to five dps? 0.00012 What is -0.00180412 rounded to five dps? -0.0018 What is -0.0000077911 rounded to six dps? -0.000008 What is 18.428027 rounded to the nearest integer? 18 What is 166663966 rounded to the nearest 100000? 166700000 Round -0.00007842625 to five decimal places. -0.00008 Round -0.0004774318 to five dps. -0.00048 Round -511.81 to the nearest one hundred. -500 What is -321836.5 rounded to the nearest one hundred thousand? -300000 Round 0.0000056071 to 6 decimal places. 0.000006 Round -0.00004264965 to six dps. -0.000043 Round -0.00366226 to 4 decimal places. -0.0037 What is -1834.503 rounded to the nearest 100? -1800 What is -0.00510382 rounded to five decimal places? -0.0051 What is 0.01330344 rounded to three dps? 0.013 What is 0.0030318 rounded to 5 dps? 0.00303 Round -0.027644 to four decimal places. -0.0276 What is -71689600 rounded to the nearest one hundred thousand? -71700000 Round -0.0000061246438 to 6 dps. -0.000006 Round -0.005295721 to six decimal places. -0.005296 Round -20867.8 to the nearest one thousand. -21000 What is -2.543709 rounded to 1 decimal place? -2.5 What is -0.023368381 rounded to three decimal places? -0.023 What is 0.0521857 rounded to three dps? 0.052 Round 0.0378042 to four decimal places. 0.0378 What is -30.535281 rounded to the nearest 10? -30 What is 143.6766 rounded to the nearest ten? 140 What is -0.048055314 rounded to 3 decimal places? -0.048 Round -0.1265582 to four dps. -0.1266 Round 40.524089 to 1 dp. 40.5 Round -0.002443555 to four dps. -0.0024 What is 0.0002285583 rounded to 6 decimal places? 0.000229 Round 1405.224 to the nearest integer. 1405 Round -0.007843536 to five decimal places. -0.00784 What is -760832000 rounded to the nearest one million? -761000000 Round -0.848569 to three decimal places. -0.849 Round -77735.63 to the nearest 1000. -78000 Round -0.00000039204298 to seven decimal places. -0.0000004 Round 3.2145 to the nearest ten. 0 Round -718.98318 to one dp. -719 Round -0.000866085 to 7 decimal places. -0.0008661 What is -0.0293098 rounded to five dps? -0.02931 Round -320.649 to the nearest integer. -321 What is 0.49038053 rounded to three dps? 0.49 What is -0.597162 rounded to 1 dp? -0.6 Round -340.521 to the nearest 100. -300 What is -1.5876993 rounded to three dps? -1.588 Round 0.14899103 to 4 dps. 0.149 What is 0.0004261897 rounded to four dps? 0.0004 What is -106.7068 rounded to the nearest 10? -110 Round 30790350 to the nearest 10000. 30790000 Round -42.34787 to 1 decimal place. -42.3 Round -2097.9754 to the nearest 100. -2100 What is 3727964 rounded to the nearest one million? 4000000 Round 2676024.5 to the nearest 100000. 2700000 Round -10478336900 to the nearest one million. -10478000000 Round -0.000001355264 to 7 dps. -0.0000014 Round -104803 to the nearest 10000. -100000 What is -0.0000436505 rounded to six decimal places? -0.000044 What is -782.29 rounded to the nearest one hundred? -800 What is -1556 rounded to the nearest 1000? -2000 What is 405795 rounded to the nearest one hundred thousand? 400000 Round -0.0326984 to two dps. -0.03 Round -1330560 to the nearest 10000. -1330000 Round -0.00000103780632 to seven dps. -0.000001 What is 8262.78 rounded to the nearest one hundred? 8300 What is 0.0000590768 rounded to 5 decimal places? 0.00006 What is -36432 rounded to the nearest one hundred thousand? 0 What is -36095520 rounded to the nearest 10000? -36100000 Round -0.000155593 to 5 dps. -0.00016 Round 155542060 to the nearest 1000000. 156000000 What is 45.72785 rounded to the nearest integer? 46 Round 415.115 to the nearest 100. 400 Round -0.0008879464 to 7 decimal places. -0.0008879 What is -38.1514 rounded to 1 decimal place? -38.2 Round 9.5516 to the nearest 100. 0 What is -46244200 rounded to the nearest 100000? -46200000 What is -4.0173 rounded to one decimal place? -4 What is 0.0000074748
{ "pile_set_name": "DM Mathematics" }
Trying to find Dale Hughes's email, wiki, phone numbers, twitter, biography, and linkedin data? You can check out HPCCOM LLC's profile on Lead411, where you can also get @hpccom.com contact information. HPCCOM LLC is an organization centered in Wilhoit, AZ, which you can find on Lead411 under the Web development category. Dale Hughes is their President. Some possible email formats for Dale Hughes are DHughes@hpccom.com, Dale.Hughes@hpccom.com, Dale@hpccom.com, and Dale_Hughes@hpccom.com. If you sign up for our free trial you will see our email@hpccom.com addresses. Does HPCCOM LLC appear to be financially stable? GREAT JOB! Poll Question Date Created Current Result Votes Similar People: International Coal Group Inc.'s Director, Organizational Development is where Dale Harper is employed. On the other hand, the organization's CEO is . They are based in Teays Valley, WV, and you can find their Lead411 profile filed under the Other industry. Dale Harper's profile contains twitter, phone numbers, linkedin, wiki, and contact information, and you can also find other International Coal Group Inc. email addresses on Lead411 with the @intlcoal.com domain. Some possible email formats for Dale Harper are DHarper@intlcoal.com, Dale.Harper@intlcoal.com, Dale@intlcoal.com, and Dale_Harper@intlcoal.com. If you sign up for our free trial you will see our email@intlcoal.com addresses. For Petro-Hunt L.L.C email addresses with the @petro-hunt.com format as well as facebook info, twitter, phone numbers, linkedin, wiki, as well as biography data for Dale Hostenske, you may check in Petro-Hunt L.L.C's Lead411 profile. Dale Hostenske is currently the Manager for Petro-Hunt L.L.C. The company's CEO is . The contact information of Petro-Hunt L.L.C on Lead411 is filed under the Oil, gas, coal. Their main offices are in Dallas, TX. Some possible email formats for Dale Hostenske are DHostenske@petro-hunt.com, Dale.Hostenske@petro-hunt.com, Dale@petro-hunt.com, and Dale_Hostenske@petro-hunt.com. If you sign up for our free trial you will see our email@petro-hunt.com addresses. Their Principal, Wildlife Biologist is Dale Herter. Their Lead411 profile is categorized under the Other biz services industry. If you're checking for Raedeke Associates, Inc email addresses, these are also available on Lead411 with the @raedeke.com email addresses and possibly Dale Herter's email. Raedeke Associates, Inc is based in Seattle, WA. You can also get Dale Herter's linkedin info, twitter data, phone numbers, wiki and biography on their Lead411 profile. Their contact information is filed under the Other biz services category. Some possible email formats for Dale Herter are DHerter@raedeke.com, Dale.Herter@raedeke.com, Dale@raedeke.com, and Dale_Herter@raedeke.com. If you sign up for our free trial you will see our email@raedeke.com addresses.
{ "pile_set_name": "Pile-CC" }
This is a response to Notice Number NOT-OD-09-058, NIH Announces the Availability of Recovery Act Funds for Competitive Revision Applications. The goal of this revision supplement is to extend our studies of nicotine vaccine efficacy in rats by introducing methods for the delivery of nicotine through inhalation of cigarette smoke. Smoking cessation medications have added substantially to our ability to treat tobacco addiction, but their efficacy is limited and new types of medications are needed. Nicotine vaccines elicit nicotine-specific antibodies which bind nicotine and alter its access to brain. Three nicotine vaccines have shown preliminary efficacy in Phase I-II clinical trials, but efficacy is closely correlated with the serum antibody titer and current vaccines do not reliably produce sufficiently high titers in all individuals. The parent grant DA10714 is using rat models of nicotine addiction to study novel means of enhancing vaccine efficacy. Like essentially all current animal studies of nicotine addiction, DA10714 models tobacco addiction using parenteral (i.v. or s.c.) administration of pure nicotine. In contrast, smokers take in nicotine by inhalation and as one of over 4,000 chemicals present in cigarette smoke. The adequacy of using such artificial dosing paradigms to model cigarette smoking is unknown and largely untested. We have adapted and characterized methods for inhalation exposure of rats to smoke simulating the smoking of 1 cigarette or periods of heavier smoking. In this revision supplement we propose to use these methods to study the effects of nicotine vaccines on the absorption and distribution of nicotine inhaled in cigarette smoke. The purposes of doing so are to 1) expand the range of preclinical models available to study nicotine vaccines, 2) assess whether inhalation models provide novel information for vaccine evaluation, 3) examine the specific role of route-specific factors such as pulmonary antibody in mediating nicotine vaccine efficacy, and 3) develop quantitative models which can be more generally used to study the contributions of the inhaled route and other smoke constituents to tobacco addiction and treatment medications development. Nicotine vaccines are an attractive initial candidate for such study because vaccination is a pharmacokinetic intervention, and accurate pharmacokinetic modeling of nicotine intake may be important in understanding and exploiting its efficacy. Aim 1 will test the hypothesis that vaccination is effective in reducing nicotine distribution to brain over a range of clinically relevant dosing conditions. Aim 2 will test the hypotheses that vaccination is more effective in reducing the distribution to brain of inhaled compared to i.v. nicotine, and that such differences are in part mediated by the presence of pulmonary mucosal or tissue antibody. Because heroin and cocaine are also often smoked, the results of this study may inform ongoing efforts to develop vaccines for these addictions as well. PUBLIC HEALTH RELEVANCE: Cigarette smoking kills 5 million people worldwide yearly. Current medications are helpful for smoking cessation but are incompletely effective. We are studying the use of a nicotine vaccine to help smokers quit, which acts by binding nicotine in blood and reducing its access to brain. Rat models of tobacco addiction use nicotine delivered intravenously, which is informative but does not accurately model the human route of intake of nicotine by inhalation from cigarette smoke. This proposal will develop and study the utility of delivering nicotine to rats via exposure to cigarette smoke, and assess whether it allows better evaluation of nicotine vaccine efficacy and can expedite its further development.
{ "pile_set_name": "NIH ExPorter" }
Q: Laravel - Disable Updated At when updating Get a problem with update using query builder on laravel 5. I've tried to disabled the updated_at but keep failing. Here is my code: $query = StockLog::where('stock_id', $id)->whereBetween('created_at', $from, $to])->update(['batch_id' => $max + 1]); I've tried 2 ways: first one at my model i set: public function setUpdatedAtAttribute($value) { /*do nothing*/ } Second one: $stocklog = new StockLog; $stocklog->timestamps = false; $query = $stocklog::where('stock_id', $id)->whereBetween('created_at', [$from, $to])->update([ 'batch_id' => $max + 1]); both of them are failed. is there anyway to disabled the updated_at? Thanks in advance A: By default, Eloquent will maintain the created_at and updated_at columns on your database table automatically. Simply add these timestamp columns to your table and Eloquent will take care of the rest. I don't really suggest removing them. But if you want use the following way. add the following to your model: public $timestamps = false; This will disable the timestamps. EDIT: it looks like you want to keep the created_at field, you can override the getUpdatedAtColumn in your model. Use the following code: public function getUpdatedAtColumn() { return null; } A: In your model, add this method: /** * @param mixed $value * @return $this */ public function setUpdatedAt($value) { return $this; } UPDATE: In Laravel 5.5: Just try to use this in your model: const CREATED_AT = null; const UPDATED_AT = null; A: The accepted answer didn't work for me, but led me in the right direction to this solution that did: class Whatever extends Model { //... const UPDATED_AT=NULL; //... Laravel 5.3
{ "pile_set_name": "StackExchange" }
/* Copyright The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ // Code generated by client-gen. DO NOT EDIT. // This package has the automatically generated typed clients. package v1
{ "pile_set_name": "Github" }
Late last week, the US named and shamed its list of the worst offenders when it comes to protecting intellectual property; China, Russia, Argentina, Chile, India, Israel, Pakistan, Thailand, and Venezuela topped the "Special 301 Report" and earned the dubious distinction of being placed on a Priority Watch List (The Pirate Bay was also singled out for special mention). Together with the countries on the regular Watch List, the report called out a grand total of 46 countries, and the reaction has been fast and furious. These Special 301 reports have been generated since the 1970s, but IP law didn't enter mainstream consciousness until the last decade or so. Now that topics like copyright and patent reform can make for front-page news, more people are complaining about the 301 process. An editorial running tomorrow morning (thanks to the miracle of time zones) in the Bangkok Post, for instance, takes on the US report, saying that the US government "has escalated its imagined dispute with Thailand far out of proportion." The US continues to object to Thailand's practices surrounding generic copies of patented drugs, but the paper argues that "drug licensing is specifically legal in a way that open sales of songs and movies are not," and it feels like Thailand was unfairly singled out in part for trying to contain a public health crisis. Taiwan, which was also named in the report, has "expressed regret" about being included in the list, according to the China Post. It was applauded for new law "aimed at ending illegal file-sharing over peer-to-peer (P2P) platforms," but the US wants to see a new law "regarding liability of Internet service providers for copyright infringements." It also wants a stricter crackdown on copyrighted material flowing through TANet, an ISP run by the country's Ministry of Education. Canadian law professor Michael Geist takes his own potshots at the 301 process, which continues to place Canada on the Watch List, and he points out that Canada's own Department of Foreign Affairs doesn't think much of Special 301. "In regard to the watch list," said a Foreign Affairs official to the House of Commons last year, "Canada does not recognize the 301 watch list process. It basically lacks reliable and objective analysis. It's driven entirely by U.S. industry. We have repeatedly raised this issue of the lack of objective analysis in the 301 watch list process with our US counterparts. I also recognize that the US industry likes to compare anyone they have a problem with, concerning their IPR regime, to China and the other big violators, but we're not on the same scale." And Israel, which made the top nine, was so upset about the possibility of remaining on the Priority Watch List that it filed an angry response with the US Trade Representative back in March, pointing out that calls for more DRM from US industries weren't required by any treaties that Israel had signed and that it saw significant problems with such measures. "The critiques and criticism of TPM [technological protection measures] both from business model perspectives and from copyright perspectives are almost endless," said the Israeli response, which wasn't enough to get the country off the hook. One of the criticisms of the 301 process is that it is driven too heavily by American business interests; indeed, the 301 Report lists "affected industry groups and other private-sector representatives" as a major source of its information. The International Intellectual Property Alliance (IIPA), which represents the BSA, ESA, MPAA, RIAA, and other major copyright business groups, has also responded to the 301 Report. Not surprisingly, it raised no significant objections and called the process "an important tool by which the US government has been able to secure improved protection and enforcement in our key markets around the world." But not even the IIPA was fully satisfied by the 46 country list; it actually called on Canada to be upgraded to the Priority Watch List due to its failure to "facilitate the development of a healthy online marketplace for copyright materials." Given the furious response to a proposed new copyright reform law in Canada that appeared ready to fill Big Content's Christmas stocking, the legislation that the IIPA wants to see may not be forthcoming soon... if it ever sees the light of day at all. The IIPA, which does much to set the agenda of the 301 process, is coming in for criticism here in the US as well. William Patry, a top copyright scholar (and Google's lead copyright attorney), said last month that "the sheer arrogance and affront to the sovereignty of foreign governments by the IIPA's annual reports and effort to penalize those governments that do not toe the IIPA's line is breathtaking." But it doesn't look like the IIPA has any plans to scale back on the rhetoric, which at the moment seems targeted especially at our neighbor to the north.
{ "pile_set_name": "OpenWebText2" }
1. Field of the Invention This invention relates to a DC high voltage supply system designed to supply a highly efficient and stable DC voltage to a load by providing a high voltage transformer circuit with a feedback control circuit, which transformer circuit is made up, in combination, of a variable magnetic leak transformer (VLT) driven under resonant conditions and a capacitor type multiple boosting circuit, and by controlling the current of the control winding of the VLT by the feedback control circuit. 2. Prior Art In recent years a combination of a boosting transformer and a capacitor type multiple boosting circuit represented by a Cockcroft-Walton circuit has been in wide use so as to reduce the size and weight of a DC power source unit used for supplying DC high voltage power to an X-ray diagnostic apparatus, an electron microscope, sputtering apparatus and which constitutes a greater part of the volume and weight of both the boosting transformer and the multiple boosting circuit instead of the conventional method in which commercial frequency is boosted by a high voltage transformer and is then rectified. But according to this combined use of the transformer and the circuit, it often happened that, viewed from the principle of the capacitor type multiple boosting circuit, the diodes used in the capacitor type multiple boosting circuit are damaged by the surge current generated during a sudden increase in load or during initial charging of the capacitor multiple boosting circuit. Also, it is possible to think of using the capacitor type multiple boosting circuit by connecting large resistors in series with load for preventing the surge current generated in this case; but this way of tackling the problem causes a decrease in voltage regulation and an increase in power loss, with the result that efficiency of the system is greatly reduced. Furthermore, slowing charging is also employed by providing a soft start circuit for preventing surge current when the power is on but it inevitably results in increased cost of the system because of the circuit being complicated.
{ "pile_set_name": "USPTO Backgrounds" }
Check out our new site Makeup Addiction Ate at a friends house Helped with the dishes
{ "pile_set_name": "OpenWebText2" }
Is an ecommerce system required? Are you looking to only take Paypal, for example? Is the site to be content managed? Do you need someone to design the site visually, rather than just code it? Do you have a budget in mind?
{ "pile_set_name": "Pile-CC" }
Tag: backlinks The importance of link building should never be underestimated. Good link building is the cornerstone of solid off-page SEO which can yield those coveted rankings and free organic traffic. Building backlinks takes time and requires patience. Making sure you maximize your efforts is vital to ensure your rewarded handsomely for your efforts. What is the Best Link Building Method Anyway? There are many different ways to build backlinks. Article directories, forum postings, press release submissions, article marketing and social bookmarking are just a few of the common methods used by internet marketers. My favorite way to build high quality backlinks is by producing unique content and submitting it to various sources that are willing to accept it in return for contextual or semi-contextual backlinks. This is by far the best link building strategy. Writing content and submitting it to relevant and good sources. You can write the content yourself or outsource it. If you need more link building power, you have to guest post… Finding blogs that relate to your content is step number one. You can search for blogs that accept guest posts in your niche by doing some smart Googling. Some blogs make their guest posting policies very clear and others don’t. Taking a few minutes to contact blog owners is worth the effort if your email is straightforward and to the point. Forget about sending a generic spammy message – blog owners get those every day from pro link builders and SEO firms asking for guest posts. Get personal and let them know about how good your articles are and how they can truly provide value to their readers. Keep it real and forthcoming and you will slowly build a list of solid blogs that are more than happy to accept your blog posts or buy seo services . Submitting articles with in-content or author bio links will greatly help your SEO efforts. Mostly because, these blogs aren’t part of link building networks or in the business of providing links to everyone via guest posting. They offer value that just can’t compare to low-quality link blogs. Good link building is about producing a wide variety of backlinks on all kind of different sources. Not just places that everyone else submits to. Finding those guest posting gems can dramatically impact your SEO efforts in ways other link building strategies simply can’t. There’s nothing quite like natural backlinks surrounded by unique relevant content on a high quality page that has some authority in your industry. That right there is the best way to build backlinks for SEO. As you can see from the examples above, using quality web content to produce backlinks will give you the most mileage when it comes to SEO and your keyword rankings. Why is the quality of content so important? Well – if your content isn’t valuable in some way – most reputable blog owners and even premium guest posting sources won’t accept them. Getting links on those really awesome sites will require good content. This might require more work but the resulting backlinks are worth every ounce.
{ "pile_set_name": "Pile-CC" }
1. Field of the Invention The present invention relates to the field of sonar signal processing and more particularly, to detecting the presence or absence of spatial random processes in physical phenomena. 2. Description of the Prior Art In some cases, it can be important or critical to know with a high probability whether data received by a sonar system is simply random noise (which may be a false alarm) or is more likely due to the detection of a vessel of interest. In either situation, it is critical to make a determination as quickly as possible. Naval sonar systems require that signals be categorized according to structure (i.e., periodic, transient, random or chaotic). A variety of large sample data processing methods such as spectral analysis, correlogram plots, and the like are available. However, a number of scenarios may also or only comprise small samples. These small samples include loss, or intermittent contact, transients, equipment failure, own ship maneuver, and the like. The existence of such sparse data sets requires methods that are appropriate for reliable and valid processing. As such, there is a need for sparse data set methods in which the methods are separate from those methods which evaluate large sample distributions. It is well known in the art that large sample methods often fail when applied to small sample data sets. The term “randomness” in regard to random noise has different meanings in science and engineering. Random (or randomness) is herein defined in terms of a “random process” as measured by a probability distribution model—namely a stochastic (Poisson) process. In naval engineering applications, waveform distributions in the time domain may be considered purely random if the distributions conform to a noise structure such as WGN (White Gaussian Noise). This determination is made regardless of the underlying generating mechanism that produced the “noise.” Pure randomness may be considered a data distribution for which no mathematical function, relation, or mapping can be constructed that provides an insight into the underlying structure. For example: no prediction model can be generated from the noise/time waveform in order to derive estimates of a target range, course, speed, depth, etc. Also, one must distinguish the term “stochastic” randomness from “deterministic” randomness (chaos) as described in U.S. Pat. No. 5,781,460. The theoretical and practical considerations relevant to the inventive process are contained in the following publications, which are incorporated herein by reference: Abramowitz, Milton and Irene Stegun. Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables. Washington, D.C. United States Government Printing Office: (1964). Feller, William. Introduction to the Theory of Probability and Its Applications. 2nd ed. Vol. I., NY: John Wiley and Sons (1957). Ruhkin, A. L. “Testing Randomness: A Suite of Statistical Procedures.” Theory of Probability and its Applications, Vol. 45, No. 1, pp. 111-132 (2000). Preparata, Franco P. and Michael I. Shamos, Computational Geometry—An Introduction, Springer Verlag (1985). Swed, F. S. and C. Eisenhart. “Tables for testing randomness of grouping in a sequence of alternatives.” The Annals of Mathematical Statistics, 14(1), pp. 66-87 (March 1943). Wald, A. and J. Wolfowitz. “On a test whether two samples are from the same population.” The Annals of Mathematical Statistics, Vol. 11, pp 147-162 (1940) Wilks, S. S. “Order statistics.” Bulletin of the American Mathematical Society. Volume 54, Number 1, Part 1, pp. 6-50 (1948). The standard approach for assessing the hypothesis of spatial randomness for large samples is outlined in the known work on probability theory by W. Feller (Ch. 6, “The Binomial and Poisson Distributions”) [Feller, William. Introduction to the Theory of Probability and Its Applications. 2nd ed. Vol. I., NY: John Wiley and Sons. 1957]. Typically, from a frequency table derived from counts of spatial data in a partitioned subspace, a Chi-square test for homogeneity of Poisson frequency levels is computed and compared to a level of statistical certainty. The Feller reference (pp. 149-154), demonstrates the utility of this procedure for several large samples of naturalistic data analyzed in finite rectangular and circular space. The noted data sets include radioactive decay measurements, micro-organism distribution on a Petri dish, and others. However, the Feller reference provides little guidance on the matter of subspace partitioning including how many partitions should be used and what should be done about non-whole subset partitions. Furthermore, most prior art randomness assessment methods are one time tests designed for one-dimensional or two-dimensional space. The methods are primarily applicable for truly random distributions. However, these quantitative techniques sometimes even fail to correctly label truly nonrandom distributions—as pointed out by Ruhkin (A. L. Ruhkin, “Testing Randomness: A Suite of Statistical Procedures”, Theory of Probability and its Applications, 2000, Vol. 45, No. 1, pp. 111-132). The following United States patents significantly improve the above-noted situation. U.S. Pat. No. 7,277,573 provides a multi-stage method for automatically characterizing data sets containing data points in which are each defined by measurements of three variables as either random or non-random. A three-dimensional Cartesian volume is sized to contain a total number N of data points in the data set which is to be characterized. The Cartesian volume is partitioned into equal-sized cubes, wherein each cube may or may not contain a data point. A predetermined route is defined that goes through every cube one time and scores each cube as a one or a zero; thereby, producing a stream of ones and zeros. The number of runs is counted and utilized to provide a Runs test which predicts if the N data points in any data set are random or non-random. Additional tests are used in conjunction with the Runs test to increase the accuracy of characterization of each data set as random or non-random. U.S. Pat. No. 7,409,323 provides a method for automatically characterizing data sets containing data points, which may be produced by measurements such as with sonar arrays, as either random or non-random. The data points for each data set are located within a Cartesian space and a polygon envelope is constructed which contains the data points. The polygon is divided into grid cells by constructing a grid over the polygon. A prediction is then made as to how many grid cells would be occupied if the data were merely a random process. The prediction becomes one of two forms depending on the sample size. For small sample sizes, an exact Poisson probability method is utilized. For large sample sizes, an approximation to the exact Poisson probability is utilized. A third test is utilized to test whether the Poisson based model is adequate to assess the data set as either random or non-random. As evidenced and in summary, the prior art does not disclose a method to provide a faster solution with greater reliability and for widely varying sizes of three-dimensional data sets. The solutions to the above-described and/or related problems have been long sought without success. Consequently, those skilled in the art will appreciate the present invention that addresses the above-described and other related problems.
{ "pile_set_name": "USPTO Backgrounds" }
Q: Call C standard library function from asm in Visual Studio I have a problem with calling C function from asm project created in visual studio (Win10 x64, Visual Studio 2015). Project consist of one asm file: .586 .model flat, stdcall option casemap:none includelib msvcrt.lib ExitProcess PROTO return:DWORD extern printf:near .data text BYTE "Text", 0 .code main PROC push offset text call printf add esp,4 invoke ExitProcess,0 main ENDP end main When I build project, linker outputs the error: Error LNK2019 unresolved external symbol _printf referenced in function _main@0 Linker output parameters: /OUT:"C:\Users\apple\Documents\SP_Lab7\Debug\SP_Lab7_Demo.exe" /MANIFEST:NO /NXCOMPAT /PDB:"C:\Users\apple\Documents\SP_Lab7\Debug\SP_Lab7_Demo.pdb" /DYNAMICBASE "kernel32.lib" "user32.lib" "gdi32.lib" "winspool.lib" "comdlg32.lib" "advapi32.lib" "shell32.lib" "ole32.lib" "oleaut32.lib" "uuid.lib" "odbc32.lib" "odbccp32.lib" /MACHINE:X86 /SAFESEH:NO /INCREMENTAL:NO /PGD:"C:\Users\apple\Documents\SP_Lab7\Debug\SP_Lab7_Demo.pgd" /SUBSYSTEM:WINDOWS /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /ManifestFile:"Debug\SP_Lab7_Demo.exe.intermediate.manifest" /ERRORREPORT:PROMPT /NOLOGO /TLBID:1 If I comment call print, then everything executes normally (even Windows API function). Is there any way to call C function from asm file without creating cpp file that includes <cstdio>? Is it possible to do? A: Microsoft refactored much of the C runtime and libraries in VS 2015. Some functions are no longer exported from the C library (some are defined in a C header file). Microsoft has some compatibility libraries like legacy_stdio_definitions.lib and legacy_stdio_wide_specifiers.lib, but you can also choose to use the older Visual Studio 2013 platform toolset with the older C libraries. To change the platform toolset: pull down the Project menu; select Properties...; go to Configuration Properties/General, and change Platform Toolset to Visual Studio 2013 (v120) A: It appears that it' possible to use the Visual Studio 2015 Toolset with a few modifications. You'll need to add these libraries to your dependencies: libcmt.lib, libvcruntime.lib, libucrt.lib, legacy_stdio_definitions.lib. Alternatively you could use includelib to include these libraries in your assembly file. Specify C calling convention for your main procedure using PROC C At the end of your file (and this is important) do not use end main, use end only. Not fixing this may cause unexpected crashes. Although we can use ExitProcess to exit our application, we can also put the return code in EAX and do a ret to return. The C runtime calls our main function, and will call the shutdown code for us upon returning. The code could look like: .586 .model flat, stdcall option casemap:none includelib libcmt.lib includelib libvcruntime.lib includelib libucrt.lib includelib legacy_stdio_definitions.lib ExitProcess PROTO return:DWORD extern printf:NEAR .data text BYTE "Text", 0 .code main PROC C ; Specify "C" calling convention push offset text call printf add esp, 4 ; invoke ExitProcess,0 ; Since the C library called main (this function) ; we can set eax to 0 and use ret`to have ; the C runtime close down and return our error ; code instead of invoking ExitProcess mov eax, 0 ret main ENDP end ; Use `end` on a line by itself ; We don't want to use `end main` as that would ; make this function our program entry point ; effectively skipping by the C runtime initialization
{ "pile_set_name": "StackExchange" }
A 21-leaf clover discovered on June 3 by Iwate prefecture farmer Shigeo Obara has shattered the Guinness world record for most leaves on a clover stem (Trifolium repens L.). The current official record is held by an 18-leaf clover that Obara found in his garden in May 2002. The record-breaking clover's 21 leaves each measure about 1 centimeter long and overlap each other like rose petals on a 3-centimeter stem. Obara, a former food crop researcher, has been conducting independent research on clovers in his garden for over 50 years. He first became interested in clover mutations after discovering an unusual patch of 4-leaf clovers in 1951. Since then, Obara has been crossbreeding the plants in his garden to research the genes associated with leaf count, color, pattern and size. Obara plans to file a new application with Guinness, although he is considering waiting a while. "We are likely to find clovers with more leaves," he says. Last month, a family member claimed to have found a 27-leaf clover, but the discovery was not confirmed. While some say that 4-leaf clovers symbolize happiness, 5-leaf clovers symbolize wealth and 6-leaf clovers symbolize fame, it is unclear what 21-leaf clovers symbolize. [Source: Yomiuri]
{ "pile_set_name": "OpenWebText2" }
##################################################################### # Dataset Name: Lottery # #Description: This is an observed/"real world" data set # consisting of 218 lottery values # from September 3, 1989 to April 14, 1990 (32 weeks). # One 3-digit random number (from 000 to 999) # is drawn per day, 7 days per week for most # weeks, but fewer days per week for some weeks. # We here use this data to test accuracy # in summary statistics calculations. # # Stat Category: Univariate: Summary Statistics # # Reference: http://www.itl.nist.gov/div898/strd/univ/lottery.html # # Data: "Real World" # 1 Response : y = 3-digit random number # 0 Predictors # 218 Observations # # Model: Lower Level of Difficulty # 2 Parameters : mu, sigma # 1 Response Variable : y # 0 Predictor Variables# # y = mu + e ##################################################################### ##################################################################### # # Certified Values # ##################################################################### mean = 518.958715596330 standardDeviation = 291.699727470969 autocorrelationCoefficient = -0.120948622967393 n = 218 ##################################################################### # # Data # ##################################################################### 162 671 933 414 788 730 817 33 536 875 670 236 473 167 877 980 316 950 456 92 517 557 956 954 104 178 794 278 147 773 437 435 502 610 582 780 689 562 964 791 28 97 848 281 858 538 660 972 671 613 867 448 738 966 139 636 847 659 754 243 122 455 195 968 793 59 730 361 574 522 97 762 431 158 429 414 22 629 788 999 187 215 810 782 47 34 108 986 25 644 829 630 315 567 919 331 207 412 242 607 668 944 749 168 864 442 533 805 372 63 458 777 416 340 436 140 919 350 510 572 905 900 85 389 473 758 444 169 625 692 140 897 672 288 312 860 724 226 884 508 976 741 476 417 831 15 318 432 241 114 799 955 833 358 935 146 630 830 440 642 356 373 271 715 367 393 190 669 8 861 108 795 269 590 326 866 64 523 862 840 219 382 998 4 628 305 747 247 34 747 729 645 856 974 24 568 24 694 608 480 410 729 947 293 53 930 223 203 677 227 62 455 387 318 562 242 428 968
{ "pile_set_name": "Github" }
Tonight at the Creative Arts Emmys, Norman Lear, 97, became the oldest Emmy winner in history with a victory in the category of Outstanding Variety Special (Live) for Live in Front of a Studio Audience: Norman Lear’s ‘All In the Family’ And ‘The Jeffersons.’ “Thank you, thank you, and holy sh*t,” laughed Lear, who executive produced and wrote the special. “I got a great reaction, saying that when opening a Christmas present when I was seven years old. It’s 90 years later, and I feel much the same.” Calling the project “one of the greatest things that ever happened to me,” the television icon gave much of the credit for the win to fellow executive producer Jimmy Kimmel, who was the driving force behind the project. “It was all Jimmy Kimmel’s idea,” he said, while also thanking “all the farmers and the ranchers across the globe that have nourished me for all these years, and gotten me here tonight.” Backstage, Lear and Kimmel talked about the next installment in the Live in Front of a Studio Audience franchise, which ABC in August picked up for two more outings — a live holiday special later this year, as well as another special in spring 2020. “We are going to do [a special] again. In fact, in December, we’re going to do something,” Kimmel said backstage. “We haven’t revealed specifically what shows we’re going to do, but we are going to do it again on ABC in December.” Lear beat out Our Planet narrator Sir David Attenborough for the title of oldest Emmy winner when the results of the evening’s final category, Outstanding Variety Special (Live), came in. Earlier in the night, Attenborough, who is 93, won a second consecutive Outstanding Narrator Emmy, this time for Our Planet. Last year, he became the oldest Emmy winner with his victory in the same category for Blue Planet II. A TV icon, Lear is a 15-time Emmy nominee with five statuettes to his name. Live in Front of a Studio Audience was a one-night-only event, which enlisted all-star casts to perform episodes of Lear’s iconic sitcoms, All in the Family and The Jeffersons. While Live in Front of a Studio Audience was nominated for three Emmys, its only win tonight was for Outstanding Variety Special (Live).
{ "pile_set_name": "OpenWebText2" }
Antimicrobial activity of conventional and plant-extract disinfectant solutions on microbial biofilms on a maxillofacial polymer surface. Dentists often note problems with infection in patients with maxillofacial prostheses. Conventional disinfection protocols are not always effective and may alter the properties of the polymer used in the prosthesis. Thus, the search for improved disinfection methods is important. The purpose of this in vitro study was to evaluate and compare the antimicrobial activity of conventional disinfectant solutions (water and neutral soap and 4% chlorhexidine) and plant extracts (Cymbopogon nardus and Hydrastis canadensis) on specimens of maxillofacial silicone contaminated with Candida albicans and Staphylococcus aureus biofilms. Seventy-two silicone (MDX4-4210) specimens were fabricated (5×2 mm) and sterilized. Thirty-six were contaminated with C albicans (10(6) cells/mL) and 36 with S aureus (10(8) cells/mL) to evaluate the antimicrobial activity of the cleaning protocols. After incubation (37°C/72 hours), the specimens were divided into 5 groups: not disinfected (positive control), soaking in saline solution for 10 minutes, soaking in 4% chlorhexidine for 10 minutes, soaking in C nardus for 10 minutes, soaking in H canadensis for 10 minutes, and washing by hand with water and neutral soap for 30 seconds. The viability of cells was evaluated by XTT (2,3-bis-(2-methoxy-4-nitro-5-sulfophenyl)-2H-tetrazolium-5-carboxanilide) assay and by scanning electron microscope analysis. The results were analyzed by ANOVA and the Tukey HSD test (α=.05). All disinfection solutions provided a statistically significant reduction in biofilm viability compared with the control group for both microorganisms (P<.05). Washing with water and neutral soap was significantly more effective in reducing biofilm viability than immersion in the disinfection solutions, with persistence of viable microorganisms between 1.05% for C albicans and 0.62% for S aureus after this cleaning protocol. Photomicrographs revealed that 4% chlorhexidine altered the surface of the polymer. Within the limitations of this in vitro study, it was concluded that the cleaning protocols with different disinfectant solutions produced a significant reduction in the viability of C albicans and S aureus biofilms on the silicone polymer. Washing with water and neutral soap was the most effective protocol against both microorganisms.
{ "pile_set_name": "PubMed Abstracts" }
ohhhhh myyyy gooooddess!! i must be blind or something!!! i didn't notice that she hid part of her leg to pretend that it happened!!! i legitimately did not see it at all!!! i'm an idiot, idk how i missed it.
{ "pile_set_name": "OpenWebText2" }
Joint statement from Russia, Germany, France and Turkey followed summit on war-torn country This article is more than 1 year old This article is more than 1 year old The leaders of Turkey, Russia, France and Germany on Saturday called for a ceasefire around the last major rebel-held bastion of Idlib in Syria to be preserved. The four nations “stressed the importance of a lasting ceasefire” according to a statement read by Turkish president, Recep Tayyip Erdoğan, at the end of a major summit in Istanbul. Erdoğan, along with Russian president Vladimir Putin, French president Emmanuel Macron and German chancellor Angela Merkel met to try to find a lasting solution to the Syrian conflict, in which more than 360,000 people have been killed since 2011. After a joint press conference, Macron urged Russia, which supports the government of Syrian president Bashar al-Assad, to exercise “very clear pressure” on Damascus for a “stable and lasting ceasefire in Idlib”. Last month Turkey, which backs the rebels, agreed with Russia to create a buffer zone around Idlib amid fears of an impending assault on the northwestern province, which many feared would lead to a humanitarian disaster. However clashes have continued in Idlib since. On Friday seven civilians were killed by Syrian regime artillery fire, the highest death toll since the ceasefire was reached. A joint statement adopted at the end of the summit called for a committee to be established to draft a new Syrian constitution before the end of the year, “paving the way for free and fair elections” in the war-torn country. The statement also spoke of “the need to ensure humanitarian organisations’ rapid, safe and unhindered access throughout Syria and immediate humanitarian assistance to reach all people in need.” It also said that conditions needed to be created “throughout the country for the safe and voluntary return of refugees and internally displaced persons”.
{ "pile_set_name": "OpenWebText2" }
Searching for Hummer H2 Axles and CV-Shafts Need to replace a failed Axle or Axle Shaft assembly on your Hummer H2? Then shop at 1A Auto for replacement front and rear Axles and Axle Shafts for your Hummer H2, at a great price. 1A Auto has a large selection of aftermarket Hummer H2 Axles and Axle Shafts, and ground shipping in the continental U.S. is always free! Visit us online or call 888-844-3393 and order today!
{ "pile_set_name": "Pile-CC" }
Establishment of a leukemic lymphocyte culture from human aqueous humor. We have recently developed new techniques for culturing chronic lymphocytic leukemia (CLL) lymphocytes from human aqueous humor. We used leukemic lymphocytes collected from the aqueous humor of a patient with CLL and leukemic glaucoma. We grew these leukemic cells in combination with a feeder cell layer and other technical refinements. Microscopy and immunoassay indicate success in obtaining a homogenous population of B-type CLL cells through the 12th passage of the culture. No significant effect on cell growth was found with either of two mitogens (PWM and PHA), or between culture with and without autologous serum. Our new techniques for culturing leukemic cells derived from the aqueous humor provide a reliable resource for the study of chronic lymphocytic leukemia lymphocytes and leukemic glaucoma.
{ "pile_set_name": "PubMed Abstracts" }
Southern African astronomy buffs have another reason to celebrate – hot on the heels of the awarding of the greater part of the Square Kilometre Array to the region, comes the news that the NamibRand Nature Reserve in Namibia has been proclaimed as the continent’s first international dark-sky reserve (IDSR), meaning that it’s one of the best places on earth to star-gaze. The proclamation falls under the dark-sky movement, an initiative of the Arizona-based non-profit International Dark-Sky Association (IDA) which was the first organisation, and is currently the largest, to embrace this concept. Located in southwestern Namibia, the privately owned NamibRand covers an area of just over 170 000 ha. It joins a handful of other areas around the world – and only three other reserves, in New Zealand, Canada and the UK – recognised by the IDA as the best places on earth to see the splendour of the universe at night, just as it was before humans came along. The dark-sky movement seeks to limit the intrusion of artificial light into the experience of observing the night skies. The adverse effect of this intrusion is known as light pollution, and it happens when artificial light is not properly shielded, causing sky glow, glare or light trespass. This limits the ability to see details of the night sky whether unaided or through a lens. Light on the ground need not be unfriendly to dark skies, if properly set up. But poorly shielded artificial light, according to the IDA, also affects the ecology of an area and impacts on nocturnal animals, predatory behaviour, migration patterns and the mating and communication habits of creatures. Gold tier status for dark-sky excellence While not the first IDSR in the world, the NamibRand is the first to achieve gold tier status. For stargazers this is good news, because it means that the reserve is unparalleled in terms of night-time viewing as any artificial lights that are present make little or no impact on the dark sky. The nearest town lies over 100km away, and the closest major city is Windhoek, some 400km distant. For ecologists, gold status is a sure indication that the NamibRand staff has made every effort to keep the impact of artificial light on the reserve’s fauna to a minimum. "We did a detailed audit of all external light fixtures on the reserve and applied corrective measures, including retrofitting, replacing fixtures or using lower wattage bulbs," said the reserve's CEO Nils Odendaal, "so that these would comply with our lighting guidelines as stipulated in our dark-sky reserve management plan." According to the reserve's dark-sky lighting guidelines, exterior lighting is kept to a minimum and where necessary, is not only fully shielded, but emits an amber or red light which is kinder to the eyes. Lights are also controlled by motion detectors or timers, where possible, to ensure they’re on for as short a time as possible. Vehicles are encouraged to use headlights (on dim) only when the light of the moon is insufficient, otherwise they use parking lights. It’s perfectly safe to drive using these less powerful lights at the slower speeds used on the reserve. Headlights on bright are only allowed on the public road C27, and vehicle lights may not be directed at buildings or tourist accommodation. The reserve will continue with its responsibility of raising awareness of the importance of preserving the night skies. “NamibRand serves on several local and national committees where we can share information with other conservation organisations and stakeholders,” said Odendaal. ”An example of this is the recently launched Nam-place project, which aims to unite landowners and custodians across large landscapes in an effort to co-manage these local landscapes for the benefit of conservation.” NamibRand is using its position on these committees, and its close ties to bodies such as the Namibia Nature Foundation, to drive awareness about light pollution and light conservation in general, said Odendaal. In addition, the NGO Namib Desert Environmental Education Trust, located in the reserve, educates visitors, schoolchildren and neighbouring communities about astronomy, the night sky, and other aspects of conservation and sustainability. The organisation hosts about 1 000 school children each year and also disseminates environmental literature to a readership base of about 18 000 readers, said Odendaal. Exceptionally dark skies NamibRand’s application was submitted to the IDA in February 2012. It was championed by retired physics and astronomy professor George Tucker from Nassau, New York. Tucker is an IDA member who had first visited Namibia back in 2003 and, he said in the application’s introduction, was amazed at the sheer volume of stars visible to him there – stars he’d never seen before. He was even able to move around, in the dark, by the light of the Milky Way alone. Over the past eight years Tucker has been conducting measurements of the darkness of the Namibian sky, using a sky quality meter – this is an instrument which gives a measure of the night sky’s brightness in terms of the magnitude per square arc-second. An arc-second is a unit of angular measurement that is equal to 1/3600 degrees of an arc – we understand that there are 360 degrees in a circle, 60 arc-minutes in a degree, and 60 arc-seconds in each arc-minute. The arc-second is a tiny measurement – for a human hair to cover one arc-second it would have to be viewed from 10 metres away. The magnitude is simply a measure of the brightness of an object. The higher the number given by the sky quality meter, the darker the sky. Tucker consistently got readings of over 22 on his meter – this means that observers will be able to see, with a help of a telescope, stars of the 22ndapparent magnitude, which are very dim. To put this in perspective, the faintest celestial object visible to a sharp naked eye is around magnitude six, and that’s in exceptionally dark conditions. The spiral galaxy M81 or Bode’s Galaxy, magnitude 6.9, is about 12-million light years away and pushes the ability of the naked eye to the limit. In terms of the Bortle Dark-Sky Scale, which measures night skies on a scale from one to nine – where one is dark enough to observe phenomena such as zodiacal light and shadows cast on the ground by the Milky Way, and nine is a brilliantly lit urban sky – the NamibRand comes in at one. The measurements haven’t changed in all of the eight years, wrote Tucker, and it was these outstanding readings that prompted him to nominate the NamibRand as an IDSR. "Viewing the pristine night sky over the NamibRand is an unforgettable experience,” he said in a statement. “Achieving this status is a significant accomplishment not just for the NamibRand, but also for Namibia and all of Africa." Because tourism forms the major part of the NamibRand’s income, it’s imperative for the sustainability of the reserve that its natural assets are protected. As an astrophotography site, it would be highly sought after. The &Beyond hospitality group maintains an observatory, which boasts not only a Meade LX200R 12-inch telescope but also a full-time astronomer, at its luxury Sossusvlei Desert Lodge, situated in the NamibRand. The group also played an important role in the reserve’s new IDSR status. According to Dark Skies Awareness, the natural sky brightness level for an unpolluted and clear starry sky is around 21.6 magnitude per square arc-second – at this level the Milky Way can be seen blazing overhead, as well as about 6 000 stars, with the naked eye. Since viewers in bright cities may count themselves lucky to see a few hundred stars in the sky on a clear night, we begin to understand how important it is to maintain these dark-sky areas. Not only are starry skies a pleasure to view, but they are an important part of human and natural life. Many animals only come out at night. Navigators have used, and still use the stars and constellations to guide them. Too much lighting leads to energy waste and the consequent release of greenhouse gases in the production of that energy. It’s even claimed that having better night vision will help to cut down on crime, as criminals will be easier to spot. The IDA has named various measures people can take to keep light pollution to a minimum in areas where it matters. They include shielding outdoor lighting; using light only when necessary and then just enough to get the job done; using dimmers and timers; and using a red- or yellow-tinted light which isn’t as harsh.
{ "pile_set_name": "Pile-CC" }
--- title: NullableOperators.( >=? )<'T> Function (F#) description: NullableOperators.( >=? )<'T> Function (F#) keywords: visual f#, f#, functional programming author: dend manager: danielfe ms.date: 05/16/2016 ms.topic: language-reference ms.prod: visual-studio-dev14 ms.technology: devlang-fsharp ms.assetid: d1336d6b-a0bf-4b91-8f52-fad25e9ac212 --- # NullableOperators.( >=? )<'T> Function (F#) The `>=` operator where a nullable value appears on the right. **Namespace/Module Path**: Microsoft.FSharp.Linq.NullableOperators **Assembly**: FSharp.Core (in FSharp.Core.dll) ## Syntax ```fsharp // Signature: ( >=? ) : 'T -> Nullable<'T> -> bool when 'T : (IComparable) and 'T : (new : unit -> 'T) and 'T : struct and 'T :> ValueType // Usage: nullableValue >=? value ``` #### Parameters *value* Type: 'T The first input value. *nullableValue* Type: **System.Nullable&#96;1**&lt;'T&gt; The second input value, as a nullable value. ## Return Value True if the first value is greater than or equal to the second value. ## Remarks If the second value is null, the result is `false`. ## Platforms Windows 8, Windows 7, Windows Server 2012, Windows Server 2008 R2 ## Version Information **F# Core Library Versions** Supported in: 4.0, Portable ## See Also [Linq.NullableOperators Module &#40;F&#35;&#41;](Linq.NullableOperators-Module-%5BFSharp%5D.md) [Microsoft.FSharp.Linq Namespace &#40;F&#35;&#41;](Microsoft.FSharp.Linq-Namespace-%5BFSharp%5D.md)
{ "pile_set_name": "Github" }
LI Staff Directory Bethany Rolando Regional Development Officer Bethany Jones, the Regional Development Officer for the Leadership Institute's South Central Region, is responsible for raising funds for LI's programs through building and cultivating relationships with supporters in eight states. During her first year at LI, she was a 2013-14 Koch Associate at the Charles Koch Institute. Prior to coming to the Leadership Institute, Bethany worked as a Student Talent and Relations intern for the Charles Koch Institute while participating in the Koch Internship Program. Previously, she graduated from The College of William and Mary (May 2013) with a Bachelor's degree in Public Policy and a minor in Management and Organizational Leadership. Bethany was born and raised in a small town outside of Charleston, South Carolina. When not working to advance liberty, she loves to read, sing, and play softball and the piano. During football season she dedicates most of her spare time to rooting for the Clemson Tigers and her fantasy football team. She currently lives in Alexandria, Virginia.
{ "pile_set_name": "Pile-CC" }
Craig Colony Craig Colony may refer to: Craig Hospital, Englewood, Colorado, US Craig Colony (New York), US
{ "pile_set_name": "Wikipedia (en)" }
Sheppard Sheppard can refer to: Places Sheppard, Wisconsin, an unincorporated community, United States Sheppard Avenue in Toronto, Canada. Hence: Sheppard subway line Sheppard West (TTC), formerly Downsview, subway station Sheppard-Yonge (TTC), formerly Sheppard, subway station Kate Sheppard House, the historic home of civil rights campaigner Kate Sheppard Sheppard Air Force Base in Texas, United States Other Sheppard (name) Sheppard (band), an Australian Brisbane-based rock band Sheppard (EP) See also Shepherd (disambiguation) Shepard (disambiguation) ru:Шепард
{ "pile_set_name": "Wikipedia (en)" }
Q: How to check that all files in a file system are can be read without errors? I'm using an external hard drive with an ext4 file system to make backups. The backup software I use (faubackup) copies the file hierarchy 1:1 into a timestamp-named folder on the hard drive, and makes incremental backups in such a way that it hardlinks new copies of files whose contents have not changed to the same file in the corresponding subfolder of the previous backup. Since I recently had a backup drive die on me, I now want to make sure that all the files written can actually be read without I/O error, so I know I can rely on my backup. One way to do so would be to read the whole partition, e.g. by dd'ing it to /dev/null. However, the disk is 3TB large, and doing so would take about 7 hours (via USB 3.0). Another way would be to use e2fsck with the -c option, but this also takes ages. I'm thinking it should be possible to speed the process up by not checking the whole disk, but only the files, which is only a fraction of the whole disk size. This could be done e.g. by writing all files to a tar archive which is not written to disk, but sent to /dev/null. Here the problem is the hard linking: If I have say 10 incremental backups, the storage for that is again just a fraction of the disk, but it appears to be about 10 times larger than that. My question: Is there a way to read only the files on the disk, and only one in each set of files hard linking to the same storage space? Or is there a way to make e2fsck -c or something similar only check the used parts of the file system (allocated blocks)? A: GNU tar does not copy the content of hard linked files multiple times. Read the first answer to this question or the official documentation on this topic. You can test this by piping the output (the archive) of tar through wc: tar cf - -C <mountpoint of your disk> . | wc -c and verify the archive size in bytes (you can compare this to the result with the tar option --hard-dereference).
{ "pile_set_name": "StackExchange" }
15% discount on all Plus size and Fat tyres for 1 day using code "PLUSTNFATYRES" This offer applies to any VEE tyre, if you do not see it in stock or listed I can get it (offer applies if bought today)
{ "pile_set_name": "Pile-CC" }
<Canvas> <Kind>42</Kind> <Name>LightSettings</Name> <IsMinified>0</IsMinified> <XPosition>230.000000000</XPosition> <YPosition>431.000000000</YPosition> </Canvas> <Widget> <Kind>2</Kind> <Name>ENABLE</Name> <Value>1</Value> </Widget> <Widget> <Kind>2</Kind> <Name>SMOOTH</Name> <Value>1</Value> </Widget> <Widget> <Kind>25</Kind> <Name>R</Name> <Value>0.500000000</Value> </Widget> <Widget> <Kind>25</Kind> <Name>G</Name> <Value>0.500000000</Value> </Widget> <Widget> <Kind>25</Kind> <Name>B</Name> <Value>0.500000000</Value> </Widget> <Widget> <Kind>25</Kind> <Name>H</Name> <Value>0.000000000</Value> </Widget> <Widget> <Kind>25</Kind> <Name>S</Name> <Value>0.000000000</Value> </Widget> <Widget> <Kind>25</Kind> <Name>V</Name> <Value>0.500000000</Value> </Widget>
{ "pile_set_name": "Github" }
Ends tomorrow. Under Armour is taking an extra 20% off outlet orders $25 or more with coupon code GETGEAR [Exp 5/22]. Plus get free 2-day shipping with ShopRunner (free signup for 2 years with PayPal). There may be some exclusions.
{ "pile_set_name": "Pile-CC" }
I'm not upset that Wagner or Wilson lost, but I think the vote should have been much closer in both cases. In fairness to RG3, the vote was held before the regular season was even over (which is really dumb, IMO), and when a lot of the votes were submitted RG3 had a strong statistical edge over Wilson. I also don't blame Andrew Luck for anything- it's not his fault that more than 25% of the voting press corps are idiots and/or just badly want to further a storyline they'd been pushing for over a year with Luck. But still, 3rd place for Wilson? That's going to look pretty dumb a couple years from now. kearly wrote:But still, 3rd place for Wilson? That's going to look pretty dumb a couple years from now. I don't think that's true. There is not such a huge gulf between the three that would make Wilson coming in 3rd such a travesty. We are a little blinded by Wilson's light, of course. But RGIII had a great year and Luck was pretty impressive too, especially with all argumentative points that have been discussed in here ad nausea. kearly wrote:But still, 3rd place for Wilson? That's going to look pretty dumb a couple years from now. I don't think that's true. There is not such a huge gulf between the three that would make Wilson coming in 3rd such a travesty. We are a little blinded by Wilson's light, of course. But RGIII had a great year and Luck was pretty impressive too, especially with all argumentative points that have been discussed in here ad nausea. As a regular season vote, there's not much to fuss about. I think there is clearly a gulf between Wilson and Luck. RG3 might actually be better than Wilson if injuries didn't factor, but they do. RG3 and especially Luck were on teams that massively overachieved last season and the Seahawks actually underachieved by two wins (FO). I think people will look back and think "really, 3 times more votes for RG3?" Not because Wilson is so much better, but because the vote should have been very close. I think people will laugh when they see that Luck got one more vote than Wilson did. His rookie year was relatively pedestrian and I suspect he's going to have an Andy Dalton type career. Wilson's climb started too late in the season...he had to shake off that game manager label which a lot of voters probably had. Is Wilson better than Luck? I think so, clearly. But not to the point of making the vote look "pretty dumb", not to mention I doubt many people look back at Rookie of the Year voting with that much interest. If Luck won OROTY I would have barfed. A sub 55% completion average is terrible. Sanchez got smoked for having a better one. I get it, Wilson came on late in the season (week 13 on), and God did he ever come on like nobody in NFL history, but it was PC and Bevell's choice to limit his playbook early on instead of letting the kid play, and nobody outside of Seattle cares that that happened. RG3 was a highlight reel from week 1, even his injurys were replayed for weeks. Of course he was going to win. It would have been so much worse if his stats for the season were flipped (if he started hot and regressed) and won the ROY. Could you imagine the let down that would be? But the ROY voting is horrible, and should be done after the season. So stupid. "When is the NFL going to start fining receivers for running routes across KAM’S MIDDLE?!?!"-bpup33 I don't mind that the others got more votes. Wilson got some votes being a third round draft pick that until the second half of the season wasn't asked to do much. All that matters to me is that Wilson is a Seahawk. I like having the chip-on-shoulder players because it gives them fire. By the end of the season, Wilson was playing better than most of the QB's in the league, not just the rookies and the two others were exceptional players with a lot more attention. We will continue to support Wilson and he will continue to do whatever he can to improve and win football games. One thing is for sure, the NFL media is giving Wilson and the Seahawks a lot more love than they are Luck and RG3. In any other year, Wilson would be ROY, he is in some pretty good company as part of the conversation. The only concern I have and I mean only is that Bevell will get cute and pass even though Lynch is gaining yardage. 1/30/2015 - loaf FlyingGreg wrote:Is Wilson better than Luck? I think so, clearly. But not to the point of making the vote look "pretty dumb", not to mention I doubt many people look back at Rookie of the Year voting with that much interest. It is already dumb and will just look dumber with time, but you are right about the latter part. It will be forgotten by most and noticed by few. It's just one of those things that if someone stumbles across this 3 years from now they'll think "wow. really?" FlyingGreg wrote:Is Wilson better than Luck? I think so, clearly. But not to the point of making the vote look "pretty dumb", not to mention I doubt many people look back at Rookie of the Year voting with that much interest. It is already dumb and will just look dumber with time, but you are right about the latter part. It will be forgotten by most and noticed by few. It's just one of those things that if someone stumbles across this 3 years from now they'll think "wow. really?" What I find dumb is the concept that Luck is an untouchable. He's already had a lot of people making convenient excuses for his poor completion percentage. He was the "ordained" from Day 1. The non-stop slobbering all over him last season was pathetic. One example - I remember the media freaking out over his "game winning TD pass" against the Lions. In essence, he threw a 4-yard dump off to Avery who did all the work. Big whoop... kearly wrote:But still, 3rd place for Wilson? That's going to look pretty dumb a couple years from now. Especially when he winds up being the most marketable player among the three legendary rookie QBs of '12. RG3 is fun to watch when healthy, but I question his long-term viability in the NFL, and Luck represents what remains of old school pocket passing QBs. Granted, Luck has mobility, it's just not a big feature in his game. The modern NFL seems to be gravitating towards an option attack with a mobile QB, which is a lot more fun to watch (to my eyes), than traditional drop back 'n pass. Wilson was a one man highlight reel for the latter 10 games of the season. To my eyes, he was just a better QB than both RG3 and Luck in that he had the same throwing ability as both while possessing mobility, elusiveness unlike anything the NFL has had to deal with in a long time, and the rare ability to progress through 3 and 4 reads on seemingly every play. RG3 is fast, but did not look as comfortable sitting in the pocket, going through his reads, as Wilson did in the latter half of the season. I feel Russell's 385 yard, 2 passing, 1 rushing TD performance in Atlanta was the face of the Russell Wilson we will see through the best parts of his career here. Accurate passing and reading the defense is still of the utmost importance in the NFL and Russell is now the prototypical mobile quarterback, though in 5 years he will probably still be the shortest. Just as deadly throwing on the run as he is in the pocket, and a threat to break a big rush, sliding before a linebacker or safety can take his head off. RG3, brilliant as he was, struck me as a flash-in-the-pan type of player, not unlike the feeling I get from Colin Kaepernick, or got from Michael Vick. Russell plays smart. He plays with the "big picture" clearly in focus. I see that quality in Luck, but not in RG3. If Russell's postseason is a fair indication of what we can expect from him in the future, this will be Russell Wilson's NFL and everyone else will merely be playing in it. He has the potential for that kind of gravitas. Feel free to contact me if you need legal assistance. I have a great lawyer that helped me with an ex who violated my privacy and kept harassing me on MySpace and Facebook. He's very good. And there is legal precedent. - linuxpro Wagner was only 15 or so sacks behind Kuechly until the Hawks went on a tear. Teams just weren't running against us as much as they were against Carolina later in the season and Kuechly just kept gaining. Imo Wagner could just as easily been the man. Stats aren't everything, the one stat that matters is wins and luck took a bad team and won 11 games, Luck had a great rookie year and has a very bright future ahead of him, I think this class will go down as the greatest QB class of all time with wilson and luck becoming HOFers and RG3 being very good but having injuries be a problem. I'm with you more often than not Kearly, but this was and is ridiculous. Find 5 Colts games from this season to watch and come back here and tell us what you found. Andrew Luck was from another planet this season, and it is one of those things you'll just never understand until you watch it. Andy Dalton? How can we quantify this for some sort of friendly wager? Once upon a time I had doubts about Luck. I have since admitted I was terribly wrong and moved on with my life. Dude, he's really, really good. Is he going to be Peyton Manning? Who knows? But Andy Dalton? How do we know Luck's team was bad? Luck supporters have an inherent contradiction in their argumentation. It goes something like this . . . Luck fan: Luck took a team that won two games in 2011 and won 11 games. Luck alone did this, the rest of the team was a two win team . . . Luck critic: that two win team won 10 games the year prior. It was injuries and bad coaching that made them play down to two wins. in reality Andrew Luck took a healthy ten win team and added a win. Luck fan: the new GM and coach completely turned over the roster in 2012. It wasn't the same team as the 2010-2011 team. The problem is the second response undermines the first. We're the 2012 Colts simply the 2011 Colts plus Andrew Luck, or were they an entirely different team? The answer is largely the latter. Ad if that's the case, why are all the wins attributed to the single player, rather than the largely new roster and entirely new coaching staff? Perhaps the win increase has as much to do with remaining vets being healthy, new players across the board stepping up (rookies Vic Ballard, TY Hilton, Dwayne Allen, veterans Donnie Avery, Samson Satele, Vontae Davis) and a coachingstaff that won coach of the year? The Colts improvement in record was not due to the addition of Luck. it was a perfect storm of health, player acquisition and improvement and quality coaching. In other words, it was a team achievement. Tical21 wrote:I'm with you more often than not Kearly, but this was and is ridiculous. Find 5 Colts games from this season to watch and come back here and tell us what you found. Andrew Luck was from another planet this season, and it is one of those things you'll just never understand until you watch it. Andy Dalton? How can we quantify this for some sort of friendly wager? Once upon a time I had doubts about Luck. I have since admitted I was terribly wrong and moved on with my life. Dude, he's really, really good. Is he going to be Peyton Manning? Who knows? But Andy Dalton? I agree that Luck is good, and probably destined for great things, but I wouldn't say he was "from another planet this season". I watched all of the rookie QBs this season, including Tannehill and Weeden, and to my eyes, Luck was easily the third best of the group. Feel free to contact me if you need legal assistance. I have a great lawyer that helped me with an ex who violated my privacy and kept harassing me on MySpace and Facebook. He's very good. And there is legal precedent. - linuxpro had he broke Peyton's record instead of just tying it, would that have pushed him into the lead... ? i feel bad for him that he didn't get the award, but as a fan i could care less, we know what we have with him.. and all those aculades will come in time. McGruff wrote:How do we know Luck's team was bad? Luck supporters have an inherent contradiction in their argumentation. It goes something like this . . . Luck fan: Luck took a team that won two games in 2011 and won 11 games. Luck alone did this, the rest of the team was a two win team . . . Luck critic: that two win team won 10 games the year prior. It was injuries and bad coaching that made them play down to two wins. in reality Andrew Luck took a healthy ten win team and added a win. Luck fan: the new GM and coach completely turned over the roster in 2012. It wasn't the same team as the 2010-2011 team. The problem is the second response undermines the first. We're the 2012 Colts simply the 2011 Colts plus Andrew Luck, or were they an entirely different team? The answer is largely the latter. Ad if that's the case, why are all the wins attributed to the single player, rather than the largely new roster and entirely new coaching staff? Perhaps the win increase has as much to do with remaining vets being healthy, new players across the board stepping up (rookies Vic Ballard, TY Hilton, Dwayne Allen, veterans Donnie Avery, Samson Satele, Vontae Davis) and a coachingstaff that won coach of the year? The Colts improvement in record was not due to the addition of Luck. it was a perfect storm of health, player acquisition and improvement and quality coaching. In other words, it was a team achievement. This. Seriously, that showman of an owner in Indy put the absolute WORST quarterbacks on the field that year. No leadership, no faith, etc by the team , they suxored pretty bad. Then voila' in comes Luck and they're awesome again? The team was great. The QB's they chose were the worst. I can't decide if it was the most underhanded scheme of all time to tank the season by your own sabotage, or the most ingenious. I don't want to take anything away from Luck but I refuse to give him the credit for "turning a losing team into a winning one". Tical21 wrote:I'm with you more often than not Kearly, but this was and is ridiculous. Find 5 Colts games from this season to watch and come back here and tell us what you found. Andrew Luck was from another planet this season, and it is one of those things you'll just never understand until you watch it. Andy Dalton? How can we quantify this for some sort of friendly wager? Once upon a time I had doubts about Luck. I have since admitted I was terribly wrong and moved on with my life. Dude, he's really, really good. Is he going to be Peyton Manning? Who knows? But Andy Dalton? I think RGIII and Wilson were lumped into being very comparable, with Luck representing the 'classic' QB. I'm not saying I agree with the logic, but it's just how I understand it... If the race was between Wilson and Luck, most of the Griffin voters would have swayed to Wilson, so I don't really see Wilson as coming in 3rd. If not for Griffin Luck would have been second and Wilson 1st.
{ "pile_set_name": "Pile-CC" }
Q: create JSON with php Sorry if this seems obvious but all the searches I am doing for this are returning complex answers and I have been struggling with this all day... I have been trying to create json data in PHP and really I want to send variables into the json data. I have not been able to do that so I followed a tutorial online and copied the code exactly yet i'm still not getting an output in my local host.. Please can somebody put me out of my misery... <?php $jsonData = new stdClass(); $people = array( array( 'name' => 'Luci', 'age' => 25, 'sex' => 'female' ), array( 'name' => 'John', 'age' => 27, 'sex' => 'male') ), array( 'name' => 'Peter', 'age' => 22,) 'sex' => 'male' ) ); $jsonData->source = "Program Knowledge"; $jsonData->published = date('Y-m-d H:s:i;'); $jsonData->status = true; $jsonData->people = $people; echo json_encode($jsonData); ?> eventually I was going to try $var = 123456; $jsonData->codeid = $var but I need to figure out just how to get it so work with data I have typed in. Many thanks..!! A: 'sex' => 'male') 'age' => 22,) Need to remove the parenthesis from these two lines, they're parsing errors
{ "pile_set_name": "StackExchange" }
Alastair Grant/Associated Press Serena Williams barely broke a sweat as she defeated Maria Sharapova 6-2, 6-4 to reach the 2015 Wimbledon women's final. Sharapova put in a thoroughly disappointing performance in the first set, as Williams took just 33 minutes to close out the opener. The Russian showed far more resilience in the second set, but Williams remained comfortable throughout. The result was confirmed by Wimbledon's official Twitter feed: Williams will face Garbine Muguruza on Saturday, as the Spaniard defeated Agnieszka Radwanska in three sets earlier on Thursday. Prior to the match, the Telegraph's Daniel Schofield described the rivalry between the pair as being "built on antipathy, mistrust and malice," but that rarely surfaced on such a one-sided occasion. This was the American's 17th consecutive victory over Sharapova in a run stretching back to the 2004 final at All England Club, when the Russian won her only Wimbledon title to date. Sharapova had not been entirely convincing throughout the tournament, with her serve in particular letting her down, and her poor record against Williams may have been weighing on her mind. Pavel Golovkin/Associated Press Three double-faults in the opening game gifted Williams a break, as Sharapova struggled for any sort of rhythm. ESPN's Brad Gilbert provided insight into the 28-year-old's troubles: A number of unforced errors into the net made things difficult for Sharapova in her second service game, but an excellent volleyed winner helped the Russian to hold. Williams used the opening game as a springboard for the rest of the set, however, breaking her opponent again as she raced into a 5-1 lead with some powerful serves and exquisite returns. As the New York Times' Ben Rothenberg noted, the 20-time Grand Slam-winner was scarcely made to work for her points: Sharapova held her serve once more to make it 5-2, but Williams cruised to the first set, serving the final game out, 40-15. As the BBC noted on its TV broadcast, the Russian made nine unforced errors and hit just four winners in the first set, compared with Williams' seven and 12, respectively. Those stats highlighted her inability to threaten Williams on the relatively rare occasions they played out significant rallies. Two strong service games for Sharapova early in the second set hinted at a comeback as she improved the depth and location of her serve significantly, but Williams always looked comfortable. At 2-2, Sharapova's resistance finally crumbled, and the 33-year-old broke her once again thanks to yet another double-fault. Suzanne Plunkett/Associated Press Along with her serve, Sharapova's groundstrokes remained a problem throughout the match, as she failed to find any real depth on her shots. However, the world No. 4 did show flashes of quality when she saved three break points to bring the score back to 4-3, admittedly courtesy of some uncharacteristically loose play from Williams. The result was never in doubt though, as Serena battled back to 5-3. Sharapova saved a match point while defending her own serve once more, but Williams closed the match out with ease on her own serve. Ultimately, the semi-final was one of the most comfortable of Williams' career. The five-time Wimbledon winner rarely had to get out of first gear to overcome Sharapova on this occasion. According to ESPN Stats & Info, Williams now has the third-most victories in women's Grand Slam competitions in the Open Era: Up next for the world No. 1 is Muguruza, who impressed in her battling victory over Radwanska on Thursday. But the 21-year-old's lack of experience will give the American a big advantage in their showdown. The youngster will have nothing to lose in her maiden Grand Slam final, but Williams is the ultimate closer, winning 20 of her 24 career finals. A sixth Wimbledon title and another big step toward another "Serena Slam" awaits for Williams.
{ "pile_set_name": "OpenWebText2" }
A helicopter pilot flying supplies to hurricane-devastated areas of the Bahamas found villagers stranded in rubble on one of the islands. NBC News reported that Justin Johnson— who owns Timberview Helicopters in Destin, Florida— was on his way to Fox Town on Little Abaco Island with his wife and a reporter. During the flight, the reporter, Vic Micolucci, pointed to a debris field and asked Johnson if anyone was down there. Johnson initially said he did not think so, but he returned the next day to find 30 villagers stranded in the rubble in the aftermath of Hurricane Dorian. “He had this intuition that he had to fly back over,” Angela Johnson, Justin’s wife, said. When Johnson landed at the site, he noticed people crawling out of the debris around him. The team immediately began handing over supplies and brought in more. MEDIC Corps, the relief organization the Johnsons had been working with, said the village was initially overlooked because it was off the main roads and many of the residents did not speak English and many did not have vehicles. The organization added that most of those rescued were Haitians who illegally migrated to the Bahamas and were afraid of being deported. They “are afraid of being deported so it is sometimes difficult to provide evacuation to these communities and they aren’t the first to show themselves,” MEDIC Corps said. The Abaco Islands received the brunt of Dorian’s impact, as it was a Category 5 storm when it devastated the area on September 1.
{ "pile_set_name": "OpenWebText2" }
![](hosplond72873-0017){#sp1 .39}
{ "pile_set_name": "PubMed Central" }
SELECT * FROM USERS INNER JOIN KNOWLEDGE_EDIT_USERS ON USERS.USER_ID = KNOWLEDGE_EDIT_USERS.USER_ID WHERE KNOWLEDGE_EDIT_USERS.KNOWLEDGE_ID = ? ORDER BY USERS.USER_ID
{ "pile_set_name": "Github" }
Q: Return a list without using NHibernate object Suppose I have an interface. public interface IBlogRepository { IList<Blog> Blogs(int pageNo, int pageSize); int TotalPosts(); } Now I created a class to implement it and use NHibernate. using NHibernate; using NHibernate.Criterion; using NHibernate.Linq; using NHibernate.Transform; using System.Collections.Generic; using System.Linq; namespace JustBlog.Core { public class BlogRepository: IBlogRepository { // NHibernate object private readonly ISession _session; public BlogRepository(ISession session) { _session = session; } public IList<Post> Posts(int pageNo, int pageSize) { var query = _session.Query<Post>() .Where(p => p.Published) .OrderByDescending(p => p.PostedOn) .Skip(pageNo * pageSize) .Take(pageSize) .Fetch(p => p.Category); query.FetchMany(p => p.Tags).ToFuture(); return query.ToFuture().ToList(); } public int TotalPosts() { return _session.Query<Post>().Where(p => p.Published).Count(); } } The above code is from somewhere on the web for creating a blog engine. However I don't know NHibernate at all, I use Entity Framework to do my job. How can I rewrite the code without using NHiberate? A: Entity Models Assuming we have our entity models like this: public class Post { public Post() { Tags = new List<Tag>(); } public int Id{ get; set; } public string Title{ get; set; } public string ShortDescription{ get; set; } public string Description{ get; set; } public string Meta{ get; set; } public string UrlSlug{ get; set; } public bool Published{ get; set; } public DateTime PostedOn{ get; set; } public DateTime? Modified{ get; set; } public int CategoryId { get; set; } public virtual Category Category{ get; set; } public virtual IList<Tag> Tags{ get; set; } } public class Tag { public int Id { get; set; } public string Name { get; set; } public string UrlSlug { get; set; } public string Description { get; set; } public virtual IList<Post> Posts { get; set; } } public class Category { public int Id { get; set; } public string Name { get; set; } public string UrlSlug { get; set; } public string Description { get; set; } public virtual IList<Post> Posts { get; set; } } Context Our context class is pretty simple. The constructor takes in the name of the connection string in web.config and we define three DbSets: public class BlogContext : DbContext { public BlogContext() : base("BlogContextConnectionStringName") { } public DbSet<Category> Categories { get; set; } public DbSet<Post> Posts { get; set; } public DbSet<Tag> Tags { get; set; } protected override void OnModelCreating(DbModelBuilder modelBuilder) { base.OnModelCreating(modelBuilder); } } Repository The interface for our blog repository is nice and easy and doesn't change much: public interface IBlogRepository { IEnumerable<Post> Posts(int pageNo, int pageSize); int TotalPosts(); } And it's the blog repository itself where things get interesting! public class BlogRepository : IBlogRepository { // NHibernate object replace with our context private readonly BlogContext _blogContext; public BlogRepository(BlogContext blogContext) { _blogContext = blogContext; } //Function to get a list of blogs public IEnumerable<Post> Posts(int pageNo, int pageSize) { //We start with the blogs db set: var query = _blogContext.Posts //Filter by Published=true .Where(p => p.Published) //Order by date they were posted .OrderByDescending(p => p.PostedOn) //Jump through the list .Skip(pageNo * pageSize) //Get the required number of blogs .Take(pageSize) //Make sure the query include all the categories .Include(b => b.Category); //Just return what we have! return query; } //Much simpler function, should be pretty self explanatory public int TotalPosts() { return _blogContext.Posts.Where(p => p.Published).Count(); } } Next Steps OK, so now we have all that set up, and a nice connection string is set up in your web.config, what do we do with it? Well let's get some blogs! var context = new BlogContext(); var repository = new BlogRepository(context); var posts = repository.Posts(0, 10); Then we can do some stuff with those blogs: foreach(var blog in blogs) { Console.WriteLine("Blog Id {0} was posted on {1} and has {2} categories", blog.Id, blog.PostedOn, blog.Categories.Count()); } Notes I didn't implement the FetchMany/ToFuture parts as they're not needed here.
{ "pile_set_name": "StackExchange" }
Q: How can I correctly output an array of objects in the reverse order from mongodb? Comments are saved in an array of objects. How can I correctly output them in reverse order (comments from newest to oldest)? My db: {"_id":{"$oid":"5e3032f14b82d14604e7cfb7"}, "videoId":"zX6bZbsZ5sU", "message":[ {"_id":{"$oid":"5e3032f14b82d14604e7cfb8"}, "user":{"$oid":"5e2571ba388ea01bcc26bc96"},"text":"1" }, {"_id":{"$oid":"5e3032f14b82d14604e7cfb9"}, "user":{"$oid":"5e2571ba388ea01bcc26bc96"},"text":"2" }, .... ] My sheme Mongoose: const schema = new Schema({ videoId: { type: String, isRequired: true }, message: [ { user: { type: Schema.Types.ObjectId, ref: 'User' }, text: { type: String } }, ] }); My code: const userComments = await Comment.find( { videoId: req.query.videoId }, { message: { $slice: [skip * SIZE_COMMENT, SIZE_COMMENT] } } ) .sort({ message: -1 }) .populate('message.user', ['avatar', 'firstName']); but sort not working; thanks in advance! A: You can simply use $reverseArray to reverse content of an array. db.collection.aggregate([ { $addFields: { message: { $reverseArray: "$message" } } } ]) Collection Data : /* 1 */ { "_id" : ObjectId("5e3032f14b82d14604e7cfb7"), "videoId" : "zX6bZbsZ5sU", "message" : [ { "_id" : ObjectId("5e3032f14b82d14604e7cfb8"), "user" : ObjectId("5e2571ba388ea01bcc26bc96"), "text" : "1" }, { "_id" : ObjectId("5e3032f14b82d14604e7cfb9"), "user" : ObjectId("5e2571ba388ea01bcc26bc96"), "text" : "2" } ] } /* 2 */ { "_id" : ObjectId("5e309318d02e05b694b0b25f"), "videoId" : "zX6bZbsZ5sUNEWWWW", "message" : [ { "_id" : ObjectId("5e3032f14b82d14604e7cfc9"), "user" : ObjectId("5e2571ba388ea01bcc26bc87"), "text" : "Old" }, { "_id" : ObjectId("5e3032f14b82d14604e7cfd0"), "user" : ObjectId("5e2571ba388ea01bcc26bc87"), "text" : "New" } ] } Result : /* 1 */ { "_id" : ObjectId("5e3032f14b82d14604e7cfb7"), "videoId" : "zX6bZbsZ5sU", "message" : [ { "_id" : ObjectId("5e3032f14b82d14604e7cfb9"), "user" : ObjectId("5e2571ba388ea01bcc26bc96"), "text" : "2" }, { "_id" : ObjectId("5e3032f14b82d14604e7cfb8"), "user" : ObjectId("5e2571ba388ea01bcc26bc96"), "text" : "1" } ] } /* 2 */ { "_id" : ObjectId("5e309318d02e05b694b0b25f"), "videoId" : "zX6bZbsZ5sUNEWWWW", "message" : [ { "_id" : ObjectId("5e3032f14b82d14604e7cfd0"), "user" : ObjectId("5e2571ba388ea01bcc26bc87"), "text" : "New" }, { "_id" : ObjectId("5e3032f14b82d14604e7cfc9"), "user" : ObjectId("5e2571ba388ea01bcc26bc87"), "text" : "Old" } ] } Your Query : You can use native MongoDB's $lookup instead of .populate() , So try below : Comments.aggregate([ { $addFields: { message: { $reverseArray: "$message" } } }, { $lookup: { from: "User", let: { ids: "$message.user" }, pipeline: [ { $match: { $expr: { $in: ["$_id", "$$ids"] } } }, { $project: { avatar: 1, firstName: 1, _id: 0 } } ], as: "userData" } } ])
{ "pile_set_name": "StackExchange" }
P53 codon 72 polymorphism and lung cancer risk: evidence from 27,958 subjects. The role of p53 codon 72 polymorphism in the development of lung cancer remains obscure due to inconsistent findings of individual case-control studies published to date. A meta-analysis was conducted to better estimate the association between the p53 codon 72 variant and lung cancer risk. All relevant publications from the PubMed, Embase, Web of Science, and Wanfang databases were retrieved. Based on the inclusion criteria, 39 publications involving 44 independent case-control studies were finally included into this meta-analysis. Data were extracted and the pooled odds ratio (OR) with the corresponding 95 % confidence interval (95 % CI) was calculated. The overall pooled ORs showed no significant relationship of the p53 codon 72 polymorphism with increased or decreased risk of lung cancer in all gene contrast models (OR Pro vs. Arg = 1.04, 95 % CI = 0.96-1.13, P OR < 0.001; OR Pro/Pro vs. Arg/Arg = 1.07, 95 % CI = 0.91-1.25, P OR < 0.001; OR Arg/Pro vs. Arg/Arg =1.04, 95 % CI = 0.94-1.15, P OR < 0.001; OR Pro/Pro + Arg/Pro vs. Arg/Arg = 1.04, 95 % CI = 0.94-1.16, P OR < 0.001; OR Pro/Pro vs. Arg/Arg + Arg/Pro = 1.07, 95 % CI = 0.93-1.23, P OR < 0.001). According to the ethnicity, no significant association was observed in subgroup analyses of the Asians, Caucasians, Africans and the mixed population. Similar finding was found in subgroup analyses of hospital-based and population-based studies. Concerning the histological types of lung cancer, the p53 codon 72 variant exerts risk effect on the lung carcinogenesis in patients with adenocarcinoma (OR Arg/Pro vs. Arg/Arg = 1.10, 95 % CI = 1.00-1.22, P OR = 0.048). Additionally, subgroup analysis by the smoking status demonstrated that the p53 codon 72 variant seemed to play a protective role in lung carcinogenesis among the non-smokers but not the smokers in the contrast model of Arg/Pro vs. Arg/Arg (OR Arg/Pro vs. Arg/Arg = 0.71, 95 % CI = 0.50-1.00, P OR = 0.049). The present meta-analysis suggests the p53 codon 72 polymorphism may weakly modify the risk for lung cancer among the adenocarcinoma patients and non-smokers. Nevertheless, this association needs further confirmation in future studies with high quality.
{ "pile_set_name": "PubMed Abstracts" }
Q: How do I find out how many arguments are being passed I am writing some C code on Linux. I want to loop through the *argv[] parameter being passed to main, but I don't know how to stop before getting a seg fault. i = 0; while (i < sizeof(argv)) { printf("%s\n", argv[i]); i ++; } This produces a seg fault because sizeof always returns 8. How do I get the actual number of elements in argv, or apply some kind of test to stop at the end of argv? A: The first argument of main is argc which is the number of arguments passed to your program. This is at least 1, which is the name of the executable. #include <stdio.h> int main(int argc, char *argv[]) { printf("%d\n", argc); return 0; } Invoked as: $ ./a.out 1 $ ./a.out 1 2 3 4 5 $ ./a.out 1 2 3 4 A: The first argument.. argc should already have the CLI argument count.. Any reason why that isn't used? int main(int argc, char** argv) { if(argc > 1) }
{ "pile_set_name": "StackExchange" }
Q: rails select records which have duplicates For my case i want to select all duplicates records from table for example table1 id | name | sub_name 1 | joe | j1 2 | tim | t1 3 | ben | b1 4 | joe | j2 5 | tim | t2 i vant to select [#<table1 id: 1, name: "joe", sub_name: "j1">, #<table1 id: 4, name: "joe", sub_name: "j2">, #<table1 id: 2, name: "tim", sub_name: "t1">, #<table1 id: 5, name: "tim", sub_name: "t2">] and then display joe - j1 - j2 tim - t1 - t2 Can anybody help me to do this using AR or SQL. i tried this query Table1.group(:name).having("count(*) > 1" but the result was [#<table1 id: 1, name: "joe", sub_name: "j1">, #<table1 id: 2, name: "tim", sub_name: "t1">] or result.count returns {["joe]=>2, ["tim"]=>2} A: Well I don't know if this is the most efficient way but you could: names_with_multiple_rows = Table1.group(:name).having("count(*) > 1") res = names_with_multiple_rows.inject({}) do |h, m| h[m.name] = Table1.select(:sub_name).where(name: m.name) h end Now res is a hash where the keys is the name and the values are the sub_names.
{ "pile_set_name": "StackExchange" }
Antioxidant activity of Crataegus aronia aqueous extract used in traditional Arab medicine in Israel. The medicinal use of extracts prepared from plant parts of the genus Crataegus dates back to ancient times. Furthermore, it has been proposed that its antioxidant constituents account for its beneficial therapeutic effects. A decoction of leaves and unripe fruits from Crataegus aronia syn. azarolus (L) (Rosaceae), the indigenous Israeli hawthorn, is used to treat cardiovascular diseases, cancer, diabetes and sexual weakness in Arab traditional medicine. Because laboratory data on the bioactivity of extracts prepared from the indigenous Israeli hawthorn is lacking, we evaluated the antioxidant and cytotoxic potentials of an extract prepared from leaves and unripe fruits in a variety of cell and cell-free in vitro assays. The antioxidant assays measured: (a) its ability to inhibit (i) oxidation of beta-carotene, (ii) 2,2'-azobis(2-amidino-propan) dihydrochloride (AAPH)-induced plasma oxidation and (iii) iron-induced lipid peroxidation in rat liver homogenates; (b) its ability to scavenge the superoxide (O2-) radical; (c) its effects on the enzyme xanthine oxidase (XO) activity; (d) its effect on the redox state of glutathione (GSH) in cultured Hep G2 cells. In addition, we also evaluated the effects of the extract on cell membrane integrity and mitochondrial respiration in cultured Hep G2 cells. Water-soluble extracts inhibited (1) oxidation of beta-carotene, (2) AAPH-induced plasma oxidation and (3) Fe(2+)-induced lipid peroxidation in rat liver homogenates. In addition, the extract (4) is an efficient scavenger of the O2- (5) increases intracellular GSH levels and (6) is not cytotoxic. Accordingly, we propose that the therapeutic benefit of Crataegus aronia can be, at least in part, attributed to its effective inhibition of oxidative processes, efficient scavenging of O2- and possible increasing GSH biosynthesis.
{ "pile_set_name": "PubMed Abstracts" }
Sequenom Sequenom () is an American company based in San Diego, California. It develops enabling molecular technologies, and highly sensitive laboratory genetic tests for NIPT. Sequenom's wholly owned subsidiarity, Sequenom Center for Molecular Medicine (SCMM), offers multiple clinical molecular genetics tests to patients, including MaterniT21, plus a noninvasive prenatal test for trisomy 21, trisomy 18, and trisomy 13, and the SensiGene RHD Fetal RHD genotyping test. In June 2014 the company sold its biosciences unit to Agena Bioscience for up to $35.8 million. In July 2016, it was announced that diagnostic and testing giant LabCorp will acquire Sequenom, paying $2.40 for every outstanding share of Sequenom stock. The acquisition was completed in September 2016. Competition Companies also offering non-invasive prenatal genetic testing include Ariosa, Ravgen, Illumina (Verinata Health),, PerkinElmer and Natera (The Panorama Prenatal Test),. Other companies and universities that are working towards developing non-invasive prenatal testing include Stanford University,. Patent litigation In January 2012, Sequenom entered a patent battle with competing companies, Ariosa and Natera, accusing them of infringing the "540 patent" (). The cases are Sequenom Inc. v. Natera Inc. 12-cv-0184, Sequenom v. Ariosa Diagnostics Inc., 12-cv-0189, U.S. District Court, Southern District of California (San Diego), and Ariosa v. Sequenom. Verinatal Health and Stanford University later filed suit against Sequenom in a dispute over the 'Quake patent'. Verinata claims that Sequenom's lawyers sent it a letter in 2010 alleging that "'the practice of non-invasive prenatal diagnostics, including diagnosis of the Down Syndrome and other genetic disorders, using cell-free nucleic acids in a sample of maternal blood infringes' the '540 patent, as well as the claims of a pending United States Patent Application." The '540 patent was invented by Isis Ltd. and expires in 2017. Stanford University owns the Quake patents and licensing rights; Verinata is its exclusive licensee. In April 2012, Sequenom acquired two pending patents from Helicos Biosciences. In consideration for the sale and transfer of the purchased assets, Sequenom paid Helicos $1.3 million. The Helicos patent applications (US Patent application 12/709,057 and 12/727,824) cover methods for detecting fetal nucleic acids and diagnosing fetal abnormalities. In July 2012, The United States District Court denied Sequenom's motion for a preliminary injunction motion against Ariosa Diagnostics. In August 2013, The Court of Appeals for the Federal Circuit vacated the District Court decision and remanded that case to the District Court. In the Ariosa litigation, the District Court (N.D.Cal.) held that the '540 patent was invalid because it claimed a natural phenomenon, the presence of cell-free fetal DNA fragments in maternal blood. On June 13, 2015, the CAFC affirmed the District Court's judgment. Finally on December 2, 2015, the Federal Circuit declined to rehear en banc. SEQureDx scandal In 2009, Sequenom Center for Molecular Medicine (SCMM) was expected to launch the SEQureDx prenatal screening tests for Down syndrome and Rhesus D. Subsequent investigation revealed significant flaws in the studies of the test's effectiveness. As a result, the board of directors of Sequenom fired CEO Harry Stylli, senior vice president of research and development Elizabeth Dragon and three other employees after a probe discovered that the company had failed to adequately supervise its Down syndrome test. CFO Paul Hawran also resigned. Board chairman Harry F. Hixson Jr. was named interim CEO and director Ronald M. Lindsay was appointed to replace Dragon. Dragon has since been charged by the Securities and Exchange Commission (SEC) because she "lied to the public about the accuracy of Sequenom's prenatal screening test for Down syndrome". She died on February 26, 2011. In 2010, Sequenom paid $14 million to settle a shareholder class-action lawsuit that arose from the errors in the development of the Down syndrome test. Sequenom executives are under investigation by the SEC for insider trading before announcement of problems with the test. On September 1, 2011 Sequenom entered into a cease-and-desist order with SEC. MaterniT21 PLUS MaterniT21 PLUS is Sequenom Center for Molecular Medicine's prenatal test for trisomy 21 (Down syndrome), trisomy 18 (Edwards syndrome) and trisomy 13 (Patau syndrome). The test operates by sampling cell-free DNA in the mother's blood, which contains some DNA from the fetus. The proportions of DNA from sequences from chromosome 21, 18, or 13 can indicate whether the fetus has trisomy in that chromosome. In a randomized controlled trial of 1,696 pregnancies at high risk for Down syndrome, the test correctly identified 98.6% of the actual cases of Down syndrome (209 out of 212), with a false positive rate of 0.2% (3 of 1471 pregnancies without Down); the test gave no result in 0.8% of the cases tested (13 of 1696). The primary advantage of MaterniT21 PLUS over the other major high accuracy tests for Down syndrome, Amniocentesis and Chorionic villus sampling, is that MaterniT21 PLUS is noninvasive. Because amniocentesis and chorionic villus sampling are invasive, they have a chance of causing miscarriage. History On August 4, 2011, Sequenom said it would call its new blood test for Down syndrome in pregnancy MaterniT21 when the product went on sale in the United States. On August 11, 2011, Sequenom announced European licensing agreement with LifeCodexx. The companies agreed to collaborate in the development and launch of a trisomy 21 laboratory-developed test and other aneuploidies testing in Germany, Austria, Switzerland, and Liechtenstein, with the potential for additional launches in other countries. Under the initial five year licensing agreement, Sequenom granted LifeCodexx licenses to key patent rights, including European Patent EP0994963B1 and pending application EP2183693A1 that enable the development and commercialization of a non-invasive aneuploidy test utilizing circulating cell-free fetal DNA in maternal plasma. On October 24, 2011 International Society of Prenatal Diagnostics (ISPD) issued a rapid response statement in response to the launch of Sequenom non-invasive Trisomy 21 (MaterniT21) test. On October 17, 2011 Sequenom announced that a clinical validation study leading to the introduction of the MaterniT21 LDT had been published in the journal Genetics in Medicine. On October 17, 2011 Sequenom Center for Molecular Medicine announced the launch of MaterniT21 Noninvasive Prenatal Test for Down Syndrome. MassARRAY Analyzer 4 Sequenom Oncomap Version 3 – "core" set interrogates ~450 mutations in 35 genes. An "extended" set interrogates ~700 mutations in 113 genes. Sequenom OncoCarta(OncoMap) identifies 396 unique "druggable" or "actionable" mutations in 33 cancer genes. In total, 417 mutations are identified. MassARRAY spectrometry is more sensitive than PreTect HPV-Proofer and Consensus PCR for type-specific detection of high-risk oncogenic human papillomavirus genotypes in cervical cancer. iPLEX ADME PGx Panel on MassARRAY System On October 4, 2011 Sequenom introduced iPLEX ADME PGx Panel on MassARRAY System, developed to genotype polymorphisms in genes associated with drug absorption, distribution, metabolism, and excretion (ADME). This Research Use Only (RUO) panel contains a set of pre-designed single nucleotide polymorphisms (SNP), insertions and deletions (INDELS) and copy number variation (CNV) assays for use in the investigation of variants with demonstrated relevance to drug metabolism. After detection on the MassARRAY (RUO) system, a proprietary software solution is then used to score and qualify polymorphisms to create a unique haplotype report. References External links Sequenom's Company Website Category:Biotechnology companies of the United States Category:Genomics companies Category:Microarrays Category:Companies listed on NASDAQ
{ "pile_set_name": "Wikipedia (en)" }
I owe it to my father to argue for death with dignity It started, as far as I know, with a twinge. A back spasm in a healthy, but prone to back spasms, 68-year-old man, who soon was reduced to near tears in the parking lot of the Thai restaurant where we’d gone to get take out. It was early summer, and I was on a short visit to the east coast. My mother was the one to worry about. She had advanced breast cancer, “stage IV” as my dad frequently called it, in a way that he felt made it easier to convey to the world the seriousness of the situation but never allowed my mother a moment of peace to forget her mortality. In those days, my dad thought only of the transience of her life, never his. He was a paragon of health, a man who had found running in his 30s and started a lifelong daily habit on the rural roads of Pennsylvania. Independent, he chopped his own wood and heated their home solely with it, scouring the landscape with a probing eye for downed trees, and driving his pickup and chainsaw to the site to claim the wood as if it was a treasure. He eschewed doctors, despite that I was one, and that his own background was one of science. As he used to say, if he didn’t outlive his own father, who died in his early 90s after a life of hard living, the world was unjust. Over the next month, the spasms became more frequent. I, in California midway through my pediatric anesthesia fellowship and weary of years of long distance medical concerns from my mother’s prolonged illness, received weekly updates via phone call but was too wrapped up in my own life to put it together. “I’m interviewing for jobs, Mom — finally the end of ten years of medical training!” “Your father has lost that 30 pounds he always wanted to!” “He’s hot all the time — I just bought new sheets so we can change them more frequently — they are damp every day.” But the pain was what he was focused on. He finally went to a doctor, because of the severe pain that left him breathless and screaming on the floor. Acute myelofibrosis was the final diagnosis. In a period of mere months, his bone marrow had turned to useless fiber. This is very rare, said the specialty Philadelphia teaching hospital where I finally brought him, desperate for answers. How long does he have? Maybe six months. The pain, though, no one could understand. It wracked him, left him screaming and shaking. As an anesthesiologist, I was accustomed to witnessing severe pain, then standing at my patient’s bedside pushing medicine until it resolved. Of his pain, I could do nothing other than to explain to physician after physician: “Until two months ago, he never even took Tylenol; he has an extremely high pain tolerance.” They shook their heads in disbelief: We don’t think this disorder causes pain. Yet it did. They treated his pain as best they could, but it was never enough. Of death, he knew, but he didn’t know. A lifelong hunter who’d grown up on a farm, he was accustomed to the cycle of life and death, yet always colored by the element of the one with control. It is difficult to describe retrospectively the stages of disbelief, rage, and acceptance he went through. Before he came back from the hospital for what would be his final time, my husband and I cleared the house of his entire gun collection. We feared my father would do something desperate; throughout his life, he would say if he ever had a serious illness he would just take his hunting rifle and end it. It seemed so wrong; society says suicide, giving up is wrong, and guns are violence. So, my husband and I acted. But now I understand this man was terrified that he had run out of options, that I had taken away what he saw as his only alternative. “Why?” He moaned over the phone after I’d left, back safely in my fellowship in California. “I would have just gone in the woods and taken care of it. I’m scared.” I started to understand. “I’ll help you, Dad. But I need to help you in a way that is legal. I’ll take a leave from fellowship, and we’ll move to a state with physician-assisted suicide, and establish residency, and they’ll help you, Dad.” But that was a pipe dream. The closest state was Vermont, at least an eight-hour drive with a man screaming in constant severe pain and a requirement to first establish residency. No, no, he said, I’ll find another way. My mother, perhaps because she was there and living with the consequences of his disease, or perhaps because the stigma of suicide ran strong in our small Pennsylvania community, focused on preventing his desired death. In the end, he died in hospice care in a hospital on Thanksgiving day, in a morphine induced haze from which he would occasionally awaken and scream. Home was not an option; no hospice provider serviced IV opiate medication in our rural area of Pennsylvania. It was his worst nightmare come true: a diapered hell. On the day before he died, he had one moment of clarity, like when you awaken from a bad dream and are trying to figure out if it is real or not. “Chrissy, am I OK?” “Dad, I’m here.” “No, I’m not OK! …” dissolved into the moans of severe pain, soon followed by the delirium of more morphine. I am haunted by the end of my father’s life, by the lack of control he felt, by his terror, and by the role I was forced into. I witnessed what I knew to be his worst fear, for someone as independent as he: “I never want to die in my own filth in a hospital somewhere.” Yet there was no option. I had taken away the only other real option for someone as incapacitated as he, to die by his own hand with a gun in his beloved woods. Yet why, in his home state, did society dictate my dad had but two, equally inhumane and unacceptable options? Why was a physician-assisted dignified death, with a semblance of choice and control in a situation that was otherwise so helpless, not available to my father? Physician-assisted suicide is currently available to terminally ill patients in four states. Nothing can change the death of my father. But especially as a physician, I owe it to him to argue for death with dignity, to advocate for patients to have the option he did not have, so other daughters do not have to feel the conflicted helplessness I can’t forget.
{ "pile_set_name": "Pile-CC" }
Induction of anti-anti-idiotype antibodies against sulfated glycosaminoglycans reduces atherosclerosis in apolipoprotein E-deficient mice. The pathogenesis of atherosclerosis is associated with the early retention of low-density lipoproteins that are trapped in the extracellular matrix of the arterial intima by interaction with glycosaminoglycan side chains of proteoglycans. Mutant mouse/human chimeric antibodies of the murine monoclonal antibody P3, which react with N-glycolyl-containing gangliosides and sulfated glycosaminoglycans, were tested for their potentially antiatherogenic properties through the induction of an idiotypic antibody network that may specifically interfere with the binding of low-density lipoproteins to proteoglycan side chains, low-density lipoprotein modification, and foam cell formation. Apolipoprotein E-deficient mice fed a high-fat, high-cholesterol diet received 5 to 6 doses of chP3R99 or chP3S98 mutant antibodies, showing high and low reactivity, respectively, against their respective antigens. Both chimeric antibodies elicited an immunodominant anti-idiotypic response in the absence of adjuvant. A striking (40%-43%) reduction (P<0.01) in total lesion areas was observed in 18-week-old mice immunized with chP3R99, but not chP3S98, compared with PBS-treated mice. The antiatherosclerotic effect was associated with increased mice sera reactivity against heparin and sulfated glycosaminoglycans, including chondroitin and dermatan sulfate. In addition, purified IgG from chP3R99-immunized mice blocked the retention of apolipoprotein B-containing lipoproteins within the arterial wall of apolipoprotein E(-/-) mice. The present study supports use of active immunization and the mounting of an idiotypic antibody network response against glycosaminoglycans as a novel approach to target atherosclerosis.
{ "pile_set_name": "PubMed Abstracts" }
Nbritton91 (8:04:40 PM) Hello, i have been contacted by a member of my clan Nbritton91 (8:04:48 PM) He is claiming harrassment by you manimal1 (8:04:49 PM) And SO Nbritton91 (8:04:53 PM) Do i need to notify wargaming? manimal1 (8:05:10 PM) If your member was manimal1 (8:05:23 PM) being talked to by me manimal1 (8:05:40 PM) HE IS A COWARD that RAN from A BATTLE Nbritton91 (8:05:54 PM) I see you are upset Nbritton91 (8:06:06 PM) But you must understand this is simply a game Nbritton91 (8:06:20 PM) and that verbal abuse, such as this, is not the correct solution to your anger Nbritton91 (8:06:36 PM) have you condsidered an anger therapy course? manimal1 (8:06:38 PM) I get UPSET when TEAMMATES run or hiode from a BATTLE manimal1 (8:06:46 PM) HIDE * Nbritton91 (8:07:01 PM) Is there something wrong with your shift key? manimal1 (8:07:23 PM) Dude who the hell u think u are manimal1 (8:07:56 PM) 8225 gaMES YOUR A SCRUB Nbritton91 (8:08:10 PM) I see manimal1 (8:08:20 PM) u see what Nbritton91 (8:08:35 PM) I see you are upset Nbritton91 (8:08:42 PM) why are you so mad? manimal1 (8:08:45 PM) LOL Nbritton91 (8:08:54 PM) I have done nothing to provoke you Nbritton91 (8:09:01 PM) Why are you insulting my statistics manimal1 (8:09:03 PM) Did you not understand my 1st reply Nbritton91 (8:09:16 PM) No i didn't Nbritton91 (8:09:22 PM) please repeat your 1st reply manimal1 (8:09:28 PM) OK LETS START OVER Nbritton91 (8:09:29 PM) I would be most appreciative Nbritton91 (8:09:31 PM) fantastic manimal1 (8:09:47 PM) Did you OR DID YOU NOT talk to me 1st Nbritton91 (8:10:18 PM) I contacted you yes, to ask you what the problem is Nbritton91 (8:10:26 PM) i do not understand the problem as you have described it Nbritton91 (8:10:33 PM) please re-explain the situation for clarity manimal1 (8:10:59 PM) I said your member was a coward and hid in a BATTLE Nbritton91 (8:11:04 PM) I am trying to understand why you are verbally assaulting a fellow clan-mate of mine Nbritton91 (8:11:13 PM) Is there a difference between a battle and a BATTLE? Nbritton91 (8:11:18 PM) for clarity manimal1 (8:11:19 PM) BRB Nbritton91 (8:11:38 PM) I'll wait manimal1 (8:13:56 PM) Difference between a battle and a battle ARE YOU IGNORANT Nbritton91 (8:14:31 PM) No i am not Nbritton91 (8:14:36 PM) I am niether a scrub Nbritton91 (8:14:39 PM) nor ignorant manimal1 (8:14:40 PM) DO YOU PLAY TO WIN Nbritton91 (8:14:44 PM) Yes Nbritton91 (8:14:45 PM) Yes i do manimal1 (8:15:21 PM) Then WHY do you let PEOPLE in your clan RECRUITER that HIDES Nbritton91 (8:15:47 PM) I am so confused. Let me explain why i contacted you. Nbritton91 (8:15:52 PM) I contacted you to give you a chance Nbritton91 (8:15:55 PM) I was being nice Nbritton91 (8:16:02 PM) I wanted you to provide me with a reason manimal1 (8:16:02 PM) HAVE you EVER spent HARD EARNED MONEY ON THIS GAME manimal1 (8:16:20 PM) Do u understand my reason Nbritton91 (8:16:21 PM) I wanted you to provide me with a reason NOT to report you Nbritton91 (8:16:26 PM) to wargaming manimal1 (8:16:36 PM) Dude. Nbritton91 (8:16:54 PM) I'm afraid i will be contacting them regarding your unaccaptable behavior manimal1 (8:17:19 PM) REPORT ME IF YOU want. I never call out anhyone that BATTLES TO WIN manimal1 (8:17:58 PM) I PLAQY TO WQIN . COWARDS need to BE REPORTED report your MEMBER Nbritton91 (8:18:04 PM) Define play to win Nbritton91 (8:18:17 PM) Provide me with proof that you "play to win" manimal1 (8:18:21 PM) NEVER RUN from a fight manimal1 (8:19:05 PM) <<<< Always 1st to a fight. MY STATS prove it Nbritton91 (8:19:19 PM) Yes they do i'm afraid manimal1 (8:19:28 PM) AND NEVER EVER HIDE ...... LIKE YOUR BOY DID Nbritton91 (8:19:36 PM) My condolences in regards to the statistics manimal1 (8:19:55 PM) 1 on 1` ME and YOU RIGHT NOW manimal1 (8:20:26 PM) Your not that good DUDE Nbritton91 (8:21:22 PM) Very well manimal1 (8:21:33 PM) NO CAP Nbritton91 (8:21:41 PM) Invite me to a training room sir manimal1 (8:21:49 PM) Invite . What tier TUFF GUY Nbritton91 (8:22:04 PM) Whatever tier you like Nbritton91 (8:22:09 PM) I would prefer 10 manimal1 (8:22:22 PM) Ok t110s Nbritton91 (8:22:55 PM) Very well, i await the invite manimal1 (8:23:29 PM) I SAID IUNVITE ME manimal1 (8:24:03 PM) How old are u Nbritton91 (8:24:43 PM) I'm sorry personal info is against wargaming code of conduct Nbritton91 (8:24:46 PM) I sent you an invite manimal1 (8:25:32 PM) LOL your PATHETIC is BEING PATHETIC agaist WARGAMING conduct to Nbritton91 (8:25:51 PM) Accept the invite please, i have a clan wars battle to attend in the near future Nbritton91 (8:26:38 PM) I have sent you three invites now Nbritton91 (8:26:54 PM) If you do not accept i will be forced to cancel our plans manimal1 (8:26:55 PM) No your IGNORANT, Report me if you belive your COWARD team mate BUT YOU BEING IGNORED
{ "pile_set_name": "OpenWebText2" }
Congratulations to Boulder County Farmers’ Markets on being our Zero Waste Business of the Month! Did you know that in 2005, the Boulder Farmers’ Market became the nation’s first ongoing Zero Waste Event? Just three years later, the Longmont Farmers’ Market followed suit! What does it mean for a farmers’ market to go for Zero Waste? It’s essentially a microcosm of how a Zero Waste Community would work -- “rules” are in place to make sure all vendors are on board to make sure all the discards generated, like food containers and packaging, are designed for reuse, recycling or the compost bin. “Infrastructure,” in the form of Zero Waste stations with recycling and compost bins, are set up throughout the markets, often with Eco-Cycle volunteers making sure guidelines are followed and questions from the public are answered. These are just a few of the principles that make it easy for market visitors to ‘do the right thing’ and practice Zero Waste.
{ "pile_set_name": "Pile-CC" }
Tax Reform Update The Tax Reform Act of 2017 made meaningful change to the Federal tax code that has helped stimulate our country’s economy. It also created some confusion and unknowns. But your Home Builders Association and other associations have been working hard with the IRS to get you some answers and create the best outcome for your businesses. IRS to Waive 2018 Withholding Penalties for Most Filers The Internal Revenue Service has announced that it is waiving the tax penalty for many home builders and other small businesses that pay estimated quarterly taxes but whose 2018 federal income tax withholding and/or estimated tax payments fell short of their total tax liability for the year. Treasury Issues Final Rule on Pass Through Entities The Treasury Department has issued final regulations for the 20% pass through entity deduction under the Tax Cuts and Jobs Act of 2017. The final regulations concern the deduction for qualified business income under Section 199A of the Internal Revenue Code. This includes individuals, partnerships, S corporations, trusts, and estates engaged in domestic trades or businesses. (the following content sourced from the National Association of REALTORS)IRS Provides Clear Test on How 20% Deduction Applies to Rental Income, Exchanges The IRS has issued final rules on the 20 percent business income deduction that was part of the 2017 Tax Reform Act. The new rule confirms that the deduction applies to your business income, as a real estate agent or broker, if you operate as a sole proprietor or owner of a partnership, S corporation, or limited liability company. It applies even if your income exceeds a threshold set in the law of $157,500 for single filers and $315,000 for joint filers. In addition, the new rule provides guidance on two other provisions: 1) whether any real estate rental income is eligible for the deduction, and 2) how the deduction applies to properties exchanged under Section 1031. Eligibility of rental income Rental property income also can qualify for the new deduction, as long as you can show that your rental operation is part of a trade or business. The IRS has released proposed guidelines that include a bright-line test for showing that rental income rises to the level of a trade or business. Under that safe harbor, you can claim the deduction if your rental activities-which include maintaining and repairing property, collecting rent, paying expenses, and conducting other typical landlord activities-total at least 250 hours a year. If your activity totals less than that, you can still try to take the deduction, but you will have to be prepared to show the IRS that your activity is part of a trade or business. Eligibility of 1031 like-kind exchanges Under earlier proposed regulations, if your income was above threshold levels set in the tax law–$157,500 for single filers, $315,000, for joint filers–and you had exchanged one property for another to defer taxes under Sec. 1031, the amount of the new deduction might be reduced because of the swap. NAR and other trade groups reached out to the IRS to change this treatment, and the IRS has made the change. Under the final rules, you can use the unadjusted basis of the depreciable portion of the property to claim at least a partial deduction.
{ "pile_set_name": "Pile-CC" }
This project is designed to investigate long-term regulation in brain of the enzyme choline acetyltransferase (CAT), catalyzing the biosynthesis of the neurotransmitter, acetylcholine. We shall study the biochemical basis for the observed multiplicity of molecular species of CAT and the mechanisms governing the regulation of its activity. To accomplish this, we shall (a) prepare and purify antibodies to the different forms of CAT found in rat brain; (b) analyze regionally in the brain the molecular nature of the different forms; (c) characterize the biochemical properties of each form; (d) determine their turnover rates; and (e) establish whether the mechanism of altered CAT activity affected by psychoactive drugs, hormones or brain lesions, as well as during growth and development, is one of induction (increase in enzyme protein) or activation (increased catalytic activity/molecule).
{ "pile_set_name": "NIH ExPorter" }
Particle formation of budesonide from alcohol-modified subcritical water solutions. Recently, subcritical water (SBCW: water that has been heated to a temperature between 100°C and 200°C at pressures of up to 70bar) has been used to dissolve several hydrophobic pharmaceutical compounds (Carr et al., 2010a). Furthermore, a number of active pharmaceutical ingredients (APIs) have been rapidly precipitated from SBCW solutions (Carr et al., 2010b,c). It is possible to alter the precipitate morphology by altering the processing variables; including the SBCW-API solution injection temperature and adding impurities (such as pharmaceutical excipients, e.g. lactose) to the precipitation chamber. The work presented in this article demonstrates that the morphology of pharmaceutical particles can be tuned by adding organic solvents (ethanol and methanol) to the SBCW-API solutions. Particle morphology has also been tuned by adding different pharmaceutical excipients (polyethylene glycol 400 and lactose) to the precipitation chamber. Different morphologies of pharmaceutical particles were produced, ranging from nanospheres of 60nm diameter to 5μm plate particles. Budesonide was used as the model API in this study. Two experimental products were spray dried to form dry powder products. The aerodynamic particle size of the powder was established by running the powder through an Andersen Cascade Impactor. It has been shown that the drug particles produced from the SBCW micronization process, when coupled with a spray drying process, are suitable for delivery to the lungs.
{ "pile_set_name": "PubMed Abstracts" }
With over 15 years experience as a Therapist, I have worked with both adults and young people. If you are experiencing difficulties in your life, then my skills and experience will enable you to explore and discover potential solutions. Using Solution Focused Hypnotherapy, we focus on what you want to achieve rather than solely on the problem(s) that prompted you to seek change.? We'd be delighted to hear from you. Your 60 minute introductory session is free, including a free CD. Welcome to Coniston Hypnotherapy. My name is Laurita Tomlinson-Smith and I trained as a Hypnotherapist in 2012, after having a course of hypnotherapy myself.I now feel happier, have more self-esteem and feel more grounded and peaceful in my life. My... Search help For the most accurate results, please enter a full postcode into the search box so we can determine your exact location. If you type in a town name or the first part of a postcode the results will be ordered by distance from the centre of that town.
{ "pile_set_name": "Pile-CC" }
1. Field of the Invention The present invention relates to a method of manufacturing a semiconductor device including an inspection process of a semiconductor element that uses semiconductor characteristics (e.g., a transistor, in particular, a field-effect transistor; typically, a metal oxide semiconductor (MOS) transistor and a thin film transistor (TFT)). More specifically, the present invention relates to a non-contact type inspection apparatus and an inspection method using the same. The present invention also relates to a method of manufacturing a semiconductor device including an inspection process of such a semiconductor element. 2. Description of the Related Art In an active matrix type liquid crystal display and an electroluminescence (EL) display, a TFT is generally provided in each pixel. In the case of a liquid crystal display, one TFT formed in each pixel functions as a switching element. In the case of an EL display, among a plurality of TFT formed in each pixel, some TFTs function as switching elements, while others control a current. It is very effective in terms of reduction in cost that an inspection process of identifying a defective product is included in an early stage in the course of manufacturing a display in which a number of TFTs is are formed before a product is completed. The reasons for this are as follows: it is not required to conduct the succeeding steps with respect to a defective product; it is easy to repair a defective product due to early finding thereof, etc. For example, in an EL display, one electrode (pixel electrode) of an EL element and a capacitor may be connected to each other with a transistor formed therebetween. It is difficult to confirm the presence of a defect until an EL display is completed and a display is actually conducted, even if there is some inconvenience in circuits or circuit elements for controlling light emission of a light-emitting element. Regarding even an EL panel that will not actually become a product, in order to be distinguished from a satisfactory product, a light-emitting element is formed, packaging is conducted, and a connector is attached to complete an EL display, whereby an inspection is conducted to the EL display. In this case, the processes of forming a light-emitting element, packaging, and attaching a connector become useless, so that time and cost cannot be reduced. Furthermore, even in the case of forming an EL panel using a multiple-chamfered substrate, the processes of packaging and attaching a connector become useless, so that time and cost cannot be also reduced. In order to detect a portion where a malfunction is caused due to the variation in pattern width of a semiconductor film, an insulating film, or a wiring (hereinafter, merely referred to as a “pattern”) and a portion where wiring is disconnected or short-circuited due to dust or defective film formation, and to confirm if circuit or circuit elements to be inspected are operated normally, an inspection is conducted. Such a defect inspection is mainly classified into an optical inspection method and a probe inspection method. According to the optical inspection method, a pattern formed on a substrate is read by a CCD or the like, and the read pattern is compared with a reference pattern to identify a defect. According to the probe inspection method, minute pins (probes) are put up at terminals on a substrate side, and a defect is identified based on the magnitude of a current or a voltage between the probes. In general, the former method is called a non-contact type inspection method, and the latter is called a stylus type inspection method. A good TFT substrate that can be used for a product is discriminated from a defective TFT substrate that cannot be used for a product by the above-mentioned inspection method in which wiring is directly connected (contact) to a TFT substrate. However, according to this method, dust is likely to adhere to a substrate during attachment and removal of connection wiring. Furthermore, according to the method of detecting a defective portion by directly bringing minute pins (probes) into contact with wiring, wiring may be damaged. This inspection method may unnecessarily increase defective products during an inspection process. Also, according to the optical inspection method, a long time is required to inspect many times.
{ "pile_set_name": "USPTO Backgrounds" }
Q: How does one effectively negotiate rent? I'm going to rent an apartment in a major urban area in the next few months. When I drive around the area where I want to rent I see numerous "For Rent" signs. Also when I look online there are numerous listings. Given the abundance of properties for rent... and from a short conversation with a real estate agent... I've come to realize that negotiating the amount of monthly rent is not out of the question (and perhaps common?). How does one effectively negotiate rent? I'm open to a variety of ideas including trading a longer lease length for a smaller monthy rent... or paying multiple months in advance in exchange for a lower monthly rent etc. Cheers! A: One of the best ways to "win" any negotiation is to make an offer in which the other guy "wins" too. Give them something you don't care much about in exchange for something you do care about. (Assuming the "something" is of value to them.) You've already hinted at this strategy in your question: Prepay months in advance. Longer lease. Longer notice time built into lease termination agreement. Do you have a car? Can you give up a parking space? Are you a non-smoker? Can you promise not to smoke in the apartment? (This is an expense for the landlord when you move out.) Do they allow cats and you don't have cats? Can you promise not to keep cats? (Another expense for the landlord.) Probably unlikely to succeed in urban areas where landlords are pros, but in other areas: can you offer to do work for rent reduction? E.g. you are handy and can help with maintenance, or you can mow the lawn, shovel the walk, paint, etc. If they have low occupancy, can you make a deal for referrals? A: I'm open to a variety of ideas including trading a longer lease length for a smaller monthy rent... or paying multiple months in advance in exchange for a lower monthly rent etc. These two ideas that you mention are good. Do your homework first. See a few places. Find out how much they are asking. Make a note of details such as square footage, appliances, amenities, etc. Then play them off each other. Be sure to point out that there are many places available in the area. Be courteous, but firm and be ready to walk away from a place you like if you are not happy with the deal they are willing (or not willing) to give you. A: Another negotiating technique is to ask for extra items to be included at the base rental rate. Examples of "extras" in an apartment rental might include monthly rental of a washer/dryer, monthly fees for fitness center access, or higher rent for a good location apartment (top floor/pool side/etc). Ask to get one or all of those items included in the rental price they offer you. If they say "yes": You win the extra items at no additional cost. If they say "no": Then challenge them that if they take away the washer/dryer you'll need to pay $50 less rent per month to meet your target price for rent.
{ "pile_set_name": "StackExchange" }
Q: Understanding the shape of tensorflow placeholders I am reading this code and I would like to understand about its implementation. One of the first things that I would like to know, is that what is the shape of some tensor objects (placeholders) such as x_init, xs, h_init, y_init, y_sample, etc. I wrote a line of code such as print(xs.shape) but it wont work. How can I understand the shape of these parameters (tensors)? And can I write something like the following in NumPy? The part of code that defines these tensors look like this: x_init = tf.placeholder(tf.float32, shape=(args.init_batch_size,) + obs_shape) xs = [tf.placeholder(tf.float32, shape=(args.batch_size, ) + obs_shape) for i in range(args.nr_gpu)] # if the model is class-conditional we'll set up label placeholders + # one-hot encodings 'h' to condition on if args.class_conditional: num_labels = train_data.get_num_labels() y_init = tf.placeholder(tf.int32, shape=(args.init_batch_size,)) h_init = tf.one_hot(y_init, num_labels) y_sample = np.split( np.mod(np.arange(args.batch_size * args.nr_gpu), num_labels), args.nr_gpu) h_sample = [tf.one_hot(tf.Variable( y_sample[i], trainable=False), num_labels) for i in range(args.nr_gpu)] A: The shape is assembled from different command line parameters: obs_shape is the shape of the input images, e.g., (32, 32, 3) args.init_batch_size and args.batch_size are the values from command line. It could be for example 30 and 40. Then shape of x_init is the concatenation of init_batch_size and obs_shape: (30, 32, 32, 3). Correspondingly, the shape of each item in xs is (40, 32, 32, 3). You couldn't evaluate xs.shape, because xs is a list of placeholders. You can evaluate xs[0].shape instead. y_sample and h_sample are the lists of tensors as well. The first one contains (batch_size, num_labels) tensors, the second one (num_labels, ).
{ "pile_set_name": "StackExchange" }
Lambros Athanassoulas Lambros Athanassoulas (born November 10, 1976 in Athens) is a Greek rally driver. Career Athanassoulas graduated from the American College of Greece and got a BEng in Mechanical Engineering and an M.A. in Finance from the University of Westminster. He started racing in 2001, after winning a competition by Greek magazine 4 Wheels and Toyota Hellas. This earnt him the chance to drive a Toyota Yaris rally car. In 2005 he won the A5 class of the Greek Rally Championship driving a Ford Puma Kit Car. This led to him becoming an official driver for Ford Motor Hellas. This included making his World Rally Championship debut in a Ford Fiesta S1600 at the 2005 Acropolis Rally. In 2006 Athanassoulas competed in the Fiesta Sporting Trophy, as part of the WRC. In 2007 and 2008 he competed at his home event of the WRC in a Group N Subaru Impreza. For the 2009 Acropolis Rally, Athanassoulas competed in a Škoda Fabia S2000, winning the PWRC class and finishing eighth overall. He followed this by winning his class on the 2009 Rally Catalunya on the debut of the new Ford Fiesta R2. He was named Greek Athlete of the Year at the end of 2009 by Status magazine. WRC results PWRC results References External links Athanassoulas Rally Team Profile at eWRC-results.com Category:1976 births Category:Greek rally drivers Category:Living people Category:World Rally Championship drivers
{ "pile_set_name": "Wikipedia (en)" }
High plasma zinc levels following oral dosing in rats and the incorporation of 3H-thymidine into deoxyribonucleic acid in rat fetuses. In fasted rats, a single oral dose of zinc, equivalent to the total daily zinc intake, caused plasma zinc levels to rise about five-fold 1.5 h after treatment. Higher values were obtained following higher dose levels. At maternal plasma zinc levels approximately 10-fold above normal, fetal viability was disturbed and incorporation of 3H-thymidine into DNA became impaired in fetal (20-d) rat brain and liver tissue.
{ "pile_set_name": "PubMed Abstracts" }
College Admissions: 3 Myths About the SATs The SAT has changed dramatically over the last 5 years... are you up to date on the best strategies? As juniors wait anxiously for SAT scores, high school hallways are abuzz with poorly founded rumors and “strategies” for the SAT. The truth is that the SAT has undergone some significant changes over the last 5 years, making it a very different test from when many parents took it in the 1970s or 1980s. And these changes have spurred a lot of misinformation. So, here are some of the biggest myths debunked to help you navigate the standardized testing maze. Myth #1: Colleges prefer the SAT over the ACT. Reality: All four year colleges in the U.S. now accept the ACT, and it is highly regarded. Some admissions professionals will even tell you that they prefer it over the SAT because it is based more on coursework and better reflects how a student will perform in college. From a student perspective, the ACT will fulfill the need for SAT I’s and SAT II’s at many colleges. I recommend that all students take the PLAN (pre-ACT) or a diagnostic ACT to see whether the SAT or ACT is a better test for them. Several sites offer free full length practice ACT tests. Myth #2: Colleges only see the scores I send them. This is often accompanied by rumors like “it doesn’t matter how many times I take the SAT because colleges won’t see it,” or “I will just focus on one section of the SAT at a time because colleges will take my highest score for each section.” Reality: Score choice was implemented by the Collegeboard in 2008, HOWEVER colleges are allowed to set their own rules on what scores they require. These guidelines are then programmed into the Collegeboard system and go into effect when you send your score reports. Some colleges do allow you to cherry pick the scores you send, but others require all scores, and some use “highest overall test date.” A few of the more popular colleges that require all test scores include: Colgate, Cornell, Columbia, GW, Georgetown, Johns Hopkins, Syracuse, Stanford, Tufts, U. Maryland, UPENN, Wesleyan and all the University of California campuses. While many of these colleges state that they will still “superscore” (use the highest section from any test date), the jury is still out on why they want all the scores. Some surmise that they want to see how many times a student took the test, and others believe that they want to see what the increase was due to test prep (so that students who can’t afford test prep are not disadvantaged in the process). Most believe it is simply in the interest of making certain that they have all the correct data to select the highest scores. Myth #3: I don’t need to take SAT II Subject Tests. Reality: First of all, you WILL need them if you are applying to the Ivy League, many small competitive liberal arts colleges (Amherst, Haverford, Williams, Vassar, etc.) or some popular universities like BC, BU, Duke, Georgetown, Tufts or UVA. Second, there is a large group of colleges that “recommend” SAT II’s, and make no mistake about it, they like to see them. These institutions include: American University, George Washington, NYU, Northwestern, UNC and USC. Finally, some colleges that don’t require SAT IIs in general, do require them for certain programs like engineering or nursing. A few even specify which subjects they require (Bio, Chem, Math Level II, etc.). The bottom line is that junior year is too early to cut off any options as to where you will be applying. And the best time to take SAT II's is at the end of sophomore or junior year for subjects you have just completed (i.e. if you just took U.S. History and Chemistry, take those tests); that way the material is fresh in your mind. To read more about SAT II Subject tests, go here. The bottom line is that you need to understand the landscape surrounding the SAT I , SAT II and ACT before you take the tests. Arm yourself with facts, not rumors, and set your testing schedule well in advance. If your scores are not where you hoped this spring, plan to spend the summer studying for the fall test date. Cristiana Quinn, M.Ed. is the founder of College Admission Advisors, LLC which provides strategic college counseling, SAT prep and athletic recruiting services www.collegeadvisorsonline.com.
{ "pile_set_name": "Pile-CC" }
Picnic Time X-Grill Virginia Tech Hokies Item #: 775-00-175-604-0 Item Rating: Availability: Usually ships in 3-4 business days Regular Price: $89.00 Easter Sale! $74.45- Save $14.55 Quantity: Select Picnic Time products are now available with your favorite licensed Collegiate Logos! We're constantly adding new schools, so if your school isn't listed call us at 866-233-6587 to see if it is available yet. Great Gift Ideas!
{ "pile_set_name": "Pile-CC" }
Testing process errors and their harms and consequences reported from family medicine practices: a study of the American Academy of Family Physicians National Research Network. Little is known about the types and outcomes of testing process errors that occur in primary care. To describe types, predictors and outcomes of testing errors reported by family physicians and office staff. Events were reported anonymously. Each office completed a survey describing their testing processes prior to event reporting. 243 clinicians and office staff of eight family medicine offices. Distribution of error types, associations with potential predictors; predictors of harm and consequences of the errors. Participants submitted 590 event reports with 966 testing process errors. Errors occurred in ordering tests (12.9%), implementing tests (17.9%), reporting results to clinicians (24.6%), clinicians responding to results (6.6%), notifying patient of results (6.8%), general administration (17.6%), communication (5.7%) and other categories (7.8%). Charting or filing errors accounted for 14.5% of errors. Significant associations (p<0.05) existed between error types and type of reporter (clinician or staff), number of labs used by the practice, absence of a results follow-up system and patients' race/ethnicity. Adverse consequences included time lost and financial consequences (22%), delays in care (24%), pain/suffering (11%) and adverse clinical consequence (2%). Patients were unharmed in 54% of events; 18% resulted in some harm, and harm status was unknown for 28%. Using multilevel logistic regression analyses, adverse consequences or harm were more common in events that were clinician-reported, involved patients aged 45-64 years and involved test implementation errors. Minority patients were more likely than white, non-Hispanic patients to suffer adverse consequences or harm. Errors occur throughout the testing process, most commonly involving test implementation and reporting results to clinicians. While significant physical harm was rare, adverse consequences for patients were common. The higher prevalence of harm and adverse consequences for minority patients is a troubling disparity needing further investigation.
{ "pile_set_name": "PubMed Abstracts" }
Q: A Borel subset of a topological space Is every Borel subset of a topological space $(X, \tau)$ either open or closed? I'm thinking that if $B \in \mathcal B(X)$, then there are many possibilities: $B$ is open for $\tau$, or it is the complement of some open set $B$ in $\tau$, and so closed, or $B = \bigcup_{n=1}^{\infty} B_n$ where each $B_n$ is open, and so open, or Some trivial case in which $B$ is obviously open/closed, or $B = \bigcap_{n=1}^{\infty} B_n$ where these are a mix of closed and open sets. Even if, ideally, all of the $B_n$'s were open, this does not necessarily give an open set, because only finite intersections yield open sets. Thus, my conclusion is no. But could you construct some counter example for this? Thank you. I encountered this confusion when trying to prove that every continuous function is Borel-measurable. I wanted to separate the proof into two cases (open, closed), but then I realized this. A: (1) You are right in believing that the answer is no. For example, in $\mathbf R$, we have the Borel set $$ B = (0,1] = \bigcap_n \left(0,\frac{n+1}n\right) $$ which is neither open nor closed. (2) The possibilities that you list, are not all. When we denote set of the open sets by $B_0(\def\R{\mathbf R}\R)$ and define inductively $$ B_n(\R) = \left\{\bigcap_k A_k\mid A_k \in B_{n-1}(\R) \text{ or } \R\setminus A_k \in B_{n-1}(\R) \right\} $$ then $B_{n-1}(\R) \subsetneq B_n(\R)$ for each $n$, and $\bigcup_n B_n(\R) \ne \textsf{Bor}(\R)$ (you have to continue the induction until $\omega_1$).
{ "pile_set_name": "StackExchange" }
Evaluating endothelial function of the common carotid artery: an in vivo human model. Flow mediated dilation (FMD) of peripheral conduit arteries is a well-established tool to evaluate endothelial function. The aims of this study are to apply the FMD model to cerebral circulation by using acetazolamide (ACZ)-induced intracranial vasodilation as a stimulus to increase common carotid artery (CCA) diameter in response to a local increase of blood flow velocity (BFV). In 15 healthy subjects, CCA end-diastolic diameter and BFV, middle cerebral artery (MCA) BFV and mean arterial blood pressure (MBP) were measured at basal conditions, after an intravenous bolus of 1g ACZ, and after placebo (saline) sublingual administration at the 15th and 20th minute. In a separate session, the same parameters were evaluated after placebo (saline) infusion instead of ACZ and after 10 microg/m(2) bs and 300 microg of glyceryl trinitrate (GTN), administered sublingually, at the 15th and 20th minute, respectively. After ACZ bolus, there was a 35% maximal MCA mean BFV increment (14th minute), together with a 22% increase of mean CCA end-diastolic BFV and a CCA diameter increment of 3.9% at the 3rd minute (p=0.024). There were no MBP significant variations up to the 15th minute (p=0.35). After GTN administration, there was a significant increment in CCA diameter (p<0.00001). ACZ causes a detectable CCA dilation in healthy individuals concomitantly with an increase in BFV. Upon demonstration that this phenomenon is endothelium dependent, this experimental model might become a valuable tool to assess endothelial function in the carotid artery.
{ "pile_set_name": "PubMed Abstracts" }
IN THE COURT OF CRIMINAL APPEALS OF TENNESSEE AT JACKSON BURTON WELCH, ) ) Petitioner, ) C. C. A. NO. 02C01-9807-CC-00222 ) vs. ) LAKE COUNTY STATE OF TENNESSEE, ) ) No. 98-7802 FILED ) Respondent. ) October 15, 1998 Cecil Crowson, Jr. Appellate C ourt Clerk ORDER This matter is before the Court upon motion of the state to affirm the judgment of the trial court by order rather than formal opinion. See Rule 20, Rules of the Court of Criminal Appeals. This case represents an appeal from the trial court’s denial of the petitioner’s petition for writ of habeas corpus. On December 17, 1996, the petitioner pled guilty to four counts of sale of cocaine under 0.5 grams and received a six year sentence. No appeal was taken. In his present petition, the petitioner claims: 1) the chancellor lacked jurisdiction to render judgment in this case; 2) his arrest was the result of an unlawful entrapment; 3) he received ineffective assistance of counsel; and 4) the convicting evidence was insufficient.1 The trial court found that these allegations are not proper subject for habeas corpus relief. Habeas corpus relief is available in Tennessee only when “it appears upon the face of the judgment or the record of the proceedings upon which the judgment is rendered” that a convicting court was without jurisdiction or authority to sentence a defendant, or that a defendant’s sentence of imprisonment or other restraint has expired. Archer v. State, 851 S.W.2d 157, 164 (Tenn. 1993). Pursuant to T.C.A. § 17-2-203 (Repealed 1997), the chancellor had jurisdiction to render the judgments in this case. Though recently repealed, this statute was in effect on the date of judgments in this 1 The petitioner makes the additional argument on appeal that the judgm ents do not specify und er wh ich st atute he wa s co nvicte d. Sin ce th e pet itione r failed to rais e the issue below , this C ourt is preclud ed from conside ring it on app eal. See T.R.A.P. 36(a). Nonetheless, this argument is without merit. The judgments clearly state the crime for which the petitioner was convicted. case. Additionally, the judgments in this case clearly establish that the convicting court had jurisdiction to convict the petitioner of an existing criminal offense, i.e., the sale of cocaine under 0.5 grams. Nor has the petitioner’s sentence expired. Review of the other claims raised by the petitioner would necessarily involve investigation beyond the face of the judgment or record, and they are, therefore, inappropriate for consideration in a habeas corpus proceeding. Accordingly, for the reasons stated above, it is hereby ORDERED that the state’s motion is granted and the judgment of the trial court is affirmed in accordance with Rule 20, Rules of the Court of Criminal Appeals. _________________________ PAUL G. SUMMERS, JUDGE _________________________ DAVID G. HAYES, JUDGE _________________________ JOE G. RILEY, JUDGE 2
{ "pile_set_name": "FreeLaw" }
Q: Control order of script executions in test_package Is it possible to control the order in which test_package executes testing scripts? I would like to start by executing some code that creates some objects shared by all individual tests. Therefore, this script needs to be executed first, before the actual test-blabla.R scripts. I could try something like test-AAA.R or test-000.R, but I am not sure if the dir function used by testthat to list the scripts in a package, returns the same (alphabetical?) order of files on all platforms. A: ?test_dir says Test files start with ‘test’ and are executed in alphabetical order (but they shouldn't have dependencies). Helper files start with ‘helper’ and loaded before any tests are run. So, use helper files. i.e. create a file with a name that begins with "helper" and put in it the code that you need to run before running all the tests.
{ "pile_set_name": "StackExchange" }
Environment Canada – The Year in Review – Canada’s Long Cold Winter Following several mild winters, Canada’s reputation as the second coldest country next to Russia was reaffirmed in winter 2013-14. While at times British Columbia and the Yukon were basking in balmy breezes and sunny skies, the rest of us were shivering our way through the coldest winter in 18 years and the third coldest in 35 years. That devilish polar vortex – a circulation of frigid, dense Arctic air − moved much farther south than normal, freezing the heart of North America from Tuktoyaktuk to Toronto to Tallahassee. At the same time, weather systems from the Pacific that help push cold lows back to the north or across to Siberia were scarce. The intensity of the cold was remarkable, but it was its duration that brought us to our knees. In most parts of the country winter came to town early and wouldn’t leave, staying from Halloween to beyond Easter. And while January did have a thaw, it was feeble and lasted hours not days or weeks. For millions of Canadians from Windsor to Quebec City, the “normal” winter period from December to February was the eighth coldest ever recorded. Even more revealing, the five months between November and March inclusive were the coldest since the start of national record-keeping in 1948. And we didn’t just feel it; we were surrounded by it in the form of record snowfalls, crippling ice storms and lingering snow cover. Snowfall records were set in Windsor, Calgary, Red Deer, Kenora and a handful of other cities across Canada. In Saskatoon, there was snow on the ground for six months – likely the longest period with continuous snow cover since 1955 when record-keeping began. The following is a sampling of how cold and snowy it was: According to weather data logged by NASA’s Curiosity Rover, residents between Regina and Rouyn woke up some mornings to temperatures colder than those on Mars. On New Year’s Day, Ottawa went from slushy puddles and melting temperatures to -23°C in fewer than 24 hours. But the city’s brutal cold had nothing on locations in Quebec, where the wind chill hit an unbearable -56 at Fermont and Normandin and exposed flesh could freeze in under two minutes. Just two days later, at Lac Benoit, the temperature dipped to a low of -47.4°C. February 26 was Hamilton’s 47th day under a cold alert, the identical number of cold alert days for the city over the past three winters combined. Freezing-degree days below 0°C were 15 per cent higher than average over southern Ontario and Quebec, which explains the unusually thick river and lake ice. Toronto experienced its c-cc-coldest winter in 20 years, which prompted the public health office to issue 36 extreme cold alerts compared to nine the previous year. Further, the city had snow on the ground for more than 100 consecutive days on top of a layer of pre-Christmas freezing rain. Kenora’s winter was the coldest and snowiest since 1938 when record-keeping began, while Kitchener-Waterloo logged 25 days below -20°C; normal is 6. Nobody had it worse than Winnipeg, arguably the coldest big city in Canada. No one alive can say they’ve lived it colder there as residents survived the coldest December to March, inclusive, since 1898 – long before urban heat islands, automobiles, heavy industry and long before global warming. Temperatures averaged -20°C when the normal average is -14.3°C, and there were 30 raw days when the temperature dropped below -30°C versus a typical average of 12. Adding to the misery, the city received an abnormally large amount of snow. The 155 cm that fell was well above the average of 100 cm and the most the city had received since the winter of 1996-97 – enough to bust the myth that it’s ever too cold to snow! The only good news was that the cold and snow were so dry it helped minimize the spring flood risk The bone-chilling, teeth-chattering weather and lasting snow had a host of negative impacts. A shortage of road salt in parts of central Canada had suppliers scrambling to find extra supplies, and consumers of propane and natural gas fumed at skyrocketing prices. The frigid weather caused record levels of power consumption as customers cranked up the thermostat to beat back the cold, and home and business owners faced up to 20 per cent billing increases. Frequent blackouts just made things worse. For the homeless, the bitter cold made a hard life harder. Hospitals were pushed to handle more cases of frostbite, hypothermia and falls from icy sidewalks and streets, and Canadian Blood Services reported a significant decline in donor numbers. The winter’s duration and difficulty also led to increased reports of depression and anxiety. At times, the intense cold created transportation nightmares. Towing companies couldn’t keep up with the calls, with waits for roadside service typically reaching five hours or more for low-priority calls. Extreme cold also caused air travel chaos when Canada’s biggest airport, Toronto’s Pearson International, shut down causing serious ripple effects for aviation across the country. For plumbers, things couldn’t have been busier as they scrambled to thaw frozen hydrants, ruptured pipes and broken sprinklers. Homeowners flooded city hot lines, complaining about burst water pipes and frozen toilets. Some residents in Winnipeg lost water for months as persistently cold temperatures froze the ground one or two metres below the surface. The city incurred the highest number of frozen water pipes in more than 35 years. Hundreds of pipes also froze in Thunder Bay and Kenora, leading the latter to issue a citywide boil-water advisory. Frozen pipes, along with big snowfalls and biting wind chills, also contributed to school snow days and event cancellations becoming regular occurrences across eastern Canada. Brutal winter conditions also took their toll on plants and animals. Winterkill was especially severe across the board with damage to golf courses costing millions of dollars in lost business and repairs. Ontario’s wine industry lost millions of dollars after cold killed 60 to 95 per cent of the grape buds. In some communities, biologists reported a record loss of trees and ornamental shrubs from winterkill, salt damage or wind desiccation. On the farm, extreme cold meant that cattle had to eat even more just to stay warm. Bee farmers reported losses in their hives as high as 50 per cent over the past winter. The severe cold and deep snow also contributed to a decline in deer populations, while brown bats literally dropped and died from the biting cold.
{ "pile_set_name": "Pile-CC" }
Q: Why does the universal cover of $GL^+_n$ not admit finite-dimensional representations? Let $GL^+_n \subset \mathbb{R}^{n \times n}$ be the subgroup of real matrices with positive determinant and $\widetilde{GL}^+_n$ be its universal cover. Why does $\widetilde{GL}^+_n$ not admit finite-dimensional representations? I would be happy with a proof or a reference. This question is relevant in spin geometry. Normally, one uses $Spin_n$, which is the universal cover of $SO_n$ (if $n \geq 3$), and a finite-dimensional representation $\rho:Spin_n \to GL(V)$ to construct the spinor bundle of a Riemannian spin manifold $(M,g)$ with spin structure $\Theta^g: Spin^g M \to SO^g M$ via $\Sigma^g M := Spin^g \times _{\rho} V$. In texts like [1], the authors work with a topological spin structure $\widetilde{GL}^+ M \to GL^+ M$ up to the point where one has to construct the spinor bundle. I like this approach, which is why I would like to better understand why this part of the construction fails. Unfortunately, I do not know much about representation theory besides what is mentioned in [1]. [1] Bär, Gauduchon, Moroianu: Generlized cylinders in Semi-Riemannian and Spin Geometry EDIT: One should require the representations to be faithful. A: By [1, Lem. II.5.23], any finite-dimensional representation $\tilde \rho: \widetilde{GL}_n^+ \to GL(V)$, $n>2$, factors through $GL^+_n$, i.e. there exists $\rho:GL^+_n \to Aut(V)$ such that $\rho \circ \vartheta = \tilde \rho$, where $\vartheta:\widetilde{GL}^+_n \to GL^+_n$ is the two-fold cover. Consequently, $\ker \tilde \rho \supset \ker \vartheta \cong Z_2 $ and $\tilde \rho$ cannot be faithful. [1] Lawson/Michelsohn: Spin Geometry
{ "pile_set_name": "StackExchange" }
Relationships between the lipid content and the rheological properties of airway secretions in cystic fibrosis. The lipid composition and rheological properties of expectorated airway secretions have been analyzed in 16 patients with cystic fibrosis (CF), separated into two groups according to whether their secretions were superinfected or not. The total lipid content was higher in the superinfected CF secretions. The content in cholesterol and GL1, GL3 and GL4 glycosphingolipid fractions were significantly higher in the superinfected in comparison to the non-superinfected group. The viscosity was two-fold higher in the CF superinfected group. A significant correlation (r = 0.72, p less than 0.01) was observed between the apparent viscosity and the total lipid content of CF secretions. Cholesterol, glycosphingolipids and sphingomyelin were the lipid components which were the most closely and positively correlated to the viscosity. On the opposite, phosphoglycerol was negatively correlated (r = -0.72, p less than 0.05) to the viscoelastic properties of CF airway secretions.
{ "pile_set_name": "PubMed Abstracts" }
Oak Cliff Blog Come and get it Saturday at Oswald apartment building in Oak Cliff Randy Johnson stands outside the apartment building at 600 Elsbeth in north Oak Cliff. Speaking of history, perhaps you’ve heard more than a few times that Jane Bryant’s apartment building is coming down at 600 Elsbeth in north Oak Cliff. That’s the place, built in 1925 at the corner of Davis and Elsbeth, where accused presidential assassin Lee Harvey Oswald and family briefly lived in 1962-63. After a long fight with the city over housing code issues, Bryant agreed earlier this year to demolish the structure by Nov. 30. The city has given her some leeway and a demolition permit. The work continues slowly with most of the brick removed from the outer walls. And many pieces — such as the bathtub, toilet and medicine cabinet — from the Oswald’s Apartment No. 2 have been sold and hauled away. But plenty of lumber, fixtures, windows and other pieces of the 10-unit building remain. And on Saturday, it’s come and get it for anyone interested. Bring your hammer, saw, crowbar, strong friends. Anyone and anything to help you take away what you want of what’s left. Oh, and bring cash. Bryant is selling the stuff. “Of course, I’m not going to charge a lot,” she told us. Time: 10 a.m. to 4 p.m. She said about 35,000 feet of yellow pine lumber (roughly 9-inch wide shiplap) and 12,000 feet of white pine flooring remained as of this morning. Bryant said she and her friend Randy Johnson have told the city they plan to have the site cleared by Jan. 5.
{ "pile_set_name": "Pile-CC" }
Q: Average slope of a curve (right side of the equation) intuition Okay, I can somehow understand why average slope of a curve is defined in such a fashion: Let $f$ be a continuous function on an interval $[a,b]$. Then, we define its average over $[a,b]$ to be the number $$ f_{\operatorname{avg}} := {1 \over b-a} \int_a^b f(t) dt $$ Let $y = F(x)$ be a curve. Fix the starting point $x_0$ and the ending point $x_1=x_0+\Delta x$. Then, the slope at a point $z$ is the derivative $F'(z)$. Thus, the "average slope'' should mean \begin{equation*} (F')_{\operatorname{avg}} = {1 \over x_1 - x_0}\int_{x_0}^{x_1} F'(t) dt = {F(x_1) - F(x_0) \over x_1-x_0}. \end{equation*} To summarize: "Avg. slope = Sum all slopes of all tangent lines over an interval $[a,b]$ and divide it by the number of slopes over that same interval." But right side of the equation somehow bothers me. I don't get intuitevely how average slope (sum of all slopes/number of slopes) equals slope of a function $F(x)$ between two endpoints $x_0$ and $x_1$. A: The right side is the slope of the line drawn between $(x_0,F(x_0))$ and $(x_1,F(x_1))$. I would see that as the definition of average slope. Averaging an infinite number of local slopes is more problematic. The right side only depends on two values of the function.
{ "pile_set_name": "StackExchange" }
Beef Wellington (disambiguation) Beef Wellington is a steak dish. Beef Wellington may also refer to: Arthur Wellesley, 1st Duke of Wellington, "The Beef" Wellington Biff Wellington, stage name of wrestler Shayne Alexander Bower Beef Wellington (wrestler), a nickname for wrestler Brad Maddox
{ "pile_set_name": "Wikipedia (en)" }
Q: Change on column based on any previous values in another column I have a dataframe like so: foo <- c(rep(FALSE, 5), TRUE, rep(FALSE, 4)) rank_order <- seq(11,20) df <- data.frame(rank_order = as.numeric(rank_order), foo = foo) What I'd like to do is add one to every value of rank_order following a row where df$foo == TRUE. That means rank_order should look like this: rank_order_target <- c(11, 12, 13, 14, 15, 17, 18, 19, 20, 21) It's easy enough to change one value of rank_order, with lag looking at one previous value of foo (like below), but how do I look at all previous values of foo? df %>% mutate(rank_order_new = case_when(lag(foo, default = FALSE) == TRUE ~ rank_order + 1, TRUE ~ rank_order)) rank_order foo rank_order_new 1 11 FALSE 11 2 12 FALSE 12 3 13 FALSE 13 4 14 FALSE 14 5 15 FALSE 15 6 16 TRUE 16 7 17 FALSE 18 8 18 FALSE 18 9 19 FALSE 19 10 20 FALSE 20 Either a base solution, or something tidyverse would be helpful. A: We can use cumsum library(dplyr) df %>% mutate(new = rank_order + cumsum(foo))
{ "pile_set_name": "StackExchange" }
/* * Copyright (c) 2004, 2005 Topspin Communications. All rights reserved. * Copyright (c) 2005, 2006, 2007, 2008 Mellanox Technologies. All rights reserved. * Copyright (c) 2005, 2006, 2007 Cisco Systems, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include <linux/sched.h> #include <linux/slab.h> #include <linux/pci.h> #include <linux/errno.h> #include <linux/mlx4/cmd.h> #include <asm/io.h> #include "mlx4.h" #define CMD_POLL_TOKEN 0xffff enum { /* command completed successfully: */ CMD_STAT_OK = 0x00, /* Internal error (such as a bus error) occurred while processing command: */ CMD_STAT_INTERNAL_ERR = 0x01, /* Operation/command not supported or opcode modifier not supported: */ CMD_STAT_BAD_OP = 0x02, /* Parameter not supported or parameter out of range: */ CMD_STAT_BAD_PARAM = 0x03, /* System not enabled or bad system state: */ CMD_STAT_BAD_SYS_STATE = 0x04, /* Attempt to access reserved or unallocaterd resource: */ CMD_STAT_BAD_RESOURCE = 0x05, /* Requested resource is currently executing a command, or is otherwise busy: */ CMD_STAT_RESOURCE_BUSY = 0x06, /* Required capability exceeds device limits: */ CMD_STAT_EXCEED_LIM = 0x08, /* Resource is not in the appropriate state or ownership: */ CMD_STAT_BAD_RES_STATE = 0x09, /* Index out of range: */ CMD_STAT_BAD_INDEX = 0x0a, /* FW image corrupted: */ CMD_STAT_BAD_NVMEM = 0x0b, /* Error in ICM mapping (e.g. not enough auxiliary ICM pages to execute command): */ CMD_STAT_ICM_ERROR = 0x0c, /* Attempt to modify a QP/EE which is not in the presumed state: */ CMD_STAT_BAD_QP_STATE = 0x10, /* Bad segment parameters (Address/Size): */ CMD_STAT_BAD_SEG_PARAM = 0x20, /* Memory Region has Memory Windows bound to: */ CMD_STAT_REG_BOUND = 0x21, /* HCA local attached memory not present: */ CMD_STAT_LAM_NOT_PRE = 0x22, /* Bad management packet (silently discarded): */ CMD_STAT_BAD_PKT = 0x30, /* More outstanding CQEs in CQ than new CQ size: */ CMD_STAT_BAD_SIZE = 0x40, /* Multi Function device support required: */ CMD_STAT_MULTI_FUNC_REQ = 0x50, }; enum { HCR_IN_PARAM_OFFSET = 0x00, HCR_IN_MODIFIER_OFFSET = 0x08, HCR_OUT_PARAM_OFFSET = 0x0c, HCR_TOKEN_OFFSET = 0x14, HCR_STATUS_OFFSET = 0x18, HCR_OPMOD_SHIFT = 12, HCR_T_BIT = 21, HCR_E_BIT = 22, HCR_GO_BIT = 23 }; enum { GO_BIT_TIMEOUT_MSECS = 10000 }; struct mlx4_cmd_context { struct completion done; int result; int next; u64 out_param; u16 token; }; static int mlx4_status_to_errno(u8 status) { static const int trans_table[] = { [CMD_STAT_INTERNAL_ERR] = -EIO, [CMD_STAT_BAD_OP] = -EPERM, [CMD_STAT_BAD_PARAM] = -EINVAL, [CMD_STAT_BAD_SYS_STATE] = -ENXIO, [CMD_STAT_BAD_RESOURCE] = -EBADF, [CMD_STAT_RESOURCE_BUSY] = -EBUSY, [CMD_STAT_EXCEED_LIM] = -ENOMEM, [CMD_STAT_BAD_RES_STATE] = -EBADF, [CMD_STAT_BAD_INDEX] = -EBADF, [CMD_STAT_BAD_NVMEM] = -EFAULT, [CMD_STAT_ICM_ERROR] = -ENFILE, [CMD_STAT_BAD_QP_STATE] = -EINVAL, [CMD_STAT_BAD_SEG_PARAM] = -EFAULT, [CMD_STAT_REG_BOUND] = -EBUSY, [CMD_STAT_LAM_NOT_PRE] = -EAGAIN, [CMD_STAT_BAD_PKT] = -EINVAL, [CMD_STAT_BAD_SIZE] = -ENOMEM, [CMD_STAT_MULTI_FUNC_REQ] = -EACCES, }; if (status >= ARRAY_SIZE(trans_table) || (status != CMD_STAT_OK && trans_table[status] == 0)) return -EIO; return trans_table[status]; } static int cmd_pending(struct mlx4_dev *dev) { u32 status = readl(mlx4_priv(dev)->cmd.hcr + HCR_STATUS_OFFSET); return (status & swab32(1 << HCR_GO_BIT)) || (mlx4_priv(dev)->cmd.toggle == !!(status & swab32(1 << HCR_T_BIT))); } static int mlx4_cmd_post(struct mlx4_dev *dev, u64 in_param, u64 out_param, u32 in_modifier, u8 op_modifier, u16 op, u16 token, int event) { struct mlx4_cmd *cmd = &mlx4_priv(dev)->cmd; u32 __iomem *hcr = cmd->hcr; int ret = -EAGAIN; unsigned long end; mutex_lock(&cmd->hcr_mutex); end = jiffies; if (event) end += msecs_to_jiffies(GO_BIT_TIMEOUT_MSECS); while (cmd_pending(dev)) { if (time_after_eq(jiffies, end)) goto out; cond_resched(); } /* * We use writel (instead of something like memcpy_toio) * because writes of less than 32 bits to the HCR don't work * (and some architectures such as ia64 implement memcpy_toio * in terms of writeb). */ __raw_writel((__force u32) cpu_to_be32(in_param >> 32), hcr + 0); __raw_writel((__force u32) cpu_to_be32(in_param & 0xfffffffful), hcr + 1); __raw_writel((__force u32) cpu_to_be32(in_modifier), hcr + 2); __raw_writel((__force u32) cpu_to_be32(out_param >> 32), hcr + 3); __raw_writel((__force u32) cpu_to_be32(out_param & 0xfffffffful), hcr + 4); __raw_writel((__force u32) cpu_to_be32(token << 16), hcr + 5); /* __raw_writel may not order writes. */ wmb(); __raw_writel((__force u32) cpu_to_be32((1 << HCR_GO_BIT) | (cmd->toggle << HCR_T_BIT) | (event ? (1 << HCR_E_BIT) : 0) | (op_modifier << HCR_OPMOD_SHIFT) | op), hcr + 6); /* * Make sure that our HCR writes don't get mixed in with * writes from another CPU starting a FW command. */ mmiowb(); cmd->toggle = cmd->toggle ^ 1; ret = 0; out: mutex_unlock(&cmd->hcr_mutex); return ret; } static int mlx4_cmd_poll(struct mlx4_dev *dev, u64 in_param, u64 *out_param, int out_is_imm, u32 in_modifier, u8 op_modifier, u16 op, unsigned long timeout) { struct mlx4_priv *priv = mlx4_priv(dev); void __iomem *hcr = priv->cmd.hcr; int err = 0; unsigned long end; down(&priv->cmd.poll_sem); err = mlx4_cmd_post(dev, in_param, out_param ? *out_param : 0, in_modifier, op_modifier, op, CMD_POLL_TOKEN, 0); if (err) goto out; end = msecs_to_jiffies(timeout) + jiffies; while (cmd_pending(dev) && time_before(jiffies, end)) cond_resched(); if (cmd_pending(dev)) { err = -ETIMEDOUT; goto out; } if (out_is_imm) *out_param = (u64) be32_to_cpu((__force __be32) __raw_readl(hcr + HCR_OUT_PARAM_OFFSET)) << 32 | (u64) be32_to_cpu((__force __be32) __raw_readl(hcr + HCR_OUT_PARAM_OFFSET + 4)); err = mlx4_status_to_errno(be32_to_cpu((__force __be32) __raw_readl(hcr + HCR_STATUS_OFFSET)) >> 24); out: up(&priv->cmd.poll_sem); return err; } void mlx4_cmd_event(struct mlx4_dev *dev, u16 token, u8 status, u64 out_param) { struct mlx4_priv *priv = mlx4_priv(dev); struct mlx4_cmd_context *context = &priv->cmd.context[token & priv->cmd.token_mask]; /* previously timed out command completing at long last */ if (token != context->token) return; context->result = mlx4_status_to_errno(status); context->out_param = out_param; complete(&context->done); } static int mlx4_cmd_wait(struct mlx4_dev *dev, u64 in_param, u64 *out_param, int out_is_imm, u32 in_modifier, u8 op_modifier, u16 op, unsigned long timeout) { struct mlx4_cmd *cmd = &mlx4_priv(dev)->cmd; struct mlx4_cmd_context *context; int err = 0; down(&cmd->event_sem); spin_lock(&cmd->context_lock); BUG_ON(cmd->free_head < 0); context = &cmd->context[cmd->free_head]; context->token += cmd->token_mask + 1; cmd->free_head = context->next; spin_unlock(&cmd->context_lock); init_completion(&context->done); mlx4_cmd_post(dev, in_param, out_param ? *out_param : 0, in_modifier, op_modifier, op, context->token, 1); if (!wait_for_completion_timeout(&context->done, msecs_to_jiffies(timeout))) { err = -EBUSY; goto out; } err = context->result; if (err) goto out; if (out_is_imm) *out_param = context->out_param; out: spin_lock(&cmd->context_lock); context->next = cmd->free_head; cmd->free_head = context - cmd->context; spin_unlock(&cmd->context_lock); up(&cmd->event_sem); return err; } int __mlx4_cmd(struct mlx4_dev *dev, u64 in_param, u64 *out_param, int out_is_imm, u32 in_modifier, u8 op_modifier, u16 op, unsigned long timeout) { if (mlx4_priv(dev)->cmd.use_events) return mlx4_cmd_wait(dev, in_param, out_param, out_is_imm, in_modifier, op_modifier, op, timeout); else return mlx4_cmd_poll(dev, in_param, out_param, out_is_imm, in_modifier, op_modifier, op, timeout); } EXPORT_SYMBOL_GPL(__mlx4_cmd); int mlx4_cmd_init(struct mlx4_dev *dev) { struct mlx4_priv *priv = mlx4_priv(dev); mutex_init(&priv->cmd.hcr_mutex); sema_init(&priv->cmd.poll_sem, 1); priv->cmd.use_events = 0; priv->cmd.toggle = 1; priv->cmd.hcr = ioremap(pci_resource_start(dev->pdev, 0) + MLX4_HCR_BASE, MLX4_HCR_SIZE); if (!priv->cmd.hcr) { mlx4_err(dev, "Couldn't map command register."); return -ENOMEM; } priv->cmd.pool = pci_pool_create("mlx4_cmd", dev->pdev, MLX4_MAILBOX_SIZE, MLX4_MAILBOX_SIZE, 0); if (!priv->cmd.pool) { iounmap(priv->cmd.hcr); return -ENOMEM; } return 0; } void mlx4_cmd_cleanup(struct mlx4_dev *dev) { struct mlx4_priv *priv = mlx4_priv(dev); pci_pool_destroy(priv->cmd.pool); iounmap(priv->cmd.hcr); } /* * Switch to using events to issue FW commands (can only be called * after event queue for command events has been initialized). */ int mlx4_cmd_use_events(struct mlx4_dev *dev) { struct mlx4_priv *priv = mlx4_priv(dev); int i; priv->cmd.context = kmalloc(priv->cmd.max_cmds * sizeof (struct mlx4_cmd_context), GFP_KERNEL); if (!priv->cmd.context) return -ENOMEM; for (i = 0; i < priv->cmd.max_cmds; ++i) { priv->cmd.context[i].token = i; priv->cmd.context[i].next = i + 1; } priv->cmd.context[priv->cmd.max_cmds - 1].next = -1; priv->cmd.free_head = 0; sema_init(&priv->cmd.event_sem, priv->cmd.max_cmds); spin_lock_init(&priv->cmd.context_lock); for (priv->cmd.token_mask = 1; priv->cmd.token_mask < priv->cmd.max_cmds; priv->cmd.token_mask <<= 1) ; /* nothing */ --priv->cmd.token_mask; priv->cmd.use_events = 1; down(&priv->cmd.poll_sem); return 0; } /* * Switch back to polling (used when shutting down the device) */ void mlx4_cmd_use_polling(struct mlx4_dev *dev) { struct mlx4_priv *priv = mlx4_priv(dev); int i; priv->cmd.use_events = 0; for (i = 0; i < priv->cmd.max_cmds; ++i) down(&priv->cmd.event_sem); kfree(priv->cmd.context); up(&priv->cmd.poll_sem); } struct mlx4_cmd_mailbox *mlx4_alloc_cmd_mailbox(struct mlx4_dev *dev) { struct mlx4_cmd_mailbox *mailbox; mailbox = kmalloc(sizeof *mailbox, GFP_KERNEL); if (!mailbox) return ERR_PTR(-ENOMEM); mailbox->buf = pci_pool_alloc(mlx4_priv(dev)->cmd.pool, GFP_KERNEL, &mailbox->dma); if (!mailbox->buf) { kfree(mailbox); return ERR_PTR(-ENOMEM); } return mailbox; } EXPORT_SYMBOL_GPL(mlx4_alloc_cmd_mailbox); void mlx4_free_cmd_mailbox(struct mlx4_dev *dev, struct mlx4_cmd_mailbox *mailbox) { if (!mailbox) return; pci_pool_free(mlx4_priv(dev)->cmd.pool, mailbox->buf, mailbox->dma); kfree(mailbox); } EXPORT_SYMBOL_GPL(mlx4_free_cmd_mailbox);
{ "pile_set_name": "Github" }
Dioryctria rossi Dioryctria rossi is a species of snout moth in the genus Dioryctria. This moth was discovered and named by Dr. Douglas Alexander Ross, chief entomologist at the Vernon Forest Entomology laboratory and research centre in Vernon BC from 1950-1970. It was described by Munroe in 1959. It is found in western North America, from southern British Columbia to northern Mexico and east to New Mexico. The wingspan is 22–34 mm. The forewings are orange brown, without an inner transverse line and with whitish markings, primarily consisting of a longitudinal streak. The hindwings are pale. The larvae feed on Pinus ponderosa, Pinus arizonica and Pinus durangensis. They feed in the cones of their host plant. The larvae are about 30 mm long and brown with dark-gray subdorsal bands and a reddish-brown head. References Category:Moths described in 1959 rossi
{ "pile_set_name": "Wikipedia (en)" }
A company called Cosmic Lifestyle Corp. has designed cocktail glasses they say will work in zero gravity. The cups look sort of like what glasses we know here on Earth, but they’ve got a system of grooves that the company says will hold the liquids more or less in place until you’re ready to put them in your face. Like any foolish endeavor, the zero gravity cups are doing a Kickstarter funding campaign. Oregon’s Ninkasi Brewing has launched vials of yeast into space, then retrieved them, and come up with an imperial stout. While this brings space beer to new heights, the real accomplishment is that the company refrained from saying that the beer is “out of this world.” If you’ve had a couple in today’s revelry, you may be the most attractive person in the bar, according to a new study. Researchers found that if a person has a glass or two of wine, their image is more attractive to other people than if they were sober. Unfortunately, this doesn’t mean double-fisting makes you sexy, it means that subtle changes in your appearance after a drink or two (read: a smile) makes you more attractive. Another downside is that if you’ve had more than a couple drinks, your attractiveness goes down. No research has been done on whether it goes back up again after 8 or 9 drinks. So, let’s just assume for now it really does help. Well guess what? It’s back! Unless you live in a fifth of the states here in the U.S. Then your nose is just boned for regular reasons. Palcohol has, once again, been approved by the Alcohol and Tobacco Tax and Trade Bureau. Four varieties can now be bought! Except approximately 10 states aren’t down with getting your alcohol buzz via your nose, so they’ve already or are planning to halt sales in their respective residences. Mind you, all of The Guys live in the same state, and at the moment, our state has not banned Palcohol. This just made the SeriouslyGuys Fantasy Football draft that much more snortier. Two U.S. Secret Service agents reportedly collided their car into a White House barrier Wednesday night. According to House Oversight Committee Chairman Jason Chaffetz (R-UT), the agents were partying in Georgetown to celebrate a Secret Service spokesman’s retirement (too soon, apparently) until called back to respond to an incident at the White House. Alcohol is suspected as a factor, which isn’t much of a secret considering that these are the same barriers that the Secret Service set up in the first place. At this point, the only secret that the Secret Service has kept protected is that they graduated from the police academy back when Commandant Lassard was in charge of it. In other news:The Guys are definitely becoming U.S. Secret Service agents. If you were looking forward to trying out some wine from the Civil War, you can give up now. A panel of wine tasters sampled one of five bottles recovered from a ship that sank in 1864 bound for the South. It was pretty much just seawater, which shouldn’t be shocking, considering that the liquid was milky and grey. At this point we should just stop getting excited about booze recovered from shipwrecks, especially if the bottles are sealed with corks that allow in seawater. Centuries from now, people will enjoy screw-top and bag wines from modern shipwrecks. Hey, remember that beer that was probably the oldest ever found? The one from the shipwreck in the Baltic? Scientists now have a better idea of what it tasted like. Since you probably don’t have the cash to spend on a brewery’s re-creation of the brew, it’s imagination time. Using chemical analysis and other techniques, researchers found that it was an amber ale, and it had an ABV of about 4.5%, which isn’t really the strength you’d think would be imported from a faraway land. Because the bacteria was still alive inside the bottle, scientists were even able to tell what kind of hops were used, but probably knew better than to list it, so they didn’t come off as beer snobs. Just grab something barrel-aged for this weekend. It will cost you less, and it will still be a trip back in time. If you’re in London this month, you’re probably excited about Annie the Owl, a pop up bar that will be open for only one week. You are also probably a crazy person. The draw of the bar is that you get to drink with owls. For those of you who don’t know, owls are huge, scary birds that can see you in the dark, and have long talons that could slice you up in a fraction of a second. A place filled with these flying death machines seems like a good place to drink, doesn’t it? Now, the bar has cut the service of all alcohol in response to concerns from animal rights groups. So you can’t even enjoy the bar for the reason you’re there: to get drunk. Drinking is good for you. And if you drink, there’s a good chance you’ll remember that fact long into your old age. According to researchers in China, a chemical in beer helps ward off degenerative diseases. A study found that xanthohumol, a chemical found in hops, might help brain cells from oxidative stress that can lead to dementia. That means that while you may not remember how many beers you had last night, there’s a better chance you’ll remember your grandchildren’s names later in life.
{ "pile_set_name": "Pile-CC" }
<!doctype html> <title>CodeMirror: YAML front matter mode</title> <meta charset="utf-8"/> <link rel=stylesheet href="../../doc/docs.css"> <link rel="stylesheet" href="../../lib/codemirror.css"> <script src="../../lib/codemirror.js"></script> <script src="../../addon/mode/overlay.js"></script> <script src="../markdown/markdown.js"></script> <script src="../gfm/gfm.js"></script> <script src="../yaml/yaml.js"></script> <script src="yaml-frontmatter.js"></script> <style>.CodeMirror { border-top: 1px solid #ddd; border-bottom: 1px solid #ddd; }</style> <div id=nav> <a href="https://codemirror.net"><h1>CodeMirror</h1><img id=logo src="../../doc/logo.png"></a> <ul> <li><a href="../../index.html">Home</a> <li><a href="../../doc/manual.html">Manual</a> <li><a href="https://github.com/codemirror/codemirror">Code</a> </ul> <ul> <li><a href="../index.html">Language modes</a> <li><a class=active href="#">YAML-Frontmatter</a> </ul> </div> <article> <h2>YAML front matter mode</h2> <form><textarea id="code" name="code"> --- receipt: Oz-Ware Purchase Invoice date: 2007-08-06 customer: given: Dorothy family: Gale items: - part_no: A4786 descrip: Water Bucket (Filled) price: 1.47 quantity: 4 - part_no: E1628 descrip: High Heeled "Ruby" Slippers size: 8 price: 100.27 quantity: 1 bill-to: &id001 street: | 123 Tornado Alley Suite 16 city: East Centerville state: KS ship-to: *id001 specialDelivery: > Follow the Yellow Brick Road to the Emerald City. Pay no attention to the man behind the curtain. --- GitHub Flavored Markdown ======================== Everything from markdown plus GFM features: ## URL autolinking Underscores_are_allowed_between_words. ## Strikethrough text GFM adds syntax to strikethrough text, which is missing from standard Markdown. ~~Mistaken text.~~ ~~**works with other formatting**~~ ~~spans across lines~~ ## Fenced code blocks (and syntax highlighting) ```javascript for (var i = 0; i &lt; items.length; i++) { console.log(items[i], i); // log them } ``` ## Task Lists - [ ] Incomplete task list item - [x] **Completed** task list item ## A bit of GitHub spice * SHA: be6a8cc1c1ecfe9489fb51e4869af15a13fc2cd2 * User@SHA ref: mojombo@be6a8cc1c1ecfe9489fb51e4869af15a13fc2cd2 * User/Project@SHA: mojombo/god@be6a8cc1c1ecfe9489fb51e4869af15a13fc2cd2 * \#Num: #1 * User/#Num: mojombo#1 * User/Project#Num: mojombo/god#1 See http://github.github.com/github-flavored-markdown/. </textarea></form> <p>Defines a mode that parses a <a href="http://jekyllrb.com/docs/frontmatter/">YAML frontmatter</a> at the start of a file, switching to a base mode at the end of that. Takes a mode configuration option <code>base</code> to configure the base mode, which defaults to <code>"gfm"</code>.</p> <script> var editor = CodeMirror.fromTextArea(document.getElementById("code"), {mode: "yaml-frontmatter"}); </script> </article>
{ "pile_set_name": "Github" }
Lola B2K/40 The Lola B2K/40 was a Le Mans Prototype developed in 2000 by Lola Cars International as a cheaper, smaller, and lighter alternative to the similar Lola B2K/10. Although specifically designed to compete in the SR2 class of the Sports Racing World Cup and Grand American Road Racing Championship, it would later be adapted to the LMP675 and LMP2 classes for the 24 Hours of Le Mans and American Le Mans Series. This design was replaced in 2005 by the Lola B05/40 and would last be used in competition in 2006. Development Although the B2K/40 shared some mechanical elements with the larger B2K/10, the design was more simplified for the sake of ease of use. Unlike the raised footbox B2K/10 and large protruding nose caused by this footbox, the B2K/40 would employ a simplified footbox and a flattened nose, although it still featured a raised element in the middle to help air flow between the cockpit and fenders. The fenders themselves were also much simpler, using a narrow wedge shape instead of the sweeping curves of the B2K/10. Intakes for the radiators would be placed inside the fenders, instead of on the outer edge as on the B2K/10. A single air intake underneath the rollbar would be employed for feeding the engine, although the B2K/40 employed a cylindrical nacelle instead of the large scoop that was sometimes used on the B2K/10. Some later chassis employed turbocharged engines, which required the addition of an intake for the turbocharger. Various designs were used over the years. For an engine, the chosen Lola unit was a small, production based V6 that was stipulated by SR2 class rules. Advanced Engine Research had an agreement with Nissan to modify their production VQ 3.0 Litre V6 for use in the SR2 class. However, as a customer car, the AER-Nissan unit was not put into all cars, and other engines have been used over the chassis' lifetime. Such choices included a Mazda turbocharged 2-rotor, a Judd V8, a Porsche Flat-6, and a Ford turbocharged Inline-4 developed by Millington. A total of sixteen B2K/40s were built. Racing history Debuting during the middle of the 2000 season, the first B2K/40s were delivered to various customers in the Grand American Road Racing Championship and American Le Mans Series in North American, and Sports Racing World Cup in Europe. Snow/Schumacher Racing, Archangel Motorsports and TRP Racing would run the USRRC season, while KnightHawk Racing and Phillips Motorsports would run ALMS with Snow/Schumacher Racing making two appearances at the end of the season. SportsRacing Team Sweden would be the only B2K/40 runner in the Sports Racing World Cup. The B2K/40s would perform well in the USRRC series, gaining five class wins, four for Archangel and one for Snow/Schumacher, and earning Archangel the SR2 class championship. In the ALMS, Snow/Schumacher would take a best result of 11th at Petit Le Mans, while SportsRacing Team Sweden would take a single win in the final round of the Sports Racing World Cup. However the best achievement of the season would be at the 24 Hours of Le Mans, where Multimatic Motorsports took the LMP675 class win, finishing 25th overall and one of only two competitors in its class to finish. For 2001, the number of B2K/40s in competition would nearly double. USRRC saw the return of Snow/Schumacher, Archangel, and TRP while gaining Northstar Racing, Porschehaus Racing, followed by Multimatic Motorsports and Rand Racing. Unfortunately, the ALMS would see only Roock-KnightHawk Racing competing regularly in the new LMP675 class, while the Sports Racing World Cup (now renamed the FIA Sportscar Championship) would see SportsRacing Team Sweden running alone again. USRRC was another dominant year for the B2K/40 with another nine class wins, six for championship winner Archangel, two for Porschehaus, and one for Rand Racing. In Europe, SportsRacing Team Sweden would take four class victories and also take the SR2 championship. Unfortunately, KnightHawk would be unable to score a single win in the ALMS season, although they did manage to take second in the LMP675 championship. They were also the only team to enter the 24 Hours of Le Mans, where they unfortunately failed to finish. 2002 would begin to see continued success for the B2K/40. The USRRC series would see only Rand Racing-Risi Competizione as consistent competitors, with Porschehaus and Archangel Motorsports only making occasional appearances. Although Rand-Risi would be amongst the few SR2 class competitors in the series, they were able to score nine wins and take the championship. However the American Le Mans Series would see a boost in competitors Archangel Motorsports moved to the series, joining occasional runners Spencer Motorsports, Essex Racing, and Kyser Racing. Archangel would take two class victories and finish third in the LMP675 championship. SportsRacing Team Sweden would continue to be the only European squad, but would not be able to repeat their previous success, finishing third in the championship without a single win. 2003 began to see a decline in results for the B2K/40. The Grand American Road Racing Association was beginning a shift towards their Daytona Prototype class, and rule restrictions left the SR2 class uncompetitive. This meant a handful of teams participated in only a select number of races. Team Essex would win class at the 24 Hours of Daytona while G&W Motorsports would win at Virginia International Raceway, earning them the championship due to points earned running a Picchio chassis in the class earlier that season. Essex Racing would be the only competitor in the American Le Mans Series, finishing third in the LMP675 championship yet not earning a single win. SportsRacing Team Sweden's B2K/40 would also not return to the FIA Sportscar Championship. 2004 would see the end of the SR2 class in the USRRC as well as the cancellation of the FIA Sportscar Championship, leaving the American Le Mans Series and new Le Mans Endurance Series as the only area of competition for the B2K/40. Miracle Motorsports, Intersport Racing, Marshall Cooke Race Car Company, and Van der Steur Racing would all regularly compete in the ALMS season, joined by Rand Racing for the 12 Hours of Sebring. B2K/40s would win eight races in the LMP2 class, with six for Intersport and three for Miracle, although it would be Miracle Motorsports who would win the championship, aided by their second car, a Courage C65. Tracsport would be the only competitor in the Le Mans Endurance Series, scoring a best finish of third in class and taking third in the championship. Intersport Racing would also go to the 24 Hours of Le Mans and successfully take the class victory, again one of only two LMP2 class competitors to finish. For 2005, a change in regulations for the 24 Hours of Le Mans and Le Mans Endurance Series would make the B2K/40 obsolete, while at the same time the new B05/40 was being launched as its replacement. In the American Le Mans Series, the B2K/40 would still be allowed. Van der Steur Racing would be the only consistent competitor, with BAT Competition appearing for a few select races and Intersport running their car for one last time at the 12 Hours of Sebring before upgrading to a B05/40. The team would finish a distant fourth in the LMP2 championship. Van der Steur, strapped for cash, would continue with their B2K/40 into 2006 at select races before finally upgrading to a Radical SR9. External links Mulsannes Corner - Lola B2K/40 technical analysis World Sports Racing Prototype - Lola Sports Car chassis index B2K 40 Category:Le Mans Prototypes Category:24 Hours of Le Mans race cars Category:Sports prototypes
{ "pile_set_name": "Wikipedia (en)" }
About Bell Media Bell Media is Canada’s leading content creation company with premier assets in television, radio, out-of-home advertising, digital media, and more. Bell Media owns 30 local television stations led by CTV, Canada’s highest-rated television network; 30 specialty channels, including TSN and RDS, and four pay TV services. Bell Media is Canada’s largest radio broadcaster, with 215 music channels including 109 licensed radio stations in 58 markets across the country, all part of the iHeartRadio brand and streaming service. Bell Media owns Astral, an out-of-home advertising network of more than 30,000 faces in five provinces. Bell Media also operates more than 200 websites; video streaming services including Crave, TSN Direct, and RDS Direct; and multi-channel network Much Studios. The company produces live theatrical shows via its partnership with Iconic Entertainment Studios; owns a majority stake in Pinewood Toronto Studios; is a partner in Just for Laughs, the live comedy event and TV producer; and owns Dome Productions Inc., one of North America’s leading production facilities providers. Bell Media is part of BCE Inc. (TSX, NYSE: BCE), Canada’s largest communications company. Learn more at www.BellMedia.ca. About Lionsgate The first major new studio in decades, Lionsgate (NYSE: LGF.A, LGF.B) is a global content leader whose films, television series, digital products and linear and over-the-top platforms reach next generation audiences around the world. In addition to its filmed entertainment leadership, Lionsgate content drives a growing presence in interactive and location-based entertainment, video games, esports and other new entertainment technologies. Lionsgate’s content initiatives are backed by a nearly 17,000-title film and television library and delivered through a global sales and distribution infrastructure. The Lionsgate brand is synonymous with original, daring and ground-breaking content created with special emphasis on the evolving patterns and diverse composition of the Company’s worldwide consumer base.
{ "pile_set_name": "OpenWebText2" }
× Expand Oliver Lang/dapd via AP Photo Missing in action: 13 Middle East experts and zero coverage of Khashoggi In 2016, two years before his assassination and two days after Donald Trump was elected president, the Saudi journalist Jamal Khashoggi was a guest speaker at the Washington Institute for Near East Policy. The think tank employs more than 30 fellows focused exclusively on Middle East affairs and has many more adjunct researchers and prominent trustees who regularly contribute to marquee media. They put out scores of policy insights, reports, and op-eds, and convene events with newsmakers on a weekly basis. But on the anniversary of Khashoggi’s bloodletting, only a single WINEP fellow weighed in. A half-year before Khashoggi’s death, WINEP trustee and longtime Middle East peace-processor Dennis Ross wrote that Washington should “get behind” Mohammed bin Salman, whom he called a “revolutionary crown prince.” Ross, for his part, received $10,000 for writing an anti-Iran op-ed in 2018, from Trump booster Elliott Broidy; only in Washington could such a corrupt revelation not be a career-breaker (though Ross has since returned the money and dissociated himself from Broidy). Having served as Obama’s senior director for Gulf affairs, Ross has written widely about how to address Middle East conflicts. So where was Ross’s op-ed on Wednesday about seeking justice for Khashoggi? This week, as journalists, researchers, and activists came together to mark the anniversary of Khashoggi’s brutal murder at the hands of the Saudi prince’s goons, a conspicuous silence fell over many of Washington’s so-called think tanks—most glaringly, on right-wing institutions that claim to focus on the promotion of democratic values. By willfully ignoring Khashoggi, they displayed a cynicism so profound that they shouldn’t be trusted to provide analysis on anything else. The Foundation for the Defense of Democracies calls itself a “research institute” and claims the mantle of “free nations,” but I couldn’t find any coverage of the heinous assault on human rights and free speech that unfolded in the Saudi consulate in Istanbul. The American Enterprise Institute, a “think tank dedicated to defending human dignity, expanding human potential, and building a freer and safer world,” might as well drop that mission statement if they aren’t interested in the dignity—much less, the life—of a Washington Post contributor. The Heritage Foundation, which has fed the ranks of the Trump administration, boasts a team of foreign-policy analysts who advise the highest rungs of power. But by ignoring the dismemberment of a journalist, they certainly are not living up to their stated goals of “Building an America where freedom, opportunity, prosperity, and civil society flourish.” The hawkish Hudson Institute, which regularly covers Middle East affairs and cheerleads for war with Iran, had no comment on their website or social media. Their commitment to “American leadership and global engagement for a secure, free, and prosperous future,” rings as hollow as hollow can be. And what of the Center for Strategic and International Studies, with its 13 Middle East experts, several of whom focus on “human rights”? Zero coverage of Khashoggi, a lacuna that belies their charge of providing “strategic insights and policy solutions to help decisionmakers chart a course toward a better world.” Your donation keeps this site free and open for all to read. Give what you can... SUPPORT THE PROSPECT These examples stand in startling contrast to how some other, not avowedly “freedom promoting” right-wing institutions have remembered the martyred journalist. In July, the Brookings Institution hosted U.N. Special Rapporteur Agnes Callamard to discuss her landmark report on the Saudi state’s crime. The Washington Post published a special section honoring their late contributor, including analysts from the Brookings Institution, Dartmouth College, the European Council on Foreign Relations, MIT, and more. The Project on Middle East Democracy, along with a dozen human rights organizations and 11 senators and congresspeople, gathered under the banner of Justice for Jamal at the Capitol last week. Fellows from the Wilson Center, the Carnegie Endowment for International Peace, New America, and the U.S. Institute for Peace also weighed in on Khashoggi—a pillar of the Saudi media whom many Washington researchers knew personally. Not all the Khashoggi indifference has been confined to right-wing institutions, however. Consider the case of the Council on Foreign Relations, the clearinghouse of foreign-policy luminaries and international political clout. Last week, on the heels of the United Nations General Assembly, they hosted an event with Saudi Minister of State Adel Al-Jubeir, who as foreign minister last year served as lead propagandist for the Khashoggi cover-up. Where was CFR’s event this week assessing what the Saudi state’s crime means for the future of geopolitics, press freedom, and diplomacy? The only thing I could find on their social media regarding the butchered journalist was a September 17 tweet citing a quote from Elizabeth Warren: “The Saudi government’s role in the brutal murder of Jamal Khashoggi and its repression of its own citizens insults all who respect human rights and calls into question its reliability as a partner.” The rot at these institutions echoes an anecdote from Warren’s book A Fighting Chance, conveying advice she had received from Larry Summers. Warren writes: I had a choice. I could be an insider or I could be an outsider. Outsiders can say whatever they want. But people on the inside don’t listen to them. Insiders, however, get lots of access and a chance to push their ideas. People—powerful people—listen to what they have to say. But insiders also understand one unbreakable rule: They don’t criticize other insiders. I had been warned. It’s now time to warn those think tanks that dismiss or assist in MBS’s cover-up: We notice what you’re saying, and what you overlook.
{ "pile_set_name": "OpenWebText2" }
The city hall spectacle following Mayor Rob Ford's crack-smoking confession became an even more bizarre sideshow Wednesday, with the appearance of a 71-year-old WWE wrestler challenging the mayor to an arm wrestle, and Ford blowing kisses to media as he escorted children through the offices for Take Our Kids to Work Day. Outside the municipal headquarters, protesters wrote chalk messages on a wall urging Ford to resign. Inside, an agitated Ford supporter wearing a red sweatshirt interrupted a councillor's media scrum to give a fiery defence of the mayor. But the appearance of WWE hall-of-famer, The Iron Sheik, carted into the building by wheelchair, brought a truly circus-like atmosphere to city hall. "I want to see Mr. Ford," The Sheik shouted during a media scrum, saying he had "no respect" for the mayor. "I just want to know: Is he a real man, or no?" Aided by an assistant, The Sheik said he would "absolutely" put Ford in his signature "Camel Clutch" chin-lock, given the opportunity. 'What kind of mayor is he?' "The man eats the cheeseburger and smokes crack. What kind of mayor is he?" he went on. The semi-retired wrestler, who is in town for a roast event, extended an invitation for the mayor to meet him for an arm wrestle in Ford's office. "What kind of role model is he for the Toronto city?" he said. The Iron Sheik was asked multiple times by city hall security to keep his voice down. "That's my voice. You want me to change my voice. I can't change it," he replied, before being asked to leave because he was too loud. The Iron Sheik isn't the first pro-wrestling legend to challenge the mayor to a physical feat of strength. In August, Hulk Hogan lost to Ford in an arm wrestling match. Defence of Ford Nation The strange day at city hall was compounded by the fact it was also Take Our Kids to Work Day. At one point, Ford walked by and blew a kiss to a throng of journalists as he took a group of youths on a tour of his offices. Ford Nation supporter Tajinder Bains says voters will continue to back Toronto's embattled mayor, even in light of his crack cocaine scandal. (CBC) Coun. Mary-Margaret McMahon said Ford was setting a poor example. "As a role model, I have some serious concerns with that. Hundreds of emails from residents, especially parents who have to explain to kids as young as six years old what crack is," she said. The mayor, who refused on Tuesday to resign or take a leave of absence — even after admitting he used crack cocaine "probably in one of my drunken stupors" — continues to retain impassioned public support. Speaking to media gathered outside the mayor's office, Tajinder Bains refuted the idea that Ford Nation was dead in the face of the mayor's scandals. "Ford Nation is every single mother out there who is having a hard time paying their bills, every senior citizen who can’t afford groceries because inflation is too high. Every dad working two jobs to put his kids through a good school," he said. "That’s who Ford Nation is and you’re saying you’re going to get rid of them? I doubt that, pal. I doubt that."
{ "pile_set_name": "OpenWebText2" }
Richard Angelo Richard Angelo (born August 29, 1962) is an American serial killer and former nurse at the Good Samaritan Hospital in West Islip, New York. In 1989, he was convicted of murdering several of his patients and sentenced to 50 years to life in prison. Early life Angelo was born on August 29, 1962, to parents who were both working in the educational sector. His mother was an economics teacher, and his father was a high school guidance counselor for the Lindenhurst school district on Long Island. He graduated from St. John the Baptist Diocesan High School in 1980 and then entered a two-year nursing program at Farmingdale State College, where he was a well-regarded honor student. Arrest Angelo first came to the attention of the public in October 1987 when he was suspected of poisoning a patient at the Good Samaritan Hospital on Long Island, where he worked as a nurse. He was accused of injecting Gerolamo Cucich with a drug into his I.V. Cucich did not feel well after the injection, so he paged another nurse to help him. Later it was found that he had Pavulon in his urine, which he had not been prescribed. He was arrested for assault on the 73-year-old patient because he was the only person to match the description given to the police (a "heavyset man with a dark beard and glasses"). Following his arrest, he confessed to having poisoned other patients with Pavulon and Anectine. As a result, as many as 30 recently deceased patients were exhumed and examined for traces of these powerful paralyzing agents. It was later concluded that he had poisoned at least 35 people at the hospital while working there for seven months. The poisoning resulted in ten deaths. He claimed that his motive was to portray himself as a hero. After poisoning his victims, he would wait until they went into cardiac arrest and then come by and save them in front of his colleagues. He was held in Suffolk County Jail for over a year, awaiting trial. He declined to pay his $50,000 bail, fearing for his safety — given the high-profile nature of the case. Conviction In December 1989, Angelo was found guilty on two counts of murder, one count of manslaughter, and one count of criminally negligent homicide. He was also convicted of assault in connection with the deaths of four other patients, and he was suspected of being responsible for several other deaths. On January 25, 1990, he was sentenced to 50-years-to-life in prison. He was 27 years old. He is serving his sentence in the Great Meadow Correctional Facility. See also List of serial killers in the United States References External links New York State - Department of Corrections and Community Supervision >Inmate Lookup Richard Angelo - Department Identification Number (DIN): 90A2242 Category:1962 births Category:Living people Category:American serial killers Category:Male serial killers Category:American people convicted of murder Category:American people convicted of manslaughter Category:Nurses convicted of killing patients Category:Male nurses Category:People convicted of murder by New York (state) Category:American people convicted of assault Category:American nurses Category:Prisoners sentenced to life imprisonment by New York (state) Category:American prisoners sentenced to life imprisonment Category:Criminals from New York City Category:20th-century American criminals Category:American male criminals Category:St. John the Baptist Diocesan High School alumni
{ "pile_set_name": "Wikipedia (en)" }
This invention relates to improvement in gear assemblies, and in particular to electrical power assisted steering assemblies which incorporate a worm and wheel gear assembly for transferring torque from an electric motor to a steering column or output shaft operatively connected thereto. It is known to provide a power steering system for a vehicle comprising an electric motor having a stator and a rotor, an input shaft operatively connected to the rotor and adapted to rotate therewith, an output shaft associated with a steering column, and a gearbox adapted to transfer torque from the input shaft to the output shaft in response to a measure of torque in the output shaft produced by a torque sensor. The motor is typically operated to apply an increasing torque to the output shaft as the measured torque increases, thus applying an assistance torque which helps to steer the vehicle. In a simple arrangement, the input shaft defines a worm gear, and the output shaft is provided with a wheel gear which is adapted to mesh with the worm gear. Whilst such a system is relatively effective, there exists a problem with noise and vibration due to incorrect meshing between the worm and gear wheel. This incorrect meshing may arise due to manufacturing tolerances, thermal changes in dimensions, distortion due to torsional loads and wear during service. According to a first aspect, the invention provides an electric power assisted steering system comprising a housing, an electric motor fixed relative to the housing having a stator and a rotor, an input shaft operatively connected to the rotor, an output shaft operatively connected to a steering column, and a torque sensor adapted to produce an output signal indicative of the torque in the output shaft, the motor being adapted to apply a torque to the output shaft dependent upon the output signal from the torque sensor through a worm gear provided on the input shaft which is adapted to mesh with a wheel gear on the output shaft, the steering system being characterised by further comprising a first bearing means which supports the input shaft relative to the housing at its end distal from the motor and a resilient biasing means adapted to act upon the first bearing means to bias the input shaft towards the wheel gear. Preferably, the input shaft is biased in a tilting movement which is centred at a second bearing means which supports the input shaft relative to the housing at its end adjacent to the motor. The biasing means may be adapted to apply a sufficient biasing force to the first bearing means to maintain a fully meshed engagement between the teeth of the worm gear and the teeth of the wheel gear over a predetermined range of torque values carried by the wheel gear. This helps to prevent gear rattle when driving straight ahead or on rough roads by ensuring both sides of the engaging teeth on the worm and wheel are in contact at substantially all times over this range of torques. Because the arrangement increases quiescent friction in the gearbox it is important to maintain control of the force applied by the biasing means over the full range of the input shaft that is required. Therefore the biasing means must have a low spring rate. The provision of the biasing means allows a controlled biasing force to be applied whilst permitting sufficient tilting movement of the input shaft to compensate for changes in dimensions due to manufacturing variations and temperature changes etc. The maximum torque value up to which the fully meshed engagement is effective is carefully chosen (by compromise) to avoid excessive friction. The biasing means may comprise a resilient spring of any type adapted to act between a portion of the housing and the first bearing means. In some configurations, it is preferred that the resilient spring comprises a leaf spring which may be attached to the housing at a first end and act upon the first bearing means at its second end. This may engage the first bearing means at the opposite side of the input shaft to the wheel gear so as to bias the worm into contact with the wheel gear. The leaf spring may be provided outside of the housing and the second end of the leaf spring may pass through an opening in the housing to engage with the first bearing means. The second end of the leaf spring may carry a seal which seals with the opening through which it passes. The input shaft may be directly connected to the motor rotor. It may extend continuously through the rotor. The input shaft may be operatively connected to the rotor through a flexible coupling which allows the worm to tilt without movement of the rotor. The flexible coupling may comprise a resilient element, for example of rubber. The motor rotor may be adapted to apply a drive force to the resilient element through one or more circumferentially spaced radially extending surfaces of the resilient element. The resilient element may in turn be adapted to apply a drive force to the input shaft through one or more other circumferentially spaced radially extending surfaces of the element. The element may comprise a spider shape having a multiple of arms presenting a number of radially extending, circumferentially spaced drive surfaces. Where the input shaft is connected to the motor rotor by a flexible coupling, a first compression means may be provided between the housing and the first bearing means at the end of the input shaft distal from the motor which applies a compressive force onto the input shaft to bias it towards the motor rotor. It may comprise a coil spring. Its function is to prevent noise and vibration due to axial free play in the second bearing means. In its compressed state, there should be free space between adjacent coils in order to avoid frictional resistance to tees tilting motion of the input shaft. A second compression means (such as a coil spring) may also be provided between the end of the input shaft adjacent the motor rotor and the motor rotor. This may be provided in a cup formed on an end of the input shaft whilst a pin extending about the rotational axis of the motor rotor and forming a part of the rotor projects into the cup to engage the spring. Thus, whilst the first compression means biases the second bearing means through the flexible coupling, the second compression means biases the rotor directly through the pin. There must be a difference between the forces provided respectively by the first and second compression means which is sufficient to bias the second bearing means in the direction of the input shaft axis by the desired amount. In an alternative, the biasing means may comprise an annular O-ring provided between the first bearing means and the housing. The O-ring may be of rubber, and may contact an outer circumference of the first bearing means and a portion of the housing. The first bearing means may therefore move relative to the housing against a resistant force applied by the O-ring as at least part of the O-ring is compressed. In such an arrangement, the wheeled gear or the worm gear may deliberately be made slightly oversize relative to the dimensions which are calculated according to the distance between their respective shaft axes. This ensures the O-ring is always under a small amount of residual compression. In an alternative to an O-ring, the resilient biasing means may comprise a resilient element which is accommodated between the first bearing means and the housing, such as a rubber spacer block. The element may be disposed between the first bearing means and the housing, opposite to the side of the input shaft which engages the gear wheel. The biasing means may act in compression or tension. The second bearing means may comprise a ball bearing which is adapted to prevent radial and axial movement of the input shaft relative to the housing whilst permitting tilting movement of the input shaft against the bias force provided by the biasing means. The second bearing means may be selected to comprise a high tolerance ball bearing assembly which is adapted by virtue of the shape of the groove in which the balls are located to substantially prevent any radial displacement of the input shaft relative to the housing as it passes through the second bearing whilst permitting the input shaft to pivot about a point on its axis which passes through a point in the plane of the second bearing means. In a preferred arrangement, the housing defines a first portion and a second portion, the first portion comprising a housing for the input shaft having at least one pair of opposed walls, and having an opening in each wall into which the first and second bearing means are respectively provided, and the second portion comprising a housing for at least part of the output shaft having at least one pair of opposed walls, an opening being provided in each wall for receiving one or more bearings adapted to secure the output shaft relative to the housing. The output shaft is preferably mounted orthogonal to the input shaft and substantially prevented from moving radially relative to the housing. A plastic lining portion may be provided around a circumferential outer face of the first bearing means which prevents contact between the first bearing means and the housing at excessive displacement. This helps to eliminate vibration noise due to metalxe2x80x94metal contact between the first bearing means and the housing. The first opening defined in the first portion of the housing (which receives the first bearing means) may comprise an elongated slot through which the input shaft passes having semicircular end portions and a central pair of parallel sides. The spacing between the parallel sides may be substantially the same as the outer diameter of the firs, bearing means. The radius of the semicircular end portions of the slot may be substantially the same as the outer radius of the first bearing means. Thus, the first bearing means may be adapted to move axially along the slot but may be prevented from moving radially perpendicular to the slot. The first opening of the first portion may comprise an annular bore having an inner surface with a diameter greater than the diameter of the first bearing means. The outer surface of the first bearing means may thus be spaced from the inner wall. The O-ring element forming the biasing means may be accommodated in this space. A groove may be provided around a circumference of the inner wall which locates the O-ring in a predetermined position relative to the wall, the depth of the groove being less then the radial thickness of the O-ring element. In this case, the first bearing means may be displaced from its rest position against the biasing force provided by the O-ring through a distance equal to the difference between the O-ring radial thickness and the depth of the groove. Thereafter, the first bearing means is prevented from further radial displacement within the opening as it engages the inner wall of the first opening. In a yet further alternative arrangement, the biasing means; may comprise a torsion bar having a first end acting upon the first bearing means and a second end fixed relative to a portion of the housing so that the torsion bar applies a biasing force against the first bearing means. The torsion bar may comprise a substantially U-shaped elongate rod having a terminal end portion on a first end of the roll which is bent through an angle of approximately ninety degrees relative to the remaining part of the first end and relative to the centre portion of the rod to engage with a portion of the first bearing means. Preferably, the terminal end portion acts directly on an outer surface of the first bearing means opposed to the wheel gear of the output shaft by pas sing through an opening channel in the housing extending radially away from the inner wall of the first opening of the first portion of the housing. The central portion of the torsion bar may be secured to a portion of the housing through one or more clamps or shackles. The second end of the torsion bar may rest upon an end face of a threaded bolt which engages with the housing. Rotation of the bolt within the threaded bore displaces the second end of the torsion bar relative to the housing. As the first end is engaged with the first bearing means this acts to increase or decrease torsion in the bar in a known manner, in turn to alter the biasing force applied to the first bearing means (i.e. for use when setting up). In a preferred arrangement, a terminal portion of the second end of the torsion bar is bent through approximately ninety degrees relative to the remaining part of the second end portion and engages within a recess in the end face of the bolt. This provides a positive location for the second end portion. In yet a further alternative arrangement, where space in the vehicle permits, the biasing means may comprise a coil spring leaving its axis substantially perpendicular to that of the wormshaft. The coil spring could be installed in a hole in the housing. A first end of the spring could apply force to the outer race of the first bearing means via a formed end of the spring or via a separate component placed between the spring and the first bearing means. A closure plug or plate at the end of the hole distal from the bearing means would provide a support to the coil spring and a means of sealing. In a most preferred arrangement, the terminal portion of the second end of the torsion bar engages with a recess in the housing. This renders the arrangement non-adjustable and tamper proof which is preferable for production versions. In a refinement, where the biasing means comprises an O-ring seal acting between the first bearing means and the housing, the centre axis of the O-ring may be offset relative to the central axis of the input shaft. This provides a different relationship between biasing force and displacement of the bearing means compared to the case where the central axes of the O-ring and shaft coincide. It is preferred that the axis of the O-ring is closer to the wheel gear than that of the input shaft where is passes through the first bearing means. To further refine the relationship between the biasing force and displacement of the first bearing means, the shape of the O-ring groove (where provided) relative to the cross-section of the O-ring may be chosen so that the compressed portion of the O-ring just completely fills the groove at a predetermined displacement corresponding to a predetermined biasing force, whereafter the rate of increase of biasing force with full displacement is significantly greater than the rate of increase of biasing force with displacement at displacements below the predetermined displacement. When the O-ring is in its normal position corresponding to zero torque on the gear wheel, the O-ring may therefore only partially fill the groove at this point. According to a second aspect, the invention provides an electric power assisted steering system comprising a housing, an electric motor fixed relative to the housing having a stator and a rotor, and input shaft operatively connected to the rotor, an output shaft operatively connected to a steering column, and a torque sensor adapted to produce an output signal indicative of the torque in the output shaft, the motor being adapted to apply a torque to the output shaft dependent upon the output signal from the torque sensor through a worm gear provided on the input shaft which is adapted to mesh with a wheel gear on the output shaft, the steering system being characterised in that the input shaft is operatively connected to the motor rotor by a flexible coupling. By providing a flexible coupling, it is possible for the input shaft to tilt relative to the motor rotor. This enables the position of the worm relative to the wheel to be adjusted to remove gear rattle without having to move the rotor. Preferably, the flexible coupling comprises a resilient element. It may be a rubber element. The element may have a plurality of substantially radially extending drive faces. One or more of the drive faces may co-operate with one or more radial drive faces defined on the rotor. One or more of the drive faces may co-operate with drive faces defined on the input shaft. The motor rotor thus can apply torque to the input shaft through these faces. Preferably, the element comprises a spider having a plurality of radially extending arms defining the drive surfaces. The input shaft may have a cup formed on its end adjacent the motor rotor. A pin located along the axis of rotation of the rotor may be adapted to be received within the cup. The flexible coupling may be provided between an end face of the cup and the rotor, perhaps around the pin. The cup may be adapted to receive a first resilient biasing element such as a spring which acts between the end of the pin and the bast of the cup to bias the rotor away from the input shaft. A second compression means may be provided which is adapted to bias the input shaft towards the rotor. This may comprise a spring located between the housing and a bearing means which supports the input shaft.
{ "pile_set_name": "USPTO Backgrounds" }
729 F.2d 1449 U.S.v.Carducci (Hugo J.) NO. 83-5480 United States Court of Appeals,third Circuit. FEB 09, 1984 1 Appeal From: W.D.Pa., 566 F.Supp. 1116 2 AFFIRMED.
{ "pile_set_name": "FreeLaw" }
Subscribe free newsletter to get latest products and discount information. Links What Type Of Solvent To Be Used For Maleic Resin? Maleic resin is a type of modified rosin that made by maleic anhydride, and then with the esterification of polyols from, maleic resin dissolved in benzene, esters and ketones and other solvents, insoluble in alcohol solvents, some soluble in petroleum Solvent, good compatibility with vegetable oil; This product has low light color, not yellowing, good thermal stability and adhesion, etc .; can improve the film gloss, hardness, fullness, abrasion resistance and has Quick-drying.
{ "pile_set_name": "Pile-CC" }
Dear Editor, We reviewed the manuscript recently published by Cumhur Cure M, et al. \[[@CR1]\], regarding colchicine and COVID-19. The authors discussed the effects of colchicine in intracellular and extracellular pH conditions and argued that low pH levels secondary to colchicine may increase the viral load of SARS-CoV-2 and, therefore, cytokine storms will be more severe. In contrast, we consider that colchicine can be a good therapeutic option because of several effects in the immunology system involved in SARS-CoV-2 infection and in the acute respiratory distress syndrome (ARDS). Colchicine has effects on the chemotaxis of inflammatory cells such as neutrophils and monocytes and on the intracellular transportation of vesicles such as endosomes and exosomes. Colchicine also inhibits the expression of E-selectin, an adhesion molecule important for binding leukocytes to endothelial cells, and the recruitment of monocytes and neutrophils to inflamed tissue. Finally, colchicine reduces neutrophil production of free radicals like superoxide \[[@CR2]\]. The recruitment of neutrophils in COVID-19 is a key factor in the severity of cases \[[@CR3]\]. Moreover, colchicine has also been associated with disrupting inflammasome activation, thereby suppressing caspase-1 activation and the subsequent release of IL-1β and IL-18 \[[@CR4]\]. In SARS-CoV-1, the inflammasome activation has been associated with this disruption \[[@CR5]\]. The authors also cited an article published in 1986 by Maurizi M, et al. \[[@CR6]\] regarding two patients who developed ARDS after treatment with toxic doses of colchicine. The first patient received approximately 80 mg of colchicine or 1.6 mg/kg, while the other patient received 15 to 20 mg or 0.25--0.3 mg/kg of colchicine for acute gout. Both patients had ARDS between 24 and 72 h after the colchicine doses \[[@CR1]\]. Both articles \[[@CR1], [@CR6]\] attributed a direct toxic action of colchicine on pneumocyte microtubules, and the inhibition of surfactant production was deemed the probable cause of death of those patients. The dose used in those cases is not recommended due the toxicity. The usual adult oral dose in acute gout is 1.2 mg/day or as a prophylactic, 0.5--1 mg/day; however, dose must be adjusted in patients with renal impairment. The high fatality rate was reported after acute ingestions exceeding 0.5 mg/kg \[[@CR7]\], and therefore, there is no support for that outcome in therapeutic doses. Classically colchicine is commonly used to treat different inflammatory diseases such as gout, as well as some autoinflammatory diseases, and cardiac conditions including viral pericardial syndromes. Recently, a patient with cardiac tamponade secondary to COVID-19 was treated with colchicine in addition to corticosteroids and antimalarials with positive clinical response \[[@CR9]\]. Other research demonstrates the positive effect of colchicine in respiratory syncytial virus replication and suppression of secondary airway inflammation given the promoted expression of IFN-α and IFN-β1 of colchicine and regulation of anti-oxidative factor production \[[@CR10], [@CR11]\]. In addition, there is a wide range of preclinical and clinical literature on the effect of colchicine inhibiting viral disease like adenoviral and adeno-associated viral \[[@CR12]\], herpes simplex virus type 1 \[[@CR13]\], Epstein-Barr virus \[[@CR14]\], and hepatitis virus \[[@CR15], [@CR16]\], among others. On the other hand, Cumhur Cure M, et al. \[[@CR1]\] discussed the interactions of colchicine with other drugs. We would agree that colchicine may indeed have interactions with other drugs; therefore, we recommend adjusting the dose in close consideration of the interactions with inhibitors of CYP3A4 as antibiotics and antivirals even though some of them are used in COVID-19 as lopinavir/ritonavir. In our experience, we reported 5 patients (age 38--61 years) with comorbidities (arterial hypertension, type 2 diabetes, among others) in treatment with colchicine for iatrogenic allogenosis 1 to 3 weeks before COVID-19 test positive. They developed mild symptoms such as headache, cough without dyspnea, and arthralgias. It should be noted that some close contacts presented severe symptoms and three of them died \[[@CR8]\]. According to a potential benefit of colchicine in COVID-19 patients, since March 26, 2020, a total of twelve studies have been registered in [www.clinicaltrials.gov](http://www.clinicaltrials.gov) and the European Union Clinical Trials Register considering the clinical utility of this well-known medication. Most of these studies correspond to randomized, open-label, phase 2 clinical trials. It is aspired to include more than 11,000 patients on three continents. The experimental arm in most studies includes an oral colchicine regimen with loading and maintenance doses plus the standard of care for COVID-19. Primary outcomes to evaluate include change in clinical condition (according to the semiquantitative ordinal scale suggested by WHO), requirement for invasive mechanical ventilation/intensive care unit, delta in the score for the Sequential Organ Failure Assessment, length of hospital stay, all-cause mortality, and change in prognostic biomarkers. The evaluation of these outcomes will be carried out in a time frame up to 30 days, and hopefully, their results will be available soon. **Publisher's note** Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. None. Written informed consent for publication of their clinical details and clinical images was obtained from the patient.
{ "pile_set_name": "PubMed Central" }
The 5-nitro drug, metronidazole (Mz), has been a mainstay of antimicrobial therapy for decades. Several of its simple derivatives such as tinidazole combine similar activity profiles with improved pharmacokinetic properties, but resistance to existing nitro drugs is increasing. Although commercial development of this drug class largely ceased decades ago, work by us and others over the last several years has shown that extensive modifications of the basic 5-nitroheterocyclic ring can lead to marked enhancement in activity against different microbes compared to existing drugs. These data suggest that Mz and other approved nitro drugs do not possess optimal activity in this drug class, yet important questions about the potential utility of novel nitro compounds must be addressed to advance their development as next-generation nitro drugs for clinical use: Is it possible to develop improved nitro drugs with broad-spectrum activity, or do enhanced activities exist only in microbe-specific fashion? Do new nitro drugs have different targets that can be exploited for overcoming resistance to existing drugs? What are the optimal pharmacokinetic properties of novel nitro drugs for maximal efficacy and potency against infections with different target microbes? Can new nitro drugs be developed with improved dosing regimens compared to existing drugs? Answers to these questions are not only critical for assessing the therapeutic potential of new nitro drugs, but are also key for identifying new leads for specific indications. The project will address these questions with a focus on two important protozoan pathogens, Trichomonas vaginalis and Giardia lamblia. We will evaluate a newly synthesized library of ~1,200 nitro drugs for activity against a broad range of drug-sensitive and drug-resistant strains of the target protozoa to identify library compounds more potent than Mz. Electrochemical approaches will be employed for determining the redox properties of the most potent nitro compounds to gain new fundamental clues about their mechanisms of action and potential toxicity. Subsequently, we will introduce new structural modifications into the top leads and evaluate them for bioactivity, cytotoxicity, electrochemical characteristics, and propensity to develop new drug resistance. Finally, we will evaluate the most promising nitro compounds for efficacy, potency, and pharmacokinetics in different murine models of protozoal infections. Upon completion of the proposed research, we expect to have elucidated broadly applicable principles that govern optimal efficacy of next-generation nitro-heterocyclic agents in the treatment of the clinically important parasitic diseases trichomoniasis and giardiasis. The comprehensive data sets to be generated will also be instrumental in selecting the most promising candidates as novel leads for the improved treatment of these infections, and potentially infections with other important pathogens, including Entamoeba histolytica, Trypanosoma cruzi, Helicobacter pylori, and Clostridium difficile, which can be treated with nitro antimicrobials.
{ "pile_set_name": "NIH ExPorter" }
DNA recombination and RNA cleavage activities of the Flp protein: roles of two histidine residues in the orientation and activation of the nucleophile for strand cleavage. Using a combination of DNA and hybrid DNA-RNA substrates, we have analyzed the mechanism of phosphoryl transfer by the Flp site-specific recombinase in three different reactions: DNA strand breakage and joining, and two types of RNA cleavage activities. These reactions were then used to characterize Flp variants altered at His309 and His345, amino acid residues that are in close proximity to two key catalytic residues (Arg308 and Tyr343). These histidine residues are important for strand cutting by Tyr343, the active-site nucleophile of Flp, but neither residue contributes to the type II RNA cleavage activity or to the strand-joining reaction in a pre-cleaved substrate. Strand cleavage reactions using small, diffusible nucleophiles indicate that this histidine pair contributes to the correct positioning and activation of Tyr343 within the shared active site of Flp. The implications of these results are evaluated against the recently solved crystal structure of Flp in association with a Holliday junction.
{ "pile_set_name": "PubMed Abstracts" }
Dust abatement Dust abatement refers to the process of inhibiting the creation of excess soil dust, a pollutant that contributes to excess levels of particulate matter. Frequently employed by local governments of arid climates such as those in the Southwest United States, dust abatement procedures may also be required in private construction as a condition of obtaining a building permit. Dust control is most commonly used in vineyards, orchards and logging roads. Dust abatement methods generally fall into four categories. The first two are petroleum-based products, such as emulsified asphalts, but they are considered environmentally hazardous, according to the report. The third category includes such non-petroleum products as lignosulfates, which are a byproduct of the wood pulping industry, but they tend to leach and run off during heavy rains, giving off odors and staining soil. Magnesium Chloride, a by product salt production, is a certified organic product that is the most environmentally friendly. Praedium sells Dustoff. The fourth category, synthetic polymers, are generally stable, durable, do not leach or give off appreciable odors, and have proven to be the most environmentally friendly. Abatement oil (an organic, lubricating and penetrating oil) is most commonly used to remove debris such as dust and asbestos. Application of this product is normally done by lathering onto the surface and then removing with a clean dry cloth. See also Air pollution SoilO Dust Control References Category:Air pollution
{ "pile_set_name": "Wikipedia (en)" }
Using detonation nanodiamond for the specific capture of glycoproteins. We demonstrate here the functionalization of detonation nanodiamond (ND) with aminophenylboronic acid (APBA) for the purpose of targeting the selective capture of glycoproteins from unfractionated protein mixtures. The reacted ND, after blending with the matrix consisting of alpha-cyano-4-hydroxy-cinnamic acid, could be applied directly for matrix-assisted laser desorption ionization (MALDI) assay. A loading capacity of approximately 350 mg of glycoprotein/g of ND could be attained on ND that has been silanized with an alkyl linker chain prior to linking with the phenylboronic acid. The role of the alkyl spacer chain is to form an exclusion shell which suppresses nonspecific binding with nonglycated proteins and to reduce steric hindrance among the bound glycoproteins. In the absence of the alkyl spacer chain, nonselective binding of proteins was obtained. This work demonstrates the usefulness of functionalized ND as a high-efficiency extraction and analysis platform for proteomics research.
{ "pile_set_name": "PubMed Abstracts" }
Batch Search Transcriptional coactivator of the p300/CBP-mediated transcription complex. Enhances SMAD-mediated transcription by strengthening the functional link between the DNA-binding SMAD transcription factors and the p300/CBP transcription coactivator complex. Stimulates estrogen-dependent transactivation activity mediated by estrogen receptors signaling; stabilizes the interaction of estrogen receptor ESR1 and histone acetyltransferase EP300. Positively regulates TGF-beta signaling through its association with the SMAD/p300/CBP-mediated transcriptional coactivator complex. Induces transcription from estrogen-responsive promoters and protection against cell death. Potentiates EGR2-mediated transcriptional activation activity from the ERBB2 promoter. Acts as an inhibitor of osteoblastic mineralization through a cAMP-dependent parathyroid hormone receptor signaling. May play a role in pigmentation of melanocytes. Associates with chromatin to the estrogen-responsive TGF-alpha promoter region in a estrogen-dependent manner. This gene encodes a member of the CREB-binding protein/p300-interacting transactivator with Asp/Glu-rich C-terminal domain (CITED) family of proteins. The encoded protein, also known as melanocyte-specific gene 1, may function as a transcriptional coactivator and may play a role in pigmentation of melanocytes. Alternatively spliced transcript variants have been described. [provided by RefSeq, Jan 2009]
{ "pile_set_name": "Pile-CC" }
(eval-when (:compile-toplevel :load-toplevel :execute) (require 'regexp2)) (defpackage :csv (:use :common-lisp :iterate :parse-number) (:export #:read-csv-file #:read-csv-stream #:write-csv-file #:write-csv-stream #:read-csv-file-and-sort)) (in-package :csv) ;;; helper function (defun parse-number-no-error (string &optional default) (let ((result (ignore-errors (parse-number:parse-number string)))) (if result result default))) ;;; cvs utilities ;;; TODO: extend parsing for quotes and linefeeds (defparameter *csv-separator* #\,) (defparameter *csv-quote* #\") (defparameter *csv-print-quote-p* nil "print \" when the element is a string?") (defparameter *csv-default-external-format* #+allegro :932 #-allegro :sjis) (defun write-csv-line (record &key stream) "Accept a record and print it in one line as a csv record. A record is a sequence of element. A element can be of any type. If record is nil, nothing will be printed. If stream is nil (default case), it will return a string, otherwise it will return nil. For efficiency reason, no intermediate string will be constructed. " (let ((result (with-output-to-string (s) (let ((*standard-output* s) (record-size (length record))) (iter (for e in-sequence record) (for i from 0) (typecase e (string (progn (if *csv-print-quote-p* (progn (write-char *csv-quote*) (write-string e) (write-char *csv-quote*)) (write-string e)))) (t (princ e))) (when (< i (1- record-size)) (write-char *csv-separator*))))))) (format stream "~&~a" result))) (defun write-csv-stream (stream table) "Accept a stream and a table and output the table as csv form to the stream. A table is a sequence of lines. A line is a sequence of elements. Elements can be any types" (iter (for l in-sequence table) (write-csv-line l :stream stream)) (write-char #\newline stream) '(ok)) (defun write-csv-file (filename table &key (external-format *csv-default-external-format*)) "Accept a filename and a table and output the table as csv form to the file. A table is a sequence of lines. A line is a sequence of elements. Elements can be any types" (with-open-file (f filename :direction :output :if-does-not-exist :create :if-exists :supersede :external-format external-format) (write-csv-stream f table))) (defun parse-csv-string (str) ;; refer RFC4180 (coerce ;; (regexp:split-re "," str) (let ((q-count (count *csv-quote* str :test #'char-equal))) (cond ((zerop q-count) (regexp:split-re *csv-separator* str)) ((evenp q-count) (macrolet ((push-f (fld flds) `(push (coerce (reverse ,fld) 'string) ,flds))) (loop with state = :at-first ;; :at-first | :data-nq | :data-q | :q-in-nq | q-in-q with field with fields for chr of-type character across str do (cond ((eq state :at-first) (setf field nil) (cond ((char-equal chr *csv-quote*) (setf state :data-q)) ((char-equal chr *csv-separator*) (push "" fields)) (t (setf state :data-nq) (push chr field)))) ((eq state :data-nq) (cond ((char-equal chr *csv-quote*) (setf state :q-in-nq)) ((char-equal chr *csv-separator*) (push-f field fields) (setf state :at-first)) (t (push chr field)))) ((eq state :q-in-nq) (cond ((char-equal chr *csv-quote*) (error "#\" inside the non quoted field")) ((char-equal chr *csv-separator*) (push-f field fields) (setf state :at-first)) (t (setf state :data-nq) (push chr field)))) ((eq state :data-q) (if (char-equal chr *csv-quote*) (setf state :q-in-q) (push chr field))) ((eq state :q-in-q) (cond ((char-equal chr *csv-quote*) (push chr field) (setf state :data-q)) ((char-equal chr *csv-separator*) (push-f field fields) (setf state :at-first)) (t (error "illegal value ( ~A ) after quotation" chr))))) finally (return (progn (push-f field fields) (reverse fields)))))) (t (error "odd number of \" ( ~A ) in a line." q-count)))) 'vector)) (defun read-csv-line (stream &key type-conv-fns map-fns (start 0) end) "Read one line from stream and return a csv record. A CSV record is a vector of elements. type-conv-fns should be a list of functions. If type-conv-fns is nil (the default case), then all will be treated as string. map-fns is a list of functions of one argument and output one result. each function in it will be applied to the parsed element. If map-fns is nil, then nothing will be applied. start and end specifies how many elements per record will be included. If start or end is negative, it counts from the end. -1 is the last element. " (declare (type (or (simple-array function *) null) type-conv-fns map-fns)) (let* ((line (read-line stream nil nil))) (when line (let* ((strs (parse-csv-string line)) (strs-size (length strs))) (when (< start 0) (setf start (+ start strs-size))) (when (and end (< end 0)) (setf end (+ end strs-size))) (setf strs (subseq strs start end)) (when type-conv-fns (unless (= (length strs) (length type-conv-fns)) (error "Number of type specifier (~a) does not match the number of elements (~a)." (length strs) (length type-conv-fns)))) (when map-fns (unless (= (length strs) (length map-fns)) (error "Number of mapping functions (~a) does not match the number of elements (~a)." (length strs) (length map-fns)))) (let ((result strs)) ;; strs is not needed so we simply overwrite it (when type-conv-fns (setf result (map 'vector #'funcall type-conv-fns result))) (when map-fns (setf result (map 'vector #'funcall map-fns result))) result))))) (defun read-csv-stream (stream &key (header t) type-spec map-fns (start 0) end) "Read from stream until eof and return a csv table. A csv table is a vector of csv records. A csv record is a vector of elements. Type spec should be a list of type specifier (symbols). If the type specifier is nil or t, it will be treated as string. If type-spec is nil (the default case), then all will be treated as string. map-fns is a list of functions of one argument and output one result. each function in it will be applied to the parsed element. If any function in the list is nil or t, it equals to #'identity. If map-fns is nil, then nothing will be applied. start and end specifies how many elements per record will be included. If start or end is negative, it counts from the end. -1 is the last element. " (let ((type-conv-fns (when type-spec (macrolet ((make-num-specifier (specifier) `(lambda (s) (let ((s (parse-number-no-error s s))) (if (numberp s) (funcall ,specifier s) s))))) (map 'vector (lambda (type) (ecase type ((t nil string) #'identity) (number #'(lambda (s) (parse-number-no-error s s))) (float (make-num-specifier #'float)) (single-float (make-num-specifier #'(lambda (s) (coerce s 'single-float)))) (double-float (make-num-specifier #'(lambda (s) (coerce s 'double-float)))) (integer (make-num-specifier #'round)) (pathname #'pathname) (symbol #'intern) (keyword (lambda (s) (intern s :keyword))))) type-spec)))) (map-fns (when map-fns (map 'vector (lambda (fn) (cond ((or (eq fn t) (eq fn nil)) #'identity) ((functionp fn) fn) ((and (symbolp fn) (not (keywordp fn))) (symbol-function fn)) (t (error "~a is not a valid function specifier." fn)))) map-fns))) (header (when header (read-csv-line stream)))) (loop for rec = (read-csv-line stream :type-conv-fns type-conv-fns :map-fns map-fns :start start :end end) while rec collect rec into result finally (return (values (coerce result 'vector) header))))) (defun read-csv-file (filename &key (header t) type-spec map-fns (external-format *csv-default-external-format*) (os :anynl-dos) (start 0) end) "Read from stream until eof and return a csv table. A csv table is a vector of csv records. A csv record is a vector of elements. Type spec should be a list of type specifier (symbols). If the type specifier is nil or t, it will be treated as string. If type-spec is nil (the default case), then all will be treated as string. map-fns is a list of functions of one argument and output one result. each function in it will be applied to the parsed element. If any function in the list is nil or t, it equals to #'identity. If map-fns is nil, then nothing will be applied. external-format (default is shift-jis) is a valid AllegroCL external-format type. OS is a set to eol-convention of the file stream. start and end specifies how many elements per record will be included. If start or end is negative, it counts from the end. -1 is the last element. " (with-open-file (f filename :external-format external-format) #+allegro (setf (excl:eol-convention f) os) (read-csv-stream f :type-spec type-spec :map-fns map-fns :start start :end end :header header))) (defun read-csv-file-and-sort (filename sort-order &key (header t) (order :ascend) type-spec map-fns (external-format *csv-default-external-format*)) (let ((table (read-csv-file filename :header header :type-spec type-spec :map-fns map-fns :external-format external-format))) (loop for i in (reverse sort-order) do (setf table (stable-sort table (ecase order (:ascend #'string<=) (:descend #'string>=)) :key (lambda (rec) (aref rec i)))) finally (return table))))
{ "pile_set_name": "Github" }
1. Field of the Invention The present invention relates to an information processing apparatus having an input device capable of searching for a plurality of designated points, and to a method and storage medium thereof. 2. Description of the Related Art In recent years, computer devices having touch panels have become widespread. Amongst these kinds of computer devices, there are some that can perform processing having simultaneously detected a plurality of points that a user designated with his or her fingers or the like on the touch panel. Furthermore, amongst these kinds of computer devices, processing for changing an enlargement factor of an image that is being displayed in accordance with a distance between two points, when the two points having been designated by the user, are simultaneously detected, has become common. Amongst these, there exist examples in which when the user touches the touch panel using two fingers (his or her index finger and his or her thumb) simultaneously, and reduces the distance between the two points by performing an operation, with the two fingers, as though he or she were pinching the surface of the touch panel, a display image is reduced, and when the two fingers are spread apart, the display image is enlarged. This kind of operation in which it is as though two fingers pinch the screen is commonly known as a pinch operation. This pinch operation is a very intuitive and easy to understand interface operation for the user because it is an operation in which an object is magnified by grabbing it with his or her fingers and spreading it apart, or in which an object is reduced by reducing a space between his or her fingers, as though he or she were handling a physical flexible object. Also, because the center point between the two points pinched by the fingers is treated at the center point for the enlargement or reduction, the user can easily and intuitively designate which part of the image on the screen he or she wishes to perform an enlargement or reduction on with the pinch operation. Also, it is possible to change a scaling factor of the screen in real-time in accordance with a movement amount of the movement of the fingers when the pinch operation is performed. For this reason, there is also an advantage in that the user can quickly set his or her desired scaling factor, without switching to a special setting screen for setting the display scaling factor. In Japanese Patent Laid-Open No. H04-222063, though not the pinch operation, a user designating in order two points and an enlargement or reduction being performed on a rectangle image similar to a rectangle that passes through the two points is described as a method for designating a plurality of points in a display screen and performing an enlargement/reduction of the screen. Normally, with a pinch operation in which a plurality of points are designated on such as screen, an enlargement or reduction of the whole image being displayed is performed. In a case where, due to the enlargement operation, the image cannot be displayed on the screen in its entirety, a part of the image is hidden. In this case, an operation such as a scroll becomes necessary in order to display the image of the part that was hidden. Also, in a case where displayed characters on the screen are small and difficult to read, when enlargement of an image is performed by a pinch operation, because the whole image is enlarged, there are problems such as the essential character part protruding from the screen. For this reason, in a case where displayed characters on the screen are small and difficult to read, there is a desire to enlarge only the character size and not the whole image. However, because normally an enlargement or reduction of the character size is performed from a separate setting screen, an intuitive and immediate operation, such as the enlargement or reduction of the image with a pinch operation cannot be performed with an operation of enlargement or reduction of the character size using the setting screen. In this case, a method can be considered in which a target of the enlargement or reduction due to the pinch operation is switched in accordance with a mode, for example, between being made to be the whole image and being made to be the character size. However, in this case, there is a problem in that because an operation is necessary to switch the mode and so arrangements of operations increase, operability is worsened.
{ "pile_set_name": "USPTO Backgrounds" }
Q: How to divide paper up into smaller pages in ConTeXt? I need to create some cards, of various sizes, cut from larger letter (8.5" x 11") pages. These are for putting in different sized pocket charts for displaying information in a classroom. All are fairly simple, with header (just like on regular pages) plus some large, vertically and horizontally centered text. Here are some sample layouts: These cards measure 8.5" x 5: _____________________ | Header | | | <--page 1 | Centered text | | | |---------------------| | Header | | | <--page 2, but would be printed on same paper | Centered text | | | |_____________________| These cards measure 11"x4" in landscape mode and the bottom 0.5" is discarded: __________________________ | Header | | Centered text | |--------------------------| | Header | | Centered text | |--------------------------| |__________________________| Is there a way to have ConTeXt setup the pages this way? A simple \pagebreak command would send documents to the next "page" while still being on the same physical page when printed, and headers appear on top like normal. Cuttings lines aren't needed, as I can measure on a paper cutter to get the cuts right. A: You could use ConTeXt's imposition mechanism. The examples have to be typeset using context --arrange. \setuparranging[2TOP] \definepapersize [card] [height=5in,width=8.5in] \setuppapersize [card] [letter] \definemakeup [custom] [align=middle,headerstate=start] \setupheadertexts[Header][] \showframe \starttext \startmakeup[custom] Centered text \stopmakeup \startmakeup[custom] Centered text \stopmakeup \stoptext For the second layout just replace in the above \definepapersize [card] [height=4in,width=11in] \setuppapersize [card] [letter,landscape]
{ "pile_set_name": "StackExchange" }
The Institute (2013 film) The Institute is a 2013 documentary film directed by Spencer McCall reconstructing the story of the "Jejune Institute", an alternate reality game set in San Francisco, through interviews with the participants and the creators. The game was produced in 2008 by Oakland-based artist Jeff Hull. Over the course of three years, it enrolled more than 10,000 players who, responding to eccentric flyers plastered all over the city, started the game by receiving their "induction" at the fake headquarters of the Institute, located in an office building in San Francisco's Financial District. References External links Category:2013 films Category:American documentary films Category:American films Category:English-language films Category:Documentary films about San Francisco Category:Alternate reality games Category:Documentary films about fandom Category:2010s documentary films
{ "pile_set_name": "Wikipedia (en)" }
News Premier Stock Car Racing Series Returns to the Ohio Track during the Late May NASCAR Weekend LEXINGTON, Ohio (October 8, 2019) – The ARCA Menards Series will return to Mid-Ohio Sports Car Course for the first time since 1965 when it competes at the 2.258-mile, 13-turn road course on May 30, 2020. The race will be held in conjunction with the annual NASCAR XFINITY Series weekend, which in previous seasons had been scheduled in the month of August. The move to a new late spring date is to accommodate television coverage of the 2020 Summer Olympics. The ARCA Menards Series last raced at Mid-Ohio in 1965, with Springfield, Ohio native Jack Bowsher earning the victory. Since then the series has raced only sporadically on road courses, with a single event on a modified oval at the defunct St. Louis International Raceway in 1986, a pair of races at Heartland Park Topeka in 1991 and 1992, a street circuit event in Des Moines, Iowa in 1994, a single event at Palm Beach International Raceway in 2010, a pair of races at Road America in 2013 and 2017, and nine races at New Jersey Motorsports Park from 2008 through 2016. The Lexington, Ohio natural terrain road course has been home to world class road racing since its inception, and since the 1980s has hosted series such as the NTT IndyCar Series, the IMSA WeatherTech SportsCar Championship, the SCCA Trans Am, plus various amateur and professional motorcycle series sanctioned by the AMA. Stock car racing returned to Mid-Ohio in 2013 when A.J. Allmendinger defeated Michael McDowell in a shootout. Other NASCAR XFINITY Seres winners include 2012 ARCA Menards Series national champion Chris Buescher, Regan Smith, former ARCA winner at Palm Beach Justin Marks, former Indianapolis 500 winner Sam Hornish, 2008 ARCA Menards Series champion Justin Allgaier, and Austin Cindric, who won August’s most recent race. "We have worked diligently with Craig Rust and his team at Mid-Ohio Sports Car Course for several seasons trying to get the ARCA Menards Series back on their schedule, and we're really happy to make it happen for the 2020 season," said ARCA President Ron Drager. "We have a lot of excitement building for 2020 and additions to our schedule like Mid-Ohio Sports Car Course are a big part of it." “We are thrilled to add the ARCA Menards Series to our NASCAR XFINITY Series weekend,” said Craig Rust, president of Mid-Ohio Sports Car Course. “Each year, it’s our objective to showcase a very diverse schedule at Mid-Ohio Sports Car Course and bring the best of motorsports to our fans. ARCA fits that billing and is a great, returning addition to the 2020 schedule.” Mid-Ohio Sports Car Course offers an exceptional family value. At all events, children 12 and under are admitted free when accompanied by a ticketed adult, plus free parking is available. The complete 2020 event schedule will be released later this fall. Visit midohio.com for more information. ABOUT ARCA The Automobile Racing Club of America (ARCA) was founded by John and Mildred Marcum in 1953 and acquired by NASCAR in April 2018. The ARCA Menards Series, now in its 67th season, will sanction the former NASCAR K&N Pro Series East and West in 2020 with the series being rebranded as the ARCA Menards Series East and West. ARCA will continue to sanction weekly races at Toledo Speedway and Flat Rock Speedway as well as with two branded late model tours, the ARCA/CRA Super Series and the ARCA Midwest Tour. ABOUT MID-OHIO SPORTS CAR COURSE A comprehensive motorsports facility in Lexington, Ohio, Mid-Ohio Sports Car Course sits on 330 acres and features a permanent road-racing circuit with two primary track configurations: 2.4-mile, 15-turn or 2.258-mile, 13-turn layout. Located 60 miles north of Columbus and 75 miles south of Cleveland near Mansfield, the natural terrain road course is commonly referred to as the “Most Competitive in the U.S.” and annually hosts a diversity of locally, regionally and nationally sanctioned race events for amateur, club and professional drivers and riders. It is also home to The Mid-Ohio School, featuring over 20 driving and riding courses, for teenagers to professional racers using Honda vehicles. Opened in 1962, Mid-Ohio Sports Car Course is owned and operated by Green Savoree Racing Promotions since 2011, just the raceway’s third private ownership group in its history. For more information, visit midohio.com, ‘like’ its Facebook page @MidOhioSportsCarCourse or follow updates on Twitter @Mid_Ohio and Instagram @officialmidohio. Media Inquiries: Steve Bidlack, Manager Marketing and Communications 419-884-4000 (office) or This email address is being protected from spambots. You need JavaScript enabled to view it.
{ "pile_set_name": "OpenWebText2" }
Campaign Bosses with Best Track Records in RI Governor’s Race See who's behind managing the 2014 RI gubernatorial races -- and how they've fared at the helm in big races to date. As the 2014 Rhode Island gubernatorial candidates continue to set their sights on November, GoLocal took a look at who's managing their operations -- and what their track records have been to date. "It is important to have campaign managers with significant electoral experience. They need to understand media relations, field operations, fundraising, and social media outreach. Those are all vital parts of contemporary campaigns. Knowing how to do each things is crucial for electoral success," said Brookings Institution Vice President and Director of Governance Studies Darrell West. West continued, "Managers with a track record of winning count for a lot because it demonstrates they have put together a strong organization. But close losses also count because it shows these individuals managed competitive races." Seasoned public affairs consultant Cara Cromwell noted as well that losing efforts weren't always in vain. "It sounds like a bad sports cliche, but you learn more on a losing campaign than a winning one simply because in a loss you rethink every decision. When you win, you focus on the outcome, not what you did to get there," said Cromwell. See Win-Loss Percentages BELOW All of the campaigns have seasoned directors at the helm, with varying degrees of experience and success among them. Danny Kedem with Mayor Angel Taveras had previously been with New York mayoral candidate Anthony Wiener, who was involved in texting scandals both while serving in Congress before ultimately resigning, and once again while running for the Big Apple's top post. Eric Hyers with Gina Raimondo had previously overseen successful David Cicilline campaigns among others, and Devin Driscoll with Clay Pell has extensive grassroots and national experience with President Obama -- but to date hasn't managed a statewide candidate campaign. Rounding out the field of Democratic candidates is Todd Giroux, who ran in 2010 as an Independent. Most of the operatives in the 2014 gubernatorial race have extensive experience, to varying degrees of success. Who can win in November? On the Republican side, seasoned political operatives Jeff Britt with Ken Block and consultant Patrick Sweeney with Allan Fung are vying for the primary nod before a run in the general election -- familiar territory to both. Veteran political analyst and pollster Victor Profughi noted however that those managing the campaign can be distinct from a consultant's role -- and might not always been the ones calling the shots. "Campaign managers often do very little of the creative work and analysis associated with the campaign. They tend to manage what others (consultants) tell them to do," said Profughi. "That's why I prefer the term consultant to manager. Many managers are long time candidate friends and/or local politicos chosen as much for friendship and local recognition as for anything else." Staying the course? While the campaigns appear to have leadership teams firmly in place, both West and Profughi weighed on what happens when a campaign opts to change leadership prior to the election, which although hasn't happened in Rhode Island to date, could be something to watch for in the coming months. "Shaking up a campaign operation mid-stream is a big problem because it shows that things are not going well and there are disagreements about vision, strategy, or operational details," said West. "It often is hard to come back with a new team because any new set of managers takes time before he or she can produce results." Profughi offered his thoughts on campaigns calling in a relief pitcher. "Changes frequently take place when a candidate is slipping or having trouble getting the campaign to click with voters. The result record is mixed. In general, campaigns should do a careful job in the initial selection, have a campaign plan in place for the entire campaign, and stick with it.," said Profughi. "If the campaign chemistry is right from the start, by all means don't change. If a change needs to happen--the sooner the better. My experience suggests that changing campaign managers and/or consultants is rarely a good idea." Hyers, who cut his teeth organizing for for now-Senator Kirsten Gillibrand when she first ran for Congress in 2006, worked as a field organizer for John Edwards for President in the 2008 Iowa caucuses, directed a statewide field operation in Wyoming, managed state legislative races in Virginia and New Jersey, and served as the Executive Director of the Connecticut Democratic Party. Hyers scored big while managing campaigns in New Jersey and Rhode Island, but came up short most recently in Massachusetts. Danny Kedem - Campaign Manager, Angel Taveras Anthony Wiener's 2013 New York Mayoral Campaign: QUIT New Haven Mayor John DeStefano's 2011 Campaign: WON Winning Percentage: .500 Kedem joined Wiener's political comeback bid after Weiner had resigned from Congress in a sexting scandal -- only to leave when Wiener once more was caught continuing the habit, spelling his downfall once again. Connecticut Magazine's Jennifer Swift delineated Kedem's campaign history as having successful managed a New Haven mayoral bid, as well as outlining Kedem's early days organizing for Missouri Democrat Claire McCaskill’s U.S. Senate campaign and Hillary Clinton's 2008 run for the presidency. Jeff Britt -- Campaign Consultant, Ken Block 2002 Bruce Bayuk RI General Assembly Write-In Campaign: LOST 2006 RI Governor Donald Carcieri Campaign: WON 2012 Mark Binder RI General Assembly Campaign: LOST Winning Percentage: .333 Seasoned political operative Britt lost close races in Bayuk and Binder, but scored big with Carcieri while working alongside Ken McKay. Following the Bayuk effort, Britt served in the "Carcieri 1" administration to work across the aisle and support GOP efforts statewide. Devin Driscoll -- Campaign Manager, Clay Pell Northeast Regional Director, Organizing for Action Communications Director, Rhode Islanders United for Marriage Winning Percentage: N/A Driscoll has headed up successful grassroots efforts and served on the Obama for America campaign in two separate stints, as Rhode Island state director for the 2012 campaign and as a field organizer in 2008. Managing Pell's run for Governor marks Driscoll's first bid to win a statewide office campaign.
{ "pile_set_name": "Pile-CC" }