text
stringlengths 1
22.8M
|
|---|
Peter Ware Higgs (born 29 May 1929) is an English theoretical physicist, Emeritus Professor at the University of Edinburgh, and Nobel Prize laureate for his work on the mass of subatomic particles.
In the 1960s, Higgs proposed that broken symmetry in electroweak theory could explain the origin of mass of elementary particles in general and of the W and Z bosons in particular. This so-called Higgs mechanism, which was proposed by several physicists besides Higgs at about the same time, predicts the existence of a new particle, the Higgs boson, the detection of which became one of the great goals of physics. On 4 July 2012, CERN announced the discovery of the boson at the Large Hadron Collider. The Higgs mechanism is generally accepted as an important ingredient in the Standard Model of particle physics, without which certain particles would have no mass.
Higgs has been honoured with a number of awards in recognition of his work, including the 1981 Hughes Medal from the Royal Society; the 1984 Rutherford Medal from the Institute of Physics; the 1997 Dirac Medal and Prize for outstanding contributions to theoretical physics from the Institute of Physics; the 1997 High Energy and Particle Physics Prize by the European Physical Society; the 2004 Wolf Prize in Physics; the 2009 Oskar Klein Memorial Lecture medal from the Royal Swedish Academy of Sciences; the 2010 American Physical Society J. J. Sakurai Prize for Theoretical Particle Physics; and a unique Higgs Medal from the Royal Society of Edinburgh in 2012. The discovery of the Higgs boson prompted fellow physicist Stephen Hawking to note that he thought that Higgs should receive the Nobel Prize in Physics for his work, which he finally did, shared with François Englert in 2013. Higgs was appointed to the Order of the Companions of Honour in the 2013 New Year Honours and in 2015 the Royal Society awarded him the Copley Medal, the world's oldest scientific prize.
Early life and education
Higgs was born in the Elswick district of Newcastle upon Tyne, England, to Thomas Ware Higgs (1898–1962) and his wife Gertrude Maude née Coghill (1895–1969). His father worked as a sound engineer for the BBC, and as a result of childhood asthma, together with the family moving around because of his father's job and later World War II, Higgs missed some early schooling and was taught at home. When his father relocated to Bedford, Higgs stayed behind in Bristol with his mother, and was largely raised there. He attended Cotham Grammar School in Bristol from 1941–46, where he was inspired by the work of one of the school's alumni, Paul Dirac, a founder of the field of quantum mechanics.
In 1946, at the age of 17, Higgs moved to City of London School, where he specialised in mathematics, then in 1947 to King's College London where he graduated with a first class honours degree in Physics in 1950 and achieved a master's degree in 1952. He was awarded an 1851 Research Fellowship from the Royal Commission for the Exhibition of 1851, and performed his doctoral research in molecular physics under the supervision of Charles Coulson and Christopher Longuet-Higgins. He was awarded a PhD degree in 1954 with a thesis entitled Some problems in the theory of molecular vibrations from King's College London.
Career and research
After finishing his doctorate, Higgs was appointed a Senior Research Fellow at the University of Edinburgh (1954–56). He then held various posts at Imperial College London, and University College London (where he also became a temporary lecturer in Mathematics). He returned to the University of Edinburgh in 1960 to take up the post of Lecturer at the Tait Institute of Mathematical Physics, allowing him to settle in the city he had enjoyed while hitchhiking to the Western Highlands as a student in 1949. He was promoted to Reader, became a Fellow of the Royal Society of Edinburgh (FRSE) in 1974 and was promoted to a Personal Chair of Theoretical Physics in 1980. He retired in 1996 and became Emeritus professor at the University of Edinburgh.
Higgs was elected Fellow of the Royal Society (FRS) in 1983 and Fellow of the Institute of Physics (FInstP) in 1991. He was awarded the Rutherford Medal and Prize in 1984. He received an honorary degree from the University of Bristol in 1997. In 2008 he received an Honorary Fellowship from Swansea University for his work in particle physics.
At Edinburgh Higgs first became interested in mass, developing the idea that particles – massless when the universe began – acquired mass a fraction of a second later as a result of interacting with a theoretical field (which became known as the Higgs field). Higgs postulated that this field permeates space, giving mass to all elementary subatomic particles that interact with it.
The Higgs mechanism postulates the existence of the Higgs field which confers mass on quarks and leptons. However this causes only a tiny portion of the masses of other subatomic particles, such as protons and neutrons. In these, gluons that bind quarks together confer most of the particle mass.
The original basis of Higgs' work came from the Japanese-born theorist and Nobel Prize laureate Yoichiro Nambu from the University of Chicago. Professor Nambu had proposed a theory known as spontaneous symmetry breaking based on what was known to happen in superconductivity in condensed matter; however, the theory predicted massless particles (the Goldstone's theorem), a clearly incorrect prediction.
Higgs is reported to have developed the fundamentals of his theory after returning to his Edinburgh New Town apartment from a failed weekend camping trip to the Highlands. He stated that there was no "eureka moment" in the development of the theory. He wrote a short paper exploiting a loophole in Goldstone's theorem (massless Goldstone particles need not occur when local symmetry is spontaneously broken in a relativistic theory) and published it in Physics Letters, a European physics journal edited at CERN, in Switzerland, in 1964.
Higgs wrote a second paper describing a theoretical model (now called the Higgs mechanism), but the paper was rejected (the editors of Physics Letters judged it "of no obvious relevance to physics"). Higgs wrote an extra paragraph and sent his paper to Physical Review Letters, another leading physics journal, which published it later in 1964. This paper predicted a new massive spin-zero boson (now known as the Higgs boson).
Other physicists, Robert Brout and François Englert and Gerald Guralnik, C. R. Hagen and Tom Kibble
had reached similar conclusions about the same time. In the published version Higgs quotes Brout and Englert and the third paper quotes the previous ones. The three papers written on this boson discovery by Higgs, Guralnik, Hagen, Kibble, Brout, and Englert were each recognised as milestone papers by Physical Review Letters 50th anniversary celebration. While each of these famous papers took similar approaches, the contributions and differences between the 1964 PRL symmetry breaking papers are noteworthy. The mechanism had been proposed in 1962 by Philip Anderson although he did not include a crucial relativistic model.
On 4 July 2012, CERN announced the ATLAS and Compact Muon Solenoid (CMS) experiments had seen strong indications for the presence of a new particle, which could be the Higgs boson, in the mass region around 126 gigaelectronvolts (GeV).
Speaking at the seminar in Geneva, Higgs commented "It's really an incredible thing that it's happened in my lifetime." Ironically, this probable confirmation of the Higgs boson was made at the same place where the editor of Physics Letters rejected Higgs' paper.
Awards and honours
Higgs has received numerous accolades including:
Civic awards
Higgs was the recipient of the Edinburgh Award for 2011. He is the fifth person to receive the Award, which was established in 2007 by the City of Edinburgh Council to honour an outstanding individual who has made a positive impact on the city and gained national and international recognition for Edinburgh.
Higgs was presented with an engraved loving cup by the Rt Hon George Grubb, Lord Provost of Edinburgh, in a ceremony held at the City Chambers on Friday 24 February 2012. The event also marked the unveiling of his handprints in the City Chambers quadrangle, where they had been engraved in Caithness stone alongside those of previous Edinburgh Award recipients.
Higgs was awarded with the Freedom of the City of Bristol in July 2013. In April 2014, he was also awarded the Freedom of the City of Newcastle upon Tyne. He was also honoured with a brass plaque installed on the Newcastle Quayside as part of the Newcastle Gateshead Initiative Local Heroes Walk of Fame.
Higgs Centre for Theoretical Physics
On 6 July 2012, Edinburgh University announced a new centre named after Professor Higgs to support future research in theoretical physics. The Higgs Centre for Theoretical Physics brings together scientists from around the world to seek "a deeper understanding of how the universe works". The centre is currently based within the James Clerk Maxwell Building, home of the University's School of Physics and Astronomy and the iGEM 2015 team (ClassAfiED). The university has also established a chair of theoretical physics in the name of Peter Higgs.
Nobel Prize in Physics
On 8 October 2013, it was announced that Higgs and François Englert would share the 2013 Nobel Prize in Physics "for the theoretical discovery of a mechanism that contributes to our understanding of the origin of mass of subatomic particles", and which recently was confirmed through the discovery of the predicted fundamental particle, by the ATLAS and CMS experiments at CERN’s Large Hadron Collider". Higgs admits he had gone out to avoid the media attention so he was informed he had been awarded the prize by an ex-neighbour on his way home, since he did not have a mobile phone.
Companion of Honour
Higgs turned down a knighthood in 1999, but in 2012 he accepted membership of The Order of the Companion of Honour. He later said that he only accepted the order because he was wrongly assured that the award was the gift of the Queen alone. He also expressed cynicism towards the honours system, and the way the system "is used for political purposes by the government in power". The order confers no title or precedence, but recipients of the order are entitled to use the post-nominal letters . In the same interview he also stated that when people ask what the after his name stands for, he replies "it means I'm an honorary Swiss." He received the order from the Queen at an investiture at Holyrood House on 1 July 2014.
Honorary Degrees
Higgs has been awarded honorary degrees from the following institutions:
DSc University of Bristol 1997
DSc University of Edinburgh 1998
DSc University of Glasgow 2002
DSc Swansea University 2008
DSc King's College London 2009
DSc University College London 2010
DSc University of Cambridge 2012
DSc Heriot-Watt University 2012
PhD SISSA, Trieste 2013
DSc University of Durham 2013
DSc University of Manchester 2013
DSc University of St Andrews 2014
DSc Free University of Brussels (ULB) 2014
DSc University of North Carolina at Chapel Hill 2015
DSc Queen's University Belfast 2015
ScD Trinity College Dublin 2016
A portrait of Higgs was painted by Ken Currie in 2008. Commissioned by the University of Edinburgh, it was unveiled on 3 April 2009 and hangs in the entrance of the James Clerk Maxwell Building of the School of Physics and Astronomy and the School of Mathematics. A large portrait by Lucinda Mackay is in the collection of the Scottish National Portrait Gallery in Edinburgh. Another portrait of Higgs by the same artist hangs in the birthplace of James Clerk Maxwell in Edinburgh, Higgs is the Honorary Patron of the James Clerk Maxwell Foundation. A portrait by Victoria Crowe was commissioned by the Royal Society of Edinburgh and unveiled in 2013.
Personal life and political views
Higgs married Jody Williamson, a fellow activist with the Campaign for Nuclear Disarmament (CND) in 1963. Their first son was born in August 1965. Higgs's family includes two sons: Chris, a computer scientist, and Jonny, a jazz musician. He has two grandchildren. The entire family lives in Edinburgh.
Higgs was an activist in the CND while in London and later in Edinburgh, but resigned his membership when the group extended its remit from campaigning against nuclear weapons to campaigning against nuclear power too. He was a Greenpeace member until the group opposed genetically modified organisms.
Higgs was awarded the 2004 Wolf Prize in Physics (sharing it with Robert Brout and François Englert), but he refused to fly to Jerusalem to receive the award because it was a state occasion attended by the then president of Israel, Moshe Katsav, and Higgs was opposed to Israel's actions in Palestine.
Higgs was actively involved in the Edinburgh University branch of the Association of University Teachers, through which he agitated for greater staff involvement in the management of the physics department.
Higgs is an atheist. He has described Richard Dawkins as having adopted a "fundamentalist" view of non-atheists. Higgs has expressed displeasure with the nickname the "God particle". Although it has been reported that he believes the term "might offend people who are religious", Higgs has stated that this is not the case, lamenting the letters he has received which claim the God particle was predicted in the Torah, the Qur'an and Buddhist scriptures. In a 2013 interview with Decca Aitkenhead, Higgs was quoted as saying:
Usually this nickname for the Higgs boson is attributed to Leon Lederman, the author of the book The God Particle: If the Universe Is the Answer, What Is the Question?, but the name is the result of the suggestion of Lederman's publisher: Lederman had originally intended to refer to it as the "goddamn particle".
References
Further reading
External links
Google Scholar List of Papers by PW Higgs
BBC profile of Peter Higgs
The god of small things – An interview with Peter Higgs in The Guardian
My Life as a Boson – A Lecture by Peter Higgs available in various formats
Physical Review Letters – 50th Anniversary Milestone Papers
In CERN Courier, Steven Weinberg reflects on spontaneous symmetry breaking
Physics World, Introducing the little Higgs
Englert-Brout-Higgs-Guralnik-Hagen-Kibble Mechanism on Scholarpedia
History of Englert-Brout-Higgs-Guralnik-Hagen-Kibble Mechanism on Scholarpedia
Sakurai Prize Videos
«I wish they hadn't dubbed it "The God Particle"» Interview with Peter Higgs
Peter Higgs: I wouldn't be productive enough for today's academic system
including the Nobel Lecture on 8 December 2013 Evading the Goldstone Theorem
1929 births
Academics of University College London
Academics of the University of Edinburgh
Alumni of the University of London
Alumni of King's College London
Living people
British Nobel laureates
British theoretical physicists
English atheists
English Nobel laureates
English physicists
English people of Scottish descent
Fellows of King's College London
Fellows of the Institute of Physics
Fellows of the Royal Society
Fellows of the Royal Society of Edinburgh
Nobel laureates in Physics
Particle physicists
People educated at the City of London School
People from Wallsend
Theoretical physicists
Wolf Prize in Physics laureates
Members of the Order of the Companions of Honour
J. J. Sakurai Prize for Theoretical Particle Physics recipients
Recipients of the Copley Medal
20th-century atheists
21st-century atheists
|
```xml
<resources>
<string name="topCenterTextMessage">Place the tag under the phone</string>
<string name="warn_key_download">Do NOT load keys from Download folder.</string>
<string name="fixed_key_found">«locked-secret» stored / loaded.</string>
<string name="fixed_key_missing">«locked-secret» missing!</string>
<string name="unfixed_key_found">«unfixed-info» stored / loaded.</string>
<string name="unfixed_key_missing">«unfixed-info» missing!</string>
<string name="nfc_unsupported">NFC support was not found!</string>
<string name="nfc_disabled">NFC is currently disabled!</string>
<string name="nfc_available">NFC is currently disabled. Enable NFC?</string>
<string name="no_decrypt_key">Decryption keys not imported</string>
<string name="filter">amiibo Filters</string>
<string name="sorting">Sorting Options</string>
<string name="display_options">Display Options</string>
<string name="amiibo_id">ID</string>
<string name="amiibo_name">Name</string>
<string name="amiibo_titles">Game Usage</string>
<string name="amiibo_game">Game Series</string>
<string name="amiibo_character">Character</string>
<string name="amiibo_series">amiibo Series</string>
<string name="amiibo_type">amiibo Type</string>
<string name="amiibo_path">File Path</string>
<string name="pref_amiibo_characters">Characters</string>
<string name="pref_amiibo_types">Amiibo Types</string>
<string name="pref_amiibo_titles">Enabled Games</string>
<string name="amiibo_name_caps">NAME</string>
<string name="amiibo_id_caps">TAGID</string>
<string name="amiibo_type_caps">AMIIBO TYPE</string>
<string name="amiibo_titles_caps">GAME USAGE</string>
<string name="amiibo_game_caps">GAME SERIES</string>
<string name="amiibo_series_caps">AMIIBO SERIES</string>
<string name="amiibo_character_caps">CHARACTER</string>
<string name="amiibo_path_caps">PATH</string>
<string name="recursive_folders">Recursive Search</string>
<string name="compact">Snappy Cards</string>
<string name="large">Banner Cards</string>
<string name="image">Icon Tile Grid</string>
<string name="decryption_keys">Decryption Keys</string>
<string name="yes">Yes</string>
<string name="no">No</string>
<string name="search">Search</string>
<string name="refresh">Refresh</string>
<string name="close">Close</string>
<string name="view">View</string>
<string name="proceed">Continue</string>
<string name="unlock">Unlock</string>
<string name="scan">Scan</string>
<string name="qr_code">QR Code</string>
<string name="upload">GATT</string>
<string name="browser">Browser</string>
<string name="retry">Retry</string>
<string name="save">Save</string>
<string name="edit">Edit</string>
<string name="export">Export</string>
<string name="cache">Cache</string>
<string name="write">Write</string>
<string name="erase">Erase</string>
<string name="cancel">Cancel</string>
<string name="dismiss">Dismiss</string>
<string name="accept">Accept</string>
<string name="submit">Submit</string>
<string name="deny">Deny</string>
<string name="start">Start</string>
<string name="delete">Delete</string>
<string name="activate">Activate</string>
<string name="download">Download</string>
<string name="parent">Parent</string>
<string name="unknown">Unknown</string>
<string name="invalid">Invalid</string>
<string name="unspecified">Unspecified</string>
<string name="install">Install</string>
<string name="sync">Sync</string>
<string name="enable">Enable</string>
<string name="replace">Replace</string>
<string name="update">Update</string>
<string name="error_caps">ERROR</string>
<string name="uid" translatable="false">UID</string>
<string name="country_code">Country Code</string>
<string name="init_date">Initialized Date</string>
<string name="mod_date">Modified Date</string>
<string name="nickname">Nickname</string>
<string name="write_count">Write Count</string>
<string name="mod_serial">Serial Number</string>
<string name="view_hex">View Hex Code</string>
<string name="hex_code">Hex Code</string>
<string name="share_qr">Share QR Code</string>
<string name="edit_props">Edit Properties</string>
<string name="bin_random">Bin Randomizer</string>
<string name="download_slot">Download Slot</string>
<string name="share_export">Share / Export</string>
<string name="flipper_save">Save as Flipper</string>
<string name="delete_binary">Delete Tag File</string>
<string name="ignore_tag_id">Ignore Tag ID #</string>
<string name="random_serial">Random Serial Number</string>
<string name="inject_game_data">Inject Save Data</string>
<string name="unlock_sparkle_cards">Unlock Sparkle Cards</string>
<string name="export_app_data">Export</string>
<string name="import_app_data">Import</string>
<string name="format_app_data">Format</string>
<string name="game_titles_view">View Compatible Games</string>
<string name="game_titles_hide">Hide Compatible Games</string>
<string name="export_flipper_files">Export as Flipper Files</string>
<string name="games_ds">3DS Compatibility</string>
<string name="games_wiiu">WiiU Compatibility</string>
<string name="games_nx">Switch Compatibility</string>
<string name="no_games_ds">No 3DS compatibility</string>
<string name="no_games_wiiu">No WiiU compatibility</string>
<string name="no_games_nx">No Switch compatibility</string>
<string name="properties">Properties</string>
<string name="mii" translatable="false">Mii</string>
<string name="mii_name">Mii Name</string>
<string name="mii_author">Mii Author</string>
<string name="app">Application</string>
<string name="app_id">App ID</string>
<string name="app_data">App Data</string>
<string name="hearts">Hearts</string>
<string name="level">Level</string>
<string name="level_cpu">CPU Level</string>
<string name="appearance">Appearance</string>
<string name="gifts">Gifts</string>
<string name="special_neutral">Special Neutral</string>
<string name="special_side">Special Side</string>
<string name="special_up">Special Up</string>
<string name="special_down">Special Down</string>
<string name="attack">Attack</string>
<string name="defense">Defense</string>
<string name="speed">Speed</string>
<string name="bonus_one">Bonus Effect 1</string>
<string name="bonus_two">Bonus Effect 2</string>
<string name="bonus_three">Bonus Effect 2</string>
<string name="report_problem">Report Problem</string>
<string name="logcat_crash">ERROR! Submit a log?</string>
<string name="send_donation">Send a Donation</string>
<string name="settings_donate">Support development with a donation</string>
<string name="google_link">Donate with Google Play</string>
<string name="sponsor_link">Sponsor on GitHub</string>
<string name="paypal_link">Donate with PayPal</string>
<string name="cancel_link">Cancel Subscription</string>
<string name="subscription_terms">By subscribing to monthly donations, a special banner will appear in settings for the duration of your subscription.\n\nYou may cancel your subscription at any time by using the button on this donation page or through Google Play.</string>
<string name="clear_foomiibo_set">Clear Foomiibo</string>
<string name="build_foomiibo_set">Build Foomiibo</string>
<string name="foomiibo_progress">Building %1$s…</string>
<string name="foomiibo_removing">Clearing %1$s…</string>
<string name="wrote_file">%1$s saved!</string>
<string name="wrote_foomiibo">%1$s Foomiibo saved!</string>
<string name="delete_file">%1$s deleted!</string>
<string name="delete_foomiibo">%1$s Foomiibo deleted!</string>
<string name="nfc_tag">NFC tag</string>
<string name="tag_detected">%1$s detected…</string>
<string name="tag_scanning">%1$s scanning…</string>
<string name="bank_writing">Writing bank %1$d of %2$d…</string>
<string name="bank_erasing">Erasing bank %1$d of %2$d…</string>
<string name="bank_select">Selecting a destination bank…</string>
<string name="chibi_robo">Chibi-Robo! Zip Lash</string>
<string name="zelda_twilight">The Legend of Zelda: Twilight Princess HD</string>
<string name="mh_stories">Monster Hunter Stories</string>
<string name="ml_paper_jam">Mario & Luigi: Paper Jam</string>
<string name="ml_superstar_saga">Mario & Luigi: Superstar Saga</string>
<string name="ms_superstars">Mario Sports Superstars</string>
<string name="mario_party_ten">Mario Party 10</string>
<string name="mario_tennis">Mario Tennis</string>
<string name="pikmin">Pikmin</string>
<string name="splatoon">Splatoon (1 & 2)</string>
<string name="splatoon_three">Splatoon 3</string>
<string name="super_smash">Super Smash Bros. (Wii U / 3DS)</string>
<string name="smash_ultimate">Super Smash Bros. Ultimate</string>
<string name="unknown_type">Unverified Tag</string>
<string name="write_raw">Write to Tag (Raw)</string>
<string name="write_auto">Write to Tag (Auto)</string>
<string name="update_tag">Update a Tag</string>
<string name="scan_tag">Scan an NFC Tag</string>
<string name="scan_data">Scan from Tag</string>
<string name="gatt_share">Upload to GATT</string>
<string name="lock_elite">Lock N2 Elite</string>
<string name="unlock_elite">Unlock N2 Elite</string>
<string name="scan_bank_no">Scan Bank %1$d</string>
<string name="scan_bank">Scan Bank %1$s</string>
<string name="scan_elite">Scan N2 Elite</string>
<string name="erase_bank">Erase Bank Tag</string>
<string name="activate_bank">Activate Bank</string>
<string name="amiibo_backup">Create Backup</string>
<string name="validate_data">Validate NTAG</string>
<string name="set_bank_count">Set Bank Count</string>
<string name="lock_amiibo">Lock N2 Elite</string>
<string name="write_collection">Write Collection</string>
<string name="erase_collection">Erase Collection</string>
<string name="feature_unavailable">Feature not implemented!</string>
<string name="activity_unavailable">Please try again later.</string>
<string name="storage_unavailable">Device storage inaccessible!</string>
<string name="guide_suggested">View the guide for more help</string>
<string name="allow">Allow</string>
<string name="camera">Camera</string>
<string name="gallery">Gallery</string>
<string name="qr_type">QR Format</string>
<string name="qr_raw">QR Text</string>
<string name="qr_bytes">QR Data</string>
<string name="qr_mii">Mii Data</string>
<string name="qr_amiibo">amiibo QR</string>
<string name="qr_invalid">Input null or invalid</string>
<string name="settings">Settings</string>
<string name="guides">Guides</string>
<string name="about">About</string>
<string name="donate">Donate</string>
<string name="menu_amiibo">Backup Browser</string>
<string name="menu_foomiibo">Foomiibo Panel</string>
<string name="menu_elite">N2 Elite Connect</string>
<string name="menu_gatt">Bluetooth Connect</string>
<string name="menu_qr_code">QR Code Toolkit</string>
<string name="menu_guides">Guides and Help</string>
<string name="menu_settings">TagMo Settings</string>
<string name="tagmo_update">TagMo Update</string>
<string name="pref_menu_return">Return to main menu</string>
<string name="prefs_configuration">Configuration</string>
<string name="pref_import_keys">Import amitool keys</string>
<string name="image_loading_details">Image data access</string>
<string name="enable_tag_validation">NTAG215 validation</string>
<string name="tag_validation_details">Not all NFC chipsets correctly verify tags.\nDisable to bypass tag compatibility checks.</string>
<string name="enable_automatic_scan">Automatic NFC Scan</string>
<string name="automatic_scan_details">Disable to prevent unexpected scans.</string>
<string name="tag_devices">Tag Devices</string>
<string name="enable_power_tag">Power Tag support</string>
<string name="power_tag_details">See Github for limitations and use.</string>
<string name="enable_elite">N2 Elite support</string>
<string name="elite_details">See Github for limitations and use.</string>
<string name="lock_elite_details">The active bank will emulate a retail statue.\nOther features will be disabled until unlocked.</string>
<string name="lock_elite_warning">Lock N2 Elite Hardware?</string>
<string name="unlock_elite_warning">Unlock N2 Elite Hardware</string>
<string name="elite_signature">N2 Elite ID: %1$s</string>
<string name="prefs_game_compat">Game Compatibility</string>
<string name="enable_ds_compat">Show 3DS Games</string>
<string name="enable_wii_compat">Show Wii U Games</string>
<string name="enable_nx_compat">Show Switch Games</string>
<string name="amiibo_info">amiibo Database</string>
<string name="pref_database_source">Preferred AmiiboAPI Database</string>
<string name="sync_amiibo_info">Sync database with AmiiboAPI</string>
<string name="import_json_details">Load amiibo database (JSON)</string>
<string name="reset_amiibo_info">Reset amiibo database cache</string>
<string name="database_stats">Database Statistics</string>
<string name="number_character">%1$d Character(s)</string>
<string name="number_type">%1$d amiibo Type(s)</string>
<string name="number_titles">%1$d Enabled Game(s)</string>
<string name="pref_advanced">Advanced</string>
<string name="pref_tagmo_theme">TagMo System Theme</string>
<string name="tagmo_theme_details">Change local theme independent of system theme settings</string>
<string name="pref_disable_debug">Disable Debug Logger</string>
<string name="disable_debug_details">May improve performance, but limits most diagnostic details</string>
<string name="pref_disclaimer">Disclaimers</string>
<string name="amiibo_info_updated">Updated amiibo database!</string>
<string name="sync_amiibo_process">Syncing with AmiiboAPI…</string>
<string name="sync_amiibo_complete">AmiiboAPI sync success!</string>
<string name="sync_amiibo_failed">AmiiboAPI sync failure!</string>
<string name="removing_amiibo_info">Clearing database cache…</string>
<string name="update_amiibo_api">AmiiboAPI database updated!</string>
<string name="locating_keys">Locating amitool Keys…</string>
<string name="refreshing_list">Searching for amiibo…</string>
<string name="device_not_found">No device or tags available!</string>
<string name="files_displayed">%1$d Files: ? / ? amiibo</string>
<string name="amiibo_collected">%1$d Files: %2$d / %3$d amiibo</string>
<string name="view_device_options">Switch to Device Options</string>
<string name="view_amiibo_details">Switch to amiibo Details</string>
<string name="write_banks">Write %1$d banks</string>
<string name="erase_banks">Erase %1$d banks</string>
<string name="edit_bank_count">Edit Bank Count for N2 Elite</string>
<string name="bank_stats">Bank %1$d Active, %2$d Banks Enabled</string>
<string name="bank_number">Bank %1$d</string>
<string name="bank_select_details">Set Bank Number</string>
<string name="elite_write_confirm">Write selections to N2 Elite?</string>
<string name="elite_erase_confirm">Erase all banks from N2 Elite?</string>
<string name="warn_delete_file">%1$s\nwill be deleted. Are you sure?</string>
<string name="delete_missing">Cannot delete a scanned tag!</string>
<string name="delete_virtual">Cannot delete a virtual tag!</string>
<string name="erase_active">Cannot erase the active bank!</string>
<string name="document_storage_root">Update Document Storage</string>
<string name="emulated_storage_root">Browse Emulated Storage</string>
<string name="physical_storage_root">Browse Physical Storage</string>
<string name="grant_file_permission">Enable "All Files" Manager</string>
<string name="force_document_storage">Enable Document Storage</string>
<string name="storage_setup">Please configure a root folder</string>
<string name="unzip_archive">Extract bins from zip file</string>
<string name="refresh_browser">Refresh Browser</string>
<string name="hint_filename">Set Binary File Name</string>
<string name="key_hex_entry">Input Hexidecimal Key</string>
<string name="save_image">Set Image File Name</string>
<string name="key_input_method">Select an input method</string>
<string name="key_input_hex">Hex Code</string>
<string name="key_input_bin">Bin File</string>
<string name="prepare_unlock">Once the N2 is scanned, it cannot be removed from the NFC field until the operation is complete.</string>
<string name="progress_unlock">Press and hold the N2 button without disconnecting from NFC.</string>
<string name="prepare_blank">N2 Elite fails verification with an empty active bank.</string>
<string name="possible_lock">Is this N2 Elite locked?</string>
<string name="possible_blank">Scanning an N2 Elite?</string>
<string name="tag_update_only">Previously written tags require \"Update\"</string>
<string name="power_tag_reset">Power Tags last used with PowerSaves require a reset.</string>
<string name="encryption_fault">This file may be corrupt or was improperly created.</string>
<string name="custom_tab_back">Back to TagMo</string>
<string name="shop_hardware">Shop amiibo hardware</string>
<string name="gatt_missing">No nearby GATT devices found!</string>
<string name="gatt_connect_fail">Failed to connect GATT service!</string>
<string name="gatt_services_null">%1$s services not found!</string>
<string name="characteristic_null">Characteristic not found!</string>
<string name="xor_invalid">Empty collection can\'t be reduced</string>
<string name="connect_device">Connect device</string>
<string name="gatt_scanning">Scanning for Bluetooth GATT…</string>
<string name="gatt_located">Connecting to Bluetooth GATT…</string>
<string name="gatt_disconnect">GATT device connection lost!</string>
<string name="gatt_count">Slot %1$s Active, %2$d Slots Occupied</string>
<string name="write_slot_file">Upload binary to GATT device</string>
<string name="write_slots">Write %1$d slots</string>
<string name="erase_slots">Erase %1$d slots</string>
<string name="slots_full">Device full</string>
<string name="slots_empty">Device empty</string>
<string name="create_blank">Create blank virtual NFC tag</string>
<string name="reset_device">Factory reset current device</string>
<string name="gatt_erase_confirm">Erase all slots from %1$s?</string>
<string name="gatt_write_confirm">Write selections to %1$s?</string>
<string name="gatt_upload">Uploading binary…</string>
<string name="gatt_remove">Wiping slot data…</string>
<string name="gatt_create">Creating blank tag…</string>
<string name="gatt_format">Formatting device…</string>
<string name="fail_bluetooth_adapter">Bluetooth adapter unavailable!</string>
<string name="fail_no_device">No GATT device connected!</string>
<string name="tiramisu_bluetooth">Android 13 requires manually enabling Bluetooth. Open Settings to enable Bluetooth?</string>
<string name="fail_permissions">Required permission denied by user!</string>
<string name="location_disclosure">Location is required for Bluetooth LE scan. In accordance with Google Play policy, TagMo is required to notify you that location data may be collected in the background. TagMo does not read, store, or share location data. As such, this notice is strictly a formality.</string>
<string name="sort_mode_set">Set device mode</string>
<string name="paired_devices">Paired Bluetooth Devices</string>
<string name="device_address">MAC: %1$s</string>
<string name="device_services">UUID: %1$s</string>
<string name="switch_devices">Switch / disconnect device</string>
<string name="clone_random">Clone with random serial</string>
<string name="data_write">Wrote main data</string>
<string name="password_write">Wrote password</string>
<string name="lock_write">Wrote lock info</string>
<string name="lock_skipped">Lock info skipped</string>
<string name="auth_response">Auth response %1$s</string>
<string name="password">Password: %1$s</string>
<string name="validation_success">Data validation successful!</string>
<string name="no_source_data">Cannot validate: No data to compare.</string>
<string name="error_no_data">No data to write</string>
<string name="error_tag_protocol">NTAG215 required, found %1$s!</string>
<string name="error_tag_unavailable">NTAG215 connection cannot be established!</string>
<string name="error_state">State error. Invalid action: %1$s</string>
<string name="error_length">Must be 8 characters</string>
<string name="error_input">Invalid input</string>
<string name="error_min_max">Must be between %1$d and %2$d</string>
<string name="error_data_write">Error while writing main data (stage 1)</string>
<string name="error_password_write">Error while setting password (stage 2)</string>
<string name="error_lock_write">Error while setting lock info (stage 3)</string>
<string name="error_tag_rewrite">Tag already an amiibo</string>
<string name="error_tag_specs">Tag is not an NTAG215</string>
<string name="error_tag_version">Tag version error</string>
<string name="error_tag_faulty">NFC tag may be corrupt.\nPlease try different tags.</string>
<string name="error_version">Version information error</string>
<string name="error_amiibo_parse">amiibo info parse error</string>
<string name="error_powertag_key">Power Tag keys not loaded</string>
<string name="error_elite_write">Error while writing data to N2 Elite</string>
<string name="error_elite_auth">Error while authenticating N2 Elite</string>
<string name="error_amiitool_init">Failed to initialize amiitool</string>
<string name="error_auth_null">Auth result was null</string>
<string name="error_archive_format">Zip encoding not supported</string>
<string name="error_archive_invalid">Archive extraction failed!</string>
<string name="error_uri_unknown">Invalid or empty file selected!</string>
<string name="error_uri_size">File is not a key or valid bin!</string>
<string name="error_unknown">An unknown error has occurred.</string>
<string name="error_tag_format">Mifare Classic 1K required!</string>
<string name="error_nxp_required">NXP chipset is required!</string>
<string name="fail_update_git">Unable to contact update server!</string>
<string name="fail_update_url">Update invalid! Please try again.</string>
<string name="fail_ssl_update">OpenSSL security provider obsolete!</string>
<string name="fail_save_file">Failed writing tag to file</string>
<string name="fail_save_data">Failed to save tag data</string>
<string name="fail_read_size">Read failed! Unexpected read size.</string>
<string name="fail_read_uid">Could not read old UID</string>
<string name="fail_mismatch_uid">Source UID does not match the target!</string>
<string name="fail_uid_invalid">Invalid UID (%1$d byte hex)</string>
<string name="fail_primary_key">Sector 0 authentication failed!</string>
<string name="fail_sector_invalid">Invalid sector (0 15)</string>
<string name="fail_invalid_size">Invalid read result size</string>
<string name="fail_powertag_keys">Failed to load Power Tag keys</string>
<string name="fail_bank_oob">Selected bank exceeds count!</string>
<string name="fail_active_oob">Active bank beyond new count!</string>
<string name="fail_parse_banks">Failed to parse banks!</string>
<string name="fail_read_amiibo">Reading amiibo data failed!</string>
<string name="fail_early_remove">amiibo removed too early!</string>
<string name="fail_amiibo_null">Please scan the amiibo again</string>
<string name="fail_read">Read failed</string>
<string name="fail_auth">Authentication failed</string>
<string name="fail_decrypt_null">Cannot decrypt empty data.</string>
<string name="fail_decrypt">Failed to decrypt tag data</string>
<string name="fail_encrypt">Failed to encrypt tag data</string>
<string name="fail_display">Failed to display tag data</string>
<string name="fail_lock">Cannot lock a blank bank!</string>
<string name="fail_unlock">Failed to prepare unlock!</string>
<string name="fail_logcat">Failed to prepare logcat!</string>
<string name="fail_bitmap">Failed to generate bitmap!</string>
<string name="fail_bank_data">Binary data empty!\nSkipping bank %1$d…</string>
<string name="fail_randomize">Randomizer failed to duplicate data!</string>
<string name="fail_firmware_api">API disabled by firmware!</string>
<string name="no_tag_loaded"><No Tag Loaded></string>
<string name="read_error"><Error Reading Tag></string>
<string name="empty_tag"><Empty Tag></string>
<string name="blank_tag"><Blank Tag></string>
<string name="blank_bank">%1$s: <Blank Bank></string>
<string name="key_not_present">Key files not loaded!</string>
<string name="key_size_error">Key file size does not match.</string>
<string name="key_size_invalid">Invalid key file size</string>
<string name="key_signature_error">Key file signature does not match.</string>
<string name="invalid_key_error">No valid key in file.</string>
<string name="key_read_error">Error reading key from storage</string>
<string name="invalid_data_null">Cannot verify empty data.</string>
<string name="invalid_tag_prefix">Data must start with 0x04.</string>
<string name="invalid_tag_lock">Lock signature mismatch.</string>
<string name="invalid_tag_cc">CC signature mismatch.</string>
<string name="invalid_tag_dynamic">Dynamic lock signature mismatch.</string>
<string name="invalid_tag_cfg_zero">CFG0 signature mismatch.</string>
<string name="invalid_tag_cfg_one">CFG1 signature mismatch.</string>
<string name="invalid_uid_length">Invalid UID length</string>
<string name="invalid_file_size">%1$s size: %2$d, Expected: %3$d</string>
<string name="invalid_data_size">Tag size: %1$d, Expected: %2$d</string>
<string name="invalid_tag_key">Key file detected. Tag skipped.</string>
<string name="invalid_app_data">Invalid app data</string>
<string name="invalid_read_result">Invalid read result size</string>
<string name="uid_key_missing">No available key for UID</string>
<string name="p10_key_missing">No available key for P10_ID</string>
<string name="power_tag_verify">Power Tag identified: %1$s</string>
<string name="amiibo_info_parse_error">Unable to parse amiibo database</string>
<string name="amiibo_info_not_loaded">amiibo database not loaded</string>
<string name="amiibo_failure_load">Failed to load amiibo database</string>
<string name="amiibo_failure_parse">Failed to parse amiibo database</string>
<string name="amiibo_failure_read">Failed to read amiibo database</string>
<string name="amiibo_failure_update">Failed to update amiibo database</string>
<string name="amiibo_failure_server">Unable to contact database server</string>
<string name="amiibo_failure_parse_default">Failed to parse default amiibo database</string>
<string name="firmware_update">N2 Elite Firmware Updated!</string>
<string name="firmware_failed">Firmware update failed. Please try again! (%1$d)</string>
<string name="firmware_obsolete">Firmware update required!</string>
<string name="speed_scan">Lift tag slightly to complete scan</string>
<string name="tag_disconnect">Tag was removed too early!</string>
<string name="unzip_item">Unzipping %1$s…</string>
<string name="mkdir_failed">Mkdir ’%1$s’ failed!</string>
<string name="downloads_hidden">Downloads are currently hidden!</string>
<string name="controller_pro" translatable="false">Pro Controller</string>
<string name="controller_rjc" translatable="false">Right Joy-Con</string>
<string name="no_controllers">No connected controller found!</string>
<string name="sub_button">%1$s Monthly Donation</string>
<string name="iap_button">%1$s One-Time Donation</string>
<string name="donation_thanks">Thanks for supporting TagMo!</string>
<string name="settings_banner">TagMo Donation Subscriber</string>
<string name="external_search">Search Amiibo</string>
<string name="conversion_title">TagMo Critical Update</string>
<string name="conversion_message">GitHub and Google Play builds are now universal.\nPlease click continue to resolve installation conflicts.</string>
<string name="disclaimer_general">No statues were harmed in making this product.</string>
<string name="duplicate_reports">Duplicate bug reports are not required</string>
</resources>
```
|
```c
/* Testcase for PR c/1501. */
double __complex__
f (void)
{
return ~(1.0 + 2.0i);
}
```
|
```shell
# $OpenBSD: putty-ciphers.sh,v 1.13 2024/02/09 08:56:59 dtucker Exp $
# Placed in the Public Domain.
tid="putty ciphers"
puttysetup
cp ${OBJ}/sshd_proxy ${OBJ}/sshd_proxy_bak
# Since there doesn't seem to be a way to set MACs on the PuTTY client side,
# we force each in turn on the server side, omitting the ones PuTTY doesn't
# support. Grepping the binary is pretty janky, but AFAIK there's no way to
# query for supported algos.
macs=""
for m in `${SSH} -Q MACs`; do
if strings "${PLINK}" | grep -E "^${m}$" >/dev/null; then
macs="${macs} ${m}"
else
trace "omitting unsupported MAC ${m}"
fi
done
ciphers=""
for c in `${SSH} -Q Ciphers`; do
if strings "${PLINK}" | grep -E "^${c}$" >/dev/null; then
ciphers="${ciphers} ${c}"
else
trace "omitting unsupported cipher ${c}"
fi
done
for c in default $ciphers; do
for m in default ${macs}; do
verbose "$tid: cipher $c mac $m"
cp ${OBJ}/.putty/sessions/localhost_proxy \
${OBJ}/.putty/sessions/cipher_$c
if [ "${c}" != "default" ]; then
echo "Cipher=$c" >> ${OBJ}/.putty/sessions/cipher_$c
fi
cp ${OBJ}/sshd_proxy_bak ${OBJ}/sshd_proxy
if [ "${m}" != "default" ]; then
echo "MACs $m" >> ${OBJ}/sshd_proxy
fi
rm -f ${COPY}
env HOME=$PWD ${PLINK} -load cipher_$c -batch -i ${OBJ}/putty.rsa2 \
cat ${DATA} > ${COPY}
if [ $? -ne 0 ]; then
fail "ssh cat $DATA failed"
fi
cmp ${DATA} ${COPY} || fail "corrupted copy"
done
done
rm -f ${COPY}
```
|
```shell
Let's play the blame game
Cache your authentication details to save time
Specify a commit by its ancestry
Interactive staging
Debug using binary search
```
|
Henrik Johannes Krebs (June 27, 1847 October 7, 1929), sometimes called Henry J. Krebs, was an American immigrant from Denmark who started the Krebs Pigments and Chemical Company to manufacture the pigment lithopone.
Biography
Krebs was born on June 27, 1847, in Lolland, Denmark, to Hans Peter Tholstrup Krebs and Johanne Louise (Wichfled) Krebs. He emigrated to the United States in 1879.
He soon became the superintendent of the Delaware Sugar Beet Factory, associated with Alfred D. Wagner. Krebs later joined Pusey & Jones Co. and built the first ice plant in Wilmington, Delaware.
In 1886, Krebs, Wagner, and William G. Pennypacker founded the Delaware Chemical Company to produce anhydrous ammonia, which later became the National Ammonia Company.
Around 1902, he formed the Krebs Pigments and Chemical Company. In 1908, he built a new plant in Newport, Delaware.
Krebs retired from Krebs Pigments and Chemical Company in 1921, when his son, August Sonin Krebs, took over as president.
In 1926, Krebs supplied $150,000 to Newport, Delaware, for the construction of a school to be named after him, on the occasion of his 50th wedding anniversary The school opened in September 1927.
For services to the Royal Polytechnic College in Copenhagen, the King of Denmark appointed Krebs Knight of the Order of the Dannebrog.
Krebs died on October 7, 1929, after a long illness. After his death, his company was purchased by DuPont.
Personal
On June 24, 1876, he married Ebba Victoria Kruse. At his death, he was survived by his wife and two children, August Sonnin, and Johanna Louise.
See also
List of people from Delaware
References
Place of death missing
1929 deaths
1847 births
20th-century American businesspeople
American company founders
American philanthropists
Danish emigrants to the United States
Businesspeople from Delaware
Knights of the Order of the Dannebrog
Patrons of schools
People from Lolland
People from Newport, Delaware
American chemical industry businesspeople
|
This is a list of heads of state of the Federal Republic of Central America from its founding in 1821/1823 until its dissolution in 1840.
Heads of State
Superior Political Chiefs
First Triumvirate
Second Triumvirate
Presidents
Political parties
See also
Federal Republic of Central America
References
List
Central America, Federal Republic
|
This is a list of places in South Africa which have standing links to local communities in other countries. In most cases, the association, especially when formalised by local government, is known as "town twinning" (usually in Europe) or "sister cities" (usually in the rest of the world).
B
Bergrivier
Heist-op-den-Berg, Belgium
Bloemfontein
Bhubaneswar, India
Nanjing, China
Buffalo City
Jinhua, China
Leiden, Netherlands
Milwaukee County, United States
Oldenburg, Germany
Qinhuangdao, China
C
Cape Town
Aachen, Germany
Accra, Ghana
Antwerp, Belgium
Bujumbura, Burundi
Dubai, United Arab Emirates
Funchal, Portugal
Haifa, Israel
Hangzhou, China
Maputo, Mozambique
Malmö, Sweden
Miami-Dade County, United States
Monterrey, Mexico
Rio de Janeiro, Brazil
Saint Petersburg, Russia
Varna, Bulgaria
D
Dordrecht
Dordrecht, Netherlands
Drakenstein
Walvis Bay, Namibia
Durban
Alexandria, Egypt
Bremen, Germany
Bulawayo, Zimbabwe
Chicago, United States
Daejeon, South Korea
Guangzhou, China
Kaohsiung, Taiwan
Leeds, England, United Kingdom
Libreville, Gabon
Maputo, Mozambique
Mersin, Turkey
Mombasa, Kenya
New Orleans, United States
Oran, Algeria
Le Port, Réunion, France
E
East London
Daqing, China
Ekurhuleni
Harbin, China
F
Franschhoek
Dilbeek, Belgium
G
George
Saint Paul, United States
Tacoma, United States
I
iLembe
Mobile, United States
Zhanjiang, China
J
Johannesburg
Addis Ababa, Ethiopia
Birmingham, England, United Kingdom
Ho Chi Minh City, Vietnam
New York City, United States
Ramallah, Palestine
Taipei, Taiwan
Verona, Italy
Windhoek, Namibia
K
King Cetshwayo
Milwaukee, United States
M
Mathopestad
Berkeley, United States
Mbombela
Baotou, China
Maia, Portugal
Mbabane, Eswatini
Mogale
Wujiang (Suzhou), China
Mossel Bay
Denpasar, Indonesia
Msunduzi
Hampton, United States
Taichung, Taiwan
N
Nelson Mandela Bay
Gothenburg, Sweden
Jacksonville, United States
Ningbo, China
Tainan, Taiwan
Newcastle
Nanchang, China
Zibo, China
O
Oudtshoorn
Alphen aan den Rijn, Netherlands
Hualien, Taiwan
Oukasie
Berkeley, United States
P
Polokwane
Reggio Emilia, Italy
Pretoria
Baku, Azerbaijan
Bucharest, Romania
Bulawayo, Zimbabwe
Hanoi, Vietnam
Kyiv, Ukraine
Port Louis, Mauritius
Taipei, Taiwan
Tehran, Iran
Washington, D.C., United States
R
Ray Nkonyeni
Oskarshamn, Sweden
S
Sol Plaatje
Changsha, China
Stellenbosch
Paju, South Korea
U
Upington
Temora, Australia
W
Witzenberg
Bozhou, China
Essen, Belgium
Worcester
Aalst, Belgium
References
South Africa
Cities in South Africa
Foreign relations of South Africa
South Africa geography-related lists
Populated places in South Africa
|
Pedn Vounder () is a tidal beach on the south coast of the Penwith peninsula in Cornwall, England. It is immediately to the west of the Logan Rock headland, below Treen Cliff. The name is from the Cornish 'pen' (head, end) and 'an vounder' (of the lane).
Access to the beach is either by boat, or on foot from the coastal path along the cliffs. The top of this path may be approached from Porthcurno, approximately to the west, or from Treen, approximately to the north. The final stage of the descent from the coastal path requires some climbing down steep rocks at the western end of the beach. It is sometimes possible to walk from Porthcurno beach along the shore, during low water of a spring tide. In common with some of the other secluded coves on the Penwith peninsula, this beach is an unofficial naturist beach. Signs on the cliff above the beach warn that swimming off the far side of the sand bar is dangerous due to strong currents.
In 1880, a wooden hut was built above the beach to house a submarine communications cable from Brest, France. It connected with a land line running via the Trereen valley, and by Burnewhall to Sheffield and on to Newlyn and Penzance. At this time Porthcurno was known world-wide as the British terminal of submarine telegraph cables.
Gallery
References
External links
Pedn Vounder Beach on cornwall-beaches.co.uk
Treen area guide on sennen-cove.com
Porthcurno and Treen information (usable on mobiles) on porthcurno.mobi
Beaches of Penwith
Nude beaches
|
Cetarthrosaurus is an extinct genus of ichthyosaur known from the eastern United Kingdom. It was collected from the Cambridge Greensand Formation, dating to late Albian or early Cenomanian stage, of the Early Cretaceous-Late Cretaceous boundary. Cetarthrosaurus was first named by Harry G. Seeley in 1873 and the type species is Cetarthrosaurus walkeri.
References
Cretaceous ichthyosaurs
Fossil taxa described in 1873
Ichthyosaurs of Europe
|
```xml
import { Route, Routes, useLocation, useParams } from 'react-router-dom';
import React from 'react';
import asyncComponent from '@erxes/ui/src/components/AsyncComponent';
import queryString from 'query-string';
const CompanyDetails = asyncComponent(
() =>
import(
/* webpackChunkName: "CompanyDetail" */ './containers/detail/CompanyDetails'
),
);
const CompaniesList = asyncComponent(
() =>
import(
/* webpackChunkName: "CompaniesList" */ './containers/CompaniesList'
),
);
const Details = () => {
const { id = '' } = useParams();
return <CompanyDetails id={id} />;
};
const List = () => {
const location = useLocation();
const queryParams = queryString.parse(location.search);
localStorage.setItem('erxes_contact_url', 'companies');
return <CompaniesList queryParams={queryParams} />;
};
const routes = () => {
return (
<Routes>
<Route
path="/companies/details/:id"
key="/companies/details/:id"
element={<Details />}
/>
<Route path="/companies" key="/companies" element={<List />} />
</Routes>
);
};
export default routes;
```
|
```shell
#!/bin/bash -e
$(dirname "$0")/install.sh
HOMEBREW_NO_INSTALL_CLEANUP=1 HOMEBREW_NO_AUTO_UPDATE=1 brew install ruby
export PATH="/usr/local/opt/ruby/bin:$PATH"
gem install xcpretty
export CODE_SIGNING_REQUIRED=NO
HOMEBREW_NO_INSTALL_CLEANUP=1 HOMEBREW_NO_AUTO_UPDATE=1 brew tap wix/brew
HOMEBREW_NO_INSTALL_CLEANUP=1 HOMEBREW_NO_AUTO_UPDATE=1 brew install applesimutils
echo 'export PATH=$PATH' >> $BASH_ENV
```
|
Heptalene is a polycyclic hydrocarbon with chemical formula , composed of two fused cycloheptatriene rings. It is an unstable, non-planar compound which is non-aromatic. The dianion, however, satisfies Hückel's rule, is thermally stable, and is planar.
See also
Benzocyclooctatetraene
References
|
```php
<?php
/*
*
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing, software
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
*/
namespace Google\Service\Aiplatform;
class GoogleCloudAiplatformV1ResourceRuntimeSpec extends \Google\Model
{
protected $raySpecType = GoogleCloudAiplatformV1RaySpec::class;
protected $raySpecDataType = '';
protected $serviceAccountSpecType = GoogleCloudAiplatformV1ServiceAccountSpec::class;
protected $serviceAccountSpecDataType = '';
/**
* @param GoogleCloudAiplatformV1RaySpec
*/
public function setRaySpec(GoogleCloudAiplatformV1RaySpec $raySpec)
{
$this->raySpec = $raySpec;
}
/**
* @return GoogleCloudAiplatformV1RaySpec
*/
public function getRaySpec()
{
return $this->raySpec;
}
/**
* @param GoogleCloudAiplatformV1ServiceAccountSpec
*/
public function setServiceAccountSpec(GoogleCloudAiplatformV1ServiceAccountSpec $serviceAccountSpec)
{
$this->serviceAccountSpec = $serviceAccountSpec;
}
/**
* @return GoogleCloudAiplatformV1ServiceAccountSpec
*/
public function getServiceAccountSpec()
{
return $this->serviceAccountSpec;
}
}
// Adding a class alias for backwards compatibility with the previous class name.
class_alias(GoogleCloudAiplatformV1ResourceRuntimeSpec::class, your_sha256_hashSpec');
```
|
```javascript
var fs = require("fs");
/* readFile */
fs.readFile("file.exp", (_, data) => {
(data : Buffer);
});
fs.readFile("file.exp", "blah", (_, data) => {
(data : string);
});
fs.readFile("file.exp", { encoding: "blah" }, (_, data) => {
(data : string);
});
fs.readFile("file.exp", {}, (_, data) => {
(data : Buffer);
});
/* readFileSync */
(fs.readFileSync("file.exp") : Buffer);
(fs.readFileSync("file.exp") : string); // error
(fs.readFileSync("file.exp", "blah") : string);
(fs.readFileSync("file.exp", "blah") : Buffer); // error
(fs.readFileSync("file.exp", { encoding: "blah" }) : string);
(fs.readFileSync("file.exp", { encoding: "blah" }) : Buffer); // error
(fs.readFileSync("file.exp", {}) : Buffer);
(fs.readFileSync("file.exp", {}) : string); // error
```
|
```javascript
import { test } from '../../test';
export default test({
compileOptions: {
dev: true // Render in dev mode to check that the validation error is not thrown
},
html: `<p>hello world</p>`
});
```
|
```javascript
import { RecoilRoot } from "recoil";
export default function MyApp({ Component, pageProps }) {
return (
<RecoilRoot>
<Component {...pageProps} />
</RecoilRoot>
);
}
```
|
The 2023–24 season is FC Petrolul Ploiești's 99th season in existence and second consecutive in the Liga I. They will also compete in the Cupa României.
Players
First-team squad
Transfers
In
Out
Overall transfer activity
Expenditure
Summer: €160,000
Winter:
Total: €160,000
Income
Summer: €125,000
Winter:
Total: €125,000
Net Totals
<small>Summer: €35,000
Winter:
Total: €35,000
Pre-season and friendlies
Competitions
Liga I
League table
Results summary
Results by round
Matches
The league fixtures were unveiled on 27 June 2023.
Cupa României
References
External links
FC Petrolul Ploiești seasons
Farul Constanta
|
```go
// Package suffix contains a kio.Filter implementation of the kustomize
// SuffixTransformer.
package suffix
```
|
```objective-c
#import "RNNComponentViewController+Utils.h"
#import "RNNTestRootViewCreator.h"
@implementation RNNComponentViewController (Utils)
+ (RNNComponentViewController *)createWithComponentId:(NSString *)componentId
initialOptions:(RNNNavigationOptions *)initialOptions {
RNNLayoutInfo *layoutInfo = [[RNNLayoutInfo alloc] init];
layoutInfo.componentId = componentId;
RNNButtonsPresenter *buttonsPresenter =
[[RNNButtonsPresenter alloc] initWithComponentRegistry:nil eventEmitter:nil];
RNNComponentViewController *component = [[RNNComponentViewController alloc]
initWithLayoutInfo:layoutInfo
rootViewCreator:[[RNNTestRootViewCreator alloc] init]
eventEmitter:nil
presenter:[[RNNComponentPresenter alloc]
initWithComponentRegistry:nil
defaultOptions:nil
buttonsPresenter:buttonsPresenter]
options:initialOptions
defaultOptions:nil];
[buttonsPresenter bindViewController:component];
return component;
}
+ (RNNComponentViewController *)createWithComponentId:(NSString *)componentId {
return [self createWithComponentId:componentId
initialOptions:RNNNavigationOptions.emptyOptions];
}
@end
```
|
```java
/*
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing, software
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*/
package org.apache.beam.sdk.extensions.avro.io;
import static org.apache.beam.sdk.transforms.display.DisplayDataMatchers.hasDisplayItem;
import static org.hamcrest.MatcherAssert.assertThat;
import static org.hamcrest.Matchers.containsInAnyOrder;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertSame;
import static org.junit.Assert.assertTrue;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Collections;
import java.util.List;
import java.util.NoSuchElementException;
import java.util.Objects;
import java.util.Random;
import java.util.stream.Collectors;
import org.apache.avro.Schema;
import org.apache.avro.file.CodecFactory;
import org.apache.avro.file.DataFileConstants;
import org.apache.avro.file.DataFileWriter;
import org.apache.avro.generic.GenericDatumReader;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
import org.apache.avro.io.DatumWriter;
import org.apache.avro.io.Decoder;
import org.apache.avro.reflect.AvroDefault;
import org.apache.avro.reflect.ReflectData;
import org.apache.avro.reflect.ReflectDatumWriter;
import org.apache.beam.sdk.coders.DefaultCoder;
import org.apache.beam.sdk.extensions.avro.coders.AvroCoder;
import org.apache.beam.sdk.io.BlockBasedSource;
import org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader;
import org.apache.beam.sdk.io.BoundedSource;
import org.apache.beam.sdk.io.BoundedSource.BoundedReader;
import org.apache.beam.sdk.io.FileBasedSource;
import org.apache.beam.sdk.io.FileSystems;
import org.apache.beam.sdk.io.fs.MatchResult.Metadata;
import org.apache.beam.sdk.options.PipelineOptions;
import org.apache.beam.sdk.options.PipelineOptionsFactory;
import org.apache.beam.sdk.testing.SourceTestUtils;
import org.apache.beam.sdk.transforms.display.DisplayData;
import org.apache.beam.sdk.util.SerializableUtils;
import org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.base.MoreObjects;
import org.checkerframework.checker.nullness.qual.Nullable;
import org.hamcrest.Matchers;
import org.junit.Rule;
import org.junit.Test;
import org.junit.rules.ExpectedException;
import org.junit.rules.TemporaryFolder;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
/** Tests for AvroSource. */
@RunWith(JUnit4.class)
public class AvroSourceTest {
private static final String VERSION_AVRO =
org.apache.avro.Schema.class.getPackage().getImplementationVersion();
@Rule public TemporaryFolder tmpFolder = new TemporaryFolder();
@Rule public ExpectedException expectedException = ExpectedException.none();
private enum SyncBehavior {
SYNC_REGULAR, // Sync at regular, user defined intervals
SYNC_RANDOM, // Sync at random intervals
SYNC_DEFAULT // Sync at default intervals (i.e., no manual syncing).
}
private static final int DEFAULT_RECORD_COUNT = 1000;
private Iterable<String> avroSupportedCodec() {
List<String> codecs = new ArrayList<>();
codecs.add(DataFileConstants.NULL_CODEC);
codecs.add(DataFileConstants.BZIP2_CODEC);
codecs.add(DataFileConstants.DEFLATE_CODEC);
codecs.add(DataFileConstants.SNAPPY_CODEC);
codecs.add(DataFileConstants.XZ_CODEC);
if (!VERSION_AVRO.equals("1.8.2")) {
codecs.add("zstandard");
}
return codecs;
}
/**
* Generates an input Avro file containing the given records in the temporary directory and
* returns the full path of the file.
*/
private <T> String generateTestFile(
String filename,
List<T> elems,
SyncBehavior syncBehavior,
int syncInterval,
AvroCoder<T> coder,
String codec)
throws IOException {
Random random = new Random(0);
File tmpFile = tmpFolder.newFile(filename);
String path = tmpFile.toString();
FileOutputStream os = new FileOutputStream(tmpFile);
DatumWriter<T> datumWriter =
coder.getType().equals(GenericRecord.class)
? new GenericDatumWriter<>(coder.getSchema())
: new ReflectDatumWriter<>(coder.getSchema());
try (DataFileWriter<T> writer = new DataFileWriter<>(datumWriter)) {
writer.setCodec(CodecFactory.fromString(codec));
writer.create(coder.getSchema(), os);
int recordIndex = 0;
int syncIndex = syncBehavior == SyncBehavior.SYNC_RANDOM ? random.nextInt(syncInterval) : 0;
for (T elem : elems) {
writer.append(elem);
recordIndex++;
switch (syncBehavior) {
case SYNC_REGULAR:
if (recordIndex == syncInterval) {
recordIndex = 0;
writer.sync();
}
break;
case SYNC_RANDOM:
if (recordIndex == syncIndex) {
recordIndex = 0;
writer.sync();
syncIndex = random.nextInt(syncInterval);
}
break;
case SYNC_DEFAULT:
default:
}
}
}
return path;
}
@Test
public void testReadWithDifferentCodecs() throws Exception {
// As Avro's default block size is 64KB, write 64K records to ensure at least one full block.
// We could make this smaller than 64KB assuming each record is at least B bytes, but then the
// test could silently stop testing the failure condition from BEAM-422.
List<Bird> expected = createRandomRecords(1 << 16);
// Test reading files generated using all codecs.
for (String codec : avroSupportedCodec()) {
String filename =
generateTestFile(
codec, expected, SyncBehavior.SYNC_DEFAULT, 0, AvroCoder.of(Bird.class), codec);
AvroSource<Bird> source = AvroSource.from(filename).withSchema(Bird.class);
List<Bird> actual = SourceTestUtils.readFromSource(source, null);
assertThat(expected, containsInAnyOrder(actual.toArray()));
}
}
@Test
public void testSplitAtFraction() throws Exception {
// A reduced dataset is enough here.
List<FixedRecord> expected = createFixedRecords(DEFAULT_RECORD_COUNT);
// Create an AvroSource where each block is 1/10th of the total set of records.
String filename =
generateTestFile(
"tmp.avro",
expected,
SyncBehavior.SYNC_REGULAR,
DEFAULT_RECORD_COUNT / 10 /* max records per block */,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
File file = new File(filename);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
List<? extends BoundedSource<FixedRecord>> splits = source.split(file.length() / 3, null);
for (BoundedSource<FixedRecord> subSource : splits) {
int items = SourceTestUtils.readFromSource(subSource, null).size();
// Shouldn't split while unstarted.
SourceTestUtils.assertSplitAtFractionFails(subSource, 0, 0.0, null);
SourceTestUtils.assertSplitAtFractionFails(subSource, 0, 0.7, null);
SourceTestUtils.assertSplitAtFractionSucceedsAndConsistent(subSource, 1, 0.7, null);
SourceTestUtils.assertSplitAtFractionSucceedsAndConsistent(
subSource, DEFAULT_RECORD_COUNT / 100, 0.7, null);
SourceTestUtils.assertSplitAtFractionSucceedsAndConsistent(
subSource, DEFAULT_RECORD_COUNT / 10, 0.1, null);
SourceTestUtils.assertSplitAtFractionFails(
subSource, DEFAULT_RECORD_COUNT / 10 + 1, 0.1, null);
SourceTestUtils.assertSplitAtFractionFails(subSource, DEFAULT_RECORD_COUNT / 3, 0.3, null);
SourceTestUtils.assertSplitAtFractionFails(subSource, items, 0.9, null);
SourceTestUtils.assertSplitAtFractionFails(subSource, items, 1.0, null);
SourceTestUtils.assertSplitAtFractionSucceedsAndConsistent(subSource, items, 0.999, null);
}
}
@Test
public void testGetProgressFromUnstartedReader() throws Exception {
List<FixedRecord> records = createFixedRecords(DEFAULT_RECORD_COUNT);
String filename =
generateTestFile(
"tmp.avro",
records,
SyncBehavior.SYNC_DEFAULT,
1000,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
File file = new File(filename);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
try (BoundedReader<FixedRecord> reader = source.createReader(null)) {
assertEquals(Double.valueOf(0.0), reader.getFractionConsumed());
}
List<? extends BoundedSource<FixedRecord>> splits = source.split(file.length() / 3, null);
for (BoundedSource<FixedRecord> subSource : splits) {
try (BoundedReader<FixedRecord> reader = subSource.createReader(null)) {
assertEquals(Double.valueOf(0.0), reader.getFractionConsumed());
}
}
}
@Test
public void testProgress() throws Exception {
// 5 records, 2 per block.
List<FixedRecord> records = createFixedRecords(5);
String filename =
generateTestFile(
"tmp.avro",
records,
SyncBehavior.SYNC_REGULAR,
2,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
try (BoundedReader<FixedRecord> readerOrig = source.createReader(null)) {
assertThat(readerOrig, Matchers.instanceOf(BlockBasedReader.class));
BlockBasedReader<FixedRecord> reader = (BlockBasedReader<FixedRecord>) readerOrig;
// Before starting
assertEquals(0.0, reader.getFractionConsumed(), 1e-6);
assertEquals(0, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// First 2 records are in the same block.
assertTrue(reader.start());
assertTrue(reader.isAtSplitPoint());
assertEquals(0, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// continued
assertTrue(reader.advance());
assertFalse(reader.isAtSplitPoint());
assertEquals(0, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// Second block -> parallelism consumed becomes 1.
assertTrue(reader.advance());
assertTrue(reader.isAtSplitPoint());
assertEquals(1, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// continued
assertTrue(reader.advance());
assertFalse(reader.isAtSplitPoint());
assertEquals(1, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// Third and final block -> parallelism consumed becomes 2, remaining becomes 1.
assertTrue(reader.advance());
assertTrue(reader.isAtSplitPoint());
assertEquals(2, reader.getSplitPointsConsumed());
assertEquals(1, reader.getSplitPointsRemaining());
// Done
assertFalse(reader.advance());
assertEquals(3, reader.getSplitPointsConsumed());
assertEquals(0, reader.getSplitPointsRemaining());
assertEquals(1.0, reader.getFractionConsumed(), 1e-6);
}
}
@Test
public void testProgressEmptySource() throws Exception {
// 0 records, 20 per block.
List<FixedRecord> records = Collections.emptyList();
String filename =
generateTestFile(
"tmp.avro",
records,
SyncBehavior.SYNC_REGULAR,
2,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
try (BoundedReader<FixedRecord> readerOrig = source.createReader(null)) {
assertThat(readerOrig, Matchers.instanceOf(BlockBasedReader.class));
BlockBasedReader<FixedRecord> reader = (BlockBasedReader<FixedRecord>) readerOrig;
// before starting
assertEquals(0.0, reader.getFractionConsumed(), 1e-6);
assertEquals(0, reader.getSplitPointsConsumed());
assertEquals(BoundedReader.SPLIT_POINTS_UNKNOWN, reader.getSplitPointsRemaining());
// confirm empty
assertFalse(reader.start());
// after reading empty source
assertEquals(0, reader.getSplitPointsConsumed());
assertEquals(0, reader.getSplitPointsRemaining());
assertEquals(1.0, reader.getFractionConsumed(), 1e-6);
}
}
@Test
public void testGetCurrentFromUnstartedReader() throws Exception {
List<FixedRecord> records = createFixedRecords(DEFAULT_RECORD_COUNT);
String filename =
generateTestFile(
"tmp.avro",
records,
SyncBehavior.SYNC_DEFAULT,
1000,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
try (BlockBasedSource.BlockBasedReader<FixedRecord> reader =
(BlockBasedSource.BlockBasedReader<FixedRecord>) source.createReader(null)) {
assertEquals(null, reader.getCurrentBlock());
expectedException.expect(NoSuchElementException.class);
expectedException.expectMessage("No block has been successfully read from");
reader.getCurrent();
}
}
@Test
public void testSplitAtFractionExhaustive() throws Exception {
// A small-sized input is sufficient, because the test verifies that splitting is non-vacuous.
List<FixedRecord> expected = createFixedRecords(20);
String filename =
generateTestFile(
"tmp.avro",
expected,
SyncBehavior.SYNC_REGULAR,
5,
AvroCoder.of(FixedRecord.class),
DataFileConstants.NULL_CODEC);
AvroSource<FixedRecord> source = AvroSource.from(filename).withSchema(FixedRecord.class);
SourceTestUtils.assertSplitAtFractionExhaustive(source, null);
}
@Test
public void testSplitsWithSmallBlocks() throws Exception {
PipelineOptions options = PipelineOptionsFactory.create();
// Test reading from an object file with many small random-sized blocks.
// The file itself doesn't have to be big; we can use a decreased record count.
List<Bird> expected = createRandomRecords(DEFAULT_RECORD_COUNT);
String filename =
generateTestFile(
"tmp.avro",
expected,
SyncBehavior.SYNC_RANDOM,
DEFAULT_RECORD_COUNT / 20 /* max records/block */,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
File file = new File(filename);
// Small minimum bundle size
AvroSource<Bird> source =
AvroSource.from(filename).withSchema(Bird.class).withMinBundleSize(100L);
// Assert that the source produces the expected records
assertEquals(expected, SourceTestUtils.readFromSource(source, options));
List<? extends BoundedSource<Bird>> splits;
int nonEmptySplits;
// Split with the minimum bundle size
splits = source.split(100L, options);
assertTrue(splits.size() > 2);
SourceTestUtils.assertSourcesEqualReferenceSource(source, splits, options);
nonEmptySplits = 0;
for (BoundedSource<Bird> subSource : splits) {
if (SourceTestUtils.readFromSource(subSource, options).size() > 0) {
nonEmptySplits += 1;
}
}
assertTrue(nonEmptySplits > 2);
// Split with larger bundle size
splits = source.split(file.length() / 4, options);
assertTrue(splits.size() > 2);
SourceTestUtils.assertSourcesEqualReferenceSource(source, splits, options);
nonEmptySplits = 0;
for (BoundedSource<Bird> subSource : splits) {
if (SourceTestUtils.readFromSource(subSource, options).size() > 0) {
nonEmptySplits += 1;
}
}
assertTrue(nonEmptySplits > 2);
// Split with the file length
splits = source.split(file.length(), options);
assertTrue(splits.size() == 1);
SourceTestUtils.assertSourcesEqualReferenceSource(source, splits, options);
}
@Test
public void testMultipleFiles() throws Exception {
String baseName = "tmp-";
List<Bird> expected = new ArrayList<>();
for (int i = 0; i < 10; i++) {
List<Bird> contents = createRandomRecords(DEFAULT_RECORD_COUNT / 10);
expected.addAll(contents);
generateTestFile(
baseName + i,
contents,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
}
AvroSource<Bird> source =
AvroSource.from(new File(tmpFolder.getRoot().toString(), baseName + "*").toString())
.withSchema(Bird.class);
List<Bird> actual = SourceTestUtils.readFromSource(source, null);
assertThat(actual, containsInAnyOrder(expected.toArray()));
}
@Test
public void testCreationWithSchema() throws Exception {
List<Bird> expected = createRandomRecords(100);
String filename =
generateTestFile(
"tmp.avro",
expected,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
// Create a source with a schema object
Schema schema = ReflectData.get().getSchema(Bird.class);
AvroSource<GenericRecord> source = AvroSource.from(filename).withSchema(schema);
List<GenericRecord> records = SourceTestUtils.readFromSource(source, null);
assertEqualsWithGeneric(expected, records);
// Create a source with a JSON schema
String schemaString = ReflectData.get().getSchema(Bird.class).toString();
source = AvroSource.from(filename).withSchema(schemaString);
records = SourceTestUtils.readFromSource(source, null);
assertEqualsWithGeneric(expected, records);
}
@Test
public void testSchemaUpdate() throws Exception {
List<Bird> birds = createRandomRecords(100);
String filename =
generateTestFile(
"tmp.avro",
birds,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
AvroSource<FancyBird> source = AvroSource.from(filename).withSchema(FancyBird.class);
List<FancyBird> actual = SourceTestUtils.readFromSource(source, null);
List<FancyBird> expected = new ArrayList<>();
for (Bird bird : birds) {
expected.add(
new FancyBird(
bird.number, bird.species, bird.quality, bird.quantity, null, "MAXIMUM OVERDRIVE"));
}
assertThat(actual, containsInAnyOrder(expected.toArray()));
}
@Test
public void testSchemaStringIsInterned() throws Exception {
List<Bird> birds = createRandomRecords(100);
String filename =
generateTestFile(
"tmp.avro",
birds,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
Metadata fileMetadata = FileSystems.matchSingleFileSpec(filename);
String schema = AvroSource.readMetadataFromFile(fileMetadata.resourceId()).getSchemaString();
// Add "" to the schema to make sure it is not interned.
AvroSource<GenericRecord> sourceA = AvroSource.from(filename).withSchema("" + schema);
AvroSource<GenericRecord> sourceB = AvroSource.from(filename).withSchema("" + schema);
assertSame(sourceA.getReaderSchemaString(), sourceB.getReaderSchemaString());
// Ensure that deserialization still goes through interning
AvroSource<GenericRecord> sourceC = SerializableUtils.clone(sourceB);
assertSame(sourceA.getReaderSchemaString(), sourceC.getReaderSchemaString());
}
@Test
public void testParseFn() throws Exception {
List<Bird> expected = createRandomRecords(100);
String filename =
generateTestFile(
"tmp.avro",
expected,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
AvroSource<Bird> source =
AvroSource.from(filename)
.withParseFn(
input ->
new Bird(
(long) input.get("number"),
input.get("species").toString(),
input.get("quality").toString(),
(long) input.get("quantity")),
AvroCoder.of(Bird.class));
List<Bird> actual = SourceTestUtils.readFromSource(source, null);
assertThat(actual, containsInAnyOrder(expected.toArray()));
}
@Test
public void testDatumReaderFactoryWithGenericRecord() throws Exception {
List<Bird> inputBirds = createRandomRecords(100);
String filename =
generateTestFile(
"tmp.avro",
inputBirds,
SyncBehavior.SYNC_DEFAULT,
0,
AvroCoder.of(Bird.class),
DataFileConstants.NULL_CODEC);
AvroSource.DatumReaderFactory<GenericRecord> factory =
(writer, reader) ->
new GenericDatumReader<GenericRecord>(writer, reader) {
@Override
protected Object readString(Object old, Decoder in) throws IOException {
return super.readString(old, in) + "_custom";
}
};
AvroSource<Bird> source =
AvroSource.from(filename)
.withParseFn(
input ->
new Bird(
(long) input.get("number"),
input.get("species").toString(),
input.get("quality").toString(),
(long) input.get("quantity")),
AvroCoder.of(Bird.class))
.withDatumReaderFactory(factory);
List<Bird> actual = SourceTestUtils.readFromSource(source, null);
List<Bird> expected =
inputBirds.stream()
.map(b -> new Bird(b.number, b.species + "_custom", b.quality + "_custom", b.quantity))
.collect(Collectors.toList());
assertThat(actual, containsInAnyOrder(expected.toArray()));
}
private void assertEqualsWithGeneric(List<Bird> expected, List<GenericRecord> actual) {
assertEquals(expected.size(), actual.size());
for (int i = 0; i < expected.size(); i++) {
Bird fixed = expected.get(i);
GenericRecord generic = actual.get(i);
assertEquals(fixed.number, generic.get("number"));
assertEquals(fixed.quality, generic.get("quality").toString()); // From Avro util.Utf8
assertEquals(fixed.quantity, generic.get("quantity"));
assertEquals(fixed.species, generic.get("species").toString());
}
}
@Test
public void testDisplayData() {
AvroSource<Bird> source =
AvroSource.from("foobar.txt").withSchema(Bird.class).withMinBundleSize(1234);
DisplayData displayData = DisplayData.from(source);
assertThat(displayData, hasDisplayItem("filePattern", "foobar.txt"));
assertThat(displayData, hasDisplayItem("minBundleSize", 1234));
}
@Test
public void testReadMetadataWithCodecs() throws Exception {
List<Bird> expected = createRandomRecords(DEFAULT_RECORD_COUNT);
// Test reading files generated using all codecs.
for (String codec : avroSupportedCodec()) {
String filename =
generateTestFile(
codec, expected, SyncBehavior.SYNC_DEFAULT, 0, AvroCoder.of(Bird.class), codec);
Metadata fileMeta = FileSystems.matchSingleFileSpec(filename);
AvroSource.AvroMetadata metadata = AvroSource.readMetadataFromFile(fileMeta.resourceId());
assertEquals(codec, metadata.getCodec());
}
}
@Test
public void testReadSchemaString() throws Exception {
List<Bird> expected = createRandomRecords(DEFAULT_RECORD_COUNT);
String codec = DataFileConstants.NULL_CODEC;
String filename =
generateTestFile(
codec, expected, SyncBehavior.SYNC_DEFAULT, 0, AvroCoder.of(Bird.class), codec);
Metadata fileMeta = FileSystems.matchSingleFileSpec(filename);
AvroSource.AvroMetadata metadata = AvroSource.readMetadataFromFile(fileMeta.resourceId());
// By default, parse validates the schema, which is what we want.
Schema schema = new Schema.Parser().parse(metadata.getSchemaString());
assertEquals(4, schema.getFields().size());
}
@Test
public void testCreateFromMetadata() throws Exception {
List<Bird> expected = createRandomRecords(DEFAULT_RECORD_COUNT);
String codec = DataFileConstants.NULL_CODEC;
String filename =
generateTestFile(
codec, expected, SyncBehavior.SYNC_DEFAULT, 0, AvroCoder.of(Bird.class), codec);
Metadata fileMeta = FileSystems.matchSingleFileSpec(filename);
AvroSource<GenericRecord> source = AvroSource.from(fileMeta);
AvroSource<Bird> sourceWithSchema = source.withSchema(Bird.class);
AvroSource<Bird> sourceWithSchemaWithMinBundleSize = sourceWithSchema.withMinBundleSize(1234);
assertEquals(FileBasedSource.Mode.SINGLE_FILE_OR_SUBRANGE, source.getMode());
assertEquals(FileBasedSource.Mode.SINGLE_FILE_OR_SUBRANGE, sourceWithSchema.getMode());
assertEquals(
FileBasedSource.Mode.SINGLE_FILE_OR_SUBRANGE, sourceWithSchemaWithMinBundleSize.getMode());
}
/**
* Class that will encode to a fixed size: 16 bytes.
*
* <p>Each object has a 15-byte array. Avro encodes an object of this type as a byte array, so
* each encoded object will consist of 1 byte that encodes the length of the array, followed by 15
* bytes.
*/
@DefaultCoder(AvroCoder.class)
public static class FixedRecord {
private byte[] value = new byte[15];
public FixedRecord() {
this(0);
}
public FixedRecord(int i) {
value[0] = (byte) i;
value[1] = (byte) (i >> 8);
value[2] = (byte) (i >> 16);
value[3] = (byte) (i >> 24);
}
public int asInt() {
return value[0] | (value[1] << 8) | (value[2] << 16) | (value[3] << 24);
}
@Override
public boolean equals(@Nullable Object o) {
if (o instanceof FixedRecord) {
FixedRecord other = (FixedRecord) o;
return this.asInt() == other.asInt();
}
return false;
}
@Override
public int hashCode() {
return toString().hashCode();
}
@Override
public String toString() {
return Integer.toString(this.asInt());
}
}
/** Create a list of count 16-byte records. */
private static List<FixedRecord> createFixedRecords(int count) {
List<FixedRecord> records = new ArrayList<>();
for (int i = 0; i < count; i++) {
records.add(new FixedRecord(i));
}
return records;
}
/** Class used as the record type in tests. */
@DefaultCoder(AvroCoder.class)
static class Bird {
long number;
String species;
String quality;
long quantity;
public Bird() {}
public Bird(long number, String species, String quality, long quantity) {
this.number = number;
this.species = species;
this.quality = quality;
this.quantity = quantity;
}
@Override
public String toString() {
return MoreObjects.toStringHelper(Bird.class)
.addValue(number)
.addValue(species)
.addValue(quantity)
.addValue(quality)
.toString();
}
@Override
public boolean equals(@Nullable Object obj) {
if (obj instanceof Bird) {
Bird other = (Bird) obj;
return Objects.equals(species, other.species)
&& Objects.equals(quality, other.quality)
&& quantity == other.quantity
&& number == other.number;
}
return false;
}
@Override
public int hashCode() {
return Objects.hash(number, species, quality, quantity);
}
}
/**
* Class used as the record type in tests.
*
* <p>Contains nullable fields and fields with default values. Can be read using a file written
* with the Bird schema.
*/
@DefaultCoder(AvroCoder.class)
public static class FancyBird {
long number;
String species;
String quality;
long quantity;
@org.apache.avro.reflect.Nullable String habitat;
@AvroDefault("\"MAXIMUM OVERDRIVE\"")
String fancinessLevel;
public FancyBird() {}
public FancyBird(
long number,
String species,
String quality,
long quantity,
String habitat,
String fancinessLevel) {
this.number = number;
this.species = species;
this.quality = quality;
this.quantity = quantity;
this.habitat = habitat;
this.fancinessLevel = fancinessLevel;
}
@Override
public String toString() {
return MoreObjects.toStringHelper(FancyBird.class)
.addValue(number)
.addValue(species)
.addValue(quality)
.addValue(quantity)
.addValue(habitat)
.addValue(fancinessLevel)
.toString();
}
@Override
public boolean equals(@Nullable Object obj) {
if (obj instanceof FancyBird) {
FancyBird other = (FancyBird) obj;
return Objects.equals(species, other.species)
&& Objects.equals(quality, other.quality)
&& quantity == other.quantity
&& number == other.number
&& Objects.equals(fancinessLevel, other.fancinessLevel)
&& Objects.equals(habitat, other.habitat);
}
return false;
}
@Override
public int hashCode() {
return Objects.hash(number, species, quality, quantity, habitat, fancinessLevel);
}
}
/** Create a list of n random records. */
private static List<Bird> createRandomRecords(long n) {
String[] qualities = {
"miserable", "forelorn", "fidgity", "squirrelly", "fanciful", "chipper", "lazy"
};
String[] species = {"pigeons", "owls", "gulls", "hawks", "robins", "jays"};
Random random = new Random(0);
List<Bird> records = new ArrayList<>();
for (long i = 0; i < n; i++) {
Bird bird = new Bird();
bird.quality = qualities[random.nextInt(qualities.length)];
bird.species = species[random.nextInt(species.length)];
bird.number = i;
bird.quantity = random.nextLong();
records.add(bird);
}
return records;
}
}
```
|
```xml
import { provideUnits } from './provideUnits';
describe('provideUnits', () => {
it('can provide units', () => {
['left', 'right', 'top', 'bottom', 'borderWidth'].forEach((property: string) => {
const testSet: (string | number)[] = [property, 1];
provideUnits(testSet, 0);
expect(testSet).toEqual([property, '1px']);
});
});
it('ignores css variables', () => {
const testSet: (string | number)[] = ['--button-fontWeight', 600];
provideUnits(testSet, 0);
expect(testSet).toEqual(['--button-fontWeight', '600']);
});
it('can provide units for directional props', () => {
['padding', 'margin', 'border'].forEach((property: string) => {
const testSet: (string | number)[] = [
property,
1,
property + 'Left',
1,
property + 'Right',
1,
property + 'Top',
1,
property + 'Bottom',
1,
];
provideUnits(testSet, 0);
provideUnits(testSet, 2);
provideUnits(testSet, 4);
provideUnits(testSet, 6);
provideUnits(testSet, 8);
expect(testSet).toEqual([
property,
'1px',
property + 'Left',
'1px',
property + 'Right',
'1px',
property + 'Top',
'1px',
property + 'Bottom',
'1px',
]);
});
});
it('ignores opacity', () => {
const testSet = ['opacity', 0];
provideUnits(testSet, 0);
expect(testSet).toEqual(['opacity', '0']);
});
});
```
|
```php
<?php
/*
* This file is part of Piplin.
*
*
* For the full copyright and license information, please view the LICENSE
* file that was distributed with this source code.
*/
return [
'label' => '',
'file_name' => '',
'file_size' => '',
];
```
|
```html
<!doctype html>
<title>CSV Data</title>
<link rel=stylesheet type=text/css href="{{ url_for('static', filename='stylesheets/bootstrap.min.css') }}">
<style>
table.dataframe, .dataframe th, .dataframe td {
border: none;
border-bottom: 1px solid #C8C8C8;
border-collapse: collapse;
text-align:center;
padding: 10px;
margin-bottom: 40px;
font-size: 0.9em;
}
.frame th {
background-color: #00ccff;
color: white;
}
tr:nth-child(odd) {
background-color:#eee;
}
tr:nth-child(even){
background-color:#fff;
}
tr:hover {
background-color: #aaa;
}
</style>
<div class="page">
<h1>{{ date_string }} </h1>
{{ df|safe }}
</div>
```
|
```css
Use `text-transform` to avoid screen-reader pronunciation errors
Use ```list-style-type``` to change the marker type in lists
Use `em` instead of `px` for sizes
Comma-separated lists
```
|
Mauro Teixeira Thorpe (born 20 November 1956) is a Brazilian former footballer who is last known to have played as a midfielder for Íbis.
Playing career
Shampoo played for Brazilian side Íbis for ten years.
Post-playing career
After retirement from professional football, Shampoo worked as a hairdresser.
In popular culture
Shampoo has a film about him called Mauro Shampoo – jogador, cabeleireiro e homem.
Personal life
Shampoo has thirteen siblings.
References
1956 births
Brazilian men's footballers
Living people
Men's association football midfielders
|
```shell
#!/bin/bash
#
#
#
# path_to_url
#
# Unless required by applicable law or agreed to in writing, software
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
#
firstlaunch=${DATA_FOLDER}/.firstlaunch
PG_CTL=$(find /usr/lib/postgresql/ -name pg_ctl)
if [ ! -d ${PGDATA} ]; then
mkdir -p ${PGDATA}
${PG_CTL} initdb
fi
echo "Starting Postgresql..."
${PG_CTL} start
RETRIES="${PG_ISREADY_RETRIES:-300}"
until pg_isready -U ${pkg.user} -d postgres --quiet || [ $RETRIES -eq 0 ]
do
echo "Connecting to Postgres, $((RETRIES--)) attempts left..."
sleep 1
done
if [ ! -f ${firstlaunch} ]; then
echo "Creating database..."
psql -U ${pkg.user} -d postgres -c "CREATE DATABASE thingsboard"
fi
echo "Postgresql is ready"
```
|
```objective-c
#ifndef CONFLUO_TYPES_ARITHMETIC_OPS_H_
#define CONFLUO_TYPES_ARITHMETIC_OPS_H_
#include <vector>
#include <cstdint>
#include <cstdlib>
#include <cstdio>
#include <cstring>
#include "exceptions.h"
#include "raw_data.h"
namespace confluo {
/**
* Unary arithmetic/bitwise operators
*/
enum unary_op_id
: uint8_t {
ASSIGN = 0, //!< ASSIGN (=)
NEGATIVE = 1, //!< NEGATIVE (-)
POSITIVE = 2, //!< POSITIVE (+)
BW_NOT = 3 //!< BW_NOT (~)
};
/**
* Binary arithmetic/bitwise operators
*/
enum binary_op_id
: uint8_t {
ADD = 0, //!< ADD (+)
SUBTRACT = 1, //!< SUBTRACT (-)
MULTIPLY = 2, //!< MULTIPLY (*)
DIVIDE = 3, //!< DIVIDE (/)
MODULO = 4, //!< MODULO (%)
BW_AND = 5, //!< BW_AND (&)
BW_OR = 6, //!< BW_OR (|)
BW_XOR = 7, //!< BW_XOR (^)
BW_LSHIFT = 8, //!< BW_LSHIFT (<<)
BW_RSHIFT = 9 //!< BW_RSHIFT (>>)
};
/** Function pointer for a unary operator */
typedef void (*unary_op_t)(void *res, const immutable_raw_data &v);
/** Function pointer for a binary operator */
typedef void (*binary_op_t)(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2);
/** List of unary operators */
typedef std::vector<unary_op_t> unary_ops_t;
/** List of binary operators */
typedef std::vector<binary_op_t> binary_ops_t;
// Unary arithmetic operators
/**
* Assigns the value of the raw immutable data to the result pointer
* @tparam T The type of data
* @param res The pointer to the resultant data
* @param v The raw immutable data to assign to the result pointer
*/
template<typename T>
inline void assign(void *res, const immutable_raw_data &v) {
*(reinterpret_cast<T *>(res)) = v.as<T>();
}
/**
* Assigns the value of the raw immutable data to the result pointer for
* string types
*
* @param res The result pointer to initialize
* @param v The value of the raw immutable data to copy
*/
template<>
inline void assign<std::string>(void *res, const immutable_raw_data &v) {
memcpy(res, v.ptr, v.size);
}
/**
* Assigns the value of the raw immutable data to the result pointer for
* void types
*
* @param res The result pointer to initialize
* @param v The raw immutable data to copy
*/
template<>
inline void assign<void>(void *, const immutable_raw_data &v) {
return;
}
/**
* Negates the raw immutable data and stores it in the result pointer
*
* @tparam T The type of data
* @param res The result pointer which will contain the negated value
* @param v The immutable raw data to negate
*/
template<typename T>
inline void negative(void *res, const immutable_raw_data &v) {
*(reinterpret_cast<T *>(res)) = -v.as<T>();
}
/**
* Negates the raw immutable data and has result pointer point
* to the negated value for the string type
*
* @param res The result pointer that will point to the negated value
* @param v The immutable raw data to be negated
*
* @throw unsupported_exception
*/
template<>
inline void negative<std::string>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "- not supported for string type");
}
/**
* Negates the raw immutable data and has the result pointer point to the
* negated value for the none type
*
* @param res The result pointer that will point to the negated value
* @param v The immutable raw data to be negated
*
* @throw unsupported_exception
*/
template<>
inline void negative<void>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "- not supported for none type");
}
/**
* Has the result pointer point to the value of the raw immutable data
*
* @tparam T The type of data
* @param res The result pointer that points to the value
* @param v The raw immutable data that contains the value
*/
template<typename T>
inline void positive(void *res, const immutable_raw_data &v) {
*(reinterpret_cast<T *>(res)) = +v.as<T>();
}
/**
* Has the result pointer point to the value of the raw immutable data
* for the string type
*
* @tparam T The type of data
* @param res The result pointer that points to the value
* @param v The raw immutable data that contains the value
* @throw unsupported_exception
*/
template<>
inline void positive<std::string>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "+ not supported for string type");
}
/**
* Has the result pointer point to the value of the raw immutable data
* for the void type
*
* @tparam T The type of data
* @param res The result pointer that points to the value
* @param v The raw immutable data that contains the value
* @throw unsupported_exception
*/
template<>
inline void positive<void>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "+ not supported for none type");
}
// Binary arithmetic operators
/**
* Performs binary addition and stores the result in the result pointer
*
* @tparam T The data type of the operands
* @param res The result of the addition
* @param v1 The first operand of the addition expression
* @param v2 The second operand of the addition expression
*/
template<typename T>
inline void add(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() + v2.as<T>();
}
/**
* Performs binary addition and stores the result in the result pointer
* for the string type
*
* @param res The result of the addition
* @param v1 The first operand of the addition expression
* @param v2 The second operand of the addition expression
*/
template<>
inline void add<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "+ not supported for string type");
}
/**
* Performs binary addition and stores the result in the result pointer
* for the none type
*
* @param res The result of the addition
* @param v1 The first operand of the addition expression
* @param v2 The second operand of the addition expression
* @throw unsupported_exception
*/
template<>
inline void add<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "+ not supported for none type");
}
/**
* Performs binary subtraction and stores the result in the result pointer
*
* @tparam The data type
* @param res The result of the subtraction
* @param v1 The first operand of the subtraction expression
* @param v2 The second operand of the subtraction expression
*/
template<typename T>
inline void subtract(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() - v2.as<T>();
}
/**
* Performs binary subtraction and stores the result in the result pointer
* for strings
*
* @param res The result of the subtraction
* @param v1 The first operand of the subtraction expression
* @param v2 The second operand of the subtraction expression
* @throw unsupported_exception
*/
template<>
inline void subtract<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "- not supported for string type");
}
/**
* Performs binary subtraction and stores the result in the result pointer
* for none type
*
* @param res The result of the subtraction
* @param v1 The first operand of the subtraction expression
* @param v2 The second operand of the subtraction expression
* @throw unsupported_exception
*/
template<>
inline void subtract<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "- not supported for none type");
}
/**
* Performs binary multiplication and stores the result in the
* result pointer
*
* @tparam T The data type of the immutable values
* @param res The result of the multiplication
* @param v1 The first operand of the multiplication expression
* @param v2 The second operand of the multiplication expression
*/
template<typename T>
inline void multiply(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() * v2.as<T>();
}
/**
* Performs binary multiplication and stores the result in the
* result pointer
*
* @tparam T The data type of the immutable values
* @param res The result of the multiplication
* @param v1 The first operand of the multiplication expression
* @param v2 The second operand of the multiplication expression
*/
template<>
inline void multiply<bool>(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<bool *>(res)) = v1.as<bool>() && v2.as<bool>();
}
/**
* Performs binary multiplication and stores the result in the
* result pointer for strings
*
* @param res The result of the multiplication
* @param v1 The first operand of the multiplication expression
* @param v2 The second operand of the multiplication expression
* @throw unsupported_exception
*/
template<>
inline void multiply<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "* not supported for string type");
}
/**
* Performs binary multiplication and stores the result in the
* result pointer for the none type
*
* @param res The result of the addition
* @param v1 The first operand of the multiplication expression
* @param v2 The second operand of the multiplication expression
* @throw unsupported_exception
*/
template<>
inline void multiply<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "* not supported for none type");
}
/**
* Performs binary division and stores the result in the
* result pointer
*
* @tparam The data type of the immutable values
* @param res The result of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
*/
template<typename T>
inline void divide(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() / v2.as<T>();
}
/**
* Performs binary division and stores the result in the
* result pointer for strings
*
* @param res The result of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void divide<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "/ not supported for string type");
}
/**
* Performs binary division and stores the result in the
* result pointer for the none type
*
* @param res The result of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void divide<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "/ not supported for none type");
}
/**
* Performs binary division and stores the remainder in the
* result pointer
*
* @tparam The data type of the immutable values
* @param res The remainder of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
*/
template<typename T>
inline void modulo(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() % v2.as<T>();
}
/**
* Performs binary division and stores the remainder in the
* result pointer for strings
*
* @param res The remainder of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void modulo<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "% not supported for string type");
}
/**
* Performs binary division and stores the remainder in the
* result pointer for the none type
*
* @param res The remainder of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void modulo<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "% not supported for none type");
}
/**
* Performs binary division and stores the remainder in the
* result pointer for the float type
*
* @param res The remainder of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void modulo<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "% not supported for float type");
}
/**
* Performs binary division and stores the remainder in the
* result pointer for the double type
*
* @param res The remainder of the division
* @param v1 The immutable value containing the dividend
* @param v2 The immutable value containing the divisor
* @throw unsupported_exception
*/
template<>
inline void modulo<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "% not supported for double type");
}
// Bitwise operators
/**
* Performs bitwise not operation and stores the result in the
* result pointer
*
* @tparam T The type of data the immutable value contains
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
*/
template<typename T>
inline void bw_not(void *res, const immutable_raw_data &v) {
*(reinterpret_cast<T *>(res)) = ~v.as<T>();
}
/**
* Performs bitwise not operation and stores the result in the
* result pointer
*
* @tparam T The type of data the immutable value contains
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
*/
template<>
inline void bw_not<bool>(void *res, const immutable_raw_data &v) {
*(reinterpret_cast<bool *>(res)) = !v.as<bool>();
}
/**
* Performs bitwise not operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
* @throw unsupported_exception
*/
template<>
inline void bw_not<std::string>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "~ not supported for string type");
}
/**
* Performs bitwise not operation and stores the result in the
* result pointer for the void type
*
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
* @throw unsupported_exception
*/
template<>
inline void bw_not<void>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "~ not supported for none type");
}
/**
* Performs bitwise not operation and stores the result in the
* result pointer for the float type
*
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
* @throw unsupported_exception
*/
template<>
inline void bw_not<float>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "~ not supported for float type");
}
/**
* Performs bitwise not operation and stores the result in the
* result pointer for the double type
*
* @param res The result of the bitwise not operation
* @param v1 The immutable value that the operator is applied to
* @throw unsupported_exception
*/
template<>
inline void bw_not<double>(void *, const immutable_raw_data &v) {
THROW(unsupported_exception, "~ not supported for double type");
}
/**
* Performs bitwise and operation and stores the result in the
* result pointer
*
* @tparam T The data type of the immutable data values
* @param res The result of the bitwise and operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
*/
template<typename T>
inline void bw_and(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() & v2.as<T>();
}
/**
* Performs bitwise and operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise and operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_and<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "& not supported for string type");
}
/**
* Performs bitwise and operation and stores the result in the
* result pointer for the none type
*
* @param res The result of the bitwise and operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_and<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "& not supported for none type");
}
/**
* Performs bitwise and operation and stores the result in the
* result pointer for floats
*
* @param res The result of the bitwise and operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_and<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "& not supported for float type");
}
/**
* Performs bitwise and operation and stores the result in the
* result pointer for doubles
*
* @param res The result of the bitwise and operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_and<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "& not supported for double type");
}
/**
* Performs bitwise or operation and stores the result in the
* result pointer
*
* @tparam T The type of data of the immutable values
* @param res The result of the bitwise or operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
*/
template<typename T>
inline void bw_or(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() | v2.as<T>();
}
/**
* Performs bitwise or operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise or operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_or<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "| not supported for string type");
}
/**
* Performs bitwise or operation and stores the result in the
* result pointer for the none type
*
* @param res The result of the bitwise or operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_or<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "| not supported for none type");
}
/**
* Performs bitwise or operation and stores the result in the
* result pointer for the float type
*
* @param res The result of the bitwise or operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_or<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "| not supported for float type");
}
/**
* Performs bitwise or operation and stores the result in the
* result pointer for the double type
*
* @param res The result of the bitwise or operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_or<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "| not supported for double type");
}
/**
* Performs bitwise xor operation and stores the result in the
* result pointer
*
* @tparam T the type of the immutable raw data
* @param res The result of the bitwise xor operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
*/
template<typename T>
inline void bw_xor(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>()
^ v2.as<T>();
}
/**
* Performs bitwise xor operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise xor operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_xor<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "^ not supported for string type");
}
/**
* Performs bitwise xor operation and stores the result in the
* result pointer for the none type
*
* @param res The result of the bitwise xor operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_xor<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "^ not supported for none type");
}
/**
* Performs bitwise xor operation and stores the result in the
* result pointer for floats
*
* @param res The result of the bitwise xor operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_xor<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "^ not supported for float type");
}
/**
* Performs bitwise xor operation and stores the result in the
* result pointer for doubles
*
* @param res The result of the bitwise xor operation
* @param v1 The first immutable value in the expression
* @param v2 The second immutable value in the expression
* @throw unsupported_exception
*/
template<>
inline void bw_xor<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "^ not supported for double type");
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer
*
* @tparam T The data type of the immutable raw data
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
*/
template<typename T>
inline void bw_lshift(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() << v2.as<T>();
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_lshift<bool>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "<< not supported for bool type");
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_lshift<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "<< not supported for string type");
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer for the none type
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_lshift<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "<< not supported for none type");
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer for floats
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_lshift<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "<< not supported for float type");
}
/**
* Performs bitwise left shift operation and stores the result in the
* result pointer for doubles
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_lshift<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, "<< not supported for double type");
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer
*
* @tparam T The data type of the immutable raw data values
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
*/
template<typename T>
inline void bw_rshift(void *res, const immutable_raw_data &v1, const immutable_raw_data &v2) {
*(reinterpret_cast<T *>(res)) = v1.as<T>() >> v2.as<T>();
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer for bools
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_rshift<bool>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, ">> not supported for bool type");
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer for strings
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_rshift<std::string>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, ">> not supported for string type");
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer for the none type
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_rshift<void>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, ">> not supported for none type");
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer for floats
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_rshift<float>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, ">> not supported for float type");
}
/**
* Performs bitwise right shift operation and stores the result in the
* result pointer for doubles
*
* @param res The result of the bitwise xor operation
* @param v1 The value to shift
* @param v2 The amount to shift by
* @throw unsupported_exception
*/
template<>
inline void bw_rshift<double>(void *, const immutable_raw_data &v1, const immutable_raw_data &v2) {
THROW(unsupported_exception, ">> not supported for double type");
}
/**
* Gets a list of the unary operators for the given type
*
* @tparam T The data type the operators act on
*
* @return A vector containing the unary operator functions
*/
template<typename T>
inline unary_ops_t init_unaryops() {
return {assign<T>, negative<T>, positive<T>, bw_not<T>};
}
/**
* Gets a list of binary operators for the given type
*
* @tparam T The data type the operators act on
*
* @return A vector containing the binary operator functions
*/
template<typename T>
inline binary_ops_t init_binaryops() {
return {add<T>, subtract<T>, multiply<T>, divide<T>, modulo<T>, bw_and<T>, bw_or<T>, bw_xor<T>, bw_lshift<T>,
bw_rshift<T>};
}
}
#endif /* CONFLUO_TYPES_ARITHMETIC_OPS_H_ */
```
|
```xml
<?xml version="1.0" encoding="UTF-8"?>
<definitions xmlns="path_to_url" xmlns:xsi="path_to_url" xmlns:xsd="path_to_url" xmlns:flowable="path_to_url" xmlns:bpmndi="path_to_url" xmlns:omgdc="path_to_url" xmlns:omgdi="path_to_url" typeLanguage="path_to_url" expressionLanguage="path_to_url" targetNamespace="path_to_url">
<process id="dogeProcess" name="DogeProcess" isExecutable="true">
<startEvent id="startEvent1"></startEvent>
<subProcess id="sid-1AA38D08-310B-4BFE-A6AD-F32488643000" name="subProcess">
<multiInstanceLoopCharacteristics isSequential="true" flowable:collection="photos" flowable:elementVariable="photo"></multiInstanceLoopCharacteristics>
<serviceTask id="sid-FCFA0F9D-8796-43E2-ACFD-13AF3465F929" name="Send to photo processing service" flowable:expression="${photoService.processPhoto(photo.id)}"></serviceTask>
<startEvent id="sid-821438D0-CD00-43F1-A847-42112E44BA12"></startEvent>
<receiveTask id="wait" name="Wait for completion"></receiveTask>
<endEvent id="sid-9BDA1151-ECDE-472B-A9E8-4F6A0E5C2901"></endEvent>
<sequenceFlow id="sid-4B4908D7-A6A3-45B6-9267-D3C355154510" sourceRef="sid-821438D0-CD00-43F1-A847-42112E44BA12" targetRef="sid-FCFA0F9D-8796-43E2-ACFD-13AF3465F929"></sequenceFlow>
<sequenceFlow id="sid-3667040B-ED6C-424E-BFAD-69718B40FD1C" sourceRef="sid-FCFA0F9D-8796-43E2-ACFD-13AF3465F929" targetRef="wait"></sequenceFlow>
<sequenceFlow id="sid-518F987E-BBEE-4565-BC94-00E02430E81A" sourceRef="wait" targetRef="sid-9BDA1151-ECDE-472B-A9E8-4F6A0E5C2901"></sequenceFlow>
</subProcess>
<sequenceFlow id="sid-13DC3678-E57E-4E0B-9F80-55E11BF7304F" sourceRef="startEvent1" targetRef="sid-1AA38D08-310B-4BFE-A6AD-F32488643000"></sequenceFlow>
<boundaryEvent id="sid-CF518067-8143-4CC3-AE33-C3FE326E9EC3" attachedToRef="sid-1AA38D08-310B-4BFE-A6AD-F32488643000" cancelActivity="true">
<timerEventDefinition>
<timeDuration>PT10M</timeDuration>
</timerEventDefinition>
</boundaryEvent>
<userTask id="sid-A51C2956-123E-44AD-9548-0239C6164596" name="Process photos manually" flowable:assignee="$INITIATOR">
<extensionElements>
<modeler:initiator-can-complete xmlns:modeler="path_to_url"><![CDATA[true]]></modeler:initiator-can-complete>
</extensionElements>
</userTask>
<sequenceFlow id="sid-F7745084-E689-431F-8F78-0FCAFE376EAE" sourceRef="sid-CF518067-8143-4CC3-AE33-C3FE326E9EC3" targetRef="sid-A51C2956-123E-44AD-9548-0239C6164596"></sequenceFlow>
<endEvent id="sid-463B9727-4C0A-4C1A-BE84-09D6842D3AF7"></endEvent>
<sequenceFlow id="sid-ECF49058-5E5D-49BB-804E-BFD9FED4D035" sourceRef="sid-A51C2956-123E-44AD-9548-0239C6164596" targetRef="sid-463B9727-4C0A-4C1A-BE84-09D6842D3AF7"></sequenceFlow>
<userTask id="sid-8E972729-8F07-4413-86A9-392A34D949EF" name="Review result" flowable:assignee="$INITIATOR">
<extensionElements>
<modeler:initiator-can-complete xmlns:modeler="path_to_url"><![CDATA[true]]></modeler:initiator-can-complete>
</extensionElements>
</userTask>
<sequenceFlow id="sid-0F909E8C-17D2-468B-8780-6DFF2C539369" sourceRef="sid-1AA38D08-310B-4BFE-A6AD-F32488643000" targetRef="sid-8E972729-8F07-4413-86A9-392A34D949EF"></sequenceFlow>
<exclusiveGateway id="sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B" default="sid-733ED6FC-9CF7-485B-A67E-B6BF8D1F90DA"></exclusiveGateway>
<sequenceFlow id="sid-952ED7E5-9F9B-42E8-9212-DECA33D7A6D2" sourceRef="sid-8E972729-8F07-4413-86A9-392A34D949EF" targetRef="sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B"></sequenceFlow>
<endEvent id="sid-DA92F2D8-10E6-4CFD-B44C-81998A8D6FAB"></endEvent>
<sequenceFlow id="sid-733ED6FC-9CF7-485B-A67E-B6BF8D1F90DA" sourceRef="sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B" targetRef="sid-DA92F2D8-10E6-4CFD-B44C-81998A8D6FAB"></sequenceFlow>
<sequenceFlow id="sid-CCE52848-81CA-478A-B0E1-B4A55D419439" sourceRef="sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B" targetRef="sid-58F18100-CA85-4D93-85F6-D61058913735">
<conditionExpression xsi:type="tFormalExpression"><![CDATA[${!approved}]]></conditionExpression>
</sequenceFlow>
<serviceTask id="sid-58F18100-CA85-4D93-85F6-D61058913735" name="Send failure email" flowable:type="mail">
<extensionElements>
<flowable:field name="to">
<flowable:string><![CDATA[test@blah.com]]></flowable:string>
</flowable:field>
<flowable:field name="from">
<flowable:string><![CDATA[no-reply@activiti.com]]></flowable:string>
</flowable:field>
<flowable:field name="html">
<flowable:string><![CDATA[Sorry, it didn't work out]]></flowable:string>
</flowable:field>
</extensionElements>
</serviceTask>
<endEvent id="sid-4A0917CE-835E-4D99-B771-8D02346009D3"></endEvent>
<sequenceFlow id="sid-D1A2FE7F-69A7-4712-9D72-4D52335388BE" sourceRef="sid-58F18100-CA85-4D93-85F6-D61058913735" targetRef="sid-4A0917CE-835E-4D99-B771-8D02346009D3"></sequenceFlow>
</process>
<bpmndi:BPMNDiagram id="BPMNDiagram_dogeProcess">
<bpmndi:BPMNPlane bpmnElement="dogeProcess" id="BPMNPlane_dogeProcess">
<bpmndi:BPMNShape bpmnElement="startEvent1" id="BPMNShape_startEvent1">
<omgdc:Bounds height="30.0" width="30.0" x="60.0" y="225.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-1AA38D08-310B-4BFE-A6AD-F32488643000" id="BPMNShape_sid-1AA38D08-310B-4BFE-A6AD-F32488643000">
<omgdc:Bounds height="268.0" width="387.0" x="167.0" y="106.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-FCFA0F9D-8796-43E2-ACFD-13AF3465F929" id="BPMNShape_sid-FCFA0F9D-8796-43E2-ACFD-13AF3465F929">
<omgdc:Bounds height="80.0" width="100.0" x="240.0" y="199.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-821438D0-CD00-43F1-A847-42112E44BA12" id="BPMNShape_sid-821438D0-CD00-43F1-A847-42112E44BA12">
<omgdc:Bounds height="30.0" width="30.0" x="180.0" y="224.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="wait" id="BPMNShape_wait">
<omgdc:Bounds height="80.0" width="100.0" x="375.0" y="199.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-9BDA1151-ECDE-472B-A9E8-4F6A0E5C2901" id="BPMNShape_sid-9BDA1151-ECDE-472B-A9E8-4F6A0E5C2901">
<omgdc:Bounds height="28.0" width="28.0" x="510.0" y="225.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-CF518067-8143-4CC3-AE33-C3FE326E9EC3" id="BPMNShape_sid-CF518067-8143-4CC3-AE33-C3FE326E9EC3">
<omgdc:Bounds height="31.0" width="31.0" x="434.49292161701123" y="359.2036040120936"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-A51C2956-123E-44AD-9548-0239C6164596" id="BPMNShape_sid-A51C2956-123E-44AD-9548-0239C6164596">
<omgdc:Bounds height="80.0" width="100.0" x="515.6531252612749" y="405.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-463B9727-4C0A-4C1A-BE84-09D6842D3AF7" id="BPMNShape_sid-463B9727-4C0A-4C1A-BE84-09D6842D3AF7">
<omgdc:Bounds height="28.0" width="28.0" x="660.6531252612749" y="431.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-8E972729-8F07-4413-86A9-392A34D949EF" id="BPMNShape_sid-8E972729-8F07-4413-86A9-392A34D949EF">
<omgdc:Bounds height="80.0" width="100.0" x="599.0" y="200.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B" id="BPMNShape_sid-CC2BFD5B-E700-4C3A-BB7D-EC47A6BA7C0B">
<omgdc:Bounds height="40.0" width="40.0" x="744.0" y="220.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-DA92F2D8-10E6-4CFD-B44C-81998A8D6FAB" id="BPMNShape_sid-DA92F2D8-10E6-4CFD-B44C-81998A8D6FAB">
<omgdc:Bounds height="28.0" width="28.0" x="829.0" y="226.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-58F18100-CA85-4D93-85F6-D61058913735" id="BPMNShape_sid-58F18100-CA85-4D93-85F6-D61058913735">
<omgdc:Bounds height="80.0" width="100.0" x="829.0" y="90.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNShape bpmnElement="sid-4A0917CE-835E-4D99-B771-8D02346009D3" id="BPMNShape_sid-4A0917CE-835E-4D99-B771-8D02346009D3">
<omgdc:Bounds height="28.0" width="28.0" x="974.0" y="116.0"></omgdc:Bounds>
</bpmndi:BPMNShape>
<bpmndi:BPMNEdge bpmnElement="sid-733ED6FC-9CF7-485B-A67E-B6BF8D1F90DA" id="BPMNEdge_sid-733ED6FC-9CF7-485B-A67E-B6BF8D1F90DA">
<omgdi:waypoint x="784.0" y="240.0"></omgdi:waypoint>
<omgdi:waypoint x="829.0" y="240.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-518F987E-BBEE-4565-BC94-00E02430E81A" id="BPMNEdge_sid-518F987E-BBEE-4565-BC94-00E02430E81A">
<omgdi:waypoint x="475.0" y="239.0"></omgdi:waypoint>
<omgdi:waypoint x="510.0" y="239.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-4B4908D7-A6A3-45B6-9267-D3C355154510" id="BPMNEdge_sid-4B4908D7-A6A3-45B6-9267-D3C355154510">
<omgdi:waypoint x="210.0" y="239.0"></omgdi:waypoint>
<omgdi:waypoint x="240.0" y="239.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-D1A2FE7F-69A7-4712-9D72-4D52335388BE" id="BPMNEdge_sid-D1A2FE7F-69A7-4712-9D72-4D52335388BE">
<omgdi:waypoint x="929.0" y="130.0"></omgdi:waypoint>
<omgdi:waypoint x="974.0" y="130.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-ECF49058-5E5D-49BB-804E-BFD9FED4D035" id="BPMNEdge_sid-ECF49058-5E5D-49BB-804E-BFD9FED4D035">
<omgdi:waypoint x="615.6531252612749" y="445.0"></omgdi:waypoint>
<omgdi:waypoint x="660.6531252612749" y="445.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-0F909E8C-17D2-468B-8780-6DFF2C539369" id="BPMNEdge_sid-0F909E8C-17D2-468B-8780-6DFF2C539369">
<omgdi:waypoint x="554.0" y="240.0"></omgdi:waypoint>
<omgdi:waypoint x="599.0" y="240.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-CCE52848-81CA-478A-B0E1-B4A55D419439" id="BPMNEdge_sid-CCE52848-81CA-478A-B0E1-B4A55D419439">
<omgdi:waypoint x="764.0" y="220.0"></omgdi:waypoint>
<omgdi:waypoint x="764.0" y="130.0"></omgdi:waypoint>
<omgdi:waypoint x="829.0" y="130.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-3667040B-ED6C-424E-BFAD-69718B40FD1C" id="BPMNEdge_sid-3667040B-ED6C-424E-BFAD-69718B40FD1C">
<omgdi:waypoint x="340.0" y="239.0"></omgdi:waypoint>
<omgdi:waypoint x="375.0" y="239.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-952ED7E5-9F9B-42E8-9212-DECA33D7A6D2" id="BPMNEdge_sid-952ED7E5-9F9B-42E8-9212-DECA33D7A6D2">
<omgdi:waypoint x="699.0" y="240.0"></omgdi:waypoint>
<omgdi:waypoint x="744.0" y="240.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-F7745084-E689-431F-8F78-0FCAFE376EAE" id="BPMNEdge_sid-F7745084-E689-431F-8F78-0FCAFE376EAE">
<omgdi:waypoint x="451.094167074073" y="390.1644338230481"></omgdi:waypoint>
<omgdi:waypoint x="455.0" y="445.0"></omgdi:waypoint>
<omgdi:waypoint x="515.6531252612749" y="445.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
<bpmndi:BPMNEdge bpmnElement="sid-13DC3678-E57E-4E0B-9F80-55E11BF7304F" id="BPMNEdge_sid-13DC3678-E57E-4E0B-9F80-55E11BF7304F">
<omgdi:waypoint x="90.0" y="240.0"></omgdi:waypoint>
<omgdi:waypoint x="167.0" y="240.0"></omgdi:waypoint>
</bpmndi:BPMNEdge>
</bpmndi:BPMNPlane>
</bpmndi:BPMNDiagram>
</definitions>
```
|
Justus Hiddes Halbertsma, West Frisian form: Joast Hiddes Halbertsma, pron. [jo.ǝst ˈhɪdəs ˈhɔlbǝtsma] (the r is silent); Dutch form: Joost Hiddes Halbertsma, pron. [joːst ˈhɪdəs ˈhalbǝrtsma] (23 October 1789 in Grou – 27 February 1869 in Deventer), was a Frisian writer, poet, minister, lexicographer and linguist. Today, he is primarily known for the poetry and short story collection De Lapekoer fan Gabe Skroar, which he wrote with his brother Eeltsje, publishing the first edition in 1822. Afterwards, this work was continually expanded, and also came to include contributions by a third brother, Tsjalling, until all the Halbertsma Brothers' prose and poetry was posthumously collected in 1871 to become the famous work Rimen en Teltsjes. Although the literary value of this collection was later disputed by some critics, it is undeniable that Rimen en Teltsjes played a role of crucial importance in the development of a new literary tradition after Western Frisian had been used almost exclusively as a spoken language for three centuries.
Life
Youth and background
Justus Hiddes Halbertsma was born on 23 October 1789 in the village of Grou, in the central part of the Dutch province of Friesland, in the house of his parents on Kowemerk ("Cow Market") street. He was the eldest son of the baker and small-time merchant Hidde Joasts Halbertsma (1756–1809) and his wife Ruerdtsje (or Riurtk) Tsjallings Binnerts (1767–1809). He had three younger brothers: Tsjalling (1792–1852), Binnert (1795–1847), and Eeltsje (1797–1858). Two children who were born later died in early childhood, the little boy in 1803, and the little girl in 1805. The four brothers were very close, possibly as a consequence of the fact that both their parents died at a relatively young age in 1809, when Justus was just twenty years of age and the others were even younger. Justus, Tsjalling and Eeltsje, who, as authors, became known as the Brothers Halbertsma later in life, were much like their father, while Binnert more resembled their mother.
Ruerdtsje Binnerts was a scion of a prominent family in Grou. Her people were Mennonites, and although her husband had been raised a Calvinist, he converted after marrying her. From Justus' letters Ruerdtsje emerges as a smart businesswoman, a loving mother, and a deeply religious person. About Hidde Halbertsma, the father, much less is known. It is thought that he might have been a mariner before his marriage. He is described as a gentle soul, who was, however, apt to take offence, and could be quite sharp-tongued in such cases. In 1784, he published a long Dutch-language poem under the title Schrikkelijke IJsgang en Overstroominge in Gelderland ("Terrible Ice-drift and Flooding in Gelderland"). From this intriguing work it is clear that his sons' literary talents were a family trait.
In 1858, Justus Halbertsma explained the difference between his parents' families when he wrote: "One cannot deny the Halbertsmas a greater skillfulness, speed of thought, a greater adroitness, quickness, and talent than the family of Ruerdtsje Tsjallings; but wat help was that to them against the poverty which overcame most of them, while Ruerdtsje Tsjallings' family for the most part maintain prosperity to this day as part of the dignified middle class?" Against the superior power of those who calmly and calculatedly grew rich, he wrote somewhere else, the Halbertsmas could only "avenge themselves through satire."
Education, clerical career, and family
Halbertsma's mother Ruerdtsje was the main force behind the thorough education her sons received. Justus she sent to the French school in the provincial capital of Leeuwarden for a year, and after that, she enrolled him in the Latin school in Leeuwarden from 1801 to 1806. At the advice of his mother and the Mennonite minister Jan Brouwer, from Leeuwarden, Justus Halbertsma chose a clerical career, for which he studied theology at the Mennonite Seminary in Amsterdam from 1807 to 1813. In that period he also immersed himself in the study of the North Germanic languages. In 1814, Halbertsma became minister in Bolsward, where he remained until 1821. Afterwards, he served the Mennonite congregation in the city of Deventer, in the province of Overijssel, from 1822 to his retirement in 1856.
Halbertsma married Johanna Iskjen Hoekema (1794–1847), a Mennonite minister's daughter from Workum, who came from a well-to-do middle-class family, on 10 May 1816. Through his wife Halbertsma attained possession of the farmstead Westerein, near Workum, where he and his family often spent a couple of weeks in summer. Apart from those holidays, he lived out his entire life in Deventer. Halbertsma and his wife had five sons from a good marriage, which was however not without its share of problems. Of these sons Petrus, Hidde, and Tsjalling studied linguistics or medicine, while Watse en Binnert became mariners. As a father, Halbertsma experienced a large amount of grief, as his son Petrus died in 1851 in a psychiatric hospital, while Binnert passed away in 1861, and Hidde took his own life in 1865.
As a literary figure
Halbertsma had been greatly influenced in his student years by his acquaintance with the well-to-do and influential Amsterdam merchant Jeronimo de Vries, who championed a national art, inspired by the Dutch Golden Age. These ideas flew in the face of more modern notions advocated by writers such as Hiëronymus van Alphen and Johannes Klinker. For Halbertsma, who was first and foremost a Frisian, and only secondly a Dutchman, this meant an orientation towards national Frisian ideals, from which emanated two goals he set himself in life. Firstly, he wanted to preserve the Western Frisian language by using it for writing again after it had been used almost exclusively as a spoken language for three centuries. From that it followed that he had to get the Frisians to read in their own language (otherwise, writing it would not be of much use), and that he should record the Frisian vocabulary in a dictionary, which would be the first dictionary ever of the Frisian language. And secondly, Halbertsma wanted to put renowned Frisian people from the past in the spotlight, to serve as an example for his contemporaries.
To attain his first goal, Halbertsma had to use his writings to penetrate the daily life of the Frisian people. This he achieved by authoring folk literature with his brother Eeltsje. The state and social status of the Western Frisian language at that time becomes clear if one observes the fact that Halbertsma and his brothers, though they laboured tirelessly for the use of Frisian as a written language for poetry and prose, lapsed into Dutch for their correspondence with each other, and apparently did not feel that was in any way strange. As Halbertsma remained the editor of Eeltsje's work for his entire life, their poetry and short fiction were strongly connected and published together from the very beginning. For that reason the linguist Foeke Buitenrust Hettema would describe Halbertsma later as the 'literary agent' of his brothers Eeltsje and Tsjalling. In 1822, their early works were collected under the title De Lapekoer fan Gabe Skroar ("Gabe Tailor's Rag Basket"; original, archaic spelling: De Lape Koer fen Gabe Skroor), a booklet consisting of 36 pages, and including six poems and one short story. This publication was attributed to the fictional 'Gabe Skroar', a lame farmer's son who became a tailor and a writer, but died young. This character was in all probability a creation of Eeltsje's. Hiding behind such a fictional author was fairly normal at that time.
The first edition of De Lapekoer fan Gabe Skroar, published in 1822, consisted of 200 copies only. Those were not sold, however, but at the expense of Halbertsma presented to acquaintances of his all over Friesland. In that way he awoke among his public a desire for more reading-matter of this nature. These stories and poems were in fact something entirely different from what the readers were used to, not only as to what language they were witten in, but also and especially concerning their informal style. The copies of the booklet often were shared around through entire networks of family and friends, and not seldom the stories and poems were transcribed by hand before the booklet was given back or passed on.
A second, expanded edition of De Lapekoer fan Gabe Skroar, consisting of 237 pages, was published in 1829, followed by a third, further expanded edition in 1834, running to almost 500 pages of a larger size. After that, additions were published separately in 1836 (De Noarger Rún oan Gabe Skroar), 1840 (Twigen út in Alde Stamme), 1854 (Leed en Wille en de Flotgerzen), and 1858 (De Jonkerboer and Teltsjes fan de Wize Mannen fan Esonstêd). By the end of his life, Halbertsma assisted in the editing of a portion of Rimen en Teltsjes (Rhymes and Tales), the comprehensive publication in which all the Halbertsma Brothers' literary works were collected. He even wrote some new pieces for it, such as the short story It Grouwe Pak ("The Thick Suit").
As a writer, Halbertsma strongly sensed what the public needed, but at the same time he was a somewhat distant and cerebral author, which tended to have an averse effect on his poetry, especially compared to the poems of his brother Eeltsje, who was a much more emotional writer. Nevertheless, at least some of Halbertsma's poems, for instance Sibbel fan De Ryp ("Sibbel from De Ryp") and Nacht ("Night") have made a lasting impression. That said, Halbertsma's short stories were much more acute thanks to his limber use of language, his lively style of writing, and the fact that he refused to mince his words, leading to a rough, scoffing, and sometimes darkly grim atmosphere. His subject-matter Halbertsma collected from various sources, including German and French oral literature, which he matchlessly revised and placed in Friesland. He was a non-conformist, whose sharp judgments of the works of others made him far from universally loved. Before everything else, Halbertsma as a writer was an essayist and a satirist, who liked to include topical political and social questions in his literary works. From the anecdotes of people who knew or met him, he emerges as a sharp-witted and extremely diligent man of capacious erudition who was a droll story-teller.
It Boalserter Nut ("The Bolsward Utility" – the name of a gentlemen's club), which was published in De Lapekoer, was Halbertsma's first short story of a more extended length. As is often the case with Halbertsma's stories, it consists of a frame-story in which several other more or less stand-alone components are absorbed, such as the essay It Jild ("The Money"). Miswier ("Continuously Bad Weather") and De Twadde Jûn ("The Second Evening") are also extensive frame-stories, and De Skearwinkel fan Joutebaas ("Boss Joute's Barbershop") is a dialogue, which includes the intriguing and completely stand-alone story Heksershol ("Hole of Witchcraft" – an old nickname for the village of Molkwerum). The only time Halbertsma wrote a literary work entirely without the participation of Eeltsje, was in 1837, when he published Oan Eölus ("To Aeolus"), a story inspired by a great storm in 1830, in which a tornado has blown thoughts and motifs together like whirling leaves. Oan Eölus could perhaps be seen as a clue that Halbertsma harboured greater literary aspirations than would fit in the De Lapekoer fan Gabe Skroar.
As a man of science
And yet, for all that, Halbertsma's literary works did not take centerstage in his life: that place was reserved for his scientific non-fiction books. More than fifty years he laboured to complete his dictionary of Western Frisian titled Lexicon Frisicum, for which he chose Latin as the descriptive language, but it remained unfinished. He organised it along the lines of the German dictionary by the Brothers Grimm, but became enmeshed in the addition of insertions and in reworkings, and in writing long semantic etymologies, a part of the work for which he especially had a predilection. In the end, he never got past the letter F, and the dictionary remained a manuscript, which was published in its unfinished form after Halbertsma's death, in 1872. In the 20th century, the material Halbertsma gathered for his Lexicon Frisicum was used by de Frisian Academy to develop several Western Frisian dictionaries, notably the 25 part Wurdboek fan de Fryske Taal ("Dictionary of the Western Frisian Language"; published 1984–2011).
Furthermore, Halbertsma was the driving force behind the commemoration of the life of the famous 17th century Frisian poet Gysbert Japiks in 1823, which is seen by some as the starting point of the Frisian Movement. Afterwards he published an extensive literary history work under the title Hulde aan Gysbert Japiks ("Hail to Gysbert Japiks") in two parts in 1824 and 1827. In 1829 Halbertsma wrote Het Geslacht der Van Harens ("The Lineage of the Van Harens"), in which he primarily described the lives and literary works of the brothers and 18th century Dutch-language writers from Friesland Onno Zwier van Haren and Willem van Haren, whose work he admired greatly.
Halbertsma was also the author of an 1836 biography of the Frisian poet Jan Cornelis Pieters Salverda (1783–1836), which almost has a naturalistic feel to it. In 1843 he published a historical work about the Mennonites, called De Doopsgezinden en Hunne Herkomst ("The Mennonites and Their Origins"), and in 1851 he followed this up with Aanteekeningen op het Vierde Deel van den Spiegel Historiael van Jacob van Maerlant ("Notes on Part Four of the Spiegel Historiael by Jacob van Maerlant). Furthermore, in 1858, at the request of Prince Louis Lucien Bonaparte, whom he had met in London, Halbertsma published It Ewangeelje fen Matthewes, which was the first Frisian translation of the Biblical Gospel of Matthew (or of any gospel for that matter). This translation evoked a great deal of criticism from Halbertsma's contemporaries, who thought it was much too realistic; today, it still has an almost modern feel to it, showing that Halbertsma was far ahead of his time.
Halbertsma had a reputation of being an erudite man, which he won at a young age. He was especially known as a scholar of the Germanic languages, and to the present day he is widely acknowledged to have been a larger-than-life figure in the field of linguistics. Shortly after returning to Friesland from his study in Amsterdam, friends from his student years who by then were occupying important posts within the civil service, introduced him to the social circles around the Frisian-minded provincial governor jonker Idsert Aebinga van Humalda. Later in life Halbertsma was acquainted with prominent Dutch literary figures, such as Willem Bilderdijk, and what's more, he maintained a lively correspondence with European greats, like the Danish linguist Rasmus Rask, the English poet Robert Southey, and the German linguist and fairytale-collector Jacob Grimm. Moreover, he was invited to join learned historical and linguistic societies in Leiden, Copenhagen, Berlin, Athens en Halle-Wittenberg.
Hence Halbertsma expected to be appointed professor of Dutch language and literature, and several times such an appointment seemed almost within his grasp, but each time it slipped away. His sharp pen probably played a role in this, but there were other reasons, too. In 1830, he lost out from the surplus of applicants washing over the Netherlands from the universities of Belgium, which at that time had just seceded from the United Kingdom of the Netherlands, and after 1840 a different type of scientist was called for, with a more specialised knowledge than Halbertsma commanded. That in 1837 he was awarded an honorary doctorate in literature by the University of Leiden, was scant consolation to him. In part because of these disappointments, Halbertsma balanced on the edge of a burn-out for a long time, and especially in 1839, when he was passed over yet again, he went through a deep crisis. It was only in the 1860s, when he was in his seventies, that he seems to have achieved a state of mental composure.
In part because of his linguistic erudition, Halbertsma has been named as the author of the infamous Oera Linda Book, a falsified work which emerged in 1867 and was written in imitated Old Frisian. It pretended to be an early history of Friesland and the Frisian people, and for a while it was considered authentic, until it was exposed as a hoax. However, it is considered much more likely that the true author of the Oera Linda Book was the librarian Eelco Verwijs, who lead the Provincial Library of Friesland and who had befriended Halbertsma, or possibly the writer François Haverschmidt or the ship carpenter and self-taught freethinker Cornelis over de Linden. If Verwijs was the culprit, it is not thought to be impossible that Halbertsma tacitly or even openly provided him with advice.
Death
After his retirement, on 26 October 1856, which Halbertsma applied for himself because his heart was not in his duties as a minister anymore, he withdrew into his upstairs apartment, where from that point on he became increasingly more lonely as his family members and friends died away, although his foreign contacts through correspondence remained intact. Justus Hiddes Halbertsma died in Deventer on 27 February 1869, when he was 79 years of age. By that time he had been a widower for 22 years. He was survived by his sons Watse and Tsjalling. In 1947, the Dr. Joast Halbertsma Award, the highest provincial decoration in the field of historical (and later more general scientific) research concerning the province of Friesland was named in his honour.
Legacy
After the death of all three Brothers Halbertsma, their short fiction and poetry was gathered under the supervision of librarian and archivist Gerben Colmjon and bookseller and historian Wopke Eekhoff. In 1871, it was published by the firm of J. de Lange in Deventer, under the title of Rimen en Teltsjes ("Rhymes and Tales"). This work is now thought of as the national book of Western Frisian literature, and although the literary value of this collection was later disputed by some critics, it is undeniable that Rimen en Teltsjes and its predecessor De Lapekoer fan Gabe Skroar played a role of crucial importance in the development of a new literary tradition after Frisian had been used almost exclusively as a spoken language for three centuries. Justus Hiddes Halbertsma's influence on subsequent Western Frisian literature was enormous and of incalculable value. In 1949, author and literary critic Anne Wadman wrote that the Halbertsmas gave "the Western Frisian people a literary monument, in which it saw its own life as a nation [...] reflected." As of today Rimen en Teltsjes still occupies the first slot in the ranking of Western Frisian literary classics.
On his travels to England, Scotland, and Ireland, in 1852 and 1853, Halbertsma was impressed by the number of museums and cabinets of antiquities there were. This provided him with the idea to transfer his own collection of antiquities (originating for a major part in Hindeloopen, while the remainder came mostly from excavations of Frisian terps) to the Province of Friesland. In 1853, the Cabinet of Antiquities of Friesland was formed out of his collection, which in 1881 became the Frisian Museum.
Halbertsma also gave to the Province of Friesland, in part when he was still alive, his extensive library of scientific books, which was added to the Provincial Library of Friesland, en is to this day available for scientific research purposes at Tresoar. A large collection of letters written by and to Halbertsma, including hundreds to his brothers Eeltsje en Tsjalling, long remained in the custody of the Halbertsma family, until they were given, in the course of the 20th century, on permanent loan to the Frisian Literary Museum and Documentation Centre (FLMD), which became part of Tresoar in 2002.
Bibliography
Literary works
(for the most part in Western Frisian)
1822 – De Lapekoer fan Gabe Skroar ("Gabe Tailor's Rag Basket"; collection of poetry and short stories, with Eeltsje Hiddes Halbertsma)
1829 – De Lapekoer fan Gabe Skroar (collection of poetry and short stories, with Eeltsje Hiddes Halbertsma; expanded version)
1830 – In Nij Liet op Grouster Merke ("New Song about the Fair in Grou"; poem)
1831 – In Skotsk Miel ("A Scottish Meal"; short story)
1834 – Foeke Sjoerds (short story about the historical 18th century Frisian writer Foeke Sjoerds)
1834 – Grêfskrift fan in Wynsûper op Him Selme ("Epitaph of a Wine-guzzler for Himself"; poem)
1834 – De Lapekoer fan Gabe Skroar (collection of poetry and short stories, with Eeltsje Hiddes Halbertsma; further expanded version)
1834 – Sûn Ferstân ("Sane of Mind"; short story)
1835 – De Skearwinkel fan Joutebaas ("Boss Joute's Barbershop"; frame-story with several poems and short stories, with Eeltsje Hiddes Halbertsma en Tsjalling Hiddes Halbertsma)
1836 – De Noarger Rún oan Gabe Skroar ("The Gelding from Norg to Gabe Tailor"; short story with a poem by Eeltsje Hiddes Halbertsma, including commentary by Justus Hiddes Halbertsma)
1836 – De Treemter fan it Sint-Anthonygasthûs ("The Conversation Room of St. Anthony's Hospital"; short story by Eeltsje Hiddes Halbertsma, including commentary by Justus Hiddes Halbertsma)
1837 – Oan Eölus ("To Aeolus"; short story)
1837 – Eölus, Grewa fan Stoarm en Onwaar ("Aeolus, Lord of Storm and Thunder"; collection of poetry and short stories, with Eeltsje Hiddes Halbertsma)
1840 – Twigen út in Alde Stamme ("Twigs from an Old Trunk"; collection of poetry and short stories, with Eeltsje Hiddes Halbertsma)
1841 – Oan Petrus, Doe't er Doktor Waard ("To Petrus, When He Took His Doctor's Degree"; poems by Eeltsje Hiddes Halbertsma and Ferdinand Hanthal, with commentary by Justus Hiddes Halbertsma)
1854 – Leed en Wille en de Flotgerzen ("Grief and Pleasure and Down Upon the Lee Shore"; collection of poetry and short stories, with Eeltsje Hiddes Halbertsma)
1871 – It Grouwe Pak ("The Thick Suit"; short story)
1871 – Rimen en Teltsjes ("Rhymes and Tales"; collection of poetry and short stories, with Eeltsje Hiddes Halbertsma en Tsjalling Hiddes Halbertsma)
18?? – De Geboarteleppel ("The Birth Spoon"; short story)
18?? – De Fûgel fan Bearn-poep ("German Bearn's Bird"; short story)
18?? – De Bear op Reis ("The Bear on a Journey"; short story)
Scientific works
(for the most part in Dutch)
1824 – Hulde aan Gysbert Japiks ("Hail to Gysbert Japiks"; part 1; about life and works of the 17th century Frisian poet Gysbert Japiks)
1827 – Hulde aan Gysbert Japiks (part 2)
1829 – Het Geslacht der Van Harens ("The Lineage of the Van Harens"; primarily about the life and works of the 18th century Frisian (Dutch-language) poets and writers Onno Zwier van Haren and Willem van Haren)
1834 – Friesche Spelling en Proeve van Vergelijking tusschen de Boven Ontworpene Spelling en Die van Gysbert Japicx ("Frisian Spelling and Comparison of the Newly Developed Spelling and That of Gysbert Japicx"; about the spelling of the Western Frisian language)
1836 – Friesic: Ancient and Modern Friesic Compared with Anglo-Sacon
1836 – Jan Cornelis Pieters Salverda (biography of the 18th century Friesian poet Jan Cornelis Pieters Salverda)
1836 – De Toenemende Wansmaak van de Nederlanders voor het Toneel ("The Increasingly Bad Taste of the Dutch in Their Theatre"; about the quality of Dutch theatre drama)
1840 – Letterkundige Naoogst ("Literary After-Harvest"; part 1)
1843 – Het Buddhisme en Zijn Stichter ("Buddhism and Its Founder"; about buddhism)
1843 – De Doopsgezinden en Hunne Herkomst ("The Mennonites and Their Origins"; about the mennonites)
1845 – Letterkundige Naoogst (part 2)
1846 – De Tongvallen in Nederland ("Dialects in the Netherlands"; about dialects spoken in the Netherlands)
1851 – Aanteekeningen op het Vierde Deel van den Spiegel Historiael van Jacob van Maerlant ("Notes on the Part Four of the Spiegel Historiael by Jacob van Maerlant"; about the Spiegel Historiael, by Jacob van Maerlant)
1858 – It Ewangeelje fen Matthewes ("The Gospel of Matthew"; translation of the Gospel of Matthew to Western Frisian)
1861 – Levensberigten der in Dit Jaar Afgestorvene Medeleden: Bericht wegens Rinse Posthumus, in Leven Kerkleeraar bij de Hervormde Gemeenten van Waaxens en Brantgum in Friesland ("Biographical Notices of the Members Who Died This Year: Notice for Rinse Posthumus, in Life Minister of the Reformed Congregations of Waaxens and Brantgum in Friesland", in: Jaarboek van de Maatschappij der Nederlandse Letterkunde, "Yearbook of the Dutch Literary Society"; obituary of the poet Rinse Posthumus)
1869 – Biografie van Deventer ("Biography of Deventer"; sarcastic description of the city of Deventer and its inhabitants)
1872 – Lexicon Frisicum (unfinished dictionary of the Western Frisian language)
1969 – Kent Gij Halbertsma van Deventer? In Tal Net Utjowne Hânskriften ("Knowest Thou Halbertsma from Deventer? A Number of Unpublished Manuscripts"; collection of unpublished works)
References
Sources
, Twataligens: Ynlieding yn Underskate Aspekten fan de Twataligens, Leeuwarden (Afûk), 1981, .
, Oer Skriuwers, Boek en Utjeften, in: , Rimen en Teltsjes, Drachten (A.J. Osinga Utjouwerij), 1993, , pp. 587–613.
, Halbertsma, Justus (Joost Hiddes), in: , Nieuwe Encyclopedie van Fryslân, Gorredijk/Leeuwarden (Utjouwerij Bornmeer/Tresoar), 2016, , pp. 1126–1127.
, Lyts Hânboek fan de Fryske Literatuer, Leeuwarden (Afûk), 1997, .
, Rimen en Teltsjes, Bolsward/Leeuwarden (A.J. Osinga Utjouwerij/R. van der Velde), 1958, no ISBN.
, Rimen en Teltsjes, Drachten (A.J. Osinga Utjouwerij), 1993, .
, Fryslân: Fêstens en Feroaring, in: , De Fryslannen, Leeuwarden (Frisian Council/Afûk), 2008, .
, Cultuur in Friesland en Friese Cultuur, 1795–1917, in: , Geschiedenis van Friesland 1750–1995, Amsterdam/Leeuwarden (Uitgeverij Boom/Fryske Akademy), 1998, , pp. 172–212.
, Spiegel van de Friese Poëzie: Van de Zeventiende Eeuw tot Heden, Amsterdam (J.M. Meulenhoff B.V.), 1994, .
, Joost Hiddes Halbertsma (1789–1869), in: Overijselse Biografieën.
, Nieuwe Encyclopedie van Fryslân, Gorredijk/Leeuwarden (Utjouwerij Bornmeer/Tresoar), 2016, .
, Tweeduizend Jaar Geschiedenis van Friesland, Leeuwarden (Uitgeverij M.Th. van Seyen), no year, no ISBN.
, Fan Fryslâns Forline, Bolsward (A.J. Osinga N.V.), 1968, no ISBN.
, Frieslands Dichters, Leiden (L. Stafleu), 1949, no ISBN.
, Foarwurd and De Bruorren Halbertsma, in: , Rimen en Teltsjes, Bolsward (A.J. Osinga N.V.), 1958, pp. 5–14.
Further reading
, J.H. Halbertsma, Gysbert Japicx en de Fryske Dialekten, Leeuwarden (Fryske Akademy), 1973, no ISBN.
, De Collectie van J.H. Halbertsma in het Fries Museum, in: De Vrije Fries, 1969, pp. 17–32.
, J.H. Halbertsma als Lexicograaf, Leeuwarden (Afûk), 2011, .
, Knooppunt Halbertsma: Joast Hiddes Halbertsma (1789–1869) en andere Europese geleerden over het Fries en andere talen, over wetenschap en over de samenleving, Hilversum (Uitgeverij Verloren), 2009, .
, Joost Halbertsma 1789-1869: Triomfen en Tragedies van een Uitmiddelpuntig Man: Een biografie, Grou (Uitgeverij Louise), 2018, .
, Joast Hiddes Halbertsma, Brekker en Bouwer, Drachten (Drukkerij J. Laverman N.V.), 1969, no ISBN.
1789 births
1869 deaths
19th-century Dutch male writers
Dutch Protestant ministers and clergy
Linguists from the Netherlands
Dutch short story writers
19th-century Dutch poets
Dutch non-fiction writers
Dutch essayists
Dutch translators
West Frisian-language writers
Translators to West Frisian
Translators of the Bible into Frisian
People from Boarnsterhim
People from Bolsward
People from Deventer
Dutch male poets
19th-century short story writers
19th-century essayists
19th-century translators
Male non-fiction writers
|
```python
def test_get_store(client, created_store_id):
response = client.get(
f"/store/{created_store_id}",
)
assert response.status_code == 200
assert response.json == {
"id": 1,
"name": "Test Store",
"items": [],
"tags": [],
}
def test_get_store_not_found(client):
response = client.get(
"/store/1",
)
assert response.status_code == 404
assert response.json == {"code": 404, "status": "Not Found"}
def test_get_store_with_item(client, created_store_id):
client.post(
"/item",
json={"name": "Test Item", "price": 10.5, "store_id": created_store_id},
)
response = client.get(
f"/store/{created_store_id}",
)
assert response.status_code == 200
assert response.json["items"] == [
{
"id": 1,
"name": "Test Item",
"price": 10.5,
}
]
def test_get_store_with_tag(client, created_store_id):
client.post(
f"/store/{created_store_id}/tag",
json={"name": "Test Tag"},
)
response = client.get(
f"/store/{created_store_id}",
)
assert response.status_code == 200
assert response.json["tags"] == [{"id": 1, "name": "Test Tag"}]
def test_create_store(client):
response = client.post(
"/store",
json={"name": "Test Store"},
)
assert response.status_code == 201
assert response.json["name"] == "Test Store"
def test_create_store_with_items(client, created_store_id):
client.post(
"/item",
json={"name": "Test Item", "price": 10.5, "store_id": 1},
)
# Get the store with id 1 and check the items contains the newly created item
response = client.get(
f"/store/{created_store_id}",
)
assert response.status_code == 200
assert response.json["items"] == [
{
"id": 1,
"name": "Test Item",
"price": 10.5,
}
]
def test_delete_store(client, created_store_id):
response = client.delete(
f"/store/{created_store_id}",
)
assert response.status_code == 200
assert response.json == {"message": "Store deleted"}
def test_delete_store_doesnt_exist(client):
response = client.delete(
"/store/1",
)
assert response.status_code == 404
assert response.json == {"code": 404, "status": "Not Found"}
def test_get_store_list_empty(client):
response = client.get(
"/store",
)
assert response.status_code == 200
assert response.json == []
def test_get_store_list_single(client):
client.post(
"/store",
json={"name": "Test Store"},
)
response = client.get(
"/store",
)
assert response.status_code == 200
assert response.json == [{"id": 1, "name": "Test Store", "items": [], "tags": []}]
def test_get_store_list_multiple(client):
client.post(
"/store",
json={"name": "Test Store"},
)
client.post(
"/store",
json={"name": "Test Store 2"},
)
response = client.get(
"/store",
)
assert response.status_code == 200
assert response.json == [
{"id": 1, "name": "Test Store", "items": [], "tags": []},
{"id": 2, "name": "Test Store 2", "items": [], "tags": []},
]
def test_get_store_list_with_items(client):
client.post(
"/store",
json={"name": "Test Store"},
)
client.post(
"/item",
json={"name": "Test Item", "price": 10.5, "store_id": 1},
)
response = client.get(
"/store",
)
assert response.status_code == 200
assert response.json == [
{
"id": 1,
"name": "Test Store",
"items": [
{
"id": 1,
"name": "Test Item",
"price": 10.5,
}
],
"tags": [],
}
]
def test_get_store_list_with_tags(client):
resp = client.post(
"/store",
json={"name": "Test Store"},
)
client.post(
f"/store/{resp.json['id']}/tag",
json={"name": "Test Tag"},
)
response = client.get(
"/store",
)
assert response.status_code == 200
assert response.json == [
{
"id": 1,
"name": "Test Store",
"items": [],
"tags": [{"id": 1, "name": "Test Tag"}],
}
]
def test_create_store_duplicate_name(client):
client.post(
"/store",
json={"name": "Test Store"},
)
response = client.post(
"/store",
json={"name": "Test Store"},
)
assert response.status_code == 400
assert response.json["message"] == "A store with that name already exists."
```
|
```sqlpl
select 1 as id
union all
select * from {{ ref('node_0') }}
union all
select * from {{ ref('node_3') }}
union all
select * from {{ ref('node_6') }}
union all
select * from {{ ref('node_8') }}
union all
select * from {{ ref('node_17') }}
union all
select * from {{ ref('node_25') }}
union all
select * from {{ ref('node_47') }}
union all
select * from {{ ref('node_174') }}
```
|
Chen Tsu-li (), also known as James T. L. Chen, (born 28 May 1933) is a Taiwanese businessman, former basketball player and coach.
National team playing career
Chen competed as a part of the Republic of China's senior national team at the 1956 Summer Olympics. He won silver medals at the 1958 Asian Games, and the 1960 FIBA Asian Championship. He also played at the 1959 FIBA World Championship, which he led in scoring average, at 20.1 points per game.
Coaching career
After he ended his basketball playing career, Chen became a basketball coach. He was a head coach in both club teams in Taiwan, and of the senior Taiwanese national team, which he coached at the 1987 FIBA Asian Championship, which was held in Bangkok.
References
External links
1932 births
Living people
Asian Games medalists in basketball
Asian Games silver medalists for Chinese Taipei
Basketball players at the 1956 Summer Olympics
Basketball players at the 1958 Asian Games
Medalists at the 1958 Asian Games
Olympic basketball players for Taiwan
Taiwanese basketball coaches
Taiwanese men's basketball players
People from Huzhou
Basketball players from Zhejiang
Taiwanese people from Zhejiang
1959 FIBA World Championship players
Republic of China men's national basketball team players
Chinese Taipei men's national basketball team coaches
|
Ryan Albert Collins (born November 1, 1975) is a former American football tight end for the Baltimore Ravens of the National Football League, the Amsterdam Admirals of NFL Europe, and the New York/New Jersey Hitmen of the XFL. He played college football at St. Thomas.
Early life
Collins was born November 1, 1975, in Minneapolis, to Albert and Barbara Collins. He attended Robbinsdale Cooper High School, where he played football, baseball, basketball, and track.
College career
Collins attended the University of St. Thomas in St. Paul, Minnesota, where he majored in communications with a minor in business. He played for the St. Thomas Tommies football and basketball teams.
In football, Collins was selected to the All-MIAC team, the AFCA Division III All-America team, and was among 36 seniors to play in the 1997 Aztec Bowl.
Professional football career
Minnesota Vikings
Collins signed with the Minnesota Vikings as a free agent on May 4, 1998. The team waived him on August 30, 1998, but signed him to the practice squad a day later, where he remained for the rest of the 1998 season. His contract expired on January 25, 1999.
Cleveland Browns
The Cleveland Browns signed Collins as a free agent on February 22, 1999. Late in the 1999 preseason, Collins suffered a shoulder injury that left him on the inactive list for the first 3 weeks of the season. He was cut by the Browns on September 27, 1999.
Baltimore Ravens
On October 12, 1999, Collins was signed to the Baltimore Ravens' practice squad. On December 3, amid a series of injuries to the Ravens' tight ends, he was activated off the practice squad to play in four games, starting three.
On December 26, 1999, Collins made three receptions for 41 yards in a 22–0 win over the Cincinnati Bengals. This was the Ravens' first shutout win in franchise history. However, during the third quarter of the game, Collins fractured his left ankle and was removed from the game. He was placed on injured reserve following the game.
He was released by the Ravens on August 28, 2000.
New York/New Jersey Hitmen
On October 30, 2000, Collins was selected with the 282nd pick of the 2001 XFL Draft by the New York/New Jersey Hitmen. He played in all 10 games with the Hitmen and recorded three receptions for 38 yards.
Europe
Amsterdam Admirals
Collins played with the Amsterdam Admirals of NFL Europe in 2002 and 2003. He started all 10 games of the 2002 season, making 34 catches for a total of 280 yards and two touchdowns. He was named to the 2002 all-NFL Europe team. In the 2003 season, he played in eight games and started six. He was again named to the 2003 all-NFL Europe team, becoming the only player to be named to the team in both 2002 and 2003.
Kansas City Chiefs
Shortly after the end of the 2002 NFL Europe season, on June 27, 2002, Collins was signed by the Kansas City Chiefs. At the time, the Chiefs were stuck in a contract stalemate with tight end Tony Gonzalez. The Chiefs released Collins on September 2, 2002.
References
Living people
1975 births
American football tight ends
Amsterdam Admirals players
Baltimore Ravens players
Cleveland Browns players
Kansas City Chiefs players
Minnesota Vikings players
New York/New Jersey Hitmen players
Players of American football from Minnesota
St. Thomas (Minnesota) Tommies football players
American expatriate players of American football
American expatriate sportspeople in the Netherlands
|
The is a wing of the Japan Air Self-Defense Force. Responsible for airborne search and rescue, it is headquartered at Iruma Air Base in Saitama Prefecture and controls squadrons and detachments across Japan. The helicopter airlift squadrons were previously under the authority of Air Support Command. It is equipped with CH-47J, UH-60J and U-125A aircraft.
History
From 2009 the Air Rescue Wing began training in air-to-air refueling with US forces, and from 2010 two of the JASDF's 401st Tactical Airlift Squadron C-130H aircraft were given air-to-air refueling equipment, allowing them to refuel UH-60J helicopters. In June 2017 UH-60JA aircraft of the wing practiced night air-to-air refueling with US forces.
On May 28, 2013 the Air Rescue Wing Naha Detachment rescued the pilot of an F-15C of the US Air Force's 44th Fighter Squadron that had crashed into the ocean off Okinawa.
On October 17, 2017 a JASDF UH-60J of the Hamamatsu Detachment of the Air Rescue Wing crashed into the sea off Shizuoka Prefecture while conducting night rescue drills. Wreckage was found but the four crew members were not located.
On June 11, 2018 the Naha Detachment rescued the pilot of another US F-15C of the 44th Fighter Squadron that had ejected over the sea off Okinawa.
Units
Air Rescue Squadron (UH-60J, U-125A)
Air Rescue Wing Chitose Detachment (Chitose Air Base)
Air Rescue Wing Matsushima Detachment (Matsushima Air Base)
Air Rescue Wing Ashiya Detachment (Ashiya Air Base)
Air Rescue Wing Akita Detachment (Akita Airport)
Air Rescue Wing Hyakuri Detachment (Hyakuri Air Base)
Air Rescue Wing Nyutabaru Detachment (Nyutabaru Air Base)
Air Rescue Wing Niigata Detachment (Niigata Airport)
Air Rescue Wing Hamamatsu Detachment (Hamamatsu Air Base)
Air Rescue Wing Naha Detachment (Naha Air Base)
Air Rescue Wing Komatsu Detachment (Komatsu Air Base)
Air Rescue Wing Komaki Detachment (Training squadron) (Komaki Air Base)
Squadrons (CH-47J)
Iruma Helicopter Airlift Squadron (Iruma Air Base)
Kasuga Helicopter Airlift Squadron (Kasuga Air Base)
Misawa Helicopter Airlift Squadron (Misawa Air Base)
Naha Helicopter Airlift Squadron (Naha Air Base)
Aircraft operated
CH-47J
UH-60J
U-125A
References
Units of the Japan Air Self-Defense Force
|
Soldier of Arete is a 1989 fantasy novel by American writer Gene Wolfe, published by Tor Books. The novel is a sequel to Soldier of the Mist.
Soldier of the Mist and Soldier of the Arete have been collected as Latro in the Mist.
Synopsis
Wolfe's forgetful protagonist Latro keeps on traveling throughout ancient Greece.
References
External links
Google books profile
Novels by Gene Wolfe
Fiction with unreliable narrators
1989 American novels
American fantasy novels
1989 fantasy novels
Fictional diaries
Tor Books books
|
```ruby
class Ipython < Formula
include Language::Python::Virtualenv
desc "Interactive computing in Python"
homepage "path_to_url"
url "path_to_url"
sha256 your_sha256_hash
license "BSD-3-Clause"
head "path_to_url", branch: "master"
bottle do
sha256 cellar: :any_skip_relocation, arm64_sonoma: your_sha256_hash
sha256 cellar: :any_skip_relocation, arm64_ventura: your_sha256_hash
sha256 cellar: :any_skip_relocation, arm64_monterey: your_sha256_hash
sha256 cellar: :any_skip_relocation, sonoma: your_sha256_hash
sha256 cellar: :any_skip_relocation, ventura: your_sha256_hash
sha256 cellar: :any_skip_relocation, monterey: your_sha256_hash
sha256 cellar: :any_skip_relocation, x86_64_linux: your_sha256_hash
end
depends_on "python@3.12"
resource "asttokens" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "decorator" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "executing" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "jedi" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "matplotlib-inline" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "parso" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "pexpect" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "prompt-toolkit" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "ptyprocess" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "pure-eval" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "pygments" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "six" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "stack-data" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "traitlets" do
url "path_to_url"
sha256 your_sha256_hash
end
resource "wcwidth" do
url "path_to_url"
sha256 your_sha256_hash
end
def install
virtualenv_install_with_resources
man1.install libexec/"share/man/man1/ipython.1"
end
test do
assert_equal "4", shell_output("#{bin}/ipython -c 'print(2+2)'").chomp
end
end
```
|
```python
# coding: utf-8
import logging
import os
import sys
import time
import traceback
import numpy as np
import pytest
import ray
import ray.cluster_utils
import ray.exceptions
import ray.experimental.channel as ray_channel
from ray.exceptions import RayChannelError, RayChannelTimeoutError
from ray.util.scheduling_strategies import NodeAffinitySchedulingStrategy
from ray.dag.compiled_dag_node import CompiledDAG
from ray._private.test_utils import get_actor_node_id
logger = logging.getLogger(__name__)
def create_driver_actor():
return CompiledDAG.DAGDriverProxyActor.options(
scheduling_strategy=NodeAffinitySchedulingStrategy(
ray.get_runtime_context().get_node_id(), soft=False
)
).remote()
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_put_local_get(ray_start_regular):
driver_actor = create_driver_actor()
chan = ray_channel.Channel(
None,
[
(driver_actor, get_actor_node_id(driver_actor)),
],
1000,
)
num_writes = 1000
for i in range(num_writes):
val = i.to_bytes(8, "little")
chan.write(val)
assert chan.read() == val
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_read_timeout(ray_start_regular):
driver_actor = create_driver_actor()
chan = ray_channel.Channel(
None,
[
(driver_actor, get_actor_node_id(driver_actor)),
],
1000,
)
with pytest.raises(RayChannelTimeoutError):
chan.read(timeout=1)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_write_timeout(ray_start_regular):
driver_actor = create_driver_actor()
chan = ray_channel.Channel(
None,
[
(driver_actor, get_actor_node_id(driver_actor)),
],
1000,
)
val = 1
bytes = val.to_bytes(8, "little")
chan.write(bytes, timeout=1)
with pytest.raises(RayChannelTimeoutError):
chan.write(bytes, timeout=1)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
@pytest.mark.parametrize("remote", [True, False])
def test_driver_as_reader(ray_start_cluster, remote):
cluster = ray_start_cluster
if remote:
# This node is for the driver. num_cpus is 1 because the
# CompiledDAG.DAGDriverProxyActor needs a place to run.
cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
# This node is for the writer actor.
cluster.add_node(num_cpus=1)
else:
# This node is for both the driver (including the
# CompiledDAG.DAGDriverProxyActor) and the writer actor.
cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def setup(self, driver_actor):
self._channel = ray_channel.Channel(
ray.get_runtime_context().current_actor,
[(driver_actor, get_actor_node_id(driver_actor))],
1000,
)
def get_channel(self):
return self._channel
def write(self):
self._channel.write(b"x")
a = Actor.remote()
ray.get(a.setup.remote(create_driver_actor()))
chan = ray.get(a.get_channel.remote())
ray.get(a.write.remote())
assert chan.read() == b"x"
@pytest.mark.parametrize("remote", [True, False])
def test_driver_as_reader_with_resize(ray_start_cluster, remote):
cluster = ray_start_cluster
if remote:
# This node is for the driver. num_cpus is 1 because the
# CompiledDAG.DAGDriverProxyActor needs a place to run.
cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
# This node is for the writer actor.
cluster.add_node(num_cpus=1)
else:
# This node is for both the driver (including the
# CompiledDAG.DAGDriverProxyActor) and the writer actor.
cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def setup(self, driver_actor):
self._channel = ray_channel.Channel(
ray.get_runtime_context().current_actor,
[(driver_actor, get_actor_node_id(driver_actor))],
1000,
)
def get_channel(self):
return self._channel
def write(self):
self._channel.write(b"x")
def write_large(self):
self._channel.write(b"x" * 2000)
a = Actor.remote()
ray.get(a.setup.remote(create_driver_actor()))
chan = ray.get(a.get_channel.remote())
ray.get(a.write.remote())
assert chan.read() == b"x"
ray.get(a.write_large.remote())
assert chan.read() == b"x" * 2000
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_set_error_before_read(ray_start_regular):
"""
Tests that if a channel is closed after a reader, a subsequent read does not block
forever.
"""
@ray.remote
class Actor:
def __init__(self):
self.arr = None
def create_channel(self, writer, reader_and_node_list):
self._channel = ray_channel.Channel(writer, reader_and_node_list, 1000)
return self._channel
def pass_channel(self, channel):
self._channel = channel
def close(self):
self._channel.close()
def write(self, arr):
self._channel.write(arr)
def read(self):
self.arr = self._channel.read()
# Keep self.arr in scope. While self.arr is in scope, its backing
# shared_ptr<MutableObjectBuffer> in C++ will also stay in scope.
# Under normal execution, this will block the next read() from
# returning, since we are still using the shared buffer.
# In this test we are checking that if the channel is closed, then
# the next read() will return an error immediately instead of
# blocking, even though we still have self.arr in scope.
return self.arr
for _ in range(10):
a = Actor.remote()
b = Actor.remote()
node_b = get_actor_node_id(b)
chan = ray.get(a.create_channel.remote(a, [(b, node_b)]))
ray.get(b.pass_channel.remote(chan))
# Use numpy to enable zero-copy deserialization.
arr = np.random.rand(100)
ray.get(a.write.remote(arr))
assert (arr == ray.get(b.read.remote())).all()
# Check that the thread does not block on the second call to read() below.
# read() acquires a lock, though if the lock is not released when
# read() fails (because the channel has been closed), then an additional
# call to read() *could* block.
# We wrap both calls to read() in pytest.raises() as both calls could
# trigger an RayChannelError exception if the channel has already been closed.
with pytest.raises(
ray.exceptions.RayTaskError, match=r"Channel closed"
) as exc_info:
ray.get([a.close.remote(), b.read.remote()])
assert isinstance(exc_info.value.as_instanceof_cause(), RayChannelError)
with pytest.raises(ray.exceptions.RayTaskError) as exc_info:
ray.get(b.read.remote())
assert isinstance(exc_info.value.as_instanceof_cause(), RayChannelError)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_errors(ray_start_regular):
"""
Tests that an exception is thrown when there are more readers than specificed in the
channel constructor.
"""
@ray.remote
class Actor:
def make_chan(self, readers, do_write=True):
self.chan = ray_channel.Channel(
ray.get_runtime_context().current_actor, readers, 1000
)
if do_write:
self.chan.write(b"hello")
return self.chan
a = Actor.remote()
# Multiple consecutive reads from the same process are fine.
driver_actor = create_driver_actor()
chan = ray.get(
a.make_chan.remote(
[(driver_actor, get_actor_node_id(driver_actor))], do_write=True
)
)
assert chan.read() == b"hello"
@ray.remote
class Reader:
def __init__(self):
pass
def read(self, chan):
return chan.read()
readers = [Reader.remote(), Reader.remote()]
# Check that an exception is thrown when there are more readers than specificed in
# the channel constructor.
chan = ray.get(
a.make_chan.remote([(readers[0], get_actor_node_id(readers[0]))], do_write=True)
)
# At least 1 reader.
with pytest.raises(ray.exceptions.RayTaskError) as exc_info:
ray.get([reader.read.remote(chan) for reader in readers])
assert "ray.exceptions.RaySystemError" in str(exc_info.value)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_put_different_meta(ray_start_regular):
driver_actor = create_driver_actor()
chan = ray_channel.Channel(
None, [(driver_actor, get_actor_node_id(driver_actor))], 1000
)
def _test(val):
chan.write(val)
read_val = chan.read()
if isinstance(val, np.ndarray):
assert np.array_equal(read_val, val)
else:
assert read_val == val
_test(b"hello")
_test("hello")
_test(1000)
_test(np.random.rand(10))
def test_multiple_channels_different_nodes(ray_start_cluster):
"""
Tests that multiple channels can be used at the same time between two nodes.
"""
cluster = ray_start_cluster
# This node is for the driver.
cluster.add_node(num_cpus=0)
ray.init(address=cluster.address)
# This node is for the Reader actors.
cluster.add_node(num_cpus=1)
@ray.remote(num_cpus=1)
class Actor:
def read(self, channel, val):
read_val = channel.read()
if isinstance(val, np.ndarray):
assert np.array_equal(read_val, val)
else:
assert read_val == val
a = Actor.remote()
node_a = get_actor_node_id(a)
chan_a = ray_channel.Channel(None, [(a, node_a)], 1000)
chan_b = ray_channel.Channel(None, [(a, node_a)], 1000)
channels = [chan_a, chan_b]
val = np.random.rand(5)
for i in range(10):
for channel in channels:
channel.write(val)
for channel in channels:
ray.get(a.read.remote(channel, val))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_resize_channel_on_same_node(ray_start_regular):
"""
Tests that the channel backing store is automatically increased when a large object
is written to it. The writer and reader are on the same node.
"""
driver_actor = create_driver_actor()
chan = ray_channel.Channel(
None, [(driver_actor, get_actor_node_id(driver_actor))], 1000
)
def _test(val):
chan.write(val)
read_val = chan.read()
if isinstance(val, np.ndarray):
assert np.array_equal(read_val, val)
else:
assert read_val == val
# `np.random.rand(100)` requires more than 1000 bytes of storage. The channel is
# allocated above with a backing store size of 1000 bytes.
_test(np.random.rand(100))
# Check that another write still works.
_test(np.random.rand(5))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_resize_channel_on_same_node_with_actor(ray_start_regular):
"""
Tests that the channel backing store is automatically increased when a large object
is written to it. The writer and reader are on the same node, and the reader is an
actor.
"""
@ray.remote
class Actor:
def __init__(self):
pass
def read(self, channel, val):
read_val = channel.read()
if isinstance(val, np.ndarray):
assert np.array_equal(read_val, val)
else:
assert read_val == val
def _test(channel, actor, val):
channel.write(val)
ray.get(actor.read.remote(channel, val))
a = Actor.remote()
node_a = get_actor_node_id(a)
chan = ray_channel.Channel(None, [(a, node_a)], 1000)
# `np.random.rand(100)` requires more than 1000 bytes of storage. The channel is
# allocated above with a backing store size of 1000 bytes.
_test(chan, a, np.random.rand(100))
# Check that another write still works.
_test(chan, a, np.random.rand(5))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_resize_channel_on_different_nodes(ray_start_cluster):
"""
Tests that the channel backing store is automatically increased when a large object
is written to it. The writer and reader are on different nodes, and the reader is an
actor.
"""
cluster = ray_start_cluster
# This node is for the driver.
cluster.add_node(num_cpus=0)
ray.init(address=cluster.address)
# This node is for the Reader actors.
cluster.add_node(num_cpus=1)
@ray.remote(num_cpus=1)
class Actor:
def __init__(self):
pass
def read(self, channel, val):
read_val = channel.read()
if isinstance(val, np.ndarray):
assert np.array_equal(read_val, val)
else:
assert read_val == val
def _test(channel, actor, val):
channel.write(val)
ray.get(actor.read.remote(channel, val))
a = Actor.remote()
node_a = get_actor_node_id(a)
chan = ray_channel.Channel(None, [(a, node_a)], 1000)
# `np.random.rand(100)` requires more than 1000 bytes of storage. The channel is
# allocated above with a backing store size of 1000 bytes.
_test(chan, a, np.random.rand(100))
# Check that another write still works.
_test(chan, a, np.random.rand(5))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
@pytest.mark.parametrize("num_readers", [1, 4])
def test_put_remote_get(ray_start_regular, num_readers):
"""
Tests that an actor can read objects/primitives of various types through a channel
when the reader is spawned with @ray.remote.
"""
@ray.remote(num_cpus=0)
class Reader:
def __init__(self):
pass
def read(self, chan, num_writes):
for i in range(num_writes):
val = i.to_bytes(8, "little")
assert chan.read() == val
for i in range(num_writes):
val = i.to_bytes(100, "little")
assert chan.read() == val
for val in [
b"hello world",
"hello again",
1000,
]:
assert chan.read() == val
num_writes = 1000
reader_and_node_list = []
for _ in range(num_readers):
handle = Reader.remote()
node = get_actor_node_id(handle)
reader_and_node_list.append((handle, node))
chan = ray_channel.Channel(None, reader_and_node_list, 1000)
chan.ensure_registered_as_writer()
done = [reader.read.remote(chan, num_writes) for reader, _ in reader_and_node_list]
for i in range(num_writes):
val = i.to_bytes(8, "little")
chan.write(val)
# Test different data size.
for i in range(num_writes):
val = i.to_bytes(100, "little")
chan.write(val)
# Test different metadata.
for val in [
b"hello world",
"hello again",
1000,
]:
chan.write(val)
ray.get(done)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
@pytest.mark.parametrize("remote", [True, False])
def test_remote_reader(ray_start_cluster, remote):
"""
Tests that an actor can read objects/primitives of various types through a channel
when the reader and writer are on the (1) same node (remote=False) along with (2)
different nodes (remote=True).
"""
num_readers = 10
num_writes = 1000
num_iterations = 3
cluster = ray_start_cluster
if remote:
# This node is for the driver.
cluster.add_node(num_cpus=0)
ray.init(address=cluster.address)
# This node is for the Reader actors.
cluster.add_node(num_cpus=num_readers)
else:
# This node is for both the driver and the Reader actors.
cluster.add_node(num_cpus=num_readers)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Reader:
def __init__(self):
pass
def get_node_id(self) -> str:
return ray.get_runtime_context().get_node_id()
def pass_channel(self, channel):
self._reader_chan = channel
def read(self, num_reads):
for i in range(num_reads):
self._reader_chan.read()
reader_and_node_list = []
for _ in range(num_readers):
handle = Reader.remote()
node = get_actor_node_id(handle)
reader_and_node_list.append((handle, node))
channel = ray_channel.Channel(None, reader_and_node_list, 1000)
# All readers have received the channel.
ray.get([reader.pass_channel.remote(channel) for reader, _ in reader_and_node_list])
for _ in range(num_iterations):
work = [reader.read.remote(num_writes) for reader, _ in reader_and_node_list]
start = time.perf_counter()
for i in range(num_writes):
channel.write(b"x")
end = time.perf_counter()
ray.get(work)
print(end - start, 10_000 / (end - start))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
@pytest.mark.parametrize("remote", [True, False])
def test_remote_reader_close(ray_start_cluster, remote):
"""
Tests that readers do not block forever on read() when they close the channel.
Specifically, the following behavior should happen:
1. Each reader calls read() on one channel.
2. Each reader calls close() on the channel on a different thread.
3. Each reader should unblock and return from read().
Tests (1) the readers and writer on the same node (remote=False) along with
different nodes (remote=True).
"""
num_readers = 10
cluster = ray_start_cluster
if remote:
# This node is for the driver.
cluster.add_node(num_cpus=0)
ray.init(address=cluster.address)
# This node is for the Reader actors.
cluster.add_node(num_cpus=num_readers)
else:
# This node is for both the driver and the Reader actors.
cluster.add_node(num_cpus=num_readers)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Reader:
def __init__(self):
pass
def get_node_id(self) -> str:
return ray.get_runtime_context().get_node_id()
def pass_channel(self, channel):
self._reader_chan = channel
def read(self):
try:
self._reader_chan.read()
except RayChannelError:
pass
def close(self):
self._reader_chan.close()
reader_and_node_list = []
for _ in range(num_readers):
handle = Reader.remote()
node = get_actor_node_id(handle)
reader_and_node_list.append((handle, node))
channel = ray_channel.Channel(None, reader_and_node_list, 1000)
# All readers have received the channel.
ray.get([reader.pass_channel.remote(channel) for reader, _ in reader_and_node_list])
reads = [
reader.read.options(concurrency_group="_ray_system").remote()
for reader, _ in reader_and_node_list
]
with pytest.raises(ray.exceptions.GetTimeoutError):
ray.get(reads, timeout=1.0)
ray.get([reader.close.remote() for reader, _ in reader_and_node_list])
ray.get(reads)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_intra_process_channel_single_reader(ray_start_cluster):
"""
(1) Test whether an actor can read/write from an IntraProcessChannel.
(2) Test whether the _SerializationContext cleans up the
data after all readers have read it.
(3) Test whether the actor can write again after reading 1 time.
"""
# This node is for both the driver and the Reader actors.
cluster = ray_start_cluster
cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def __init__(self):
pass
def pass_channel(self, channel):
self._chan = channel
def read(self):
return self._chan.read()
def write(self, value):
self._chan.write(value)
def get_ctx_buffer_size(self):
ctx = ray_channel.ChannelContext.get_current().serialization_context
return len(ctx.intra_process_channel_buffers)
actor = Actor.remote()
channel = ray_channel.IntraProcessChannel(num_readers=1)
ray.get(actor.pass_channel.remote(channel))
ray.get(actor.write.remote("hello"))
assert ray.get(actor.read.remote()) == "hello"
# The _SerializationContext should clean up the data after a read.
assert ray.get(actor.get_ctx_buffer_size.remote()) == 0
# Write again after reading num_readers times.
ray.get(actor.write.remote("world"))
assert ray.get(actor.read.remote()) == "world"
# The _SerializationContext should clean up the data after a read.
assert ray.get(actor.get_ctx_buffer_size.remote()) == 0
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_intra_process_channel_multi_readers(ray_start_cluster):
"""
(1) Test whether an actor can read/write from an IntraProcessChannel.
(2) Test whether the _SerializationContext cleans up the
data after all readers have read it.
(3) Test whether the actor can write again after reading num_readers times.
(4) Test whether an exception is raised when calling write() before all readers
have read the data.
"""
# This node is for both the driver and the Reader actors.
cluster = ray_start_cluster
cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def __init__(self):
pass
def pass_channel(self, channel):
self._chan = channel
def read(self):
return self._chan.read()
def write(self, value):
self._chan.write(value)
def get_ctx_buffer_size(self):
ctx = ray_channel.ChannelContext.get_current().serialization_context
return len(ctx.intra_process_channel_buffers)
actor = Actor.remote()
channel = ray_channel.IntraProcessChannel(num_readers=2)
ray.get(actor.pass_channel.remote(channel))
ray.get(actor.write.remote("hello"))
# first read
assert ray.get(actor.read.remote()) == "hello"
assert ray.get(actor.get_ctx_buffer_size.remote()) == 1
# second read
assert ray.get(actor.read.remote()) == "hello"
assert ray.get(actor.get_ctx_buffer_size.remote()) == 0
# Write again after reading num_readers times.
ray.get(actor.write.remote("world"))
# first read
assert ray.get(actor.read.remote()) == "world"
assert ray.get(actor.get_ctx_buffer_size.remote()) == 1
# second read
assert ray.get(actor.read.remote()) == "world"
assert ray.get(actor.get_ctx_buffer_size.remote()) == 0
# Write again
ray.get(actor.write.remote("hello world"))
# first read
assert ray.get(actor.read.remote()) == "hello world"
assert ray.get(actor.get_ctx_buffer_size.remote()) == 1
with pytest.raises(ray.exceptions.RayTaskError):
ray.get(actor.write.remote("world hello"))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_composite_channel_single_reader(ray_start_cluster):
"""
(1) The driver can write data to CompositeChannel and an actor can read it.
(2) An actor can write data to CompositeChannel and the actor itself can read it.
(3) An actor can write data to CompositeChannel and another actor can read it.
(4) An actor can write data to CompositeChannel and the driver can read it.
"""
# This node is for both the driver and the Reader actors.
cluster = ray_start_cluster
cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def __init__(self):
pass
def pass_channel(self, channel):
self._chan = channel
def create_composite_channel(self, writer, reader_and_node_list):
self._chan = ray_channel.CompositeChannel(writer, reader_and_node_list)
return self._chan
def read(self):
return self._chan.read()
def write(self, value):
self._chan.write(value)
actor1 = Actor.remote()
actor2 = Actor.remote()
node1 = get_actor_node_id(actor1)
node2 = get_actor_node_id(actor2)
# Create a channel to communicate between driver process and actor1.
driver_to_actor1_channel = ray_channel.CompositeChannel(None, [(actor1, node1)])
ray.get(actor1.pass_channel.remote(driver_to_actor1_channel))
driver_to_actor1_channel.write("hello")
assert ray.get(actor1.read.remote()) == "hello"
# Create a channel to communicate between two tasks in actor1.
ray.get(actor1.create_composite_channel.remote(actor1, [(actor1, node1)]))
ray.get(actor1.write.remote("world"))
assert ray.get(actor1.read.remote()) == "world"
# Create a channel to communicate between actor1 and actor2.
actor1_to_actor2_channel = ray.get(
actor1.create_composite_channel.remote(actor1, [(actor2, node2)])
)
ray.get(actor2.pass_channel.remote(actor1_to_actor2_channel))
ray.get(actor1.write.remote("hello world"))
assert ray.get(actor2.read.remote()) == "hello world"
# Create a channel to communicate between actor2 and driver process.
driver_actor = create_driver_actor()
actor2_to_driver_channel = ray.get(
actor2.create_composite_channel.remote(
actor2, [(driver_actor, get_actor_node_id(driver_actor))]
)
)
ray.get(actor2.write.remote("world hello"))
assert actor2_to_driver_channel.read() == "world hello"
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_composite_channel_multiple_readers(ray_start_cluster):
"""
Test the behavior of CompositeChannel when there are multiple readers.
(1) The driver can write data to CompositeChannel and two actors can read it.
(2) An actor can write data to CompositeChannel and another actor, as well as
itself, can read it.
(3) An actor writes data to CompositeChannel and two Ray tasks on the same
actor read it. This is not supported and should raise an exception.
"""
# This node is for both the driver and the Reader actors.
cluster = ray_start_cluster
cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
@ray.remote(num_cpus=1)
class Actor:
def __init__(self):
pass
def pass_channel(self, channel):
self._chan = channel
def create_composite_channel(self, writer, reader_and_node_list):
self._chan = ray_channel.CompositeChannel(writer, reader_and_node_list)
return self._chan
def read(self):
return self._chan.read()
def write(self, value):
self._chan.write(value)
actor1 = Actor.remote()
actor2 = Actor.remote()
node1 = get_actor_node_id(actor1)
node2 = get_actor_node_id(actor2)
# The driver writes data to CompositeChannel and actor1 and actor2 read it.
driver_output_channel = ray_channel.CompositeChannel(
None, [(actor1, node1), (actor2, node2)]
)
ray.get(actor1.pass_channel.remote(driver_output_channel))
ray.get(actor2.pass_channel.remote(driver_output_channel))
driver_output_channel.write("hello")
assert ray.get([actor1.read.remote(), actor2.read.remote()]) == ["hello"] * 2
# actor1 writes data to CompositeChannel and actor1 and actor2 read it.
actor1_output_channel = ray.get(
actor1.create_composite_channel.remote(
actor1, [(actor1, node1), (actor2, node2)]
)
)
ray.get(actor2.pass_channel.remote(actor1_output_channel))
ray.get(actor1.write.remote("world"))
assert ray.get([actor1.read.remote(), actor2.read.remote()]) == ["world"] * 2
actor1_output_channel = ray.get(
actor1.create_composite_channel.remote(
actor1, [(actor1, node1), (actor1, node1)]
)
)
ray.get(actor1.write.remote("hello world"))
assert ray.get(actor1.read.remote()) == "hello world"
assert ray.get(actor1.read.remote()) == "hello world"
with pytest.raises(ray.exceptions.RayTaskError):
# actor1_output_channel has two readers, so it can only be read twice.
# The third read should raise an exception.
ray.get(actor1.read.remote())
"""
TODO (kevin85421): Add tests for the following cases:
(1) actor1 writes data to CompositeChannel and two Ray tasks on actor2 read it.
(2) actor1 writes data to CompositeChannel and actor2 and the driver reads it.
Currently, (1) is not supported, and (2) is blocked by the reference count issue.
"""
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_put_error(ray_start_cluster):
cluster = ray_start_cluster
# This node is for both the driver (including the CompiledDAG.DAGDriverProxyActor)
# and the writer actor.
cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
def _wrap_exception(exc):
backtrace = ray._private.utils.format_error_message(
"".join(traceback.format_exception(type(exc), exc, exc.__traceback__)),
task_exception=True,
)
wrapped = ray.exceptions.RayTaskError(
function_name="do_exec_tasks",
traceback_str=backtrace,
cause=exc,
)
return wrapped
@ray.remote(num_cpus=1)
class Actor:
def setup(self, reader_and_node_list):
self._channel = ray_channel.Channel(
ray.get_runtime_context().current_actor,
reader_and_node_list,
1000,
)
def get_channel(self):
return self._channel
def write(self, write_error):
if write_error:
try:
raise ValueError("")
except Exception as exc:
self._channel.write(_wrap_exception(exc))
else:
self._channel.write(b"x")
a = Actor.remote()
driver_actor = create_driver_actor()
ray.get(a.setup.remote([(driver_actor, get_actor_node_id(driver_actor))]))
chan = ray.get(a.get_channel.remote())
# Putting a bytes object multiple times is okay.
for _ in range(3):
ray.get(a.write.remote(write_error=False))
assert chan.read() == b"x"
# Putting an exception multiple times is okay.
for _ in range(3):
ray.get(a.write.remote(write_error=True))
try:
assert chan.read()
except Exception as exc:
assert isinstance(exc, ValueError)
assert isinstance(exc, ray.exceptions.RayTaskError)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_payload_large(ray_start_cluster):
cluster = ray_start_cluster
# This node is for the driver.
first_node_handle = cluster.add_node(num_cpus=1)
# This node is for the reader.
second_node_handle = cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
cluster.wait_for_nodes()
nodes = [first_node_handle.node_id, second_node_handle.node_id]
# We want to check that there are two nodes. Thus, we convert `nodes` to a set and
# then back to a list to remove duplicates. Then we check that the length of `nodes`
# is 2.
nodes = list(set(nodes))
assert len(nodes) == 2
@ray.remote(num_cpus=1)
class Actor:
def get_node_id(self):
return ray.get_runtime_context().get_node_id()
def read(self, channel, val):
assert channel.read() == val
def create_actor(node):
return Actor.options(
scheduling_strategy=NodeAffinitySchedulingStrategy(node, soft=False)
).remote()
driver_node = ray.get_runtime_context().get_node_id()
actor_node = nodes[0] if nodes[0] != driver_node else nodes[1]
assert driver_node != actor_node
a = create_actor(actor_node)
node_a = ray.get(a.get_node_id.remote())
assert driver_node != ray.get(a.get_node_id.remote())
# Ray sets the gRPC payload max size to 512 MiB. We choose a size in this test that
# is a bit larger.
size = 1024 * 1024 * 600
ch = ray_channel.Channel(None, [(a, node_a)], size)
val = b"x" * size
ch.write(val)
ray.get(a.read.remote(ch, val))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_payload_resize_large(ray_start_cluster):
cluster = ray_start_cluster
# This node is for the driver.
first_node_handle = cluster.add_node(num_cpus=1)
# This node is for the reader.
second_node_handle = cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
cluster.wait_for_nodes()
nodes = [first_node_handle.node_id, second_node_handle.node_id]
# We want to check that there are two nodes. Thus, we convert `nodes` to a set and
# then back to a list to remove duplicates. Then we check that the length of `nodes`
# is 2.
nodes = list(set(nodes))
assert len(nodes) == 2
@ray.remote(num_cpus=1)
class Actor:
def get_node_id(self):
return ray.get_runtime_context().get_node_id()
def read(self, channel, val):
assert channel.read() == val
def create_actor(node):
return Actor.options(
scheduling_strategy=NodeAffinitySchedulingStrategy(node, soft=False)
).remote()
driver_node = ray.get_runtime_context().get_node_id()
actor_node = nodes[0] if nodes[0] != driver_node else nodes[1]
assert driver_node != actor_node
a = create_actor(actor_node)
assert driver_node != ray.get(a.get_node_id.remote())
ch = ray_channel.Channel(None, [(a, actor_node)], 1000)
# Ray sets the gRPC payload max size to 512 MiB. We choose a size in this test that
# is a bit larger.
size = 1024 * 1024 * 600
val = b"x" * size
ch.write(val)
ray.get(a.read.remote(ch, val))
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_readers_on_different_nodes(ray_start_cluster):
cluster = ray_start_cluster
# This node is for the driver (including the CompiledDAG.DAGDriverProxyActor) and
# one of the readers.
first_node_handle = cluster.add_node(num_cpus=2)
# This node is for the other reader.
second_node_handle = cluster.add_node(num_cpus=1)
ray.init(address=cluster.address)
cluster.wait_for_nodes()
nodes = [first_node_handle.node_id, second_node_handle.node_id]
# We want to check that there are two nodes. Thus, we convert `nodes` to a set and
# then back to a list to remove duplicates. Then we check that the length of `nodes`
# is 2.
nodes = list(set(nodes))
assert len(nodes) == 2
@ray.remote(num_cpus=1)
class Actor:
def get_node_id(self):
return ray.get_runtime_context().get_node_id()
def create_actor(node):
return Actor.options(
scheduling_strategy=NodeAffinitySchedulingStrategy(node, soft=False)
).remote()
a = create_actor(nodes[0])
b = create_actor(nodes[1])
actors = [a, b]
nodes_check = ray.get([act.get_node_id.remote() for act in actors])
a_node = nodes_check[0]
b_node = nodes_check[1]
assert a_node != b_node
driver_actor = create_driver_actor()
driver_node = get_actor_node_id(driver_actor)
with pytest.raises(
ValueError, match="All reader actors must be on the same node.*"
):
ray_channel.Channel(
None, [(driver_actor, driver_node), (a, a_node), (b, b_node)], 1000
)
@pytest.mark.skipif(
sys.platform != "linux" and sys.platform != "darwin",
reason="Requires Linux or Mac.",
)
def test_bunch_readers_on_different_nodes(ray_start_cluster):
cluster = ray_start_cluster
# This node is for the driver (including the DriverHelperActor) and two of the
# readers.
first_node_handle = cluster.add_node(num_cpus=3)
# This node is for the other two readers.
second_node_handle = cluster.add_node(num_cpus=2)
ray.init(address=cluster.address)
cluster.wait_for_nodes()
nodes = [first_node_handle.node_id, second_node_handle.node_id]
# We want to check that the readers are on different nodes. Thus, we convert `nodes`
# to a set and then back to a list to remove duplicates. Then we check that the
# length of `nodes` is 2.
nodes = list(set(nodes))
assert len(nodes) == 2
@ray.remote(num_cpus=1)
class Actor:
def get_node_id(self):
return ray.get_runtime_context().get_node_id()
def create_actor(node):
return Actor.options(
scheduling_strategy=NodeAffinitySchedulingStrategy(node, soft=False)
).remote()
a = create_actor(nodes[0])
b = create_actor(nodes[0])
c = create_actor(nodes[1])
d = create_actor(nodes[1])
actors = [a, b, c, d]
nodes_check = ray.get([act.get_node_id.remote() for act in actors])
a_node = nodes_check[0]
b_node = nodes_check[1]
c_node = nodes_check[2]
d_node = nodes_check[3]
assert a_node == b_node
assert b_node != c_node
assert c_node == d_node
driver_actor = create_driver_actor()
driver_node = get_actor_node_id(driver_actor)
with pytest.raises(
ValueError, match="All reader actors must be on the same node.*"
):
ray_channel.Channel(
None,
[
(driver_actor, driver_node),
(a, a_node),
(b, b_node),
(c, c_node),
(d, d_node),
],
1000,
)
if __name__ == "__main__":
if os.environ.get("PARALLEL_CI"):
sys.exit(pytest.main(["-n", "auto", "--boxed", "-vs", __file__]))
else:
sys.exit(pytest.main(["-sv", __file__]))
```
|
```xml
import { Component, OnInit, Input } from '@angular/core';
import { BsModalRef } from 'ngx-bootstrap/modal';
@Component({
selector: 'app-alert-modal',
templateUrl: './alert-modal.component.html',
styleUrls: ['./alert-modal.component.scss']
})
export class AlertModalComponent implements OnInit {
@Input() type = 'success';
@Input() message: string;
constructor(public bsModalRef: BsModalRef) { }
ngOnInit() {
}
onClose() {
this.bsModalRef.hide();
}
}
```
|
Hackpen White Horse is a chalk hill figure of a white horse on Hackpen Hill, located below The Ridgeway on the edge of the Marlborough Downs, two miles south east of Broad Hinton, Wiltshire, England. It is one of nine white horse hill figures located in Wiltshire. It is also known as the Broad Hinton White Horse due to its proximity to the village. Supposedly cut by local parish clerk Henry Eatwell in 1838 to commemorate the coronation of Queen Victoria, the horse is wide by 90 feet (27m) high. The horse is regularly scoured and maintained.
Origins and early history
The origin of the horse is uncertain, and is sometimes said to be the only 19th century white horse to have little of its history known. It is generally regarded that the horse was cut in 1838 by Henry Eatwell, a parish clerk of Broad Hinton, assisted by a local pub landlord. It is said to commemorate the coronation of Queen Victoria.
Description and location
The horse is cut of chalk, is square, making it the only square-dimension horse in England, and faces west-northwest. Although Hackpen Hill is high (), it is a gentle slope, especially when compared to the hills of most other Wiltshire horses. Because the hill is gentle, the horse is partly banked up and slightly raised from the surrounding grass to make it more easily visible. The head was initially elevated to help with the foreshortening. The best view of the horse is said to be from the nearby B4041 road, whilst the A361 road near Broad Hinton also provides a clear view.
At the top of the hill is a car park where the Ridgeway crosses the B4041 road, and a footpath stretches from there down to the horse, making the horse accessible to the public. Ironically, many real horses often roam the field. It has been suggested that the stones for Stonehenge and Avebury may have come from a field of sarsen stones just to the south east of its location. The expression "as different as chalk and cheese" is sometimes believed to refer to the land divided by Hackpen Hill. The hill forms the boundary between the high chalk downs to the south of it and the clay cattle country to the north, where cheese is a product of the milk from the cattle, so the two areas "are as different as chalk and cheese." Hackpen White Horse was not the only hillside shape cut to commemorate Queen Victoria; in 1887, for her Golden Jubilee, a hillside row of trees were planted in the shape of a "V" in Westmeston, Sussex.
The horse ties "neck-and-neck" with Broad Town White Horse as the closest white horse to Swindon.
Scouring and recent history
The horse is regularly scoured (cleaned and maintained). In either May or June 2000, John Wain cleaned it single-handedly, taking him some five hours. He later flew David Brewer over the area to photograph the village of Broad Hinton and the white horse for brewers's book Images of a Wiltshire Downland Village: Broad Hinton and Uffcott. Wain cleaned it annually until Bevan Pope cleaned the horse single-handedly on 23 September 2004. Wain cleaned the horse again with the help of a group of friends on 1 February 2011 and 4 February 2012. On both occasions, they illuminated the newly cleaned horse when the work was completed. Although to illuminate a white horse has been sporadic tradition for other horses in Wiltshire, those occasions marked the first times it had been done for Hackpen White Horse.
In March 2009, the horse was transformed into a "red horse" for the Comic Relief charity's Red Nose Day campaign; this was achieved by covering the horse's shape with numerous red sheets and fabrics. The White Horse pub, located half a mile away in Winterbourne Bassett, features an illustration resembling the horse as its logo. The pub itself was named after the eight horses in Wiltshire.
The horse has also featured in several artworks, including a stained glass window made by Berry Stained Glass, Benoit Philppe's The Hackpen White Horse oil on canvas painting, and a silver necklace created in 2015 by Devizes-based jeweller Daniel Pike. In 2005, the horse appeared in episode 1 of series 6 of Top Gear, and, in 2012, for a Pukka Pies sponsorship advert for ITV travel series Ade in Britain, Pukka Pies modified a photograph of the location to include a hill figure of one of their pies instead of the horse.
See also
Wiltshire white horses
Westbury White Horse
Pewsey White Horse
Devizes White Horse
Broad Town White Horse
Cherhill White Horse
Marlborough White Horse
Alton Barnes White Horse
Other white horses
Uffington White Horse
Osmington White Horse
Kilburn White Horse
Woolbury White Horse
Litlington White Horse
References
White horses (hill figures) in England
History of Wiltshire
1838 works
Monuments and memorials to Queen Victoria
|
```objective-c
/*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
*
* You should have received a copy of the GNU Lesser General Public
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#ifndef AVUTIL_TIME_INTERNAL_H
#define AVUTIL_TIME_INTERNAL_H
#include <time.h>
#include "config.h"
#if !HAVE_GMTIME_R && !defined(gmtime_r)
static inline struct tm *gmtime_r(const time_t* clock, struct tm *result)
{
struct tm *ptr = gmtime(clock);
if (!ptr)
return NULL;
*result = *ptr;
return result;
}
#endif
#if !HAVE_LOCALTIME_R && !defined(localtime_r)
static inline struct tm *localtime_r(const time_t* clock, struct tm *result)
{
struct tm *ptr = localtime(clock);
if (!ptr)
return NULL;
*result = *ptr;
return result;
}
#endif
#endif /* AVUTIL_TIME_INTERNAL_H */
```
|
```objective-c
// This file is part of Eigen, a lightweight C++ template library
// for linear algebra.
//
//
// This Source Code Form is subject to the terms of the Mozilla
// with this file, You can obtain one at path_to_url
#ifndef EIGEN_CWISE_BINARY_OP_H
#define EIGEN_CWISE_BINARY_OP_H
namespace Eigen {
namespace internal {
template<typename BinaryOp, typename Lhs, typename Rhs>
struct traits<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >
{
// we must not inherit from traits<Lhs> since it has
// the potential to cause problems with MSVC
typedef typename remove_all<Lhs>::type Ancestor;
typedef typename traits<Ancestor>::XprKind XprKind;
enum {
RowsAtCompileTime = traits<Ancestor>::RowsAtCompileTime,
ColsAtCompileTime = traits<Ancestor>::ColsAtCompileTime,
MaxRowsAtCompileTime = traits<Ancestor>::MaxRowsAtCompileTime,
MaxColsAtCompileTime = traits<Ancestor>::MaxColsAtCompileTime
};
// even though we require Lhs and Rhs to have the same scalar type (see CwiseBinaryOp constructor),
// we still want to handle the case when the result type is different.
typedef typename result_of<
BinaryOp(
const typename Lhs::Scalar&,
const typename Rhs::Scalar&
)
>::type Scalar;
typedef typename cwise_promote_storage_type<typename traits<Lhs>::StorageKind,
typename traits<Rhs>::StorageKind,
BinaryOp>::ret StorageKind;
typedef typename promote_index_type<typename traits<Lhs>::StorageIndex,
typename traits<Rhs>::StorageIndex>::type StorageIndex;
typedef typename Lhs::Nested LhsNested;
typedef typename Rhs::Nested RhsNested;
typedef typename remove_reference<LhsNested>::type _LhsNested;
typedef typename remove_reference<RhsNested>::type _RhsNested;
enum {
Flags = cwise_promote_storage_order<typename traits<Lhs>::StorageKind,typename traits<Rhs>::StorageKind,_LhsNested::Flags & RowMajorBit,_RhsNested::Flags & RowMajorBit>::value
};
};
} // end namespace internal
template<typename BinaryOp, typename Lhs, typename Rhs, typename StorageKind>
class CwiseBinaryOpImpl;
/** \class CwiseBinaryOp
* \ingroup Core_Module
*
* \brief Generic expression where a coefficient-wise binary operator is applied to two expressions
*
* \tparam BinaryOp template functor implementing the operator
* \tparam LhsType the type of the left-hand side
* \tparam RhsType the type of the right-hand side
*
* This class represents an expression where a coefficient-wise binary operator is applied to two expressions.
* It is the return type of binary operators, by which we mean only those binary operators where
* both the left-hand side and the right-hand side are Eigen expressions.
* For example, the return type of matrix1+matrix2 is a CwiseBinaryOp.
*
* Most of the time, this is the only way that it is used, so you typically don't have to name
* CwiseBinaryOp types explicitly.
*
* \sa MatrixBase::binaryExpr(const MatrixBase<OtherDerived> &,const CustomBinaryOp &) const, class CwiseUnaryOp, class CwiseNullaryOp
*/
template<typename BinaryOp, typename LhsType, typename RhsType>
class CwiseBinaryOp :
public CwiseBinaryOpImpl<
BinaryOp, LhsType, RhsType,
typename internal::cwise_promote_storage_type<typename internal::traits<LhsType>::StorageKind,
typename internal::traits<RhsType>::StorageKind,
BinaryOp>::ret>,
internal::no_assignment_operator
{
public:
typedef typename internal::remove_all<BinaryOp>::type Functor;
typedef typename internal::remove_all<LhsType>::type Lhs;
typedef typename internal::remove_all<RhsType>::type Rhs;
typedef typename CwiseBinaryOpImpl<
BinaryOp, LhsType, RhsType,
typename internal::cwise_promote_storage_type<typename internal::traits<LhsType>::StorageKind,
typename internal::traits<Rhs>::StorageKind,
BinaryOp>::ret>::Base Base;
EIGEN_GENERIC_PUBLIC_INTERFACE(CwiseBinaryOp)
typedef typename internal::ref_selector<LhsType>::type LhsNested;
typedef typename internal::ref_selector<RhsType>::type RhsNested;
typedef typename internal::remove_reference<LhsNested>::type _LhsNested;
typedef typename internal::remove_reference<RhsNested>::type _RhsNested;
EIGEN_DEVICE_FUNC
EIGEN_STRONG_INLINE CwiseBinaryOp(const Lhs& aLhs, const Rhs& aRhs, const BinaryOp& func = BinaryOp())
: m_lhs(aLhs), m_rhs(aRhs), m_functor(func)
{
EIGEN_CHECK_BINARY_COMPATIBILIY(BinaryOp,typename Lhs::Scalar,typename Rhs::Scalar);
// require the sizes to match
EIGEN_STATIC_ASSERT_SAME_MATRIX_SIZE(Lhs, Rhs)
eigen_assert(aLhs.rows() == aRhs.rows() && aLhs.cols() == aRhs.cols());
}
EIGEN_DEVICE_FUNC
EIGEN_STRONG_INLINE Index rows() const {
// return the fixed size type if available to enable compile time optimizations
if (internal::traits<typename internal::remove_all<LhsNested>::type>::RowsAtCompileTime==Dynamic)
return m_rhs.rows();
else
return m_lhs.rows();
}
EIGEN_DEVICE_FUNC
EIGEN_STRONG_INLINE Index cols() const {
// return the fixed size type if available to enable compile time optimizations
if (internal::traits<typename internal::remove_all<LhsNested>::type>::ColsAtCompileTime==Dynamic)
return m_rhs.cols();
else
return m_lhs.cols();
}
/** \returns the left hand side nested expression */
EIGEN_DEVICE_FUNC
const _LhsNested& lhs() const { return m_lhs; }
/** \returns the right hand side nested expression */
EIGEN_DEVICE_FUNC
const _RhsNested& rhs() const { return m_rhs; }
/** \returns the functor representing the binary operation */
EIGEN_DEVICE_FUNC
const BinaryOp& functor() const { return m_functor; }
protected:
LhsNested m_lhs;
RhsNested m_rhs;
const BinaryOp m_functor;
};
// Generic API dispatcher
template<typename BinaryOp, typename Lhs, typename Rhs, typename StorageKind>
class CwiseBinaryOpImpl
: public internal::generic_xpr_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >::type
{
public:
typedef typename internal::generic_xpr_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >::type Base;
};
/** replaces \c *this by \c *this - \a other.
*
* \returns a reference to \c *this
*/
template<typename Derived>
template<typename OtherDerived>
EIGEN_STRONG_INLINE Derived &
MatrixBase<Derived>::operator-=(const MatrixBase<OtherDerived> &other)
{
call_assignment(derived(), other.derived(), internal::sub_assign_op<Scalar,typename OtherDerived::Scalar>());
return derived();
}
/** replaces \c *this by \c *this + \a other.
*
* \returns a reference to \c *this
*/
template<typename Derived>
template<typename OtherDerived>
EIGEN_STRONG_INLINE Derived &
MatrixBase<Derived>::operator+=(const MatrixBase<OtherDerived>& other)
{
call_assignment(derived(), other.derived(), internal::add_assign_op<Scalar,typename OtherDerived::Scalar>());
return derived();
}
} // end namespace Eigen
#endif // EIGEN_CWISE_BINARY_OP_H
```
|
```xml
import React from 'react';
import './fast-refresh';
export declare function App(): React.JSX.Element;
//# sourceMappingURL=qualified-entry.d.ts.map
```
|
```javascript
/**
* @license Apache-2.0
*
*
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing, software
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*/
'use strict';
// MODULES //
var tape = require( 'tape' );
var Float32Array = require( '@stdlib/array/float32' );
var sfill = require( './../lib/ndarray.js' );
// TESTS //
tape( 'main export is a function', function test( t ) {
t.ok( true, __filename );
t.strictEqual( typeof sfill, 'function', 'main export is a function' );
t.end();
});
tape( 'the function has an arity of 5', function test( t ) {
t.strictEqual( sfill.length, 5, 'has expected arity' );
t.end();
});
tape( 'the function fills a strided array', function test( t ) {
var expected;
var x;
x = new Float32Array([
4.0,
2.0,
-3.0,
5.0,
-1.0,
2.0,
-5.0,
6.0
]);
expected = new Float32Array([
5.0,
5.0,
5.0,
5.0,
5.0,
5.0,
5.0,
5.0
]);
sfill( x.length, 5.0, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
x = new Float32Array( [ 1.0, 2.0 ] );
expected = new Float32Array( [ 5.0, 5.0 ] );
sfill( x.length, 5.0, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
tape( 'the function returns a reference to the input array', function test( t ) {
var out;
var x;
x = new Float32Array( [ 1.0, 2.0, 3.0, 4.0, 5.0 ] );
out = sfill( x.length, 3.0, x, 1, 0 );
t.strictEqual( out, x, 'same reference' );
t.end();
});
tape( 'if provided an `N` parameter less than or equal to `0`, the function returns the input array unchanged', function test( t ) {
var expected;
var x;
x = new Float32Array( [ 3.0, -4.0, 1.0 ] );
expected = new Float32Array( [ 3.0, -4.0, 1.0 ] );
sfill( 0, 10.0, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
sfill( -4, 10.0, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
tape( 'the function supports specifying a stride', function test( t ) {
var expected;
var x;
x = new Float32Array([
2.0, // 0
-3.0,
-5.0, // 1
7.0,
6.0 // 2
]);
expected = new Float32Array([
5.0, // 0
-3.0,
5.0, // 1
7.0,
5.0 // 2
]);
sfill( 3, 5.0, x, 2, 0 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
tape( 'the function supports specifying a negative stride', function test( t ) {
var expected;
var x;
x = new Float32Array([
2.0, // 2
-3.0,
-5.0, // 1
7.0,
6.0 // 0
]);
expected = new Float32Array([
5.0, // 2
-3.0,
5.0, // 1
7.0,
5.0 // 0
]);
sfill( 3, 5.0, x, -2, x.length-1 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
tape( 'the function supports an offset parameter', function test( t ) {
var expected;
var x;
x = new Float32Array([
1.0,
2.0, // 0
3.0,
4.0, // 1
5.0,
6.0 // 2
]);
expected = new Float32Array([
1.0,
5.0, // 0
3.0,
5.0, // 1
5.0,
5.0 // 2
]);
sfill( 3, 5.0, x, 2, 1 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
tape( 'if `stride` is equal to `1`, the function efficiently fills a strided array', function test( t ) {
var expected;
var alpha;
var x;
var i;
alpha = 3.0;
x = new Float32Array( 100 );
expected = new Float32Array( x.length );
for ( i = 0; i < x.length; i++ ) {
x[ i ] = i;
expected[ i ] = alpha;
}
sfill( x.length, alpha, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
x = new Float32Array( 240 );
expected = new Float32Array( x.length );
for ( i = 0; i < x.length; i++ ) {
x[ i ] = i;
expected[ i ] = alpha;
}
sfill( x.length, alpha, x, 1, 0 );
t.deepEqual( x, expected, 'returns expected value' );
t.end();
});
```
|
```smalltalk
/****************************************************************************
*
* path_to_url
* path_to_url
* path_to_url
* path_to_url
****************************************************************************/
using UnityEngine;
namespace QFramework
{
public class MDAsset : TextAsset
{
}
}
```
|
```objective-c
/*
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
* OTHER DEALINGS IN THE SOFTWARE.
*
* Authors: AMD
*
*/
#ifndef __DAL_HW_FACTORY_DCN21_H__
#define __DAL_HW_FACTORY_DCN21_H__
/* Initialize HW factory function pointers and pin info */
void dal_hw_factory_dcn21_init(struct hw_factory *factory);
#endif /* __DAL_HW_FACTORY_DCN20_H__ */
```
|
The Recreation Ground is a cricket ground in Kington, Herefordshire. The first recorded match on the ground was in 1996, when Herefordshire played Wales Minor Counties in the grounds first Minor Counties Championship match. From 1996 to present, the ground has hosted 6 Minor Counties Championship matches and 3 MCCA Knockout Trophy matches.
The ground also hosted a single List-A match in 1999, when Herefordshire played Yorkshire in the NatWest Trophy.
In local domestic cricket, The Recreation Ground is the home ground of Kington Cricket Club.
References
External links
The Recreation Ground on CricketArchive
The Recreation Ground on Cricinfo
Cricket grounds in Herefordshire
Sports venues completed in 1996
Herefordshire County Cricket Club
|
The 2022 Melbourne Summer Set 1 was one of two tournaments of the Melbourne Summer Set. The Melbourne Summer Set 1 featured an ATP Tour 250 tournament and a WTA 250 tournament, both played on hardcourts and as part of the 2022 WTA Tour.
Rafael Nadal won the men's singles title to earn his 89th career ATP title and his first in Australia since 2009. Simona Halep won the women's singles title to earn her 23rd career WTA title and first title in Australia.
Champions
Men's singles
Rafael Nadal def. Maxime Cressy 7–6(8–6), 6–3
Women's singles
Simona Halep def. Veronika Kudermetova 6–2, 6–3
Men's doubles
Wesley Koolhof / Neal Skupski def. Aleksandr Nedovyesov / Aisam-ul-Haq Qureshi 6–4, 6–4
Women's doubles
Asia Muhammad / Jessica Pegula def. Sara Errani / Jasmine Paolini 6–3, 6–1
Points and prize money
Point distribution
*per team
Prize money
ATP singles main draw entrants
Seeds
1 Rankings are as of 27 December 2021
Other entrants
The following players received wildcards into the singles main draw:
Nick Kyrgios
Andy Murray
Christopher O'Connell
The following player received entry as an alternate:
Emil Ruusuvuori
The following players received entry from the qualifying draw:
Ričardas Berankis
Maxime Cressy
Rinky Hijikata
Andreas Seppi
The following players received entry as lucky losers:
Sebastián Báez
Henri Laaksonen
Withdrawals
Before the tournament
Alexander Bublik → replaced by Peter Gojowczyk
Lloyd Harris → replaced by Alex Molčan
Ilya Ivashka → replaced by Sebastián Báez
Nick Kyrgios → replaced by Henri Laaksonen
Kei Nishikori → replaced by Emil Ruusuvuori
During the tournament
Tallon Griekspoor
Retirements
Benoît Paire
ATP doubles main-draw entrants
Seeds
1 Rankings are as of 27 December 2021
Other entrants
The following pairs received wildcards into the doubles main draw:
Norbert Gombos / Alex Molčan
Marc Polmans / Alexei Popyrin
The following pair received entry using a protected ranking:
Altuğ Çelikbilek / Yannick Maden
The following pairs received entry as alternates:
Sebastián Báez / Tomás Martín Etcheverry
Facundo Bagnis / Bernabé Zapata Miralles
Altuğ Çelikbilek / Yannick Maden
Rinky Hijikata / Christopher O'Connell
Jozef Kovalík / Sergiy Stakhovsky
Jaume Munar / Rafael Nadal
Withdrawals
Before the tournament
Romain Arneodo / Benoît Paire → replaced by Jaume Munar / Rafael Nadal
Alexander Bublik / Mackenzie McDonald → replaced by Mackenzie McDonald / Reilly Opelka
Marco Cecchinato / Andreas Seppi → replaced by Altuğ Çelikbilek / Yannick Maden
Marcus Daniell / Marcelo Demoliner → replaced by Marcus Daniell / Denis Kudla
Lloyd Harris / Alexei Popyrin → replaced by Rinky Hijikata / Christopher O'Connell
Ilya Ivashka / Andrei Vasilevski → replaced by Jozef Kovalík / Sergiy Stakhovsky
Fabrice Martin / Andreas Mies → replaced by Romain Arneodo / Andreas Mies
Adrian Mannarino / Hugo Nys → replaced by Facundo Bagnis / Bernabé Zapata Miralles
Denys Molchanov / David Vega Hernández → replaced by Ričardas Berankis / Denys Molchanov
Marc Polmans / Alexei Popyrin → replaced by Sebastián Báez / Tomás Martín Etcheverry
During the tournament
Romain Arneodo / Andreas Mies
Jaume Munar / Rafael Nadal
WTA singles main-draw entrants
Seeds
1 Rankings as of 27 December 2021
Other entrants
The following players received wildcards into the singles main draw:
Seone Mendez
Lizette Cabrera
Arina Rodionova
The following player received entry as an alternate:
Lauren Davis
The following players received entry from the qualifying draw:
Destanee Aiava
Anna Bondár
Nao Hibino
Viktória Kužmová
Lesley Pattinama Kerkhove
Zheng Qinwen
The following player received entry as a lucky loser:
Mai Hontama
Withdrawals
Before the tournament
Camila Giorgi → replaced by Lauren Davis
Camila Osorio → replaced by Mai Hontama
Emma Raducanu → replaced by Elena-Gabriela Ruse
Jil Teichmann → replaced by Maryna Zanevska
During the tournament
Naomi Osaka
WTA doubles main draw entrants
Seeds
1 Rankings as of 27 December 2021
Other entrants
The following pairs received wildcards into the doubles main draw:
Destanee Aiava / Lizette Cabrera
Gabriella Da Silva-Fick / Olivia Tjandramulia
The following pairs received entry as alternates into the doubles main draw:
Desirae Krawczyk / Christina McHale
Aliaksandra Sasnovich / Anastasija Sevastova
Withdrawals
Before the tournament
Veronika Kudermetova / Elise Mertens → replaced by Desirae Krawczyk / Christina McHale
Elixane Lechemia / Ingrid Neel → replaced by Vivian Heisen / Ingrid Neel
Wang Xinyu / Zheng Saisai → replaced by Aliaksandra Sasnovich / Anastasija Sevastova
During the tournament
Daria Kasatkina / Liudmila Samsonova
See also
2022 Melbourne Summer Set 2
References
External links
ATP tournament Official website
WTA tournament Official website
2022 ATP Tour
2022 WTA Tour
2022 in Australian tennis
January 2022 sports events in Australia
Melbourne Summer Set
|
A pasty is a British baked pastry.
Pasty or Pastie may also refer to:
Pastie, a large, round patéd pie eaten in Northern Ireland
Pasties, adhesive coverings applied to cover a person's nipples
Pasty (horse), a racehorse
Pasty Harris (born 1944), English cricketer (from Cornwall)
a pale and unhealthy appearance; pallor
an implementation of Pastebin
See also
Pate (disambiguation)
|
Street Symphony may refer to:
Street Symphony (producer), American music producer
"Street Symphony" (song), 1999 song by Monica
|
```yaml
id: Microsoft 365 Defender - Emails Indicators Hunt
version: -1
contentitemexportablefields:
contentitemfields: {}
name: Microsoft 365 Defender - Emails Indicators Hunt
description: |-
This playbook retrieves email data based on the "URLDomain", "SHA256" and "IPAddress" inputs.
SHA256 - Emails with attachments matching the "SHA256" input are retrieved.
URLDomain - If the "URLDomain" value is found as a substring of URL(s) in the body of the email, the email is retrieved.
IPAddress - Emails with "SenderIPv4"/SenderIPv6" or URLs (in the body) matching the "IPAddress" input are retrieved.
starttaskid: "0"
tasks:
"0":
id: "0"
taskid: 4e50d59c-2f80-42a2-877d-c95d20e20047
type: start
task:
id: 4e50d59c-2f80-42a2-877d-c95d20e20047
version: -1
name: ""
iscommand: false
brand: ""
description: ''
nexttasks:
'#none#':
- "1"
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 40,
"y": -120
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"1":
id: "1"
taskid: 4c477a33-925d-4946-83e3-e237dfa465a2
type: condition
task:
id: 4c477a33-925d-4946-83e3-e237dfa465a2
version: -1
name: Check if Microsoft 365 Defender is enabled
description: Returns 'yes' if integration brand is available. Otherwise returns 'no'.
scriptName: IsIntegrationAvailable
type: condition
iscommand: false
brand: ""
nexttasks:
"no":
- "4"
"yes":
- "2"
- "3"
- "13"
scriptarguments:
brandname:
simple: Microsoft 365 Defender
results:
- brandInstances
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 40,
"y": 60
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"2":
id: "2"
taskid: d7b8bb79-3028-4f41-85b4-52cd5b913070
type: condition
task:
id: d7b8bb79-3028-4f41-85b4-52cd5b913070
version: -1
name: URLDomain input exists?
type: condition
iscommand: false
brand: ""
description: ""
nexttasks:
'#default#':
- "4"
URLDomain Exists:
- "11"
separatecontext: false
conditions:
- label: URLDomain Exists
condition:
- - operator: isNotEmpty
left:
value:
complex:
root: inputs.URLDomain
iscontext: true
continueonerrortype: ""
view: |-
{
"position": {
"x": 770,
"y": 270
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"3":
id: "3"
taskid: 18e4ddb8-4e5f-4abe-8272-7c05f414081d
type: condition
task:
id: 18e4ddb8-4e5f-4abe-8272-7c05f414081d
version: -1
name: SHA256 input exists?
type: condition
iscommand: false
brand: ""
description: ""
nexttasks:
'#default#':
- "4"
SHA256 Exists:
- "12"
separatecontext: false
conditions:
- label: SHA256 Exists
condition:
- - operator: isNotEmpty
left:
value:
complex:
root: inputs.SHA256
iscontext: true
continueonerrortype: ""
view: |-
{
"position": {
"x": 310,
"y": 270
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"4":
id: "4"
taskid: de06f46a-17b8-427d-83fe-d58c41119de7
type: title
task:
id: de06f46a-17b8-427d-83fe-d58c41119de7
version: -1
name: Done
type: title
iscommand: false
brand: ""
description: ''
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 40,
"y": 1290
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"5":
id: "5"
taskid: 7ef19685-5d03-44e8-8718-bdcaa1891658
type: regular
task:
id: 7ef19685-5d03-44e8-8718-bdcaa1891658
version: -1
name: Retrieve emails data (URLDomain)
description: 'Advanced hunting is a threat-hunting tool that uses specially constructed queries to examine the past 30 days of event data in Microsoft 365 Defender. Details on how to write queries: path_to_url
script: Microsoft 365 Defender|||microsoft-365-defender-advanced-hunting
type: regular
iscommand: true
brand: Microsoft 365 Defender
nexttasks:
'#none#':
- "10"
scriptarguments:
limit:
complex:
root: inputs.ResultsLimit
query:
simple: let _start = now(-${inputs.SearchTimeframe}d); EmailEvents | where Timestamp >= _start | where RecipientEmailAddress != "${inputs.ListenerMailbox}" | join EmailUrlInfo on NetworkMessageId | where Url has_any ("${URLDomainParsed}") or UrlDomain has_any ("${URLDomainParsed}") | project-away AdditionalFields, EmailActionPolicy, UserLevelPolicy, EmailActionPolicyGuid, OrgLevelPolicy, UserLevelAction, EmailClusterId, ReportId, ReportId1, Timestamp1, OrgLevelAction, SenderObjectId, NetworkMessageId1
timeout:
simple: ${inputs.Timeout}
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 770,
"y": 720
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"6":
id: "6"
taskid: 1386f67d-0a38-438e-8a03-7435ffe62954
type: regular
task:
id: 1386f67d-0a38-438e-8a03-7435ffe62954
version: -1
name: Retrieve emails data (SHA256)
description: 'Advanced hunting is a threat-hunting tool that uses specially constructed queries to examine the past 30 days of event data in Microsoft 365 Defender. Details on how to write queries: path_to_url
script: Microsoft 365 Defender|||microsoft-365-defender-advanced-hunting
type: regular
iscommand: true
brand: Microsoft 365 Defender
nexttasks:
'#none#':
- "10"
scriptarguments:
limit:
complex:
root: inputs.ResultsLimit
query:
simple: let _start = now(-${inputs.SearchTimeframe}d); EmailEvents | where Timestamp >= _start | where RecipientEmailAddress != "${inputs.ListenerMailbox}" | join EmailAttachmentInfo on NetworkMessageId | where SHA256 has_any ("${SHA256Parsed}") | project-away AdditionalFields, EmailActionPolicy, UserLevelPolicy, EmailActionPolicyGuid, OrgLevelPolicy, UserLevelAction, EmailClusterId, ReportId, ReportId1, Timestamp1, RecipientEmailAddress1, SenderDisplayName1, NetworkMessageId1, DetectionMethods1, SenderFromAddress1, RecipientObjectId1, ThreatNames1, ThreatTypes1, SenderObjectId1, OrgLevelAction, SenderObjectId
timeout:
simple: ${inputs.Timeout}
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 310,
"y": 720
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"10":
id: "10"
taskid: 8bd4c50d-26ba-48ac-8d71-a918feb08ecb
type: regular
task:
id: 8bd4c50d-26ba-48ac-8d71-a918feb08ecb
version: -1
name: Save matching email results (with URLs or attachments)
description: |-
Set a value in context under the key you entered. If no value is entered, the script doesn't do anything.
This automation runs using the default Limited User role, unless you explicitly change the permissions.
For more information, see the section about permissions here:
path_to_url
scriptName: SetAndHandleEmpty
type: regular
iscommand: false
brand: ""
nexttasks:
'#none#':
- "16"
scriptarguments:
key:
simple: Microsoft365Defender.RetrievedEmails
value:
complex:
root: Microsoft365Defender.Hunt.results
filters:
- - operator: isNotEmpty
left:
value:
simple: Microsoft365Defender.Hunt.results.SHA256
iscontext: true
- operator: isNotEmpty
left:
value:
simple: Microsoft365Defender.Hunt.results.Url
iscontext: true
transformers:
- operator: uniq
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 1000,
"y": 920
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"11":
id: "11"
taskid: a4a3ec81-21db-43a8-885b-bdd11db950ea
type: regular
task:
id: a4a3ec81-21db-43a8-885b-bdd11db950ea
version: -1
name: Set URLDomain search argument
description: Set a value in context under the key you entered.
scriptName: Set
type: regular
iscommand: false
brand: ""
nexttasks:
'#none#':
- "5"
scriptarguments:
key:
simple: URLDomainParsed
value:
complex:
root: inputs.URLDomain
transformers:
- operator: join
args:
separator:
value:
simple: ', '
- operator: replace
args:
limit: {}
replaceWith:
value:
simple: '", "'
toReplace:
value:
simple: ', '
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 770,
"y": 510
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"12":
id: "12"
taskid: 28263a51-7d7a-4213-8d41-0d01ffdf09b2
type: regular
task:
id: 28263a51-7d7a-4213-8d41-0d01ffdf09b2
version: -1
name: Set SHA256 search argument
description: Set a value in context under the key you entered.
scriptName: Set
type: regular
iscommand: false
brand: ""
nexttasks:
'#none#':
- "6"
scriptarguments:
key:
simple: SHA256Parsed
value:
complex:
root: inputs.SHA256
transformers:
- operator: join
args:
separator:
value:
simple: ', '
- operator: replace
args:
limit: {}
replaceWith:
value:
simple: '","'
toReplace:
value:
simple: ','
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 310,
"y": 510
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"13":
id: "13"
taskid: bf851c18-5e4c-4594-8ec7-6a3b2d8bae6c
type: condition
task:
id: bf851c18-5e4c-4594-8ec7-6a3b2d8bae6c
version: -1
name: IPAddress input exists?
type: condition
iscommand: false
brand: ""
description: ""
nexttasks:
'#default#':
- "4"
IPAddress Exists:
- "14"
separatecontext: false
conditions:
- label: IPAddress Exists
condition:
- - operator: isNotEmpty
left:
value:
complex:
root: inputs.IPAddress
iscontext: true
right:
value: {}
continueonerrortype: ""
view: |-
{
"position": {
"x": 1450,
"y": 270
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"14":
id: "14"
taskid: 22329bf3-20bb-4f0a-8247-0d247a7672ad
type: regular
task:
id: 22329bf3-20bb-4f0a-8247-0d247a7672ad
version: -1
name: Set IPAddress search argument
description: Set a value in context under the key you entered.
scriptName: Set
type: regular
iscommand: false
brand: ""
nexttasks:
'#none#':
- "15"
- "18"
scriptarguments:
key:
simple: IPAddressParsed
value:
complex:
root: inputs.IPAddress
transformers:
- operator: join
args:
separator:
value:
simple: ', '
- operator: replace
args:
limit: {}
replaceWith:
value:
simple: '", "'
toReplace:
value:
simple: ', '
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 1450,
"y": 510
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"15":
id: "15"
taskid: a9b33515-9725-4e37-8188-2120f8fb69e3
type: regular
task:
id: a9b33515-9725-4e37-8188-2120f8fb69e3
version: -1
name: Retrieve emails data (IPAddress) - EmailEvents table
script: Microsoft 365 Defender|||microsoft-365-defender-advanced-hunting
description: 'This task will retrieve emails data based on the EmailEvents table. It will look for the given IP in the "SenderIPv4" and in the "SenderIPv46" properties.'
type: regular
iscommand: true
brand: Microsoft 365 Defender
nexttasks:
'#none#':
- "10"
scriptarguments:
limit:
complex:
root: inputs.ResultsLimit
query:
simple: let _start = now(-${inputs.SearchTimeframe}d); EmailEvents | where Timestamp >= _start | where RecipientEmailAddress != "${inputs.ListenerMailbox}" | where SenderIPv4 has_any ("${IPAddressParsed}") or SenderIPv6 in ("${IPAddressParsed}") | project-away AdditionalFields, EmailActionPolicy, UserLevelPolicy, EmailActionPolicyGuid, OrgLevelPolicy, UserLevelAction, EmailClusterId, ReportId, OrgLevelAction, SenderObjectId
timeout:
simple: ${inputs.Timeout}
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 1230,
"y": 720
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"16":
id: "16"
taskid: ea249095-3e24-4994-8a27-030a0e0d9da1
type: regular
task:
id: ea249095-3e24-4994-8a27-030a0e0d9da1
version: -1
name: Save matching email results (based on sender IP)
description: |-
Set a value in context under the key you entered. If no value is entered, the script doesn't do anything.
This automation runs using the default Limited User role, unless you explicitly change the permissions.
For more information, see the section about permissions here:
path_to_url
scriptName: SetAndHandleEmpty
type: regular
iscommand: false
brand: ""
nexttasks:
'#none#':
- "4"
scriptarguments:
append:
simple: "true"
key:
simple: Microsoft365Defender.RetrievedEmails
value:
complex:
root: Microsoft365Defender.Hunt.results
filters:
- - operator: notIn
left:
value:
simple: Microsoft365Defender.Hunt.results.NetworkMessageId
iscontext: true
right:
value:
simple: Microsoft365Defender.RetrievedEmails.NetworkMessageId
iscontext: true
transformers:
- operator: uniq
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 1000,
"y": 1100
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
"18":
id: "18"
taskid: 42a24672-f182-4b50-83b7-b08a6b79c4c6
type: regular
task:
id: 42a24672-f182-4b50-83b7-b08a6b79c4c6
version: -1
name: Retrieve emails data (IPAddress) - EmailUrlInfo table
description: 'This task will retrieve email data based on the EmailUrlInfo table. It will look for the given IP in the "UrlDomain" property since the IP can be part of a URL.'
script: Microsoft 365 Defender|||microsoft-365-defender-advanced-hunting
type: regular
iscommand: true
brand: Microsoft 365 Defender
nexttasks:
'#none#':
- "10"
scriptarguments:
limit:
complex:
root: inputs.ResultsLimit
query:
simple: let _start = now(-${inputs.SearchTimeframe}d); EmailEvents | where Timestamp >= _start | where RecipientEmailAddress != "${inputs.ListenerMailbox}" | join EmailUrlInfo on NetworkMessageId | where UrlDomain has_any ("${IPAddressParsed}") | project-away AdditionalFields, EmailActionPolicy, UserLevelPolicy, EmailActionPolicyGuid, OrgLevelPolicy, UserLevelAction, EmailClusterId, ReportId, ReportId1, Timestamp1, OrgLevelAction, SenderObjectId, NetworkMessageId1
timeout:
simple: ${inputs.Timeout}
separatecontext: false
continueonerrortype: ""
view: |-
{
"position": {
"x": 1670,
"y": 720
}
}
note: false
timertriggers: []
ignoreworker: false
skipunavailable: false
quietmode: 0
isoversize: false
isautoswitchedtoquietmode: false
system: true
view: |-
{
"linkLabelsPosition": {
"13_4_#default#": 0.16,
"1_13_yes": 0.88,
"1_2_yes": 0.89,
"1_4_no": 0.12,
"2_4_#default#": 0.12,
"3_4_#default#": 0.15
},
"paper": {
"dimensions": {
"height": 1475,
"width": 2010,
"x": 40,
"y": -120
}
}
}
inputs:
- key: URLDomain
value: {}
required: false
description: Domain or URL to search within emails. Can be a single domain or URL or an array of domains or URLs to search. The search looks for the exact Domain or URL.
playbookInputQuery:
- key: SHA256
value: {}
required: false
description: The SHA256 hash file or an array of hashes to search within emails.
playbookInputQuery:
- key: IPAddress
value: {}
required: false
description: The source or destination IP address to search. Can be a single address or an array of IP addresses.
playbookInputQuery:
- key: Timeout
value:
simple: "180"
required: false
description: The time limit in seconds for the HTTP request to run. Default is 180.
playbookInputQuery:
- key: SearchTimeframe
value:
simple: "7"
required: false
description: Number of days past to search. Default is 7.
playbookInputQuery:
- key: ResultsLimit
value:
simple: "50"
required: false
description: Number of retrieved entries. Enter -1 for unlimited query. 50 is the default.
playbookInputQuery:
- key: ListenerMailbox
value: {}
required: false
description: The mailbox of the listening integration. In case it is provided, the emails found in it will be ignored.
playbookInputQuery:
outputs:
- contextPath: Microsoft365Defender.RetrievedEmails
description: Email objects containing relevant fields.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.InternetMessageId
description: Internet Message ID of the email.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SenderFromDomain
description: Sender domain.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.EmailDirection
description: Email direction (inbound/outbound).
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.DeliveryLocation
description: Delivery location.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.AuthenticationDetails
description: Authentication details (SPF, DKIM, DMARC, CompAuth).
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.DeliveryAction
description: Delivery action.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.Subject
description: Email subject.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.AttachmentCount
description: Number of attachments.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.ThreatNames
description: Threat names.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.RecipientEmailAddress
description: Recipient email address.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.EmailAction
description: Email action.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.EmailLanguage
description: Email language.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SenderFromAddress
description: Sender address.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.Timestamp
description: Timestamp.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SenderDisplayName
description: Sender display name.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SenderIPv4
description: Sender IPv4
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.ConfidenceLevel
description: Confidence level.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.ThreatTypes
description: Threat types.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SHA256
description: SHA256 of the attachments (if exists in the email).
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.Url
description: URLs found in the email's body.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.UrlCount
description: Number of URLs found in the email's body.
type: string
- contextPath: Microsoft365Defender.RetrievedEmails.SenderIPv6
description: Sender IPv6.
type: unknown
tests:
- No tests (auto formatted)
fromversion: 6.2.0
```
|
```c
/* packet-dcerpc-ubikvote.c
*
* Routines for DCE DFS Ubik Voting routines.
* This information is based off the released idl files from opengroup.
* ftp://ftp.opengroup.org/pub/dce122/dce/src/file.tar.gz file/ncsubik/ubikvote_proc.idl
*
* Wireshark - Network traffic analyzer
* By Gerald Combs <gerald@wireshark.org>
*
* This program is free software; you can redistribute it and/or
* as published by the Free Software Foundation; either version 2
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
*
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
*/
#include "config.h"
#include <epan/packet.h>
#include "packet-dcerpc.h"
void proto_register_ubikvote (void);
void proto_reg_handoff_ubikvote (void);
static int proto_ubikvote = -1;
static int hf_ubikvote_opnum = -1;
static gint ett_ubikvote = -1;
static e_guid_t uuid_ubikvote = { 0x4d37f2dd, 0xed43, 0x0003, { 0x02, 0xc0, 0x37, 0xcf, 0x1e, 0x00, 0x00, 0x00 } };
static guint16 ver_ubikvote = 4;
static dcerpc_sub_dissector ubikvote_dissectors[] = {
{ 0, "Beacon", NULL, NULL},
{ 1, "Debug", NULL, NULL},
{ 2, "SDebug", NULL, NULL},
{ 3, "GetServerInterfaces", NULL, NULL},
{ 4, "GetSyncSite", NULL, NULL},
{ 5, "DebugV2", NULL, NULL},
{ 6, "SDebugV2", NULL, NULL},
{ 7, "GetSyncSiteIdentity", NULL, NULL},
{ 0, NULL, NULL, NULL }
};
void
proto_register_ubikvote (void)
{
static hf_register_info hf[] = {
{ &hf_ubikvote_opnum,
{ "Operation", "ubikvote.opnum", FT_UINT16, BASE_DEC,
NULL, 0x0, NULL, HFILL }}
};
static gint *ett[] = {
&ett_ubikvote,
};
proto_ubikvote = proto_register_protocol ("DCE DFS FLDB UBIKVOTE", "UBIKVOTE", "ubikvote");
proto_register_field_array (proto_ubikvote, hf, array_length (hf));
proto_register_subtree_array (ett, array_length (ett));
}
void
proto_reg_handoff_ubikvote (void)
{
/* Register the protocol as dcerpc */
dcerpc_init_uuid (proto_ubikvote, ett_ubikvote, &uuid_ubikvote, ver_ubikvote, ubikvote_dissectors, hf_ubikvote_opnum);
}
/*
* Editor modelines - path_to_url
*
* Local variables:
* c-basic-offset: 8
* tab-width: 8
* indent-tabs-mode: t
* End:
*
* vi: set shiftwidth=8 tabstop=8 noexpandtab:
* :indentSize=8:tabSize=8:noTabs=false:
*/
```
|
```html
@if(registrarService.inNewRegistrarMode()) {
<app-new-registrar />
} @else {
<div class="console-app__registrars">
<div class="console-app__registrars-header">
<h1 class="mat-headline-4">Registrars</h1>
<button
mat-flat-button
color="primary"
(click)="openNewRegistrar()"
aria-label="Add new registrar"
>
<mat-icon>add</mat-icon>
Add new registrar
</button>
</div>
<div class="console-app__scrollable-wrapper">
<div class="console-app__scrollable">
<mat-form-field class="console-app__registrars-filter">
<mat-label>Search</mat-label>
<input
matInput
(keyup)="applyFilter($event)"
placeholder="..."
type="search"
/>
<mat-icon matPrefix>search</mat-icon>
</mat-form-field>
<mat-table
[dataSource]="dataSource"
class="mat-elevation-z0"
class="console-app__registrars-table"
matSort
>
<ng-container
*ngFor="let column of columns"
[matColumnDef]="column.columnDef"
>
<mat-header-cell *matHeaderCellDef>
{{ column.header }}
</mat-header-cell>
<mat-cell
*matCellDef="let row"
[innerHTML]="column.cell(row)"
></mat-cell>
</ng-container>
<mat-header-row *matHeaderRowDef="displayedColumns"></mat-header-row>
<mat-row
*matRowDef="let row; columns: displayedColumns"
(click)="openDetails(row.registrarId)"
></mat-row>
</mat-table>
<mat-paginator
class="mat-elevation-z0"
[pageSizeOptions]="[5, 10, 20]"
showFirstLastButtons
></mat-paginator>
</div>
</div>
</div>
}
```
|
```go
//
// path_to_url
//
// Unless required by applicable law or agreed to in writing, software
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
package model
import (
"encoding/json"
"fmt"
"sort"
"strconv"
"strings"
)
var (
// ZeroSample is the pseudo zero-value of Sample used to signal a
// non-existing sample. It is a Sample with timestamp Earliest, value 0.0,
// and metric nil. Note that the natural zero value of Sample has a timestamp
// of 0, which is possible to appear in a real Sample and thus not suitable
// to signal a non-existing Sample.
ZeroSample = Sample{Timestamp: Earliest}
)
// Sample is a sample pair associated with a metric. A single sample must either
// define Value or Histogram but not both. Histogram == nil implies the Value
// field is used, otherwise it should be ignored.
type Sample struct {
Metric Metric `json:"metric"`
Value SampleValue `json:"value"`
Timestamp Time `json:"timestamp"`
Histogram *SampleHistogram `json:"histogram"`
}
// Equal compares first the metrics, then the timestamp, then the value. The
// semantics of value equality is defined by SampleValue.Equal.
func (s *Sample) Equal(o *Sample) bool {
if s == o {
return true
}
if !s.Metric.Equal(o.Metric) {
return false
}
if !s.Timestamp.Equal(o.Timestamp) {
return false
}
if s.Histogram != nil {
return s.Histogram.Equal(o.Histogram)
}
return s.Value.Equal(o.Value)
}
func (s Sample) String() string {
if s.Histogram != nil {
return fmt.Sprintf("%s => %s", s.Metric, SampleHistogramPair{
Timestamp: s.Timestamp,
Histogram: s.Histogram,
})
}
return fmt.Sprintf("%s => %s", s.Metric, SamplePair{
Timestamp: s.Timestamp,
Value: s.Value,
})
}
// MarshalJSON implements json.Marshaler.
func (s Sample) MarshalJSON() ([]byte, error) {
if s.Histogram != nil {
v := struct {
Metric Metric `json:"metric"`
Histogram SampleHistogramPair `json:"histogram"`
}{
Metric: s.Metric,
Histogram: SampleHistogramPair{
Timestamp: s.Timestamp,
Histogram: s.Histogram,
},
}
return json.Marshal(&v)
}
v := struct {
Metric Metric `json:"metric"`
Value SamplePair `json:"value"`
}{
Metric: s.Metric,
Value: SamplePair{
Timestamp: s.Timestamp,
Value: s.Value,
},
}
return json.Marshal(&v)
}
// UnmarshalJSON implements json.Unmarshaler.
func (s *Sample) UnmarshalJSON(b []byte) error {
v := struct {
Metric Metric `json:"metric"`
Value SamplePair `json:"value"`
Histogram SampleHistogramPair `json:"histogram"`
}{
Metric: s.Metric,
Value: SamplePair{
Timestamp: s.Timestamp,
Value: s.Value,
},
Histogram: SampleHistogramPair{
Timestamp: s.Timestamp,
Histogram: s.Histogram,
},
}
if err := json.Unmarshal(b, &v); err != nil {
return err
}
s.Metric = v.Metric
if v.Histogram.Histogram != nil {
s.Timestamp = v.Histogram.Timestamp
s.Histogram = v.Histogram.Histogram
} else {
s.Timestamp = v.Value.Timestamp
s.Value = v.Value.Value
}
return nil
}
// Samples is a sortable Sample slice. It implements sort.Interface.
type Samples []*Sample
func (s Samples) Len() int {
return len(s)
}
// Less compares first the metrics, then the timestamp.
func (s Samples) Less(i, j int) bool {
switch {
case s[i].Metric.Before(s[j].Metric):
return true
case s[j].Metric.Before(s[i].Metric):
return false
case s[i].Timestamp.Before(s[j].Timestamp):
return true
default:
return false
}
}
func (s Samples) Swap(i, j int) {
s[i], s[j] = s[j], s[i]
}
// Equal compares two sets of samples and returns true if they are equal.
func (s Samples) Equal(o Samples) bool {
if len(s) != len(o) {
return false
}
for i, sample := range s {
if !sample.Equal(o[i]) {
return false
}
}
return true
}
// SampleStream is a stream of Values belonging to an attached COWMetric.
type SampleStream struct {
Metric Metric `json:"metric"`
Values []SamplePair `json:"values"`
Histograms []SampleHistogramPair `json:"histograms"`
}
func (ss SampleStream) String() string {
valuesLength := len(ss.Values)
vals := make([]string, valuesLength+len(ss.Histograms))
for i, v := range ss.Values {
vals[i] = v.String()
}
for i, v := range ss.Histograms {
vals[i+valuesLength] = v.String()
}
return fmt.Sprintf("%s =>\n%s", ss.Metric, strings.Join(vals, "\n"))
}
func (ss SampleStream) MarshalJSON() ([]byte, error) {
if len(ss.Histograms) > 0 && len(ss.Values) > 0 {
v := struct {
Metric Metric `json:"metric"`
Values []SamplePair `json:"values"`
Histograms []SampleHistogramPair `json:"histograms"`
}{
Metric: ss.Metric,
Values: ss.Values,
Histograms: ss.Histograms,
}
return json.Marshal(&v)
} else if len(ss.Histograms) > 0 {
v := struct {
Metric Metric `json:"metric"`
Histograms []SampleHistogramPair `json:"histograms"`
}{
Metric: ss.Metric,
Histograms: ss.Histograms,
}
return json.Marshal(&v)
} else {
v := struct {
Metric Metric `json:"metric"`
Values []SamplePair `json:"values"`
}{
Metric: ss.Metric,
Values: ss.Values,
}
return json.Marshal(&v)
}
}
func (ss *SampleStream) UnmarshalJSON(b []byte) error {
v := struct {
Metric Metric `json:"metric"`
Values []SamplePair `json:"values"`
Histograms []SampleHistogramPair `json:"histograms"`
}{
Metric: ss.Metric,
Values: ss.Values,
Histograms: ss.Histograms,
}
if err := json.Unmarshal(b, &v); err != nil {
return err
}
ss.Metric = v.Metric
ss.Values = v.Values
ss.Histograms = v.Histograms
return nil
}
// Scalar is a scalar value evaluated at the set timestamp.
type Scalar struct {
Value SampleValue `json:"value"`
Timestamp Time `json:"timestamp"`
}
func (s Scalar) String() string {
return fmt.Sprintf("scalar: %v @[%v]", s.Value, s.Timestamp)
}
// MarshalJSON implements json.Marshaler.
func (s Scalar) MarshalJSON() ([]byte, error) {
v := strconv.FormatFloat(float64(s.Value), 'f', -1, 64)
return json.Marshal([...]interface{}{s.Timestamp, string(v)})
}
// UnmarshalJSON implements json.Unmarshaler.
func (s *Scalar) UnmarshalJSON(b []byte) error {
var f string
v := [...]interface{}{&s.Timestamp, &f}
if err := json.Unmarshal(b, &v); err != nil {
return err
}
value, err := strconv.ParseFloat(f, 64)
if err != nil {
return fmt.Errorf("error parsing sample value: %s", err)
}
s.Value = SampleValue(value)
return nil
}
// String is a string value evaluated at the set timestamp.
type String struct {
Value string `json:"value"`
Timestamp Time `json:"timestamp"`
}
func (s *String) String() string {
return s.Value
}
// MarshalJSON implements json.Marshaler.
func (s String) MarshalJSON() ([]byte, error) {
return json.Marshal([]interface{}{s.Timestamp, s.Value})
}
// UnmarshalJSON implements json.Unmarshaler.
func (s *String) UnmarshalJSON(b []byte) error {
v := [...]interface{}{&s.Timestamp, &s.Value}
return json.Unmarshal(b, &v)
}
// Vector is basically only an alias for Samples, but the
// contract is that in a Vector, all Samples have the same timestamp.
type Vector []*Sample
func (vec Vector) String() string {
entries := make([]string, len(vec))
for i, s := range vec {
entries[i] = s.String()
}
return strings.Join(entries, "\n")
}
func (vec Vector) Len() int { return len(vec) }
func (vec Vector) Swap(i, j int) { vec[i], vec[j] = vec[j], vec[i] }
// Less compares first the metrics, then the timestamp.
func (vec Vector) Less(i, j int) bool {
switch {
case vec[i].Metric.Before(vec[j].Metric):
return true
case vec[j].Metric.Before(vec[i].Metric):
return false
case vec[i].Timestamp.Before(vec[j].Timestamp):
return true
default:
return false
}
}
// Equal compares two sets of samples and returns true if they are equal.
func (vec Vector) Equal(o Vector) bool {
if len(vec) != len(o) {
return false
}
for i, sample := range vec {
if !sample.Equal(o[i]) {
return false
}
}
return true
}
// Matrix is a list of time series.
type Matrix []*SampleStream
func (m Matrix) Len() int { return len(m) }
func (m Matrix) Less(i, j int) bool { return m[i].Metric.Before(m[j].Metric) }
func (m Matrix) Swap(i, j int) { m[i], m[j] = m[j], m[i] }
func (mat Matrix) String() string {
matCp := make(Matrix, len(mat))
copy(matCp, mat)
sort.Sort(matCp)
strs := make([]string, len(matCp))
for i, ss := range matCp {
strs[i] = ss.String()
}
return strings.Join(strs, "\n")
}
```
|
```python
# This file is part of rdiff-backup.
#
# rdiff-backup is free software; you can redistribute it and/or modify
# option) any later version.
#
# rdiff-backup is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
#
# along with rdiff-backup; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301, USA
"""Manage logging, displaying and recording messages with required verbosity"""
import datetime
import os # needed to grab verbosity as environment variable
import re
import shutil
import sys
import textwrap
import typing
import traceback
from rdiff_backup import Globals
LOGFILE_ENCODING = "utf-8"
# type definitions
Verbosity = typing.Literal[0, 1, 2, 3, 4, 5, 6, 7, 8, 9] # : typing.TypeAlias
InputVerbosity = typing.Union[int, str] # : typing.TypeAlias
# we need to define constants
NONE: Verbosity = 0 # are always output as-is on stdout
ERROR: Verbosity = 1
WARNING: Verbosity = 2
NOTE: Verbosity = 3
INFO: Verbosity = 5
DEBUG: Verbosity = 8
TIMESTAMP: Verbosity = 9 # for adding the timestamp
# mapping from severity to prefix (must be less than 9 characters)
_LOG_PREFIX: dict[Verbosity, str] = {
NONE: "",
ERROR: "ERROR:",
WARNING: "WARNING:",
NOTE: "NOTE:",
INFO: "*",
DEBUG: "DEBUG:",
}
class LoggerError(Exception):
pass
class Logger:
"""All functions which deal with logging"""
def __init__(self):
self.log_file_open = None
self.log_file_local = None
# if something wrong happens during initialization, we want to know
self.file_verbosity: Verbosity = NONE
self.term_verbosity: Verbosity = WARNING
def __call__(self, message, verbosity):
"""
Log message that has verbosity importance
message can be a string or bytes
"""
if verbosity > self.file_verbosity and verbosity > self.term_verbosity:
return
if not isinstance(message, (bytes, str)):
raise TypeError(
"You can only log bytes or str, and not {lt}".format(lt=type(message))
)
if verbosity <= self.file_verbosity:
self.log_to_file(message, verbosity)
if verbosity <= self.term_verbosity:
self.log_to_term(message, verbosity)
# @API(Log.log_to_file, 200)
def log_to_file(self, message, verbosity=None):
"""Write the message to the log file, if possible"""
if self.log_file_open:
if self.log_file_local:
tmpstr = self._format(message, self.file_verbosity, verbosity)
self.logfp.write(_to_bytes(tmpstr))
self.logfp.flush()
else:
self.log_file_conn.log.Log.log_to_file(message, verbosity)
def log_to_term(self, message, verbosity):
"""Write message to stdout/stderr"""
if verbosity in {ERROR, WARNING} or Globals.server:
termfp = sys.stderr
else:
termfp = sys.stdout
tmpstr = self._format(message, self.term_verbosity, verbosity)
# if the verbosity is below 9 and the string isn't deemed
# pre-formatted by newlines (we ignore the last character)
if self.file_verbosity <= DEBUG and "\n" not in tmpstr[:-1]:
termfp.write(
textwrap.fill(
tmpstr,
subsequent_indent=" " * 9,
break_long_words=False,
break_on_hyphens=False,
width=shutil.get_terminal_size().columns - 1,
)
+ "\n"
)
else:
termfp.write(tmpstr)
def conn(self, direction, result, req_num):
"""Log some data on the connection
The main worry with this function is that something in here
will create more network traffic, which will spiral to
infinite regress. So, for instance, logging must only be done
to the terminal, because otherwise the log file may be remote.
"""
if self.term_verbosity <= DEBUG:
return
if type(result) is bytes:
result_repr = repr(result)
else:
result_repr = str(result)
# shorten the result to a max size of 720 chars with ellipsis if needed
# result_repr = result_repr[:720] + (result_repr[720:] and '[...]') # noqa: E265
if Globals.server:
conn_str = "Server"
else:
conn_str = "Client"
self.log_to_term(
"{cs} {di} ({rn}): {rr}".format(
cs=conn_str, di=direction, rn=req_num, rr=result_repr
),
DEBUG,
)
def FatalError(self, message, return_code=1):
"""Log a fatal error and exit"""
self.log_to_term("Fatal Error: {em}".format(em=message), ERROR)
sys.exit(return_code)
def exception(self, only_terminal=0, verbosity=INFO):
"""Log an exception and traceback
If only_terminal is zero, log normally.
If it is 1, then only log to disk if log file is local
If it is 2, don't log to disk at all.
"""
assert only_terminal in (
0,
1,
2,
), "Variable only_terminal '{ot}' must be one of [012]".format(ot=only_terminal)
if only_terminal == 0 or (only_terminal == 1 and self.log_file_open):
logging_func = self.__call__
else:
logging_func = self.log_to_term
if verbosity >= self.term_verbosity:
return
exception_string = self._exception_to_string()
try:
logging_func(exception_string, verbosity)
except OSError:
print("OS error while trying to log exception!")
print(exception_string)
# @API(Log.set_verbosity, 300)
def set_verbosity(
self,
file_verbosity: InputVerbosity,
term_verbosity: typing.Union[InputVerbosity, None] = None,
) -> int:
"""
Set verbosity levels, logfile and terminal. Takes numbers or strings.
The function makes sure that verbosities are only modified if both
input values are correct.
If not provided, the terminal verbosity is set from the logfile one.
Returns an integer code.
"""
try:
# we set a temporary verbosity to make sure we overwrite the
# actual one only if both values are correct
tmp_verbosity: Verbosity = self.validate_verbosity(file_verbosity)
if term_verbosity is None:
self.term_verbosity = tmp_verbosity
else:
self.term_verbosity = self.validate_verbosity(term_verbosity)
except ValueError:
return Globals.RET_CODE_ERR
else:
self.file_verbosity = tmp_verbosity
return Globals.RET_CODE_OK
def open_logfile(self, log_rp):
"""Inform all connections of an open logfile.
log_rp.conn will write to the file, and the others will pass
write commands off to it.
"""
assert not self.log_file_open, "Can't open an already opened logfile"
log_rp.conn.log.Log.open_logfile_local(log_rp)
for conn in Globals.connections:
conn.log.Log.open_logfile_allconn(log_rp.conn)
# @API(Log.open_logfile_allconn, 200)
def open_logfile_allconn(self, log_file_conn):
"""Run on all connections to signal log file is open"""
self.log_file_open = 1
self.log_file_conn = log_file_conn
# @API(Log.open_logfile_local, 200)
def open_logfile_local(self, log_rp):
"""Open logfile locally - should only be run on one connection"""
assert (
log_rp.conn is Globals.local_connection
), "Action only foreseen locally and not over {conn}".format(conn=log_rp.conn)
try:
self.logfp = log_rp.open("ab")
except OSError as exc:
raise LoggerError(
"Unable to open logfile {lf} due to "
"exception '{ex}'".format(lf=log_rp, ex=exc)
)
self.log_file_local = 1
def close_logfile(self):
"""Close logfile and inform all connections"""
if self.log_file_open:
for conn in Globals.connections:
conn.log.Log.close_logfile_allconn()
self.log_file_conn.log.Log.close_logfile_local()
# @API(Log.close_logfile_allconn, 200)
def close_logfile_allconn(self):
"""Run on every connection"""
self.log_file_open = None
# @API(Log.close_logfile_local, 200)
def close_logfile_local(self):
"""Run by logging connection - close logfile"""
assert (
self.log_file_conn is Globals.local_connection
), "Action only foreseen locally and not over {lc}".format(
lc=self.log_file_conn
)
self.logfp.close()
self.log_file_local = None
def _exception_to_string(self):
"""Return string version of current exception"""
type, value, tb = sys.exc_info()
s = "Exception '%s' raised of class '%s':\n%s" % (
value,
type,
"".join(traceback.format_tb(tb)),
)
return s
def _format(self, message, verbosity, msg_verbosity):
"""Format the message, possibly adding date information"""
if verbosity <= DEBUG:
# pre-formatted informative messages are returned as such
if msg_verbosity in {NONE, INFO} and "\n" in message[:-1]:
return "{msg}\n".format(msg=message)
else:
return "{pre:<9}{msg}\n".format(
pre=_LOG_PREFIX[msg_verbosity], msg=message
)
else:
timestamp = (
datetime.datetime.now(datetime.timezone.utc)
.astimezone()
.strftime("%F %H:%M:%S.%f %z")
)
if Globals.server:
role = "SERVER"
else:
role = "CLIENT"
return "{time} <{role}-{pid}> {pre} {msg}\n".format(
time=timestamp,
role=role,
pid=os.getpid(),
pre=_LOG_PREFIX[msg_verbosity],
msg=message,
)
@classmethod
def validate_verbosity(cls, input_verbosity: InputVerbosity) -> Verbosity:
"""
Validate verbosity and returns its value as integer.
The input value can be a string or an integer, between 0 and 9.
Any wrong value raises a ValueError exception.
"""
try:
verbosity = int(input_verbosity)
except ValueError:
Log(
"Verbosity must be a number, received '{vb}' "
"instead".format(vb=input_verbosity),
ERROR,
)
raise ValueError
if verbosity in typing.get_args(Verbosity):
return typing.cast(Verbosity, verbosity)
else:
Log(
"Verbosity must be between 0 and 9, received '{vb}' "
"instead".format(vb=verbosity),
ERROR,
)
raise ValueError
Log = Logger()
class ErrorLog:
"""
Log each recoverable error in error_log file
There are three types of recoverable errors: ListError, which
happens trying to list a directory or stat a file, UpdateError,
which happen when trying to update a changed file, and
SpecialFileError, which happen when a special file cannot be
created. See the error policy file for more info.
"""
_log_fileobj = None
@classmethod
def open(cls, data_dir, time_string, compress=True):
"""Open the error log, prepare for writing"""
assert not cls._log_fileobj, "Log already open, can't be reopened"
base_rp = data_dir.append("error_log.%s.data" % time_string)
if compress: # FIXME extract MaybeGzip from rpath and make it utils?
from rdiff_backup import rpath
cls._log_fileobj = rpath.MaybeGzip(base_rp)
else:
cls._log_fileobj = base_rp.open("wb", compress=0)
@classmethod
# @API(ErrorLog.isopen, 200)
def isopen(cls):
"""True if the error log file is currently open"""
if Globals.isbackup_writer or not Globals.backup_writer:
return cls._log_fileobj is not None
else:
return Globals.backup_writer.log.ErrorLog.isopen()
@classmethod
# @API(ErrorLog.write_if_open, 200)
def write_if_open(cls, error_type, rp, exc):
"""Call cls._write(...) if error log open, only log otherwise"""
if not Globals.isbackup_writer and Globals.backup_writer:
return Globals.backup_writer.log.ErrorLog.write_if_open(error_type, rp, exc)
if cls.isopen():
cls._write(error_type, rp, exc)
else:
Log(cls._get_log_string(error_type, rp, exc), WARNING)
@classmethod
def close(cls):
"""Close the error log file"""
if cls.isopen():
cls._log_fileobj.close()
cls._log_fileobj = None
@classmethod
def _get_log_string(cls, error_type, rp, exc):
"""Return log string to put in error log"""
assert (
error_type == "ListError"
or error_type == "UpdateError"
or error_type == "SpecialFileError"
), "Unknown error type {et}".format(et=error_type)
return "{et}: '{rp}' {ex}".format(et=error_type, rp=rp, ex=exc)
@classmethod
def _write(cls, error_type, rp, exc):
"""Add line to log file indicating error exc with file rp"""
logstr = cls._get_log_string(error_type, rp, exc)
Log(logstr, WARNING)
if Globals.null_separator:
logstr += "\0"
else:
logstr = re.sub("\n", " ", logstr)
logstr += "\n"
cls._log_fileobj.write(_to_bytes(logstr))
def _to_bytes(logline, encoding=LOGFILE_ENCODING):
"""
Convert string into bytes for logging into file.
"""
assert logline, "There must be a text to encode"
assert isinstance(logline, str), "Text to encode must be str and not {lt}".format(
lt=type(logline)
)
return logline.encode(encoding, "backslashreplace")
```
|
Lu Yen-hsun was the defending champion, and defended his title defeating Luca Vanni 6–7(7–9), 6–4, 6–4 in the final.
Seeds
Lu Yen-hsun (champion)
Blaž Kavčič (withdrew due to exhaustion)
Go Soeda (quarterfinals)
Yūichi Sugita (semifinals)
Alexander Kudryavtsev (semifinals, retired)
Marco Chiudinelli (quarterfinals)
Thomas Fabbiano (quarterfinals)
Denys Molchanov (quarterfinals)
Draw
Finals
Top half
Bottom half
References
Main Draw
Qualifying Draw
OEC Kaohsiung - Singles
2014 Singles
2014 in Taiwanese tennis
|
Martin Hofmann (born 31 March 1978) is a Czech actor.
Selected filmography
Film
Pleasant Moments (2006)
A Prominent Patient (2016)
The Affair (2019)
Women on the Run (2019)
Shotgun Justice (2019)
Havel (2020)
Bet on Friendship (2021)
Borders of Love (2022)
Střídavka (2022)
Television
Ulice (2005)
Terapie (2011)
Trpaslík (2017)
Rédl (2018)
Most! (2019)
The Sleepers (2019)
Odznak Vysočina (2022)
References
External links
1978 births
Living people
People from České Budějovice
Czech male film actors
Czech male stage actors
Academy of Performing Arts in Prague alumni
Czech male television actors
21st-century Czech male actors
|
Zulema Fátima Yoma (born 18 December 1942) was the First Lady of Argentina from 1989 until 1991, when she divorced President Carlos Menem.
Biography
A native of Nonogasta in La Rioja Province, Yoma was married for twenty-five years (1966–91) to Carlos Saúl Menem, who served as President of Argentina from July 1989 to December 1999. Her parents were Syrian Muslims, as Menem's were. They had two children, a son, Carlos Saúl Facundo Menem Yoma, who died in a helicopter crash in 1995, and a daughter, Zulemita, who, starting in 1991, fulfilled the role of First Lady at formal occasions for the remaining eight years of her father's presidency.
Controversies
Her son, Carlos Saúl Facundo Menem, died in a helicopter crash in 1995. Although it was ruled an accident, there are conspiracy theories that say he was actually murdered, of which Zulema Yoma and her daughter Zulemita are convinced. Yoma felt her son's death was politically motivated and even stated that the corpse buried in the tomb of the Islamic cemetery is not that of his son, requesting an exhumation.
Her son had an illegitimate daughter named Antonella Pinetta, who was born in 1988. Yoma did not recognize her as her granddaughter and refused to have her DNA examined, requesting that the test should be made from the body of her son, to support her request for the exhumation of the body. While President Menem agreed to participate in the DNA testing, he refused to proceed with the exhumation.
References
1942 births
Living people
Argentine Muslims
First ladies and gentlemen of Argentina
Argentine people of Syrian descent
Argentine people of Arab descent
People from La Rioja Province, Argentina
Menem family
|
```objective-c
This program is free software; you can redistribute it and/or modify
the Free Software Foundation
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
with this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. */
#include <io.h>
```
|
Red Barricades: ASL Historical Module 1 is a board game published in 1990 by Avalon Hill.
Contents
Red Barricades is a game in which an area of Stalingrad is depicted to scale.
Reception
Mike Siggins reviewed Red Barricades for Games International magazine, and gave it a rating of 9 out of 10, and stated that "For non-ASL players this one won't be on the purchase list. For the fans, it's a must buy."
The Chicago Tribune said that "The game includes more than 400 additional counters, a new rules chapter and several campaign games. The Advanced Squad Leader game system requires a serious commitment of time and money and is for specialists only."
Reviews
Casus Belli #57
References
Advanced Squad Leader
Avalon Hill games
Board games introduced in 1990
|
The Mitsunobu reaction is an organic reaction that converts an alcohol into a variety of functional groups, such as an ester, using triphenylphosphine and an azodicarboxylate such as diethyl azodicarboxylate (DEAD) or diisopropyl azodicarboxylate (DIAD). Although DEAD and DIAD are most commonly used, there are a variety of other azodicarboxylates available which facilitate an easier workup and/or purification and in some cases, facilitate the use of more basic nucleophiles. It was discovered by Oyo Mitsunobu (1934–2003). In a typical protocol, one dissolves the alcohol, the carboxylic acid, and triphenylphosphine in tetrahydrofuran or other suitable solvent (e.g. diethyl ether), cool to 0 °C using an ice-bath, slowly add the DEAD dissolved in THF, then stir at room temperature for several hours. The alcohol reacts with the phosphine to create a good leaving group then undergoes an inversion of stereochemistry in classic SN2 fashion as the nucleophile displaces it. A common side-product is produced when the azodicarboxylate displaces the leaving group instead of the desired nucleophile. This happens if the nucleophile is not acidic enough (pKa larger than 13) or is not nucleophilic enough due to steric or electronic constraints. A variation of this reaction utilizing a nitrogen nucleophile is known as a Fukuyama–Mitsunobu.
Several reviews have been published.
Reaction mechanism
The reaction mechanism of the Mitsunobu reaction is fairly complex. The identity of intermediates and the roles they play has been the subject of debate.
Initially, the triphenyl phosphine (2) makes a nucleophilic attack upon diethyl azodicarboxylate (1) producing a betaine intermediate 3, which deprotonates the carboxylic acid (4) to form the ion pair 5. The formation of the ion pair 5 is very fast.
The second phase of the mechanism is proposed to be phosphorus-centered, the DEAD having been converted to the hydrazine. The ratio and interconversion of intermediates 8–11 depend on the carboxylic acid pKa and the solvent polarity. Although several phosphorus intermediates are present, the attack of the carboxylate anion upon intermediate 8 is the only productive pathway forming the desired product 12 and triphenylphosphine oxide (13).
The formation of the oxyphosphonium intermediate 8 is slow and facilitated by the alkoxide. Therefore, the overall rate of reaction is controlled by carboxylate basicity and solvation.
Order of addition of reagents
The order of addition of the reagents of the Mitsunobu reaction can be important. Typically, one dissolves the alcohol, the carboxylic acid, and triphenylphosphine in tetrahydrofuran or other suitable solvent (e.g. diethyl ether), cool to 0 °C using an ice-bath, slowly add the DEAD dissolved in THF, then stir at room temperature for several hours. If this is unsuccessful, then preforming the betaine may give better results. To preform the betaine, add DEAD to triphenylphosphine in tetrahydrofuran at 0 °C, followed by the addition of the alcohol and finally the acid.
Variations
Other nucleophilic functional groups
Many other functional groups can serve as nucleophiles besides carboxylic acids. For the reaction to be successful, the nucleophile must have a pKa less than 15.
Modifications
Several modifications to the original reagent combination have been developed in order to simplify the separation of the product and avoid production of so much chemical waste. One variation of the Mitsunobu reaction uses resin-bound triphenylphosphine and uses di-tert-butylazodicarboxylate instead of DEAD. The oxidized triphenylphosphine resin can be removed by filtration, and the di-tert-butylazodicarboxylate byproduct is removed by treatment with trifluoroacetic acid. Bruce H. Lipshutz has developed an alternative to DEAD, di-(4-chlorobenzyl)azodicarboxylate (DCAD) where the hydrazine by-product can be easily removed by filtration and recycled back to DCAD.
A modification has also been reported in which DEAD can be used in catalytic versus stoichiometric quantities, however this procedure requires the use of stoichiometric (diacetoxyiodo)benzene to oxidise the hydrazine by-product back to DEAD.
Denton and co-workers have reported a redox-neutral variant of the Mitsunobu reaction which employs a phosphorus(III) catalyst to activate the substrate, ensuring inversion in the nucleophilic attack, and uses a Dean-Stark trap to remove the water by-product.
Phosphorane reagents
Tsunoda et al. have shown that one can combine the triphenylphosphine and the diethyl azodicarboxylate into one reagent: a phosphorane ylide. Both (cyanomethylene)trimethylphosphorane (CMMP, R = Me) and (cyanomethylene)tributylphosphorane (CMBP, R = Bu) have proven particularly effective.
The ylide acts as both the reducing agent and the base. The byproducts are acetonitrile (6) and the trialkylphosphine oxide (8).
Uses
The Mitsunobu reaction has been applied in the synthesis of aryl ethers:
With these particular reactants the conversion with DEAD fails because the hydroxyl group is only weakly acidic. Instead the related 1,1'-(azodicarbonyl)dipiperidine (ADDP) is used of which the betaine intermediate is a stronger base. The phosphine is a polymer-supported triphenylphosphine (PS-PPh3).
The reaction has been used to synthesize quinine, colchicine, sarain, morphine, stigmatellin, eudistomin, oseltamivir, strychnine, and nupharamine.
References
See also
Appel reaction
Substitution reactions
Name reactions
|
```c++
#include <ATen/FunctionalTensorWrapper.h>
#include <ATen/FunctionalInverses.h>
#include <ATen/TensorUtils.h>
#include <ATen/WrapDimUtils.h>
#include <ATen/core/IListRef.h>
#include <ATen/core/LegacyTypeDispatch.h>
#include <c10/util/Exception.h>
#include <c10/util/irange.h>
#ifndef AT_PER_OPERATOR_HEADERS
#include <ATen/Functions.h>
#else
#include <ATen/ops/_propagate_xla_data.h>
#include <ATen/ops/_to_copy.h>
#endif
namespace at {
void FunctionalTensorWrapper::set_constructor_metadata() {
TORCH_INTERNAL_ASSERT(value_.defined());
// Note: "level" is a concept that we don't know how to compute in core.
// For now I'm retroactively setting this in functorch,
// but once Open Multiple Dispatch lands we should be able to calculate this in core.
level_ = -1;
// mirror all of the generic tensor metadata onto the wrapper
copy_generic_tensor_metadata(value_.getIntrusivePtr().get(), this);
refresh_numel();
refresh_contiguous();
storage_access_should_throw_ = false;
// In general, the sizes/stride metadata on a tensor can change as it is mutated,
// and these changes need to be reflected in the metadata of the wrapper.
set_allow_tensor_metadata_change(true);
key_set_ = c10::DispatchKeySet(c10::DispatchKey::Functionalize) | value_.key_set();
// All of the keys corresponding to functorch transforms should not be copied over.
// Functorch transforms all have their own wrapper tensors (e.g. BatchedTensorImpl) which expect
// to participate in the functorch transforms.
key_set_ = key_set_ - c10::functorch_transforms_ks - c10::python_ks;
// We override a bunch of _custom(), so make sure they get called
// TODO: metadata copying may not actually be necessary then
set_custom_sizes_strides(SizesStridesPolicy::CustomSizes);
set_custom_device(true);
// E.g. when running torch.compile under inference mode, we need to make sure that
// for any inputs that were created outside of inference mode (so they are not inference tensors),
// then the functional wrappers that we wrap them with should also not be inference tensors.
version_counter_ = value_.unsafeGetTensorImpl()->version_counter();
}
FunctionalTensorWrapper::FunctionalTensorWrapper(const Tensor& value)
: c10::TensorImpl(
c10::Storage(c10::make_intrusive<functionalization::FunctionalStorageImpl>(value)),
c10::DispatchKeySet(DispatchKey::Functionalize) | value.key_set(),
value.dtype()
),
value_(value)
{
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(value_));
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
set_constructor_metadata();
}
void FunctionalTensorWrapper::freeze_storage() const {
functional_storage_impl()->freeze();
}
// Note [Functionalization: Alias Removal]
// When someone calls a view() op during the functionalization pass, e.g. 'b = a.view(...)',
// we link `b` and `a` to a shared Alias object to preserve the aliasing relationship.
//
// How do we do that?
//
// Every FunctionalTensorWrapper contains a dummy FunctionalStorageImpl, which subclasses from c10::StorageImpl.
// It doesn't contain any data (similar to MetaTensor storage), but it contains an Alias object that knows about the base tensor.
// When a tensor is created through a view operation, both the new and old tensor point to the same FunctionalStorageImpl.
//
// As mutations are applied to any of the views, we also queue each mutation up on the Alias object, so we can replay them.
// When the user requests a tensor that's had a view taken, we check if it's up to date.
// If it's not up to date, we first replay all of the queued up mutations onto the alias, and then re-apply the current view
// on top of the newly updated alias.
//
// Why do we queue up and lazily run mutations on the alias, instead of updating the alias eagerly?
// This behavior was taken from pytorch/xla, which the alias-removal logic was inspired from.
// One benefit of the laziness is that we save work in the cases where a user has multiple views and mutates one of them,
// but never uses the other views later in the program (in which case we'll never update the alias).
// It also has downsides though: repeatedly applying mutations to the same view without syncing
// will silently use up more and more memory as more mutations are queued up.
//
// Corresponding diagram:
//
// b = a.view(...)
//
// a b
// | | If the user asks for b and its out of date,
// \/ \/ We regenerate b by replaying its views from the alias.
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - .
// | FunctionalTensorWrapper | | FunctionalTensorWrapper |
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - .
// | value | storage | | storage | Value |
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - .
// | \ / |
// | \ / |
// | . - - - - - - - - - - - - . |
// | | FunctionalStorageImpl | |
// | . - - - - - - - - - - - - . |
// | | Alias | |
// | . - - - - - - - - - - - - . |
// | / mutations to a or b |
// | / are queued onto Alias |
// | / |
// \/ / \/
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - - - .
// | TensorImpl | | TensorImpl |
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - - - .
// | value | storage | | storage | Value |
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - - - .
// | |
// | |
// | |
// | In this picture the two tensor views their own storages, |
// | have their own storages, but backends like functorch |
// \/ are allowed to re-alias underneath the pass \/
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - - - .
// | underyling_storage | | underyling_storage |
// . - - - - - - - - - - - - - . . - - - - - - - - - - - - - - - .
//
// This constructor is only used by view ops.
// - view_value: The output tensor that we need to wrap.
// - base: The "base" of the view that `view_value` was generated from.
// See Note [Functionalization: Alias Removal Part 2] for more details on the mutation replay logic.
FunctionalTensorWrapper::FunctionalTensorWrapper(const Tensor& view_value, const FunctionalTensorWrapper* base, const functionalization::ViewMeta& meta)
: c10::TensorImpl(
c10::DispatchKeySet(DispatchKey::Functionalize),
view_value.dtype(),
view_value.device()
),
value_(view_value),
is_multi_output_view_(base->is_multi_output_view_ || meta.is_multi_output),
was_storage_changed_(base->was_storage_changed_),
is_symbolic_(base->is_symbolic_)
{
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(value_));
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
set_constructor_metadata();
// Copy the original tensor's ViewMeta vector and push the current one.
if (!base->view_metas_.empty()) {
view_metas_ = base->view_metas_; // copy
}
view_metas_.push_back(meta);
maybe_mark_symbolic(meta);
storage_ = base->storage_; // alias this tensor's storage with the base tensor's
}
functionalization::FunctionalStorageImpl* FunctionalTensorWrapper::functional_storage_impl() const {
return static_cast<functionalization::FunctionalStorageImpl*>(storage_.unsafeGetStorageImpl());
}
void FunctionalTensorWrapper::commit_update() {
auto storage_impl = functional_storage_impl();
storage_impl->add_update(value_, view_metas_);
// As an optimization, we used to mark the tensor here as "up-to-date",
// That way, code like:
// x = torch.ones(1'000'000)
// x[0].add_(1)
// doesn't result in an unnecessary materialization of the base.
// This optimization results in the slice temporarily haven't incorrect
// stride/storage_offset though, and DCE should handle that optimization anyway.
// generation_ = storage_impl->generation();
}
bool FunctionalTensorWrapper::is_up_to_date() const {
auto alias_generation = functional_storage_impl()->generation();
return generation_ == alias_generation;
}
// See Note [Functionalization Pass - Inplace View Ops]
void FunctionalTensorWrapper::mutate_view_meta(const at::functionalization::ViewMeta& meta) {
view_metas_.push_back(meta);
// Manually track the fact that this tensor recieved a metadata mutation!
has_metadata_mutation_ = true;
// Mark this tensor as being symbolic if there are any symbolic inputs used by the view operation.
maybe_mark_symbolic(meta);
// Note [Functionalization Pass - Inplace View Ops]
// So, these ops are special - they're mutation AND view ops. They get special codegen.
// An example is transpose_, e.g. `a.transpose_()`
// Calling transpose_() should ensure that a gets an alias, and append the new ViewMeta to a's current list of ViewMetas.
at::AutoDispatchSkipFunctionalize guard;
value_ = meta.forward_fn(value_, meta.out_index);
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
}
// Note [Functionalization: Mutation Removal]
// Mutation removal is used to take a program like this:
//
// a.add_(b)
//
// and replace it with a slightly different program that has the same semantics:
//
// tmp = a.add(b)
// a.replace_(tmp)
//
// Where the replace_() call is implemented directly in the functionalization pass, so it is transparent to the backend.
// This is useful for backends that aren't able to handle certain types of mutations, like functorch.
//
// Why do we need to wrap every tensor in a FunctionalTensorWrapper? Consider this program:
//
// Before:
// tensor.add_(batched_tensor)
//
// After:
// tmp = tensor.add(batched_tensor)
// tensor.replace_(tmp)
//
// In the above, tmp is a batched tensor (because adding a normal tensor to a batched tensor does broadcasting and creates a batched tensor).
// But we can't just replace the underlying memory backing `tensor` with `tmp` - a batched tensor takes up more space!
// Instead, every input, intermediate and output of the program is wrapped in a FunctionalTensorImpl, which wraps the underlying tensor.
void FunctionalTensorWrapper::replace_(const Tensor& other, bool from_lazy_regenerate) {
// TODO: going to need to change this if we want nested functionalize() transforms.
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(other));
value_ = other;
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
// out= ops are allowed to resize the output tensors, mutating both the data and metadata of the tensor.
// We need to propagate that metadata mutation to the wrapper (new size).
auto sizes_ = value_.sym_sizes();
auto strides_ = value_.sym_strides();
auto storage_offset_ = value_.sym_storage_offset();
set_sizes_and_strides(sizes_, strides_, storage_offset_);
if (dtype() != value_.unsafeGetTensorImpl()->dtype() || layout() != value_.unsafeGetTensorImpl()->layout()) {
// .to() should not re-entrantly go through functionalization.
at::AutoDispatchSkipFunctionalize guard;
// and we want _to_copy() to show up in the graph, not the composite .to() operator
// (this can happen if autograd has already run by the time we enter this code)
value_ = at::_to_copy(value_, c10::TensorOptions().dtype(dtype()).layout(layout()));
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
}
// might not be until after the no_grad region is exited.
// Therefore, replace_() is not unconditionally safe to check the current no_grad state.
// If this is a lazy regeneration, then it is guaranteed that we have already
// done the mutation for the storage alias (when we originally performed the mutation),
// so no counter update may be needed.
// Example: if a mutation happens to a view under a no_grad,
// we won't call replace_() on the other alias until the alias is later used, which
if (!from_lazy_regenerate) {
mark_mutation();
if (!at::GradMode::is_enabled() || InferenceMode::is_enabled()) {
// This mutation happened under no_grad or inference_mode
mark_mutation_during_no_grad_or_inference_mode();
}
}
}
bool FunctionalTensorWrapper::has_data_mutation() {
// Current tensor's data was mutated if its storage saw any mutations.
return functional_storage_impl()->generation() > 0;
}
void FunctionalTensorWrapper::set__impl(const FunctionalTensorWrapper* other) {
// self.set_(src) will cause self to have all of the tensor properties of self.
value_ = other->value_;
generation_ = other->generation_;
view_metas_ = other->view_metas_;
is_symbolic_ = other->is_symbolic_;
// FREEZE the old storage, preventing mutations to it.
// this is a huge pain to handle properly in all cases, so we ban it.
functional_storage_impl()->freeze();
// Unsafely swap out the storage with other's storage,
// disconnecting `self` with its view chain
storage_ = other->storage_;
/// explicitly mark the tensor as having its storage changed from set_()
// Otherwise, we don't actually have a 100% accurate way to check this.
// (We could check if the updated value has a new storage than the original value,
// but this won't also let us uniquely determine if the tensor **also**
// experienced a data mutation).
was_storage_changed_ = true;
auto sizes_ = value_.sym_sizes();
auto strides_ = value_.sym_strides();
auto storage_offset_ = value_.sym_storage_offset();
set_sizes_and_strides(sizes_, strides_, storage_offset_);
}
void FunctionalTensorWrapper::storage_resize_(const c10::SymInt& new_size) {
auto curr_storage_size = value_.unsafeGetTensorImpl()->unsafe_storage().unsafeGetStorageImpl()->sym_nbytes();
// storage resizing is severely limited: we only support resizing either to zero, or from zero bytes.
TORCH_CHECK(new_size == 0 || curr_storage_size == 0, "new_size: ", new_size, ". curr_storage_size: ", curr_storage_size);
// The "functionalization rule" for storage resizing is a giant no-op, mainly because we don't want
// resize_() calls to actualy emit any ops in the functional graph.
// How does it work?
// Resizing up (old size == 0):
// We do nothing in this case.
// The expection is that for the user code to be valid, the next op that should run against the current tensor "x"
// will be a x.copy_(y) (or similar), that will fully overwrite the data of x.
// If there are any outstanding aliases of x, we expect them not to be used until after the copy_() call
// (otherwise the eager code would be invalid),
// and therefore functionalization will regenerate the aliases off of the result of `x.copy(y)`.
// Resizing down (new size == 0):
// We also do nothing in this case. The assumption is that after resizing a tensor down,
// it is fully unused in the program (unless it is later resized back up first, has data copied in)
// Although it might be saved for backward, which happens in FSDP.
// The expected pattern is that the param will then be resized back up from zero in the backward.
// Mark the tensor as having its storage resized.
// This is so we can detect it for inputs in AOTAutograd and error / emit
// an input mutation resize_() appropriately
functional_storage_impl()->mark_inductor_storage_resize(new_size);
}
void FunctionalTensorWrapper::maybe_replace_storage(const Tensor& other) {
// Note [resize_() in functionalization pass]
// resize_() is a special operator in functionalization because it can reallocate its underlying storage.
// This function is only ever called in the case that resize_() needs to reallocate its storage to a larger size.
//
// However, functionalization currently bans the following code:
// a = torch.ones(2)
// b = a.view(2)
// b.resize_(4) # b is a view tensor, that we are trying to increase the storage size of
//
// Why is this code difficult to handle?
// The functionalization pass currently keeps aliases in sync by making the following assumptions:
// - The base tensor always refers to all of the data
// - Whenever you have b = view_op(a), b should always refer to a subset of as memory.
//
// The code above breaks that assumption b.resize_(4) actually needs to update "a"
// to tell it that it is now actually some slice of a pre-existing larger storage.
// We're also no longer re-generate "b" fully from "a" anymore, since "a" refers to a slice of "b"'s data.
//
// This is probably fixable in theory, but:
// - the fix would likey complicated the functionalization logic quite a bit.
// - the primary use case for resize_() today is resizing zero-sized tensors in out= variants of operators
// - resize_() also can give you weird results today if you try to resize_() a weirdly strided tensor.
//
// Given all of the above, for now we're just banning the above usage.
TORCH_CHECK(storage().use_count() == 1, "Attempted to resize a view tensor to a larger size. This is not allowed in the functionalization pass");
TORCH_CHECK(view_metas_.empty(), "Attempted to resize a view tensor to a larger size. This is not allowed in the functionalization pass");
// If this tensor is not a view (and has no outstanding views taken out on it),
// Then it's safe to throw out the old storage and replace it with the new, larger one.
storage_ = c10::Storage(c10::make_intrusive<functionalization::FunctionalStorageImpl>(other));
value_ = other;
TORCH_INTERNAL_ASSERT(!value_.key_set().has(c10::DispatchKey::Functionalize));
generation_ = 0;
// And update the metadata on the wrapper to reflect the new sizes and strides
set_sizes_and_strides(value_.sizes(), value_.strides());
refresh_numel();
// (Technically we should be guaranteed that the tensor was already contiguous,
// since it's guaranteed not to have been a view. Doesnt hurt to run though)
refresh_contiguous();
// Swapping out the storage of a tensor (aka from a resize_() call) will update the sizes and strides of the tensor,
// so we need to record the fact that metadata was mutated.
has_metadata_mutation_ = true;
}
void FunctionalTensorWrapper::_unsafe_reset_storage() {
// Reset the storage with the current value_ tensor as the base
storage_ = c10::Storage(c10::make_intrusive<functionalization::FunctionalStorageImpl>(value_));
// Reset the generation so that it matches the new storage
generation_ = 0;
// Clear any pre-existing view metas so that base and value_ are semantically the same
view_metas_.clear();
}
void FunctionalTensorWrapper::sync_() {
if (is_up_to_date()) {
return;
}
apply_updates();
regenerate_from_base();
}
Tensor FunctionalTensorWrapper::apply_view_metas(const Tensor& base) {
auto t = base;
// Reapply views to get the viewed tensor from the base in alias_
for (auto& view_meta: view_metas_) {
t = view_meta.forward_fn(t, view_meta.out_index);
}
return t;
}
void FunctionalTensorWrapper::regenerate_from_base() {
at::AutoDispatchSkipFunctionalize guard;
auto storage_impl = functional_storage_impl();
auto t = storage_impl->base();
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(t));
t = apply_view_metas(t);
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(t));
replace_(t, /*from_lazy_regenerate=*/true);
generation_ = storage_impl->generation();
}
bool FunctionalTensorWrapper::apply_updates() {
// Apply all updates on alias_
auto storage_impl = functional_storage_impl();
return storage_impl->apply_updates();
}
const char* FunctionalTensorWrapper::tensorimpl_type_name() const {
return "FunctionalTensorWrapper";
}
void FunctionalTensorWrapper::copy_tensor_metadata(
const FunctionalTensorWrapper* src_impl,
FunctionalTensorWrapper* dest_impl,
const c10::VariableVersion& version_counter,
bool allow_tensor_metadata_change) {
TensorImpl::copy_tensor_metadata(
src_impl,
dest_impl,
version_counter,
allow_tensor_metadata_change);
// FunctionalTensorWrapper-specific fields.
dest_impl->value_ = src_impl->value_;
dest_impl->level_ = src_impl->level_;
dest_impl->has_metadata_mutation_ = src_impl->has_metadata_mutation_;
dest_impl->is_multi_output_view_ = src_impl->is_multi_output_view_;
dest_impl->was_storage_changed_ = src_impl->was_storage_changed_;
dest_impl->is_symbolic_ = src_impl->is_symbolic_;
dest_impl->generation_ = src_impl->generation_;
dest_impl->view_metas_ = src_impl->view_metas_;
}
void FunctionalTensorWrapper::copy_tensor_metadata_and_refresh(
const FunctionalTensorWrapper* src_impl,
FunctionalTensorWrapper* dest_impl,
const c10::VariableVersion& version_counter,
bool allow_tensor_metadata_change) const {
copy_tensor_metadata(src_impl, dest_impl, version_counter, allow_tensor_metadata_change);
dest_impl->refresh_numel();
dest_impl->refresh_contiguous();
}
template <typename VariableVersion>
c10::intrusive_ptr<TensorImpl> FunctionalTensorWrapper::shallow_copy_and_detach_core(
VariableVersion&& version_counter,
bool allow_tensor_metadata_change) const {
if (key_set_.has(DispatchKey::Python) &&
!c10::impl::tls_is_dispatch_key_excluded(DispatchKey::Python)) {
auto r = pyobj_slot_.load_pyobj_interpreter()->detach(this);
if (r) {
r->set_version_counter(std::forward<VariableVersion>(version_counter));
r->set_allow_tensor_metadata_change(allow_tensor_metadata_change);
return r;
}
}
auto impl = c10::make_intrusive<FunctionalTensorWrapper>(value_);
copy_tensor_metadata_and_refresh(
/*src_impl=*/this,
/*dest_impl=*/impl.get(),
/*version_counter=*/std::forward<VariableVersion>(version_counter),
/*allow_tensor_metadata_change=*/allow_tensor_metadata_change);
return impl;
}
c10::intrusive_ptr<TensorImpl> FunctionalTensorWrapper::shallow_copy_and_detach(
const c10::VariableVersion& version_counter,
bool allow_tensor_metadata_change) const {
return shallow_copy_and_detach_core(
version_counter, allow_tensor_metadata_change);
}
c10::intrusive_ptr<TensorImpl> FunctionalTensorWrapper::shallow_copy_and_detach(
c10::VariableVersion&& version_counter,
bool allow_tensor_metadata_change) const {
return shallow_copy_and_detach_core(
std::move(version_counter), allow_tensor_metadata_change);
}
void FunctionalTensorWrapper::shallow_copy_from(const c10::intrusive_ptr<TensorImpl>& impl) {
AT_ASSERT(has_compatible_shallow_copy_type(impl->key_set()));
auto functional_impl =
static_cast<FunctionalTensorWrapper*>(impl.get());
copy_tensor_metadata_and_refresh(
/*src_impl=*/functional_impl,
/*dest_impl=*/this,
/*version_counter=*/version_counter(),
/*allow_tensor_metadata_change=*/allow_tensor_metadata_change());
}
c10::Device FunctionalTensorWrapper::device_custom() const {
return value_.unsafeGetTensorImpl()->device();
}
at::IntArrayRef FunctionalTensorWrapper::sizes_custom() const {
return value_.unsafeGetTensorImpl()->sizes();
}
at::IntArrayRef FunctionalTensorWrapper::strides_custom() const {
return value_.unsafeGetTensorImpl()->strides();
}
int64_t FunctionalTensorWrapper::dim_custom() const {
return value_.unsafeGetTensorImpl()->dim();
}
int64_t FunctionalTensorWrapper::numel_custom() const {
return value_.unsafeGetTensorImpl()->numel();
}
bool FunctionalTensorWrapper::is_contiguous_custom(at::MemoryFormat memory_format) const {
return value_.unsafeGetTensorImpl()->is_contiguous(memory_format);
}
c10::SymIntArrayRef FunctionalTensorWrapper::sym_sizes_custom() const {
return value_.unsafeGetTensorImpl()->sym_sizes();
}
c10::SymIntArrayRef FunctionalTensorWrapper::sym_strides_custom() const {
return value_.unsafeGetTensorImpl()->sym_strides();
}
c10::SymInt FunctionalTensorWrapper::sym_size_custom(int64_t d) const {
return value_.unsafeGetTensorImpl()->sym_size(d);
}
c10::SymInt FunctionalTensorWrapper::sym_storage_offset_custom() const {
return value_.unsafeGetTensorImpl()->sym_storage_offset();
}
c10::Layout FunctionalTensorWrapper::layout_impl() const {
return value_.unsafeGetTensorImpl()->layout();
}
namespace functionalization {
namespace impl {
Tensor to_functional_tensor(const Tensor& tensor) {
// Note [Wrapped Numbers <> Functionalization]
if (tensor.unsafeGetTensorImpl()->is_wrapped_number()) {
return tensor;
}
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(!isFunctionalTensor(tensor));
return at::detail::make_tensor<FunctionalTensorWrapper>(tensor);
}
std::optional<Tensor> to_functional_tensor(const std::optional<Tensor>& tensor) {
if (tensor.has_value()) {
return std::make_optional<Tensor>(to_functional_tensor(*tensor));
}
return std::nullopt;
}
c10::List<::std::optional<Tensor>> to_functional_tensor(const c10::List<::std::optional<Tensor>>& t_list) {
c10::List<::std::optional<Tensor>> outputs;
outputs.reserve(t_list.size());
for (const auto i : c10::irange(t_list.size())) {
outputs.push_back(to_functional_tensor(t_list[i]));
}
return outputs;
}
std::vector<Tensor> to_functional_tensor(ITensorListRef t_list) {
std::vector<Tensor> outputs;
outputs.reserve(t_list.size());
for (const auto& tensor : t_list) {
outputs.push_back(to_functional_tensor(tensor));
}
return outputs;
}
Tensor from_functional_tensor(const Tensor& tensor, bool assert_functional) {
// Note [Wrapped Numbers <> Functionalization]
if (!tensor.defined() || tensor.unsafeGetTensorImpl()->is_wrapped_number()) {
return tensor;
}
if (isFunctionalTensor(tensor)) {
auto impl = unsafeGetFunctionalWrapper(tensor);
return impl->value();
} else {
// If the current tensor is not functional, then raise an error
// if assert_functional is true. Otherwise, return the input.
TORCH_INTERNAL_ASSERT(!assert_functional)
return tensor;
}
}
std::optional<Tensor> from_functional_tensor(const std::optional<Tensor>& t, bool assert_functional) {
if (t.has_value()) {
return std::make_optional<Tensor>(from_functional_tensor(*t, assert_functional));
}
return std::nullopt;
}
std::vector<Tensor> from_functional_tensor(ITensorListRef t_list) {
std::vector<Tensor> outputs;
outputs.reserve(t_list.size());
for (const auto& tensor : t_list) {
// from_functional_tensor(Tensor) has asserts to make sure you don't accidentally call
// it on a non-functional input,
// but from_functional_tensor(TensorList) can recieve a list containing both
// functional and non-functional tensors.
// Example of when that can happen: torch.cat(function_input_tensor, global_state_tensor).
// When that happens, we're okay with only unwrapping the functional tensors.
outputs.push_back(from_functional_tensor(tensor, /*assert_functional=*/false));
}
return outputs;
}
c10::List<::std::optional<Tensor>> from_functional_tensor(const c10::List<::std::optional<Tensor>>& t_list) {
c10::List<::std::optional<Tensor>> outputs;
outputs.reserve(t_list.size());
for (const auto i : c10::irange(t_list.size())) {
outputs.push_back(from_functional_tensor(t_list[i], /*assert_functional=*/false));
}
return outputs;
}
void sync(const Tensor& t) {
if (t.unsafeGetTensorImpl()->is_wrapped_number()) {
// Note [Wrapped Numbers <> Functionalization]
// Unfortunately, we can't easily guarantee that wrapped numbers (scalar-tensors)
// get wrapped up in a FunctionalTensorWrapper object, since they skip the dispatcher.
// That shouldn't matter, since I don't think we're allowed to assign to wrapped numbers anyway.
return;
}
// Not every tensor that hits a functionalization kernel is necessarily a functional tensor.
// For example, xla_tensor.copy_(cpu_tensor) needs to hit the functionalization kernel
// to sync xla_tensor, but not cpu_tensor.
if (!at::functionalization::impl::isFunctionalTensor(t)) {
return;
}
auto functional_impl = at::functionalization::impl::unsafeGetFunctionalWrapper(t);
functional_impl->sync_();
}
void sync(const std::optional<Tensor>& t) {
if (t.has_value()) {
sync(*t);
}
}
void sync(ITensorListRef t_list) {
for (const auto& t : t_list) {
sync(t);
}
}
void sync(const c10::List<::std::optional<Tensor>>& t_list) {
for (const auto i : c10::irange(t_list.size())) {
sync(t_list[i]);
}
}
void replace_(const Tensor& functional_tensor, const Tensor& other) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(isFunctionalTensor(functional_tensor));
unsafeGetFunctionalWrapper(functional_tensor)->replace_(other);
}
void replace_(const ITensorListRef functional_tensor, ITensorListRef other) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(functional_tensor.size() == other.size());
auto functional_tensor_it = functional_tensor.begin();
auto other_it = other.begin();
for (C10_UNUSED const auto i : c10::irange(functional_tensor.size())) {
replace_(*functional_tensor_it++, *other_it++);
}
}
void propagate_xla_data(const Tensor& functional_tensor, const Tensor& other) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(isFunctionalTensor(functional_tensor));
if (functional_tensor.key_set().has(c10::DispatchKey::XLA)) {
at::_propagate_xla_data(at::functionalization::impl::unsafeGetFunctionalWrapper(functional_tensor)
->value(), other);
}
}
void propagate_xla_data(const ITensorListRef functional_tensor, ITensorListRef other) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(functional_tensor.size() == other.size());
auto functional_tensor_it = functional_tensor.begin();
auto other_it = other.begin();
for (C10_UNUSED const auto i : c10::irange(functional_tensor.size())) {
propagate_xla_data(*functional_tensor_it++, *other_it++);
}
}
void propagate_xla_data_direct(const Tensor& tensor, const Tensor& other) {
if (tensor.key_set().has(c10::DispatchKey::XLA)) {
at::_propagate_xla_data(tensor, other);
}
}
void propagate_xla_data_direct(const ITensorListRef tensor,
ITensorListRef other) {
auto tensor_it = tensor.begin();
auto other_it = other.begin();
for (C10_UNUSED const auto i : c10::irange(tensor.size())) {
propagate_xla_data_direct(*tensor_it++, *other_it++);
}
}
void commit_update(const Tensor& functional_tensor) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(isFunctionalTensor(functional_tensor));
unsafeGetFunctionalWrapper(functional_tensor)->commit_update();
}
void commit_update(ITensorListRef functional_tensor) {
for (const auto& t : functional_tensor) {
commit_update(t);
}
}
void unsafe_reset_storage(const Tensor& functional_tensor) {
TORCH_INTERNAL_ASSERT_DEBUG_ONLY(isFunctionalTensor(functional_tensor));
unsafeGetFunctionalWrapper(functional_tensor)->_unsafe_reset_storage();
}
void mark_mutation_hidden_from_autograd(const Tensor& functional_tensor) {
TORCH_CHECK(isFunctionalTensor(functional_tensor));
unsafeGetFunctionalWrapper(functional_tensor)->mark_mutation_hidden_from_autograd();
}
bool are_all_mutations_hidden_from_autograd(const Tensor& functional_tensor) {
TORCH_CHECK(isFunctionalTensor(functional_tensor));
return unsafeGetFunctionalWrapper(functional_tensor)->are_all_mutations_hidden_from_autograd();
}
bool are_all_mutations_under_no_grad_or_inference_mode(const Tensor& functional_tensor) {
TORCH_CHECK(isFunctionalTensor(functional_tensor));
return unsafeGetFunctionalWrapper(functional_tensor)->are_all_mutations_under_no_grad_or_inference_mode();
}
bool isFunctionalTensor(const at::Tensor& tensor) {
return tensor.unsafeGetTensorImpl()->key_set().has(c10::DispatchKey::Functionalize);
}
bool isFunctionalTensor(const std::optional<Tensor>& t) {
if (t.has_value()) {
return isFunctionalTensor(*t);
} else {
return false;
}
}
bool isFunctionalTensor(const c10::List<::std::optional<Tensor>>& t_list) {
if (t_list.empty()) return false;
auto functional_count = 0;
for (const auto i : c10::irange(t_list.size())) {
if (!t_list[i].has_value() || !t_list[i]->defined()) continue;
if (isFunctionalTensor(t_list[i])) {
++functional_count;
}
}
return functional_count > 0;
}
template <typename T>
bool isFunctionalTensorIListRef(c10::IListRef<T> list) {
if (list.size() == 0) return false;
auto functional_count = 0;
for (const auto& tensor : list) {
if (!tensor.defined()) continue;
if (isFunctionalTensor(tensor)) {
++functional_count;
}
}
return functional_count > 0;
}
bool isFunctionalTensor(ITensorListRef list) {
return isFunctionalTensorIListRef(list);
}
void freeze_functional_tensor(const Tensor& tensor) {
TORCH_INTERNAL_ASSERT(at::functionalization::impl::isFunctionalTensor(tensor));
auto functional_base_impl = at::functionalization::impl::unsafeGetFunctionalWrapper(tensor);
functional_base_impl->freeze_storage();
}
Tensor create_functional_tensor_with_view_meta(const at::Tensor& view_to_wrap, const at::Tensor& base, functionalization::ViewMeta meta, int64_t out_idx) {
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(view_to_wrap));
TORCH_INTERNAL_ASSERT(at::functionalization::impl::isFunctionalTensor(base));
auto functional_base_impl = at::functionalization::impl::unsafeGetFunctionalWrapper(base);
if (out_idx != 0) {
// Note [out_idx in ViewMeta]
// When a view op outputs multiple tensors, each output needs its own separate ViewMeta.
// Each ViewMeta also tracks the index of the particular output tensor, which is needed in the reverse function.
meta = meta.to_out_idx(out_idx);
}
return at::detail::make_tensor<FunctionalTensorWrapper>(view_to_wrap, functional_base_impl, meta);
}
std::vector<Tensor> create_functional_tensor_with_view_meta(ITensorListRef view_to_wrap, const at::Tensor& base, const functionalization::ViewMeta& meta) {
std::vector<Tensor> outputs(view_to_wrap.size());
int64_t i = 0;
for (const auto& tensor : view_to_wrap) {
outputs[i] = create_functional_tensor_with_view_meta(tensor, base, meta, i);
i++;
}
return outputs;
}
void mutate_view_meta(const at::Tensor& self, const functionalization::ViewMeta& meta) {
TORCH_INTERNAL_ASSERT(at::functionalization::impl::isFunctionalTensor(self));
auto self_impl = at::functionalization::impl::unsafeGetFunctionalWrapper(self);
self_impl->mutate_view_meta(meta);
}
// Note [Propagating strides in the functionalization pass]
// In order to properly compute stride information, the functionalization pass
// calls each {view} reference implementations with meta tensors.
// The output meta tensor's stride info serves as a reference for what the correct strides should be.
void set_sizes_strides_offset(const Tensor& out, const Tensor& reference_out) {
out.unsafeGetTensorImpl()->set_sizes_and_strides(reference_out.sym_sizes(), reference_out.sym_strides(), reference_out.sym_storage_offset());
}
void set_sizes_strides_offset(const std::vector<Tensor>& outs, const std::vector<Tensor>& reference_outs) {
TORCH_INTERNAL_ASSERT(outs.size() == reference_outs.size());
for (const auto i : c10::irange(reference_outs.size())) {
set_sizes_strides_offset(outs[i], reference_outs[i]);
}
}
thread_local bool _functionalizationReapplyViews;
bool getFunctionalizationReapplyViewsTLS() {
return _functionalizationReapplyViews;
}
void setFunctionalizationReapplyViewsTLS(bool reapply_views) {
_functionalizationReapplyViews = reapply_views;
}
} // namespace impl
// Given an **out-of-place** op that might internally call view/inplace ops,
// This function will "functionalize" it.
// That is, it will call the operator, but removing any intermediate views/mutations
// that are performed inside of it.
// This is useful for LTC/XLA, which would like to re-use some of our composite kernels
// from pytorch core but not have to worry about the view ops that they might call.
// e.g. at::block_diag
void functionalize_op_helper(const c10::OperatorHandle& op, torch::jit::Stack* stack) {
const auto& schema = op.schema();
const auto num_arguments = schema.arguments().size();
const auto arguments_begin = stack->size() - num_arguments;
auto arguments = torch::jit::last(stack, num_arguments);
// Wrap all tensor-like inputs into FunctionalTensorWrappers.
// When we re-invoke the dispatcher, this will automatically enable the functionalization pass.
for (uint64_t idx = 0; idx < num_arguments; ++idx) {
const auto& ivalue = arguments[idx];
if (ivalue.isTensor()) {
const auto& t = ivalue.toTensor();
if (t.defined()) {
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(t),
"The composite op functionalization fallback expects its inputs all not to be functional tensors");
auto t_new = c10::IValue(at::functionalization::impl::to_functional_tensor(t));
(*stack)[arguments_begin + idx] = t_new;
}
} else if (ivalue.isTensorList()) {
auto tensors = ivalue.toTensorList();
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(tensors),
"The composite op functionalization fallback expects its inputs all not to be functional tensors");
auto t_new = c10::IValue(at::functionalization::impl::to_functional_tensor(tensors));
(*stack)[arguments_begin + idx] = t_new;
} else if (ivalue.isOptionalTensorList()) {
auto opt_tensors = ivalue.toOptionalTensorList();
TORCH_INTERNAL_ASSERT(!at::functionalization::impl::isFunctionalTensor(opt_tensors),
"The composite op functionalization fallback expects its inputs all not to be functional tensors");
auto t_new = c10::IValue(at::functionalization::impl::to_functional_tensor(opt_tensors));
(*stack)[arguments_begin + idx] = t_new;
}
}
{
// Today when you call at::empty(device=lazy), the lazy backend decides whether or not to wrap
// the output in a functional tensor based on TLS.
// In this code, we're re-entrantly entering functionalization in the same call-stack,
// so we need to manually fix up TLS as if it hadn't already been called.
auto curr_tls = c10::impl::tls_local_dispatch_key_set();
auto tls_reenable_functionalize = c10::impl::PODLocalDispatchKeySet();
tls_reenable_functionalize.set_included(curr_tls.included_);
tls_reenable_functionalize.set_excluded(curr_tls.excluded_.remove(c10::DispatchKey::Functionalize));
c10::impl::ForceDispatchKeyGuard guard_(tls_reenable_functionalize);
// So, we should probably provide a way to directly call a kernel registered to
// the `CompositeExplicitAutograd` key.
// We can't do that today, so this should be a reasonably good proxy
// (It won't work in cases where an op has both a CompositeExplicitAutograd kernel
// AND a dedicated meta kernel, but that probably shouldn't ever happen).
op.redispatchBoxed(c10::DispatchKeySet(c10::DispatchKey::Meta), stack);
}
const auto num_returns = schema.returns().size();
const auto returns_begin = stack->size() - num_returns;
auto returns = torch::jit::last(stack, num_returns);
for (const auto idx : c10::irange(num_returns)) {
const auto& ivalue = returns[idx];
if (ivalue.isTensor()) {
const auto& t = ivalue.toTensor();
if (!t.defined()) continue;
at::functionalization::impl::sync(t);
auto t_new = c10::IValue(at::functionalization::impl::from_functional_tensor(t));
(*stack)[returns_begin + idx] = t_new;
} else if (ivalue.isTensorList()) {
auto tensors = ivalue.toTensorList();
at::functionalization::impl::sync(tensors);
auto t_new = c10::IValue(at::functionalization::impl::from_functional_tensor(tensors));
(*stack)[returns_begin + idx] = t_new;
} else if (ivalue.isOptionalTensorList()) {
auto opt_tensors = ivalue.toOptionalTensorList();
at::functionalization::impl::sync(opt_tensors);
auto t_new = c10::IValue(at::functionalization::impl::from_functional_tensor(opt_tensors));
(*stack)[returns_begin + idx] = t_new;
}
}
}
} // namespace functionalization
} // namespace at
```
|
```go
//
//
// path_to_url
//
// Unless required by applicable law or agreed to in writing, software
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
package prometheus
import (
"context"
"fmt"
"reflect"
"strconv"
"strings"
"time"
"github.com/google/uuid"
appsv1 "k8s.io/api/apps/v1"
v1 "k8s.io/api/core/v1"
apierrors "k8s.io/apimachinery/pkg/api/errors"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
monitoringv1 "github.com/prometheus-operator/prometheus-operator/pkg/apis/monitoring/v1"
"github.com/prometheus-operator/prometheus-operator/pkg/informers"
"github.com/prometheus-operator/prometheus-operator/pkg/operator"
)
// Config defines the operator's parameters for the Prometheus controllers.
// Whenever the value of one of these parameters is changed, it triggers an
// update of the managed statefulsets.
type Config struct {
LocalHost string
ReloaderConfig operator.ContainerConfig
PrometheusDefaultBaseImage string
ThanosDefaultBaseImage string
Annotations operator.Map
Labels operator.Map
}
type StatusReporter struct {
Kclient kubernetes.Interface
Reconciliations *operator.ReconciliationTracker
SsetInfs *informers.ForResource
Rr *operator.ResourceReconciler
}
func KeyToStatefulSetKey(p monitoringv1.PrometheusInterface, key string, shard int) string {
keyParts := strings.Split(key, "/")
return fmt.Sprintf("%s/%s", keyParts[0], statefulSetNameFromPrometheusName(p, keyParts[1], shard))
}
func statefulSetNameFromPrometheusName(p monitoringv1.PrometheusInterface, name string, shard int) string {
if shard == 0 {
return fmt.Sprintf("%s-%s", Prefix(p), name)
}
return fmt.Sprintf("%s-%s-shard-%d", Prefix(p), name, shard)
}
func NewTLSAssetSecret(p monitoringv1.PrometheusInterface, config Config) *v1.Secret {
s := &v1.Secret{
Data: map[string][]byte{},
}
operator.UpdateObject(
s,
operator.WithLabels(config.Labels),
operator.WithAnnotations(config.Annotations),
operator.WithManagingOwner(p),
operator.WithName(TLSAssetsSecretName(p)),
operator.WithNamespace(p.GetObjectMeta().GetNamespace()),
)
return s
}
// ValidateRemoteWriteSpec checks that mutually exclusive configurations are not
// included in the Prometheus remoteWrite configuration section, while also validating
// the RemoteWriteSpec child fields.
// Reference:
// path_to_url#remote_write
func ValidateRemoteWriteSpec(spec monitoringv1.RemoteWriteSpec) error {
var nonNilFields []string
for k, v := range map[string]interface{}{
"basicAuth": spec.BasicAuth,
"oauth2": spec.OAuth2,
"authorization": spec.Authorization,
"sigv4": spec.Sigv4,
"azureAd": spec.AzureAD,
} {
if reflect.ValueOf(v).IsNil() {
continue
}
nonNilFields = append(nonNilFields, fmt.Sprintf("%q", k))
}
if len(nonNilFields) > 1 {
return fmt.Errorf("%s can't be set at the same time, at most one of them must be defined", strings.Join(nonNilFields, " and "))
}
if spec.AzureAD != nil {
if spec.AzureAD.ManagedIdentity == nil && spec.AzureAD.OAuth == nil && spec.AzureAD.SDK == nil {
return fmt.Errorf("must provide Azure Managed Identity or Azure OAuth or Azure SDK in the Azure AD config")
}
if spec.AzureAD.ManagedIdentity != nil && spec.AzureAD.OAuth != nil {
return fmt.Errorf("cannot provide both Azure Managed Identity and Azure OAuth in the Azure AD config")
}
if spec.AzureAD.OAuth != nil && spec.AzureAD.SDK != nil {
return fmt.Errorf("cannot provide both Azure OAuth and Azure SDK in the Azure AD config")
}
if spec.AzureAD.ManagedIdentity != nil && spec.AzureAD.SDK != nil {
return fmt.Errorf("cannot provide both Azure Managed Identity and Azure SDK in the Azure AD config")
}
if spec.AzureAD.OAuth != nil {
_, err := uuid.Parse(spec.AzureAD.OAuth.ClientID)
if err != nil {
return fmt.Errorf("the provided Azure OAuth clientId is invalid")
}
}
}
return nil
}
// Process will determine the Status of a Prometheus resource (server or agent) depending on its current state in the cluster.
func (sr *StatusReporter) Process(ctx context.Context, p monitoringv1.PrometheusInterface, key string) (*monitoringv1.PrometheusStatus, error) {
commonFields := p.GetCommonPrometheusFields()
pStatus := monitoringv1.PrometheusStatus{
Paused: commonFields.Paused,
}
var (
availableStatus monitoringv1.ConditionStatus = monitoringv1.ConditionTrue
availableReason string
availableCondition = monitoringv1.Condition{
Type: monitoringv1.Available,
LastTransitionTime: metav1.Time{
Time: time.Now().UTC(),
},
ObservedGeneration: p.GetObjectMeta().GetGeneration(),
}
messages []string
replicas = 1
)
if commonFields.Replicas != nil {
replicas = int(*commonFields.Replicas)
}
for shard := range ExpectedStatefulSetShardNames(p) {
ssetName := KeyToStatefulSetKey(p, key, shard)
obj, err := sr.SsetInfs.Get(ssetName)
if err != nil {
if apierrors.IsNotFound(err) {
// Statefulset hasn't been created or is already deleted.
availableStatus = monitoringv1.ConditionFalse
availableReason = "StatefulSetNotFound"
messages = append(messages, fmt.Sprintf("shard %d: statefulset %s not found", shard, ssetName))
pStatus.ShardStatuses = append(
pStatus.ShardStatuses,
monitoringv1.ShardStatus{
ShardID: strconv.Itoa(shard),
})
continue
}
return nil, fmt.Errorf("failed to retrieve statefulset: %w", err)
}
sset := obj.(*appsv1.StatefulSet).DeepCopy()
if sr.Rr.DeletionInProgress(sset) {
continue
}
stsReporter, err := operator.NewStatefulSetReporter(ctx, sr.Kclient, sset)
if err != nil {
return nil, fmt.Errorf("failed to retrieve statefulset state: %w", err)
}
pStatus.Replicas += int32(len(stsReporter.Pods))
pStatus.UpdatedReplicas += int32(len(stsReporter.UpdatedPods()))
pStatus.AvailableReplicas += int32(len(stsReporter.ReadyPods()))
pStatus.UnavailableReplicas += int32(len(stsReporter.Pods) - len(stsReporter.ReadyPods()))
pStatus.ShardStatuses = append(
pStatus.ShardStatuses,
monitoringv1.ShardStatus{
ShardID: strconv.Itoa(shard),
Replicas: int32(len(stsReporter.Pods)),
UpdatedReplicas: int32(len(stsReporter.UpdatedPods())),
AvailableReplicas: int32(len(stsReporter.ReadyPods())),
UnavailableReplicas: int32(len(stsReporter.Pods) - len(stsReporter.ReadyPods())),
},
)
if len(stsReporter.ReadyPods()) >= replicas {
// All pods are ready (or the desired number of replicas is zero).
continue
}
switch {
case len(stsReporter.ReadyPods()) == 0:
availableReason = "NoPodReady"
availableStatus = monitoringv1.ConditionFalse
case availableCondition.Status != monitoringv1.ConditionFalse:
availableReason = "SomePodsNotReady"
availableStatus = monitoringv1.ConditionDegraded
}
for _, p := range stsReporter.Pods {
if m := p.Message(); m != "" {
messages = append(messages, fmt.Sprintf("shard %d: pod %s: %s", shard, p.Name, m))
}
}
}
pStatus.Conditions = operator.UpdateConditions(
pStatus.Conditions,
monitoringv1.Condition{
Type: monitoringv1.Available,
Status: availableStatus,
Reason: availableReason,
Message: strings.Join(messages, "\n"),
LastTransitionTime: metav1.Time{
Time: time.Now().UTC(),
},
ObservedGeneration: p.GetObjectMeta().GetGeneration(),
},
sr.Reconciliations.GetCondition(key, p.GetObjectMeta().GetGeneration()),
)
return &pStatus, nil
}
```
|
```xml
import { graphql } from "react-relay";
import { Environment } from "relay-runtime";
import {
commitMutationPromiseNormalized,
createMutation,
MutationInput,
} from "coral-framework/lib/relay";
import { ChangeReportStatusMutation as MutationTypes } from "coral-admin/__generated__/ChangeReportStatusMutation.graphql";
let clientMutationId = 0;
const ChangeReportStatusMutation = createMutation(
"changeReportStatus",
(environment: Environment, input: MutationInput<MutationTypes>) => {
const result = commitMutationPromiseNormalized<MutationTypes>(environment, {
mutation: graphql`
mutation ChangeReportStatusMutation(
$input: ChangeDSAReportStatusInput!
) {
changeDSAReportStatus(input: $input) {
dsaReport {
id
status
history {
id
createdBy {
username
}
createdAt
body
type
status
}
}
clientMutationId
}
}
`,
variables: {
input: {
userID: input.userID,
reportID: input.reportID,
status: input.status,
clientMutationId: (clientMutationId++).toString(),
},
},
});
return result;
}
);
export default ChangeReportStatusMutation;
```
|
Phanariots, Phanariotes, or Fanariots (, , ) were members of prominent Greek families in Phanar (Φανάρι, modern Fener), the chief Greek quarter of Constantinople where the Ecumenical Patriarchate is located, who traditionally occupied four important positions in the Ottoman Empire: Voivode of Moldavia, Voivode of Wallachia, Grand Dragoman of the Porte and Grand Dragoman of the Fleet. Despite their cosmopolitanism and often-Western education, the Phanariots were aware of their Greek ancestry and culture; according to Nicholas Mavrocordatos' Philotheou Parerga, "We are a race completely Hellenic".
They emerged as a class of wealthy Greek merchants (of mostly noble Byzantine descent) during the second half of the 16th century, and were influential in the administration of the Ottoman Empire's Balkan domains in the 18th century. The Phanariots usually built their houses in the Phanar quarter to be near the court of the Patriarch, who (under the Ottoman millet system) was recognized as the spiritual and secular head (millet-bashi) of the Orthodox subjects—the Rum Millet, or "Roman nation" of the empire, except those under the spiritual care of the Patriarchs of Antioch, Jerusalem, Alexandria, Ohrid and Peć—often acting as archontes of the Ecumenical See. They dominated the administration of the patriarchate, often intervening in the selection of hierarchs (including the Ecumenical Patriarch of Constantinople).
Overview
Many members of Phanariot families (who had acquired great wealth and influence during the 17th century) occupied high political and administrative posts in the Ottoman Empire. From 1669 until the Greek War of Independence in 1821, Phanariots made up the majority of the dragomans to the Ottoman government (the Porte) and foreign embassies due to the Greeks' higher level of education than the general Ottoman population. With the church dignitaries, local notables from the provinces and the large Greek merchant class, Phanariots represented the better-educated members of Greek society during Ottoman rule until the 1821 start of the Greek War of Independence. During the war, Phanariots influenced decisions by the Greek National Assembly (the representative body of Greek revolutionaries, which met six times between 1821 and 1829). Between 1711–1716 and 1821, a number of Phanariots were appointed Hospodars (voivodes or princes) in the Danubian Principalities (Moldavia and Wallachia) (usually as a promotion from the offices of Dragoman of the Fleet and Dragoman of the Porte); the period is known as the Phanariot epoch in Romanian history.
Ottoman Empire
After the fall of Constantinople, Mehmet II deported the city's Christian population, leaving only the Jewish inhabitants of Balat, repopulating the city with Christians and Muslims from throughout the whole empire and the newly conquered territories. Phanar was repopulated with Greeks from Mouchlion in the Peloponnese and, after 1461, with citizens of the Empire of Trebizond.
The roots of Greek ascendancy can be traced to the Ottoman need for skilled, educated negotiators as their empire declined and they relied on treaties rather than force. During the 17th century, the Ottomans began having problems in foreign relations and difficulty dictating terms to their neighbours; for the first time, the Porte needed to participate in diplomatic negotiations.
With the Ottomans traditionally ignoring Western European languages and cultures, officials were at a loss. The Porte assigned those tasks to the Greeks, who had a long mercantile and educational tradition and the necessary skills. The Phanariots, Greek and Hellenized families primarily from Constantinople, occupied high posts as secretaries and interpreters for Ottoman officials.
Diplomats and patriarchs
As a result of Phanariot and ecclesiastical administration, the Greeks expanded their influence in the 18th-century empire while retaining their Greek Orthodox faith and Hellenism. This had not always been the case in the Ottoman realm. During the 16th century, the South Slavs—the most prominent in imperial affairs—converted to Islam to enjoy the full rights of Ottoman citizenship (especially in the Eyalet of Bosnia; Serbs tended to occupy high military positions.
A Slavic presence in Ottoman administration gradually became hazardous for its rulers, since the Slavs tended to support Habsburg armies during the Great Turkish War. By the 17th century the Greek Patriarch of Constantinople was the religious and administrative ruler of the empire's Orthodox subjects, regardless of ethnic background. All formerly-independent Orthodox patriarchates, including the Serbian Patriarchate renewed in 1557, came under the authority of the Greek Orthodox Church. Most of the Greek patriarchs were drawn from the Phanariots.
Two Greek social groups emerged, challenging the leadership of the Greek Church: the Phanariots in Constantinople and the local notables in the Helladic provinces (kodjabashis, dimogerontes and prokritoi). According to 19th-century Greek historian Constantine Paparrigopoulos, the Phanariots initially sought the most important secular offices of the patriarchal court and could frequently intervene in the election of bishops and influence crucial decisions by the patriarch. Greek merchants and clergy of Byzantine aristocratic origin, who acquired economic and political influence and were later known as Phanariots, settled in extreme northwestern Constantinople (which had become central to Greek interests after the establishment of the patriarch's headquarters in 1461, shortly after Hagia Sophia was converted into a mosque).
Patriarchate
After the 1453 fall of Constantinople, when the Sultan replaced de jure the Byzantine Emperor for subjugated Christians, he recognized the Ecumenical Patriarch as the religious and national leader (ethnarch) of the Greeks and other ethnic groups in the Greek Orthodox Millet. The Patriarchate had primary importance, occupying this key role for Christians of the Empire because the Ottomans did not legally distinguish between nationality and religion and considered the empire's Orthodox Christians a single entity.
The position of the Patriarchate in the Ottoman state encouraged Greek renaissance projects centering on the resurrection and revitalization of the Byzantine Empire. The Patriarch and his church dignitaries constituted the first centre of power for the Greeks in the Ottoman state, which infiltrated Ottoman structures and attracted the former Byzantine nobility.
Merchant middle class
The wealth of the extensive Greek merchant class provided the material basis for the intellectual revival featured in Greek life for more than half a century before 1821. Greek merchants endowed libraries and schools. On the eve of the Greek War of Independence, the three most important centres of Greek learning (schools-cum-universities) were in the commercial centres of Chios, Smyrna and Aivali. The first Greek millionaire of the Ottoman era was Michael "Şeytanoğlu" Kantakouzenos, who earned 60,000 ducats a year from his control of the fur trade from Muscovy.
Civil servants
During the 18th century, the Phanariots were a hereditary clerical−aristocratic group who managed the affairs of the patriarchate and the dominant political power of the Ottoman Greek community. They became a significant political factor in the empire and, as diplomatic agents, played a role in the affairs of Great Britain, France and the Russian Empire.
The Phanariots competed for the most important administrative offices in the Ottoman administration; these included collecting imperial taxes, monopolies on commerce, working under contract in a number of enterprises, supplying the court and ruling the Danubian Principalities. They engaged in private trade, controlling the crucial wheat trade on the Black Sea. The Phanariots expanded their commercial activities into the Kingdom of Hungary and then to the other Central European states. Their activities intensified their contacts with Western nations, and they became familiar with Western languages and cultures.
Before the beginning of the Greek War of Independence, the Phanariots were firmly established as the political elite of Hellenism. According to Greek historian Constantine Paparrigopoulos, this was a natural evolution given the Phanariots' education and experience in supervising large parts of the empire. According to Nikos Svoronos argued, the Phanariots subordinated their national identity to their class identity and tried to peacefully co−exist with the Ottomans; they did not enrich the Greek national identity and lost ground to groups which flourished through their confrontation with the Ottoman Empire (the klephts and armatoloi).
Danubian principalities
A Greek presence had established itself in both Danubian Principalities of Moldavia and Wallachia, resulting in the appointment of Greek princes before the 18th century. After the Phanariot era, some Phanariot families in Wallachia and Moldavia identified themselves as Romanian in Romanian society (including the Rosetti family; C. A. Rosetti represented the radical, nationalist cause during and after the 1848 Wallachian revolution.)
Phanariot attention focused on occupying the most favorable offices the empire could offer non-Muslims and the principalities of Moldavia and Wallachia, which were still relatively rich and—more importantly—autonomous (despite having to pay tribute as vassal states). Many Greeks had found favorable conditions there for commercial activities, in comparison with the Ottoman Empire, and an opportunity for political power; they entered Wallachian and Moldavian boyar nobility by marriage.
Reigns of local princes were not excluded on principle. Several hellenized Romanian noble families, such as the Callimachis (originally Călmașul), the Racovițăs and the Albanian Ghicas penetrated the Phanar nucleus to increase their chances of occupying the thrones and maintain their positions.
Most sources agree that 1711 was when the gradual erosion of traditional institutions reached its zenith, but characteristics ascribed to the Phanariot era had made themselves felt long before it. The Ottomans enforced their choice of hospodars as far back as the 15th century, and foreign (usually Greek or Levantine) boyars competed with local ones since the late 16th century. Rulers since Dumitraşcu Cantacuzino in Moldavia and George Ducas (a prince of Greek origin) in Wallachia, both in 1673, were forced to surrender their family members as hostages in Constantinople. The traditional elective system in the principalities, resulting in long periods of political disorder, was dominated by a small number of ambitious families who competed violently for the two thrones and monopolized land ownership.
1711–1715
A change in policy was indicated by the fact that autonomous Wallachia and Moldavia had entered a period of skirmishes with the Ottomans, due to the insubordination of local princes associated with the rise of Imperial Russia's power under Peter the Great and the firm presence of the Habsburg Empire on the Carpathian border with the principalities. Dissidence in the two countries became dangerous for the Turks, who were confronted with the attraction on the population of protection by a fellow Eastern Orthodox state. This became obvious with Mihai Racoviță's second rule in Moldavia, when the prince plotted with Peter to have Ottoman rule overthrown. His replacement, Nicholas Mavrocordatos, was the first official Phanariot in his second reign in Moldavia and replaced Ștefan Cantacuzino in Wallachia as the first Phanariot ruler of that country.
A crucial moment was the Russo−Turkish War of 1710−1713, when Dimitrie Cantemir sided with Russia and agreed to Russian tutelage of his country. After Russia experienced a major defeat and Cantemir went into exile, the Ottomans took charge of the succession to the throne of Moldavia. This was followed by similar measures in Wallachia, prompted by Ștefan Cantacuzino's alliance with Habsburg commander Prince Eugene of Savoy in the closing stages of the Great Turkish War.
Rulers and retinues
The person raised to the office of prince was usually the chief dragoman of the Porte, well-versed in contemporary politics and Ottoman statecraft. The new prince, who obtained his office in exchange for a generous bribe, proceeded to the country he was selected to govern (whose language he usually did not know). When the new princes were appointed, they were escorted to Iași or Bucharest by retinues composed of their families, favourites and creditors (from whom they had borrowed the bribes). The prince and his appointees counted on recouping these in as short a time as possible, amassing an amount sufficient to live on after their brief time in office.
Thirty-one princes, from eleven families, ruled the two principalities during the Phanariot epoch. When the choice became limited to a few families due to princely disloyalty to the Porte, rulers would be moved from one principality to the other; the prince of Wallachia (the richer of the two principalities) would pay to avert his transfer to Iaşi, and the prince of Moldavia would bribe supporters in Constantinople to appoint him to Wallachia. Constantine Mavrocordatos ruled a total of ten times in Moldavia and Wallachia. The debt was owed to several creditors, rather than to the Sultan; the central institutions of the Ottoman Empire generally seemed determined to maintain their rule over the principalities and not exploit them irrationally. In an early example, Ahmed III paid part of Nicholas Mavrocordatos' sum.
Administration and boyars
The Phanariot epoch was initially characterized by fiscal policies driven by Ottoman needs and the ambitions of some hospodars, who (mindful of their fragile status) sought to pay back their creditors and increase their wealth while in a position of power. To make the reigns lucrative while raising funds to satisfy the needs of the Porte, princes channeled their energies into taxing the inhabitants into destitution. The most odious taxes (such as the văcărit first imposed by Iancu Sasul in the 1580s), mistakenly identified with the Phanariots in modern Romanian historiography, were much older.
The mismanagement of many Phanariot rulers contrasts with the achievements and projects of others, such as Constantine Mavrocordatos (who abolished serfdom in Wallachia in 1746 and Moldavia in 1749) and Alexander Ypsilantis, who were inspired by Habsburg serf policy. Ypsilantis tried to reform legislation and impose salaries for administrative offices in an effort to halt the depletion of funds the administrators, local and Greek alike, were using for their own maintenance; it was, by then, more profitable to hold office than to own land. His Pravilniceasca condică, a relatively modern legal code, met stiff boyar resistance.
The focus of such rules was often the improvement of state structure against conservative wishes. Contemporary documents indicate that, despite the change in leadership and boyar complaints, about 80 percent of those seated in the Divan (an institution roughly equivalent to the estates of the realm) were members of local families. This made endemic the social and economic issues of previous periods, since the inner circle of boyars blocked initiatives (such as Alexander Ypsilantis') and obtained, extended and preserved tax exemptions.
Russian influence
The Phanariots copied Russian and Habsburg institutions; during the mid-18th century they made noble rank dependent on state service, as Peter I of Russia did. After the Treaty of Kuchuk-Kainarji (1774) allowed Russia to intervene on the side of Ottoman Eastern Orthodox subjects, most of the Porte's tools of political pressure became ineffective. They had to offer concessions to maintain a hold on the countries as economic and strategic assets. The treaty made any increase in tribute impossible, and between 1774 and the 1820s it plummeted from about 50,000 to 20,000 gold coins (equivalent to Austrian gold currency) in Wallachia and to 3,100 in Moldavia.
Immediately afterward, Russia forcefully used its new prerogative. The deposition of Constantine Ypsilantis (in Wallachia) and Alexander Mourousis (in Moldavia) by Selim III, called on by French Empire's ambassador to the Ottoman Empire Horace Sébastiani (whose fears of pro−Russian conspiracies in Bucharest were partially confirmed), was the casus belli for the 1806–1812 conflict, and Russian general Mikhail Andreyevich Miloradovich swiftly reinstated Ypsilantis during his military expedition to Wallachia.
Such gestures began a period of effective Russian supervision, culminating with the Organic Statute administration of the 1830s. The Danubian principalities grew in strategic importance with the Napoleonic Wars and the decline of the Ottoman Empire, as European states became interested in halting Russian southward expansion (which included the 1812 annexation of Bessarabia). New consulates in the two countries' capitals, ensuring the observation of developments in Russian−Ottoman relations, had an indirect impact on the local economy as rival diplomats began awarding protection and sudit status to merchants competing with local guilds. Nicholas I of Russia pressured Wallachia and Moldavia into granting constitutions (in 1831 and 1832, respectively) to weaken native rulers.
The boyars began a petition campaign against the princes in power; addressed to the Porte and the Habsburg monarchy, they primarily demanded Russian supervision. Although they referred to incidents of corruption and misrule, the petitions indicate their signers' conservatism. The boyars tend to refer to (fictitious) "capitulations" which either principality would have signed with the Ottomans, demanding that rights guaranteed through them be restored. They viewed reform attempts by princes as illegitimate; in alternative proposals (usually in the form of constitutional projects), the boyars expressed desire for an aristocratic republic.
Greek War of Independence and legacy
The active part taken by Greek princes in revolts after 1820 and the disorder provoked by the Filiki Eteria (of which the Ghica, Văcărescu and Golescu families were active members after its uprising against the Ottoman Empire in Moldavia and Tudor Vladimirescu's Wallachian uprising) led to the disappearance of promotions from the Phanar community; the Greeks were no longer trusted by the Porte. Amid tense relations between boyars and princes, Vladimirescu's revolt was primarily the result of compromise between Oltenian pandurs and the regency of boyars attempting to block the ascension of Scarlat Callimachi (the last Phanariot ruler in Bucharest). Ioan Sturdza's rule in Moldavia and Grigore IV Ghica's in Wallachia are considered the first of the new period, although the new regime abruptly ended in Russian occupation during another Russo−Turkish War and the subsequent period of Russian influence.
Most Phanariots were patrons of Greek culture, education and printing. They founded academies which attracted teachers and pupils from throughout the Orthodox commonwealth, and there was awareness of intellectual trends in Habsburg Europe. Many of the Phanariot princes were capable, farsighted rulers. As prince of Wallachia in 1746 and Moldavia in 1749, Constantine Mavrocordatos abolished serfdom and Alexandros Ipsilantis of Wallachia (reigned 1774–1782) initiated extensive administrative and legal reforms. Ipsilanti's reign coincided with subtle shifts in economic and social life and the emergence of spiritual and intellectual aspirations which pointed to the West and reform.
Condemnation of the Phanariots is a focus of Romanian nationalism, usually integrated into a general resentment of foreigners. The tendency unifies pro− and anti−modernisation attitudes; Phanariot Greeks are painted as reactionary elements (by Communist Romania) and agents of brutal, opportunistic change (as in Mihai Eminescu's Scrisoarea a III-a).
Extant Phanariot families
Here is a non-exhaustive list of Phanariot families:
Angelos, imperial family originally from central Philadelphia, Byzantine emperors.
Argyropoulos, noble family of the Byzantine Empire also known as Argyros, see John Argyropoulos.
Athanasovici
Callimachi family, also known as Călmașu, Kalmaşu or Kallimaşu, originally a Romanian boyar family from Moldavia.
Callivazis, originally from Trebizond, relocated to the Russian Empire.
Canellou
Cantacuzino, claimed to be originated from the Byzantine noble family Kantakouzenos.
Caradjas, also known as Caragea or Karatzas.
Caratheodoris family, see also Constantin Carathéodory
Cariophyllis
Chrisoscoleos
Chrisovergis, also known as Hrisovergis, from the Peloponnese
Diamandis
Doukas, also known as Dukas, imperial family originally from Paphlagonia, Despotate of Epirus despots.
Evalidis, also known as Evaoglous, Hadjievalidis, from Trebizond
Gerakas, from Kefalonia
Geralis, from Mytilene and Kefalonia
Ghica family, originally Albanians from Macedonia
Hantzeris, also known as Handjeri, Hançeri, Pıçakçı and Hançeroglou, see Constantine Hangerli
Kavadas, from Chios
Komnenos, also known as Komnenus or Comnenos, including its cadet branches of Axouch, Axouchos or Afouxechos, from Trebizond, Byzantine and Trebizond emperors.
Lambrinos
Lapithis, from Crete
Lazaridis, also known as Lazarević, a Serbianized family originally from Montenegro.
Lermis, also known as Lermioglous or Lermilis, from Pontus.
Levidis officials on the Patriarchate and dignitaries in the Imperial Court (Sublime Porte).
Mamonas
Manos, originated from Kastoria, see Aspasia Manos
Mavrocordatos, from Chios, see Alexandros Mavrokordatos.
Mavrogenis, from Paros, see Manto Mavrogenous.
Mavroudis
Mourouzis family, see Alexander Mourouzis
Musurus, see Marcus Musurus
Palaiologos, imperial family originally from central Asia minor, later marquesses of Montferrat
Photeinos
Philanthropenos, noble family of the Byzantine Empire.
Rallis, from Chios, later a political family in the Hellenic Republic.
Rizos Rangavis, see Alexandros Rizos Rangavis
Racoviță, also known as Racovitza, Romanian noble family from Moldavia and Wallachia.
Ramalo
Rodocanachi
Romalo
Rosetti family, also known as Ruset or Russeti, Moldavian Boyar family of Byzantine and Genoan origins.
Scanavis
Schinas
Sereslis
Soutzos family, also known as Suțu, Sutzu or Sütçü, see Michael Soutzos.
Tzanavarakis, also known Tzanavaris, Çanavaris or Canavaroğulları.
Venturas
Vlachoutzis
Văcărescu family, Romanian boyars from Wallachia and the first poets in Romanian literature
Vlastos, from Crete
Ypsilantis, from Trebizond, see Alexander Ypsilantis and Demetrios Ypsilantis
Extinct Phanariot families
Aristarchis
Ballasakis
Cananos
Caryophyles
Dimakis
Eupragiotes
Iancoleos (della Rocca)
Moronas
Negris
Paladas, from Crete
Plaginos
Rizos Neroulos
Ramadan
Souldjaroglou
Tzoukes
See also
Ottoman Greeks
Diafotismos
Greeks in Romania
Bulgarian Exarchate
Early Modern Romania
Danubian Principalities
List of rulers of Moldavia
List of rulers of Wallachia
History of the Russo-Turkish wars
State organisation of the Ottoman Empire
Notes
References
Mihai Berza, "Haraciul Moldovei și al Țării Românești în sec. XV–XIX", in Studii și Materiale de Istorie Medie, II, 1957, p. 7–47
Alex Drace-Francis, The Making of Modern Romanian Culture, London & New York, 2006,
Neagu Djuvara, Între Orient și Occident. Țările române la începutul epocii moderne, Humanitas, Bucharest, 1995
Vlad Georgescu, Istoria ideilor politice românești (1369–1878), Munich, 1987
Eric Hobsbawm, Age of Revolutions, section "Greek War of Independence"
Konstantinos Paparrigopoulos (Pavlos Karolidis), History of the Hellenic Nation (Volume Eb), Eleftheroudakis, Athens, 1925
L. S. Stavrianos, The Balkans Since 1453
Ottoman culture
Christianity in the Ottoman Empire
Demographics of the Ottoman Empire
People from the Ottoman Empire by ethnic or national origin
Government of the Ottoman Empire
Greek diaspora
History of Moldavia (1711–1822)
History of Wallachia (1714–1821)
Ottoman period in the Balkans
Greece–Turkey relations
Eastern Orthodox Christian culture
|
I with ogonek (majuscule: Į, minuscule: į) is a letter of the Latin alphabet formed by addition of the ogonek to the letter I. It is used in Lithuanian, Western Apache, Chipewyan, Mescalero-Chiricahua, Muscogee, Dadibi, Dalecarlian, Gwichʼin, Hän, Iñapari, Kaska, Navajo, Sierra Otomi, Sekani, Tagish, Tlingit, Tutchone, Winnebago, and Ixtlán Zapotec.
Usage
In Lithuanian, it is the 14th letter of the alphabet, and is pronounced as long close front unrounded vowel ([iː]). In the past, the letter was used to denote the nasalized close front unrounded vowel ([ĩ]). Currently, it appears in the words that used to be nasalized in the past, for example in įkalnė, which means uphill.
was also used in the Latin alphabet of the Khakas language between 1929-1939, representing the sound . The current Cyrillic alphabet uses the dotted I for the same sound.
The letter also appears in various Indigenous languages of North America, which are: Western Apache, Chipewyan, Mescalero-Chiricahua, Muscogee, Dadibi, Dalecarlian, Gwichʼin, Hän, Iñapari, Kaska, Navajo, Sierra Otomi, Sekani, Tagish, Tlingit, Tutchone, Winnebago, and Ixtlán Zapotec. In most of them, the letter represent the nasalized close front unrounded vowel ([ĩ]).
Encoding
References
Latin letters with diacritics
|
Mysore painting () is an important form of classical South Indian painting style that originated in and around the town of Mysore in Karnataka encouraged and nurtured by the Mysore rulers. Painting in Karnataka has a long and illustrious history, tracing its origins back to the Ajanta Caves period (2nd century BC to 6th century AD). The distinct school of Mysore painting evolved from the paintings during the Vijayanagara Empire period, the rulers of Vijayanagara and their feudatories encouraged literature, art and architecture as well as religious and philosophical discussions. With the fall of the Vijayanagara Empire after the 1565 Battle of Talikota, the artists who were until then under royal patronage migrated to various places such as Mysore, Tanjore and Surpur among others. Absorbing the local artistic traditions and customs, the erstwhile Vijayanagara school of painting gradually evolved into the many styles of painting in South India, including the Mysore and Tanjore schools of painting.
Mysore paintings are known for their elegance, muted colours, and attention to detail. The themes for most of these paintings are Hindu gods and goddesses and scenes from Hindu mythology.
History
The fall of the Vijayanagar Empire in 1565 AD and the sack of Hampi in the Battle of Talikota resulted initially in distress for scores of families of painters who had been dependent on the patronage of the empire. As Dr. Charita points out, these families of artists, called Chitrakaras, migrated to various pockets (feudatories) of the Vijayanagara Empire. As scholar A.L. Narasimhan traces, some of these surviving paintings are witnessed at Shravanabelagola, Sira, Keregoodirangapura, Srirangapattana, Nippani, Sibi, Naragunda, Bettadapura, Hardanahalli, Mudukutore, Mysore, Chitradurga, Kollegala, Raichur, Hiriyur, Benakanakere, Anegundi, Yalladahalli, Lepakshi and many other places located in Karnataka.
Raja Wodeyar I (1578–1617 A.D) provided a vital service to the cause of painting by rehabilitating several families of painters of the Vijayanagara School at Srirangapatna.
The successors of Raja Wodeyar continued to patronize the art of painting by commissioning temples and palaces to be painted with mythological scenes. However, none of these paintings have survived due to the ravages of war between the British on the one side and Hyder Ali and Tippu Sultan on the other. Hyder and Tippu who bested the Wodeyars took over the reins of Mysore for a brief period. However, the artists (Chitragars) continued to be patronised and flourished under the reign of Tipu and Hyder too. The Narasimha swamy temple in Seebi on the highway between Tumkur and Sira was built by Nallappa who was in the service of both Haidar Ali and Tipu Sultan, during Tipu's reign and has several wonderful wall frescoes in the Vijayanagar style which gradually evolved into the Mysore and Tanjore schools of painting. The murals detailing the Battle of Polilur and other painted work at the Daria Daulat Bagh palace of Tipu Sultan in Ganjam, Srirangapatna are also prime examples of the Mysore school of painting.
After the death of Tipu Sultan in 1799 AD, the state was restored back to the Wodeyars of Mysore and its ruler Mummadi Krishnaraja Wodeyar III (1799-1868 AD) who was contemporaneous with Serfoji II of Thanjavur. This ushered in a new era by reviving the ancient traditions of Mysore and extending patronage to music, sculpture, painting, dancing and literature. Most of the traditional paintings of the Mysore School, which have survived until today, belong to this reign. Furthermore, Krishnaraja Wodeyar provided new fillip to the artists of the Mysore school through his Magnum Opus Sritattvanidhi, which would remain the ready reckoner on Mysore style for many years to come. On the walls of Jagan Mohan Palace, Mysore (Karnataka), the fascinating range of paintings which flourished under Krishnaraja Wodeyar can be seen: from portraits of the Mysore rulers, their family members and important personages in Indian history, through self-portraits of the artists themselves which Krishnaraja Wodeyar coaxed them to paint, to murals depicting the Hindu pantheon and Puranic and mythological scenes.
Literary and Inscriptional
The most famous of the manuscripts detailing the various nuances of the Mysore school and listing out the various Gods and Goddesses, is the Sritattvanidhi, a voluminous
work of 1500 pages prepared under the patronage of Mummadi Krishnaraja Wodeyar. This pictorial digest is a compendium of illustrations of gods, goddesses and mythological figures with instructions to painters on an incredible range of topics concerning composition placement, colour choice, individual attributes and mood. The ragas, seasons, eco-happenings, animals, and plant world are also effectively depicted in these paintings as co-themes or contexts.
Other Sanskrit literary sources such as the Visnudharmottara Purana, Abhilasitarthacintamani and Sivatatvaratnakara also throw light on the objectives and principles of painting, methods of preparing pigments, brushes and the carrier, qualifications of the chitrakar (traditional community of painters) the principles of painting and the technique to be followed.
Materials
The ancient painters in Mysore prepared their own materials. The colours were from natural sources and were of vegetable or mineral substances such as leaves, stones and flowers. Brushes were made with squirrel hairs for delicate work, but for drawing superfine lines, a brush made of pointed blades of a special variety of grass had to be used. Due to the long-lasting quality of the stone- and plant-based colours used, the original Mysore paintings retain their freshness and lustre even today.
Technique and characteristic
Mysore Paintings are characterized by delicate lines, intricate brush strokes,
graceful delineation of figures and the discreet use of bright vegetable colours
and lustrous gold leaf. More than mere decorative pieces, the paintings are
designed to inspire feelings of devotion and humility in the viewer. The painter’s
individual skill in giving expression to various emotions is therefore of paramount
importance to this style of painting.
The first stage of Mysore Painting was to prepare the ground; paper, wood, cloth or wall grounds were variously used. The paper board was made of paper pulp or waste paper, which was dried in the sun and then rubbed smooth with a polished quartz pebble. If the ground was cloth it was pasted on a wooden board using a paste composed of dry white lead (safeda) mixed with gum and a small quantity of gruel (ganji). The board was then dried and burnished. Wood surfaces were prepared by
applying dry white lead, yellow ochre and gum, and walls were treated with yellow ochre, chalk and gum. After preparation of the ground a rough sketch of the picture was drawn with crayon prepared from the straight twigs of the tamarind tree. The next step was to paint the furthest objects such as sky, hill and river and then gradually animal and human figures were approached in greater detail. After colouring the figures, the artists would turn to elaboration of the faces, dress and ornaments including the gesso work (gold covering), which is an important feature of Mysore painting.
Gesso work
Gesso work was the hallmark of all traditional paintings of Karnataka. Gesso refers to the paste mixture of white lead powder, gambose and glue which is used as an embossing material and covered with gold foil. The gesso work in Mysore paintings is low in relief and intricate as compared to the thick gold relief work of the Tanjore School. Gesso was used in Mysore painting for depicting intricate designs of clothes, jewellery and architectural details on pillars and arches that usually framed the deities. The work was taken up in the morning when the base of the gold work on the painting was still moist so as to hold the gold foil firmly. After allowing the painting to dry, glazing was carried out by covering the painting with thin paper and rubbing over it with a soft glazing stone known as kaslupada kallu.
When the thin paper was removed the painting shone brightly and looked resplendent with the combination of gold and a variety of colours.
See also
Mysore pak
Mysore Agarbathi
Mysore Sandal Soap
Mysore Sandalwood Oil
Channapatna toys
References
Further reading
(see index: p. 148-152)
External links
Karnataka year book, Government of Karnataka.
History and facts about Mysore paintings
Schools of Indian painting
Culture of Karnataka
Culture of Mysore
Geographical indications in Karnataka
|
Aspella producta is a species of sea snail, a marine gastropod mollusc in the family Muricidae, the murex snails or rock snails.
Description
Distribution
References
Gastropods described in 1861
Aspella
|
The 2013 Regional Women's Twenty20 Championship was the second season of the women's Twenty20 cricket competition played in the West Indies. It took place in August 2013, with 8 teams taking part and all matches taking place in Grenada. Jamaica won the tournament, beating Barbados in the final to claim their second T20 title.
The tournament followed the 2013 Regional Women's Championship.
Competition format
The eight teams were divided into two groups of four, playing in a round-robin format. Matches were played using a Twenty20 format. The top two teams in each group progressed to the semi-finals, whilst the bottom two teams in each group played-off in two play-off matches.
The group worked on a points system with positions being based on the total points. Points were awarded as follows:
Win: 5 points
Loss: 0 points.
Abandoned/No Result: 3 points.
Points tables
Group A
Group B
Source: Windies Cricket
Knock-Out Stage
Play-Offs
Semi-finals
Final
Statistics
Most runs
Source: CricketArchive
Most wickets
Source: CricketArchive
References
External links
Series home at Windies Cricket
Twenty20 Blaze
2013 in West Indian cricket
|
```objective-c
#ifndef __CARES_RULES_H
#define __CARES_RULES_H
*
* Permission to use, copy, modify, and distribute this software and its
* documentation for any purpose and without fee is hereby granted, provided
* that the above copyright notice appear in all copies and that both that
* copyright notice and this permission notice appear in supporting
* documentation, and that the name of M.I.T. not be used in advertising or
* publicity pertaining to distribution of the software without specific,
* written prior permission. M.I.T. makes no representations about the
* suitability of this software for any purpose. It is provided "as is"
* without express or implied warranty.
*/
/* ================================================================ */
/* COMPILE TIME SANITY CHECKS */
/* ================================================================ */
/*
* NOTE 1:
* -------
*
* All checks done in this file are intentionally placed in a public
* header file which is pulled by ares.h when an application is
* being built using an already built c-ares library. Additionally
* this file is also included and used when building the library.
*
* If compilation fails on this file it is certainly sure that the
* problem is elsewhere. It could be a problem in the ares_build.h
* header file, or simply that you are using different compilation
* settings than those used to build the library.
*
* Nothing in this file is intended to be modified or adjusted by the
* c-ares library user nor by the c-ares library builder.
*
* Do not deactivate any check, these are done to make sure that the
* library is properly built and used.
*
* You can find further help on the c-ares development mailing list:
* path_to_url
*
* NOTE 2
* ------
*
* Some of the following compile time checks are based on the fact
* that the dimension of a constant array can not be a negative one.
* In this way if the compile time verification fails, the compilation
* will fail issuing an error. The error description wording is compiler
* dependent but it will be quite similar to one of the following:
*
* "negative subscript or subscript is too large"
* "array must have at least one element"
* "-1 is an illegal array size"
* "size of array is negative"
*
* If you are building an application which tries to use an already
* built c-ares library and you are getting this kind of errors on
* this file, it is a clear indication that there is a mismatch between
* how the library was built and how you are trying to use it for your
* application. Your already compiled or binary library provider is the
* only one who can give you the details you need to properly use it.
*/
/*
* Verify that some macros are actually defined.
*/
#ifndef CARES_TYPEOF_ARES_SOCKLEN_T
# error "CARES_TYPEOF_ARES_SOCKLEN_T definition is missing!"
Error Compilation_aborted_CARES_TYPEOF_ARES_SOCKLEN_T_is_missing
#endif
#ifndef CARES_SIZEOF_ARES_SOCKLEN_T
# error "CARES_SIZEOF_ARES_SOCKLEN_T definition is missing!"
Error Compilation_aborted_CARES_SIZEOF_ARES_SOCKLEN_T_is_missing
#endif
/*
* Macros private to this header file.
*/
#define CareschkszEQ(t, s) sizeof(t) == s ? 1 : -1
#define CareschkszGE(t1, t2) sizeof(t1) >= sizeof(t2) ? 1 : -1
/*
* Verify that the size previously defined and expected for
* ares_socklen_t is actually the the same as the one reported
* by sizeof() at compile time.
*/
typedef char
__cares_rule_02__
[CareschkszEQ(ares_socklen_t, CARES_SIZEOF_ARES_SOCKLEN_T)];
/*
* Verify at compile time that the size of ares_socklen_t as reported
* by sizeof() is greater or equal than the one reported for int for
* the current compilation.
*/
typedef char
__cares_rule_03__
[CareschkszGE(ares_socklen_t, int)];
/* ================================================================ */
/* EXTERNALLY AND INTERNALLY VISIBLE DEFINITIONS */
/* ================================================================ */
/*
* Get rid of macros private to this header file.
*/
#undef CareschkszEQ
#undef CareschkszGE
/*
* Get rid of macros not intended to exist beyond this point.
*/
#undef CARES_PULL_WS2TCPIP_H
#undef CARES_PULL_SYS_TYPES_H
#undef CARES_PULL_SYS_SOCKET_H
#undef CARES_TYPEOF_ARES_SOCKLEN_T
#endif /* __CARES_RULES_H */
```
|
The surname Guillemin can refer to:
Philippe François Zéphirin Guillemin (1814-1886), Roman Catholic bishop
Amédée Guillemin (1826–1893), French science author and journalist
Amédée Henri Guillemin (1860-1941), French WWI general
Anne-Marie Guillemin (1868–1963), French translator of Latin texts
Ernst Guillemin (1898–1970), American electrical engineer and computer scientist
Francisco Romano Guillemin (1884–1950), Mexican painter
Jean Baptiste Antoine Guillemin (1796–1842), French botanist
Jean-Philippe Guillemin (born 1972), French Open-source computer programmer and musician
Jeanne Guillemin (1943–2019), American medical anthropologist
Roger Guillemin (born 1924), neurologist and Nobel laureate in medicine
Victor Guillemin (born 1937), American mathematician
Robert Charles Guillemin (born 1939), ephemeral artist known as Sidewalk Sam
|
```xml
import { Token } from './token'
import { ValueToken } from './value-token'
import { TokenKind } from '../parser'
export class RangeToken extends Token {
constructor (
public input: string,
public begin: number,
public end: number,
public lhs: ValueToken,
public rhs: ValueToken,
public file?: string
) {
super(TokenKind.Range, input, begin, end, file)
}
}
```
|
Peter A. Freeman is the founding dean of Georgia Tech's College of Computing, a position he held from 1990 to 2002. Freeman was assistant director of the National Science Foundation from 2002 to 2007.
Freeman has been emeritus dean of the Georgia Tech College of Computing since 2007. He is currently the director of the Washington Advisory Group. Freeman is a Fellow of the Institute of Electrical and Electronics Engineers, the Association for Computing Machinery, and the American Association for the Advancement of Science.
Education
Freeman received a Bachelor of Arts in physics and mathematics from Rice University in 1963, a Master of Arts in mathematics and psychology from the University of Texas at Austin in 1965, and a Ph.D. in computer science from Carnegie Mellon University in 1970.
Early career
After graduation from Carnegie Mellon, Freeman was a researcher, professor, and administrator at the University of California, Irvine's Department of Information & Computer Science from 1971 to 1987, where he focused on artificial intelligence and software engineering research, funded primarily by the National Science Foundation. Freeman is one of the founders of software engineering education, starting a course on the subject in 1974. During the course of his career at UC Irvine, Freeman also administered the undergraduate and graduate programs and served as the department chair.
In 1987, Freeman became the director of the National Science Foundation's Division of Computer and Computation Research; in this post, he managed about $20 million in grants each year, and helped develop the High Performance Computing and Communications program.
For a year after his post at NSF, Freeman served as a Visiting Distinguished Professor at George Mason University (1989–1990) to assist in the creation of a computing program at that university.
Recent career
From 1992 to 1995, Freeman became the chief information officer of the Georgia Institute of Technology, and oversaw their Office of Information Technology (OIT) as it prepared for the 1996 Summer Olympics. At the time, OIT had an annual budget of $10 million.
When Georgia Tech was reorganized in 1988 under the administration of John Patrick Crecine (see History of Georgia Tech), the School of Information and Computer Science was promoted to college status, on the same administrative level as Georgia Tech's mainstay, the College of Engineering. Thus the Georgia Tech College of Computing was created. Freeman was selected as the college's founding dean and oversaw the program's transition and subsequent rapid growth; he would hold this post until 2002. Under his administration, the school added three research centers (including the Georgia Tech Information Security Center), increased research funding from $2 million annually to $10 million annually, hired 55 faculty members, and secured millions in funding for the construction of the College of Computing Building.
From 2002 to 2007 Freeman was held the position of Georgia Tech professor while he served as one of seven assistant directors of the National Science Foundation. At NSF, he led the $1 billion Information Technology Research program, helped form the GENI Project, led the Cyber-enabled Discovery and Innovation Initiative, and led initiatives to revitalize computer science education and attract minorities to computing.
In 2007, Freeman was named Emeritus Dean of the Georgia Tech College of Computing; in this position he consults with current administrators, faculty, and students; gives lectures and writes op-eds promoting issues in computer science.
Since 2007, Freeman has been the director of the Washington Advisory Group, where he advises on policy issues related to secondary education and research in STEM fields to companies, universities, governments, and non-profit organizations.
References
Living people
Rice University alumni
University of Texas at Austin alumni
Carnegie Mellon University alumni
Georgia Tech faculty
University of California, Irvine faculty
United States National Science Foundation officials
George Mason University faculty
American computer scientists
Fellow Members of the IEEE
Fellows of the Association for Computing Machinery
Fellows of the American Association for the Advancement of Science
1941 births
Chief information officers
|
The 2000 FA Women's Charity Shield was the first Women's FA Community Shield, as with its male equivalent, the Community Shield is an annual football match played between the winners of the previous season's league and the previous season's Women's FA Cup. The match was contested between Arsenal and Charlton Athletic, it ended 1-1. It is the only time that the trophy was shared between two teams.
References
Women's FA Community Shield
Community Shield
Community Shield
Community Shield
Community Shield
|
```objective-c
/**
*
*/
#pragma once
#include "soc.h"
#ifdef __cplusplus
extern "C" {
#endif
#define REG_TIMG_BASE(i) (DR_REG_TIMERGROUP0_BASE + i*0x1000)
#define TIMG_T0CONFIG_REG(i) (REG_TIMG_BASE(i) + 0x0000)
/* TIMG_T0_EN : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: When set timer 0 time-base counter is enabled*/
#define TIMG_T0_EN (BIT(31))
#define TIMG_T0_EN_M (BIT(31))
#define TIMG_T0_EN_V 0x1
#define TIMG_T0_EN_S 31
/* TIMG_T0_INCREASE : R/W ;bitpos:[30] ;default: 1'h1 ; */
/*description: When set timer 0 time-base counter increment. When cleared timer
0 time-base counter decrement.*/
#define TIMG_T0_INCREASE (BIT(30))
#define TIMG_T0_INCREASE_M (BIT(30))
#define TIMG_T0_INCREASE_V 0x1
#define TIMG_T0_INCREASE_S 30
/* TIMG_T0_AUTORELOAD : R/W ;bitpos:[29] ;default: 1'h1 ; */
/*description: When set timer 0 auto-reload at alarming is enabled*/
#define TIMG_T0_AUTORELOAD (BIT(29))
#define TIMG_T0_AUTORELOAD_M (BIT(29))
#define TIMG_T0_AUTORELOAD_V 0x1
#define TIMG_T0_AUTORELOAD_S 29
/* TIMG_T0_DIVIDER : R/W ;bitpos:[28:13] ;default: 16'h1 ; */
/*description: Timer 0 clock (T0_clk) prescale value.*/
#define TIMG_T0_DIVIDER 0x0000FFFF
#define TIMG_T0_DIVIDER_M ((TIMG_T0_DIVIDER_V)<<(TIMG_T0_DIVIDER_S))
#define TIMG_T0_DIVIDER_V 0xFFFF
#define TIMG_T0_DIVIDER_S 13
/* TIMG_T0_EDGE_INT_EN : R/W ;bitpos:[12] ;default: 1'h0 ; */
/*description: When set edge type interrupt will be generated during alarm*/
#define TIMG_T0_EDGE_INT_EN (BIT(12))
#define TIMG_T0_EDGE_INT_EN_M (BIT(12))
#define TIMG_T0_EDGE_INT_EN_V 0x1
#define TIMG_T0_EDGE_INT_EN_S 12
/* TIMG_T0_LEVEL_INT_EN : R/W ;bitpos:[11] ;default: 1'h0 ; */
/*description: When set level type interrupt will be generated during alarm*/
#define TIMG_T0_LEVEL_INT_EN (BIT(11))
#define TIMG_T0_LEVEL_INT_EN_M (BIT(11))
#define TIMG_T0_LEVEL_INT_EN_V 0x1
#define TIMG_T0_LEVEL_INT_EN_S 11
/* TIMG_T0_ALARM_EN : R/W ;bitpos:[10] ;default: 1'h0 ; */
/*description: When set alarm is enabled*/
#define TIMG_T0_ALARM_EN (BIT(10))
#define TIMG_T0_ALARM_EN_M (BIT(10))
#define TIMG_T0_ALARM_EN_V 0x1
#define TIMG_T0_ALARM_EN_S 10
#define TIMG_T0LO_REG(i) (REG_TIMG_BASE(i) + 0x0004)
/* TIMG_T0_LO : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Register to store timer 0 time-base counter current value lower 32 bits.*/
#define TIMG_T0_LO 0xFFFFFFFF
#define TIMG_T0_LO_M ((TIMG_T0_LO_V)<<(TIMG_T0_LO_S))
#define TIMG_T0_LO_V 0xFFFFFFFF
#define TIMG_T0_LO_S 0
#define TIMG_T0HI_REG(i) (REG_TIMG_BASE(i) + 0x0008)
/* TIMG_T0_HI : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Register to store timer 0 time-base counter current value higher 32 bits.*/
#define TIMG_T0_HI 0xFFFFFFFF
#define TIMG_T0_HI_M ((TIMG_T0_HI_V)<<(TIMG_T0_HI_S))
#define TIMG_T0_HI_V 0xFFFFFFFF
#define TIMG_T0_HI_S 0
#define TIMG_T0UPDATE_REG(i) (REG_TIMG_BASE(i) + 0x000c)
/* TIMG_T0_UPDATE : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Write any value will trigger a timer 0 time-base counter value
update (timer 0 current value will be stored in registers above)*/
#define TIMG_T0_UPDATE 0xFFFFFFFF
#define TIMG_T0_UPDATE_M ((TIMG_T0_UPDATE_V)<<(TIMG_T0_UPDATE_S))
#define TIMG_T0_UPDATE_V 0xFFFFFFFF
#define TIMG_T0_UPDATE_S 0
#define TIMG_T0ALARMLO_REG(i) (REG_TIMG_BASE(i) + 0x0010)
/* TIMG_T0_ALARM_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Timer 0 time-base counter value lower 32 bits that will trigger the alarm*/
#define TIMG_T0_ALARM_LO 0xFFFFFFFF
#define TIMG_T0_ALARM_LO_M ((TIMG_T0_ALARM_LO_V)<<(TIMG_T0_ALARM_LO_S))
#define TIMG_T0_ALARM_LO_V 0xFFFFFFFF
#define TIMG_T0_ALARM_LO_S 0
#define TIMG_T0ALARMHI_REG(i) (REG_TIMG_BASE(i) + 0x0014)
/* TIMG_T0_ALARM_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Timer 0 time-base counter value higher 32 bits that will trigger the alarm*/
#define TIMG_T0_ALARM_HI 0xFFFFFFFF
#define TIMG_T0_ALARM_HI_M ((TIMG_T0_ALARM_HI_V)<<(TIMG_T0_ALARM_HI_S))
#define TIMG_T0_ALARM_HI_V 0xFFFFFFFF
#define TIMG_T0_ALARM_HI_S 0
#define TIMG_T0LOADLO_REG(i) (REG_TIMG_BASE(i) + 0x0018)
/* TIMG_T0_LOAD_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Lower 32 bits of the value that will load into timer 0 time-base counter*/
#define TIMG_T0_LOAD_LO 0xFFFFFFFF
#define TIMG_T0_LOAD_LO_M ((TIMG_T0_LOAD_LO_V)<<(TIMG_T0_LOAD_LO_S))
#define TIMG_T0_LOAD_LO_V 0xFFFFFFFF
#define TIMG_T0_LOAD_LO_S 0
#define TIMG_T0LOADHI_REG(i) (REG_TIMG_BASE(i) + 0x001c)
/* TIMG_T0_LOAD_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: higher 32 bits of the value that will load into timer 0 time-base counter*/
#define TIMG_T0_LOAD_HI 0xFFFFFFFF
#define TIMG_T0_LOAD_HI_M ((TIMG_T0_LOAD_HI_V)<<(TIMG_T0_LOAD_HI_S))
#define TIMG_T0_LOAD_HI_V 0xFFFFFFFF
#define TIMG_T0_LOAD_HI_S 0
#define TIMG_T0LOAD_REG(i) (REG_TIMG_BASE(i) + 0x0020)
/* TIMG_T0_LOAD : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Write any value will trigger timer 0 time-base counter reload*/
#define TIMG_T0_LOAD 0xFFFFFFFF
#define TIMG_T0_LOAD_M ((TIMG_T0_LOAD_V)<<(TIMG_T0_LOAD_S))
#define TIMG_T0_LOAD_V 0xFFFFFFFF
#define TIMG_T0_LOAD_S 0
#define TIMG_T1CONFIG_REG(i) (REG_TIMG_BASE(i) + 0x0024)
/* TIMG_T1_EN : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: When set timer 1 time-base counter is enabled*/
#define TIMG_T1_EN (BIT(31))
#define TIMG_T1_EN_M (BIT(31))
#define TIMG_T1_EN_V 0x1
#define TIMG_T1_EN_S 31
/* TIMG_T1_INCREASE : R/W ;bitpos:[30] ;default: 1'h1 ; */
/*description: When set timer 1 time-base counter increment. When cleared timer
1 time-base counter decrement.*/
#define TIMG_T1_INCREASE (BIT(30))
#define TIMG_T1_INCREASE_M (BIT(30))
#define TIMG_T1_INCREASE_V 0x1
#define TIMG_T1_INCREASE_S 30
/* TIMG_T1_AUTORELOAD : R/W ;bitpos:[29] ;default: 1'h1 ; */
/*description: When set timer 1 auto-reload at alarming is enabled*/
#define TIMG_T1_AUTORELOAD (BIT(29))
#define TIMG_T1_AUTORELOAD_M (BIT(29))
#define TIMG_T1_AUTORELOAD_V 0x1
#define TIMG_T1_AUTORELOAD_S 29
/* TIMG_T1_DIVIDER : R/W ;bitpos:[28:13] ;default: 16'h1 ; */
/*description: Timer 1 clock (T1_clk) prescale value.*/
#define TIMG_T1_DIVIDER 0x0000FFFF
#define TIMG_T1_DIVIDER_M ((TIMG_T1_DIVIDER_V)<<(TIMG_T1_DIVIDER_S))
#define TIMG_T1_DIVIDER_V 0xFFFF
#define TIMG_T1_DIVIDER_S 13
/* TIMG_T1_EDGE_INT_EN : R/W ;bitpos:[12] ;default: 1'h0 ; */
/*description: When set edge type interrupt will be generated during alarm*/
#define TIMG_T1_EDGE_INT_EN (BIT(12))
#define TIMG_T1_EDGE_INT_EN_M (BIT(12))
#define TIMG_T1_EDGE_INT_EN_V 0x1
#define TIMG_T1_EDGE_INT_EN_S 12
/* TIMG_T1_LEVEL_INT_EN : R/W ;bitpos:[11] ;default: 1'h0 ; */
/*description: When set level type interrupt will be generated during alarm*/
#define TIMG_T1_LEVEL_INT_EN (BIT(11))
#define TIMG_T1_LEVEL_INT_EN_M (BIT(11))
#define TIMG_T1_LEVEL_INT_EN_V 0x1
#define TIMG_T1_LEVEL_INT_EN_S 11
/* TIMG_T1_ALARM_EN : R/W ;bitpos:[10] ;default: 1'h0 ; */
/*description: When set alarm is enabled*/
#define TIMG_T1_ALARM_EN (BIT(10))
#define TIMG_T1_ALARM_EN_M (BIT(10))
#define TIMG_T1_ALARM_EN_V 0x1
#define TIMG_T1_ALARM_EN_S 10
#define TIMG_T1LO_REG(i) (REG_TIMG_BASE(i) + 0x0028)
/* TIMG_T1_LO : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Register to store timer 1 time-base counter current value lower 32 bits.*/
#define TIMG_T1_LO 0xFFFFFFFF
#define TIMG_T1_LO_M ((TIMG_T1_LO_V)<<(TIMG_T1_LO_S))
#define TIMG_T1_LO_V 0xFFFFFFFF
#define TIMG_T1_LO_S 0
#define TIMG_T1HI_REG(i) (REG_TIMG_BASE(i) + 0x002c)
/* TIMG_T1_HI : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Register to store timer 1 time-base counter current value higher 32 bits.*/
#define TIMG_T1_HI 0xFFFFFFFF
#define TIMG_T1_HI_M ((TIMG_T1_HI_V)<<(TIMG_T1_HI_S))
#define TIMG_T1_HI_V 0xFFFFFFFF
#define TIMG_T1_HI_S 0
#define TIMG_T1UPDATE_REG(i) (REG_TIMG_BASE(i) + 0x0030)
/* TIMG_T1_UPDATE : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Write any value will trigger a timer 1 time-base counter value
update (timer 1 current value will be stored in registers above)*/
#define TIMG_T1_UPDATE 0xFFFFFFFF
#define TIMG_T1_UPDATE_M ((TIMG_T1_UPDATE_V)<<(TIMG_T1_UPDATE_S))
#define TIMG_T1_UPDATE_V 0xFFFFFFFF
#define TIMG_T1_UPDATE_S 0
#define TIMG_T1ALARMLO_REG(i) (REG_TIMG_BASE(i) + 0x0034)
/* TIMG_T1_ALARM_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Timer 1 time-base counter value lower 32 bits that will trigger the alarm*/
#define TIMG_T1_ALARM_LO 0xFFFFFFFF
#define TIMG_T1_ALARM_LO_M ((TIMG_T1_ALARM_LO_V)<<(TIMG_T1_ALARM_LO_S))
#define TIMG_T1_ALARM_LO_V 0xFFFFFFFF
#define TIMG_T1_ALARM_LO_S 0
#define TIMG_T1ALARMHI_REG(i) (REG_TIMG_BASE(i) + 0x0038)
/* TIMG_T1_ALARM_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Timer 1 time-base counter value higher 32 bits that will trigger the alarm*/
#define TIMG_T1_ALARM_HI 0xFFFFFFFF
#define TIMG_T1_ALARM_HI_M ((TIMG_T1_ALARM_HI_V)<<(TIMG_T1_ALARM_HI_S))
#define TIMG_T1_ALARM_HI_V 0xFFFFFFFF
#define TIMG_T1_ALARM_HI_S 0
#define TIMG_T1LOADLO_REG(i) (REG_TIMG_BASE(i) + 0x003c)
/* TIMG_T1_LOAD_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Lower 32 bits of the value that will load into timer 1 time-base counter*/
#define TIMG_T1_LOAD_LO 0xFFFFFFFF
#define TIMG_T1_LOAD_LO_M ((TIMG_T1_LOAD_LO_V)<<(TIMG_T1_LOAD_LO_S))
#define TIMG_T1_LOAD_LO_V 0xFFFFFFFF
#define TIMG_T1_LOAD_LO_S 0
#define TIMG_T1LOADHI_REG(i) (REG_TIMG_BASE(i) + 0x0040)
/* TIMG_T1_LOAD_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: higher 32 bits of the value that will load into timer 1 time-base counter*/
#define TIMG_T1_LOAD_HI 0xFFFFFFFF
#define TIMG_T1_LOAD_HI_M ((TIMG_T1_LOAD_HI_V)<<(TIMG_T1_LOAD_HI_S))
#define TIMG_T1_LOAD_HI_V 0xFFFFFFFF
#define TIMG_T1_LOAD_HI_S 0
#define TIMG_T1LOAD_REG(i) (REG_TIMG_BASE(i) + 0x0044)
/* TIMG_T1_LOAD : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Write any value will trigger timer 1 time-base counter reload*/
#define TIMG_T1_LOAD 0xFFFFFFFF
#define TIMG_T1_LOAD_M ((TIMG_T1_LOAD_V)<<(TIMG_T1_LOAD_S))
#define TIMG_T1_LOAD_V 0xFFFFFFFF
#define TIMG_T1_LOAD_S 0
#define TIMG_WDTCONFIG0_REG(i) (REG_TIMG_BASE(i) + 0x0048)
/* TIMG_WDT_EN : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: When set SWDT is enabled*/
#define TIMG_WDT_EN (BIT(31))
#define TIMG_WDT_EN_M (BIT(31))
#define TIMG_WDT_EN_V 0x1
#define TIMG_WDT_EN_S 31
/* TIMG_WDT_STG0 : R/W ;bitpos:[30:29] ;default: 1'd0 ; */
/*description: Stage 0 configuration. 0: off 1: interrupt 2: reset CPU 3: reset system*/
#define TIMG_WDT_STG0 0x00000003
#define TIMG_WDT_STG0_M ((TIMG_WDT_STG0_V)<<(TIMG_WDT_STG0_S))
#define TIMG_WDT_STG0_V 0x3
#define TIMG_WDT_STG0_S 29
/* TIMG_WDT_STG1 : R/W ;bitpos:[28:27] ;default: 1'd0 ; */
/*description: Stage 1 configuration. 0: off 1: interrupt 2: reset CPU 3: reset system*/
#define TIMG_WDT_STG1 0x00000003
#define TIMG_WDT_STG1_M ((TIMG_WDT_STG1_V)<<(TIMG_WDT_STG1_S))
#define TIMG_WDT_STG1_V 0x3
#define TIMG_WDT_STG1_S 27
/* TIMG_WDT_STG2 : R/W ;bitpos:[26:25] ;default: 1'd0 ; */
/*description: Stage 2 configuration. 0: off 1: interrupt 2: reset CPU 3: reset system*/
#define TIMG_WDT_STG2 0x00000003
#define TIMG_WDT_STG2_M ((TIMG_WDT_STG2_V)<<(TIMG_WDT_STG2_S))
#define TIMG_WDT_STG2_V 0x3
#define TIMG_WDT_STG2_S 25
/* TIMG_WDT_STG3 : R/W ;bitpos:[24:23] ;default: 1'd0 ; */
/*description: Stage 3 configuration. 0: off 1: interrupt 2: reset CPU 3: reset system*/
#define TIMG_WDT_STG3 0x00000003
#define TIMG_WDT_STG3_M ((TIMG_WDT_STG3_V)<<(TIMG_WDT_STG3_S))
#define TIMG_WDT_STG3_V 0x3
#define TIMG_WDT_STG3_S 23
/* TIMG_WDT_EDGE_INT_EN : R/W ;bitpos:[22] ;default: 1'h0 ; */
/*description: When set edge type interrupt generation is enabled*/
#define TIMG_WDT_EDGE_INT_EN (BIT(22))
#define TIMG_WDT_EDGE_INT_EN_M (BIT(22))
#define TIMG_WDT_EDGE_INT_EN_V 0x1
#define TIMG_WDT_EDGE_INT_EN_S 22
/* TIMG_WDT_LEVEL_INT_EN : R/W ;bitpos:[21] ;default: 1'h0 ; */
/*description: When set level type interrupt generation is enabled*/
#define TIMG_WDT_LEVEL_INT_EN (BIT(21))
#define TIMG_WDT_LEVEL_INT_EN_M (BIT(21))
#define TIMG_WDT_LEVEL_INT_EN_V 0x1
#define TIMG_WDT_LEVEL_INT_EN_S 21
/* TIMG_WDT_CPU_RESET_LENGTH : R/W ;bitpos:[20:18] ;default: 3'h1 ; */
/*description: length of CPU reset selection. 0: 100ns 1: 200ns 2: 300ns
3: 400ns 4: 500ns 5: 800ns 6: 1.6us 7: 3.2us*/
#define TIMG_WDT_CPU_RESET_LENGTH 0x00000007
#define TIMG_WDT_CPU_RESET_LENGTH_M ((TIMG_WDT_CPU_RESET_LENGTH_V)<<(TIMG_WDT_CPU_RESET_LENGTH_S))
#define TIMG_WDT_CPU_RESET_LENGTH_V 0x7
#define TIMG_WDT_CPU_RESET_LENGTH_S 18
/* TIMG_WDT_SYS_RESET_LENGTH : R/W ;bitpos:[17:15] ;default: 3'h1 ; */
/*description: length of system reset selection. 0: 100ns 1: 200ns 2: 300ns
3: 400ns 4: 500ns 5: 800ns 6: 1.6us 7: 3.2us*/
#define TIMG_WDT_SYS_RESET_LENGTH 0x00000007
#define TIMG_WDT_SYS_RESET_LENGTH_M ((TIMG_WDT_SYS_RESET_LENGTH_V)<<(TIMG_WDT_SYS_RESET_LENGTH_S))
#define TIMG_WDT_SYS_RESET_LENGTH_V 0x7
#define TIMG_WDT_SYS_RESET_LENGTH_S 15
/* TIMG_WDT_FLASHBOOT_MOD_EN : R/W ;bitpos:[14] ;default: 1'h1 ; */
/*description: When set flash boot protection is enabled*/
#define TIMG_WDT_FLASHBOOT_MOD_EN (BIT(14))
#define TIMG_WDT_FLASHBOOT_MOD_EN_M (BIT(14))
#define TIMG_WDT_FLASHBOOT_MOD_EN_V 0x1
#define TIMG_WDT_FLASHBOOT_MOD_EN_S 14
#define TIMG_WDTCONFIG1_REG(i) (REG_TIMG_BASE(i) + 0x004c)
/* TIMG_WDT_CLK_PRESCALE : R/W ;bitpos:[31:16] ;default: 16'h1 ; */
/*description: SWDT clock prescale value. Period = 12.5ns * value stored in this register*/
#define TIMG_WDT_CLK_PRESCALE 0x0000FFFF
#define TIMG_WDT_CLK_PRESCALE_M ((TIMG_WDT_CLK_PRESCALE_V)<<(TIMG_WDT_CLK_PRESCALE_S))
#define TIMG_WDT_CLK_PRESCALE_V 0xFFFF
#define TIMG_WDT_CLK_PRESCALE_S 16
#define TIMG_WDTCONFIG2_REG(i) (REG_TIMG_BASE(i) + 0x0050)
/* TIMG_WDT_STG0_HOLD : R/W ;bitpos:[31:0] ;default: 32'd26000000 ; */
/*description: Stage 0 timeout value in SWDT clock cycles*/
#define TIMG_WDT_STG0_HOLD 0xFFFFFFFF
#define TIMG_WDT_STG0_HOLD_M ((TIMG_WDT_STG0_HOLD_V)<<(TIMG_WDT_STG0_HOLD_S))
#define TIMG_WDT_STG0_HOLD_V 0xFFFFFFFF
#define TIMG_WDT_STG0_HOLD_S 0
#define TIMG_WDTCONFIG3_REG(i) (REG_TIMG_BASE(i) + 0x0054)
/* TIMG_WDT_STG1_HOLD : R/W ;bitpos:[31:0] ;default: 32'h7ffffff ; */
/*description: Stage 1 timeout value in SWDT clock cycles*/
#define TIMG_WDT_STG1_HOLD 0xFFFFFFFF
#define TIMG_WDT_STG1_HOLD_M ((TIMG_WDT_STG1_HOLD_V)<<(TIMG_WDT_STG1_HOLD_S))
#define TIMG_WDT_STG1_HOLD_V 0xFFFFFFFF
#define TIMG_WDT_STG1_HOLD_S 0
#define TIMG_WDTCONFIG4_REG(i) (REG_TIMG_BASE(i) + 0x0058)
/* TIMG_WDT_STG2_HOLD : R/W ;bitpos:[31:0] ;default: 32'hfffff ; */
/*description: Stage 2 timeout value in SWDT clock cycles*/
#define TIMG_WDT_STG2_HOLD 0xFFFFFFFF
#define TIMG_WDT_STG2_HOLD_M ((TIMG_WDT_STG2_HOLD_V)<<(TIMG_WDT_STG2_HOLD_S))
#define TIMG_WDT_STG2_HOLD_V 0xFFFFFFFF
#define TIMG_WDT_STG2_HOLD_S 0
#define TIMG_WDTCONFIG5_REG(i) (REG_TIMG_BASE(i) + 0x005c)
/* TIMG_WDT_STG3_HOLD : R/W ;bitpos:[31:0] ;default: 32'hfffff ; */
/*description: Stage 3 timeout value in SWDT clock cycles*/
#define TIMG_WDT_STG3_HOLD 0xFFFFFFFF
#define TIMG_WDT_STG3_HOLD_M ((TIMG_WDT_STG3_HOLD_V)<<(TIMG_WDT_STG3_HOLD_S))
#define TIMG_WDT_STG3_HOLD_V 0xFFFFFFFF
#define TIMG_WDT_STG3_HOLD_S 0
#define TIMG_WDTFEED_REG(i) (REG_TIMG_BASE(i) + 0x0060)
/* TIMG_WDT_FEED : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: Write any value will feed SWDT*/
#define TIMG_WDT_FEED 0xFFFFFFFF
#define TIMG_WDT_FEED_M ((TIMG_WDT_FEED_V)<<(TIMG_WDT_FEED_S))
#define TIMG_WDT_FEED_V 0xFFFFFFFF
#define TIMG_WDT_FEED_S 0
#define TIMG_WDTWPROTECT_REG(i) (REG_TIMG_BASE(i) + 0x0064)
/* TIMG_WDT_WKEY : R/W ;bitpos:[31:0] ;default: 32'h50d83aa1 ; */
/*description: If change its value from default then write protection is on.*/
#define TIMG_WDT_WKEY 0xFFFFFFFF
#define TIMG_WDT_WKEY_M ((TIMG_WDT_WKEY_V)<<(TIMG_WDT_WKEY_S))
#define TIMG_WDT_WKEY_V 0xFFFFFFFF
#define TIMG_WDT_WKEY_S 0
#define TIMG_RTCCALICFG_REG(i) (REG_TIMG_BASE(i) + 0x0068)
/* TIMG_RTC_CALI_START : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: */
#define TIMG_RTC_CALI_START (BIT(31))
#define TIMG_RTC_CALI_START_M (BIT(31))
#define TIMG_RTC_CALI_START_V 0x1
#define TIMG_RTC_CALI_START_S 31
/* TIMG_RTC_CALI_MAX : R/W ;bitpos:[30:16] ;default: 15'h1 ; */
/*description: */
#define TIMG_RTC_CALI_MAX 0x00007FFF
#define TIMG_RTC_CALI_MAX_M ((TIMG_RTC_CALI_MAX_V)<<(TIMG_RTC_CALI_MAX_S))
#define TIMG_RTC_CALI_MAX_V 0x7FFF
#define TIMG_RTC_CALI_MAX_S 16
/* TIMG_RTC_CALI_RDY : RO ;bitpos:[15] ;default: 1'h0 ; */
/*description: */
#define TIMG_RTC_CALI_RDY (BIT(15))
#define TIMG_RTC_CALI_RDY_M (BIT(15))
#define TIMG_RTC_CALI_RDY_V 0x1
#define TIMG_RTC_CALI_RDY_S 15
/* TIMG_RTC_CALI_CLK_SEL : R/W ;bitpos:[14:13] ;default: 2'h1 ; */
/*description: */
#define TIMG_RTC_CALI_CLK_SEL 0x00000003
#define TIMG_RTC_CALI_CLK_SEL_M ((TIMG_RTC_CALI_CLK_SEL_V)<<(TIMG_RTC_CALI_CLK_SEL_S))
#define TIMG_RTC_CALI_CLK_SEL_V 0x3
#define TIMG_RTC_CALI_CLK_SEL_S 13
/* TIMG_RTC_CALI_START_CYCLING : R/W ;bitpos:[12] ;default: 1'd1 ; */
/*description: */
#define TIMG_RTC_CALI_START_CYCLING (BIT(12))
#define TIMG_RTC_CALI_START_CYCLING_M (BIT(12))
#define TIMG_RTC_CALI_START_CYCLING_V 0x1
#define TIMG_RTC_CALI_START_CYCLING_S 12
#define TIMG_RTCCALICFG1_REG(i) (REG_TIMG_BASE(i) + 0x006c)
/* TIMG_RTC_CALI_VALUE : RO ;bitpos:[31:7] ;default: 25'h0 ; */
/*description: */
#define TIMG_RTC_CALI_VALUE 0x01FFFFFF
#define TIMG_RTC_CALI_VALUE_M ((TIMG_RTC_CALI_VALUE_V)<<(TIMG_RTC_CALI_VALUE_S))
#define TIMG_RTC_CALI_VALUE_V 0x1FFFFFF
#define TIMG_RTC_CALI_VALUE_S 7
#define TIMG_LACTCONFIG_REG(i) (REG_TIMG_BASE(i) + 0x0070)
/* TIMG_LACT_EN : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_EN (BIT(31))
#define TIMG_LACT_EN_M (BIT(31))
#define TIMG_LACT_EN_V 0x1
#define TIMG_LACT_EN_S 31
/* TIMG_LACT_INCREASE : R/W ;bitpos:[30] ;default: 1'h1 ; */
/*description: */
#define TIMG_LACT_INCREASE (BIT(30))
#define TIMG_LACT_INCREASE_M (BIT(30))
#define TIMG_LACT_INCREASE_V 0x1
#define TIMG_LACT_INCREASE_S 30
/* TIMG_LACT_AUTORELOAD : R/W ;bitpos:[29] ;default: 1'h1 ; */
/*description: */
#define TIMG_LACT_AUTORELOAD (BIT(29))
#define TIMG_LACT_AUTORELOAD_M (BIT(29))
#define TIMG_LACT_AUTORELOAD_V 0x1
#define TIMG_LACT_AUTORELOAD_S 29
/* TIMG_LACT_DIVIDER : R/W ;bitpos:[28:13] ;default: 16'h1 ; */
/*description: */
#define TIMG_LACT_DIVIDER 0x0000FFFF
#define TIMG_LACT_DIVIDER_M ((TIMG_LACT_DIVIDER_V)<<(TIMG_LACT_DIVIDER_S))
#define TIMG_LACT_DIVIDER_V 0xFFFF
#define TIMG_LACT_DIVIDER_S 13
/* TIMG_LACT_EDGE_INT_EN : R/W ;bitpos:[12] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_EDGE_INT_EN (BIT(12))
#define TIMG_LACT_EDGE_INT_EN_M (BIT(12))
#define TIMG_LACT_EDGE_INT_EN_V 0x1
#define TIMG_LACT_EDGE_INT_EN_S 12
/* TIMG_LACT_LEVEL_INT_EN : R/W ;bitpos:[11] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_LEVEL_INT_EN (BIT(11))
#define TIMG_LACT_LEVEL_INT_EN_M (BIT(11))
#define TIMG_LACT_LEVEL_INT_EN_V 0x1
#define TIMG_LACT_LEVEL_INT_EN_S 11
/* TIMG_LACT_ALARM_EN : R/W ;bitpos:[10] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_ALARM_EN (BIT(10))
#define TIMG_LACT_ALARM_EN_M (BIT(10))
#define TIMG_LACT_ALARM_EN_V 0x1
#define TIMG_LACT_ALARM_EN_S 10
/* TIMG_LACT_LAC_EN : R/W ;bitpos:[9] ;default: 1'h1 ; */
/*description: */
#define TIMG_LACT_LAC_EN (BIT(9))
#define TIMG_LACT_LAC_EN_M (BIT(9))
#define TIMG_LACT_LAC_EN_V 0x1
#define TIMG_LACT_LAC_EN_S 9
/* TIMG_LACT_CPST_EN : R/W ;bitpos:[8] ;default: 1'h1 ; */
/*description: */
#define TIMG_LACT_CPST_EN (BIT(8))
#define TIMG_LACT_CPST_EN_M (BIT(8))
#define TIMG_LACT_CPST_EN_V 0x1
#define TIMG_LACT_CPST_EN_S 8
/* TIMG_LACT_RTC_ONLY : R/W ;bitpos:[7] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_RTC_ONLY (BIT(7))
#define TIMG_LACT_RTC_ONLY_M (BIT(7))
#define TIMG_LACT_RTC_ONLY_V 0x1
#define TIMG_LACT_RTC_ONLY_S 7
#define TIMG_LACTRTC_REG(i) (REG_TIMG_BASE(i) + 0x0074)
/* TIMG_LACT_RTC_STEP_LEN : R/W ;bitpos:[31:6] ;default: 26'h0 ; */
/*description: */
#define TIMG_LACT_RTC_STEP_LEN 0x03FFFFFF
#define TIMG_LACT_RTC_STEP_LEN_M ((TIMG_LACT_RTC_STEP_LEN_V)<<(TIMG_LACT_RTC_STEP_LEN_S))
#define TIMG_LACT_RTC_STEP_LEN_V 0x3FFFFFF
#define TIMG_LACT_RTC_STEP_LEN_S 6
#define TIMG_LACTLO_REG(i) (REG_TIMG_BASE(i) + 0x0078)
/* TIMG_LACT_LO : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_LO 0xFFFFFFFF
#define TIMG_LACT_LO_M ((TIMG_LACT_LO_V)<<(TIMG_LACT_LO_S))
#define TIMG_LACT_LO_V 0xFFFFFFFF
#define TIMG_LACT_LO_S 0
#define TIMG_LACTHI_REG(i) (REG_TIMG_BASE(i) + 0x007c)
/* TIMG_LACT_HI : RO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_HI 0xFFFFFFFF
#define TIMG_LACT_HI_M ((TIMG_LACT_HI_V)<<(TIMG_LACT_HI_S))
#define TIMG_LACT_HI_V 0xFFFFFFFF
#define TIMG_LACT_HI_S 0
#define TIMG_LACTUPDATE_REG(i) (REG_TIMG_BASE(i) + 0x0080)
/* TIMG_LACT_UPDATE : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_UPDATE 0xFFFFFFFF
#define TIMG_LACT_UPDATE_M ((TIMG_LACT_UPDATE_V)<<(TIMG_LACT_UPDATE_S))
#define TIMG_LACT_UPDATE_V 0xFFFFFFFF
#define TIMG_LACT_UPDATE_S 0
#define TIMG_LACTALARMLO_REG(i) (REG_TIMG_BASE(i) + 0x0084)
/* TIMG_LACT_ALARM_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_ALARM_LO 0xFFFFFFFF
#define TIMG_LACT_ALARM_LO_M ((TIMG_LACT_ALARM_LO_V)<<(TIMG_LACT_ALARM_LO_S))
#define TIMG_LACT_ALARM_LO_V 0xFFFFFFFF
#define TIMG_LACT_ALARM_LO_S 0
#define TIMG_LACTALARMHI_REG(i) (REG_TIMG_BASE(i) + 0x0088)
/* TIMG_LACT_ALARM_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_ALARM_HI 0xFFFFFFFF
#define TIMG_LACT_ALARM_HI_M ((TIMG_LACT_ALARM_HI_V)<<(TIMG_LACT_ALARM_HI_S))
#define TIMG_LACT_ALARM_HI_V 0xFFFFFFFF
#define TIMG_LACT_ALARM_HI_S 0
#define TIMG_LACTLOADLO_REG(i) (REG_TIMG_BASE(i) + 0x008c)
/* TIMG_LACT_LOAD_LO : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_LOAD_LO 0xFFFFFFFF
#define TIMG_LACT_LOAD_LO_M ((TIMG_LACT_LOAD_LO_V)<<(TIMG_LACT_LOAD_LO_S))
#define TIMG_LACT_LOAD_LO_V 0xFFFFFFFF
#define TIMG_LACT_LOAD_LO_S 0
#define TIMG_LACTLOADHI_REG(i) (REG_TIMG_BASE(i) + 0x0090)
/* TIMG_LACT_LOAD_HI : R/W ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_LOAD_HI 0xFFFFFFFF
#define TIMG_LACT_LOAD_HI_M ((TIMG_LACT_LOAD_HI_V)<<(TIMG_LACT_LOAD_HI_S))
#define TIMG_LACT_LOAD_HI_V 0xFFFFFFFF
#define TIMG_LACT_LOAD_HI_S 0
#define TIMG_LACTLOAD_REG(i) (REG_TIMG_BASE(i) + 0x0094)
/* TIMG_LACT_LOAD : WO ;bitpos:[31:0] ;default: 32'h0 ; */
/*description: */
#define TIMG_LACT_LOAD 0xFFFFFFFF
#define TIMG_LACT_LOAD_M ((TIMG_LACT_LOAD_V)<<(TIMG_LACT_LOAD_S))
#define TIMG_LACT_LOAD_V 0xFFFFFFFF
#define TIMG_LACT_LOAD_S 0
#define TIMG_INT_ENA_TIMERS_REG(i) (REG_TIMG_BASE(i) + 0x0098)
/* TIMG_LACT_INT_ENA : R/W ;bitpos:[3] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_INT_ENA (BIT(3))
#define TIMG_LACT_INT_ENA_M (BIT(3))
#define TIMG_LACT_INT_ENA_V 0x1
#define TIMG_LACT_INT_ENA_S 3
/* TIMG_WDT_INT_ENA : R/W ;bitpos:[2] ;default: 1'h0 ; */
/*description: Interrupt when an interrupt stage timeout*/
#define TIMG_WDT_INT_ENA (BIT(2))
#define TIMG_WDT_INT_ENA_M (BIT(2))
#define TIMG_WDT_INT_ENA_V 0x1
#define TIMG_WDT_INT_ENA_S 2
/* TIMG_T1_INT_ENA : R/W ;bitpos:[1] ;default: 1'h0 ; */
/*description: interrupt when timer1 alarm*/
#define TIMG_T1_INT_ENA (BIT(1))
#define TIMG_T1_INT_ENA_M (BIT(1))
#define TIMG_T1_INT_ENA_V 0x1
#define TIMG_T1_INT_ENA_S 1
/* TIMG_T0_INT_ENA : R/W ;bitpos:[0] ;default: 1'h0 ; */
/*description: interrupt when timer0 alarm*/
#define TIMG_T0_INT_ENA (BIT(0))
#define TIMG_T0_INT_ENA_M (BIT(0))
#define TIMG_T0_INT_ENA_V 0x1
#define TIMG_T0_INT_ENA_S 0
#define TIMG_INT_RAW_TIMERS_REG(i) (REG_TIMG_BASE(i) + 0x009c)
/* TIMG_LACT_INT_RAW : RO ;bitpos:[3] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_INT_RAW (BIT(3))
#define TIMG_LACT_INT_RAW_M (BIT(3))
#define TIMG_LACT_INT_RAW_V 0x1
#define TIMG_LACT_INT_RAW_S 3
/* TIMG_WDT_INT_RAW : RO ;bitpos:[2] ;default: 1'h0 ; */
/*description: Interrupt when an interrupt stage timeout*/
#define TIMG_WDT_INT_RAW (BIT(2))
#define TIMG_WDT_INT_RAW_M (BIT(2))
#define TIMG_WDT_INT_RAW_V 0x1
#define TIMG_WDT_INT_RAW_S 2
/* TIMG_T1_INT_RAW : RO ;bitpos:[1] ;default: 1'h0 ; */
/*description: interrupt when timer1 alarm*/
#define TIMG_T1_INT_RAW (BIT(1))
#define TIMG_T1_INT_RAW_M (BIT(1))
#define TIMG_T1_INT_RAW_V 0x1
#define TIMG_T1_INT_RAW_S 1
/* TIMG_T0_INT_RAW : RO ;bitpos:[0] ;default: 1'h0 ; */
/*description: interrupt when timer0 alarm*/
#define TIMG_T0_INT_RAW (BIT(0))
#define TIMG_T0_INT_RAW_M (BIT(0))
#define TIMG_T0_INT_RAW_V 0x1
#define TIMG_T0_INT_RAW_S 0
#define TIMG_INT_ST_TIMERS_REG(i) (REG_TIMG_BASE(i) + 0x00a0)
/* TIMG_LACT_INT_ST : RO ;bitpos:[3] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_INT_ST (BIT(3))
#define TIMG_LACT_INT_ST_M (BIT(3))
#define TIMG_LACT_INT_ST_V 0x1
#define TIMG_LACT_INT_ST_S 3
/* TIMG_WDT_INT_ST : RO ;bitpos:[2] ;default: 1'h0 ; */
/*description: Interrupt when an interrupt stage timeout*/
#define TIMG_WDT_INT_ST (BIT(2))
#define TIMG_WDT_INT_ST_M (BIT(2))
#define TIMG_WDT_INT_ST_V 0x1
#define TIMG_WDT_INT_ST_S 2
/* TIMG_T1_INT_ST : RO ;bitpos:[1] ;default: 1'h0 ; */
/*description: interrupt when timer1 alarm*/
#define TIMG_T1_INT_ST (BIT(1))
#define TIMG_T1_INT_ST_M (BIT(1))
#define TIMG_T1_INT_ST_V 0x1
#define TIMG_T1_INT_ST_S 1
/* TIMG_T0_INT_ST : RO ;bitpos:[0] ;default: 1'h0 ; */
/*description: interrupt when timer0 alarm*/
#define TIMG_T0_INT_ST (BIT(0))
#define TIMG_T0_INT_ST_M (BIT(0))
#define TIMG_T0_INT_ST_V 0x1
#define TIMG_T0_INT_ST_S 0
#define TIMG_INT_CLR_TIMERS_REG(i) (REG_TIMG_BASE(i) + 0x00a4)
/* TIMG_LACT_INT_CLR : WO ;bitpos:[3] ;default: 1'h0 ; */
/*description: */
#define TIMG_LACT_INT_CLR (BIT(3))
#define TIMG_LACT_INT_CLR_M (BIT(3))
#define TIMG_LACT_INT_CLR_V 0x1
#define TIMG_LACT_INT_CLR_S 3
/* TIMG_WDT_INT_CLR : WO ;bitpos:[2] ;default: 1'h0 ; */
/*description: Interrupt when an interrupt stage timeout*/
#define TIMG_WDT_INT_CLR (BIT(2))
#define TIMG_WDT_INT_CLR_M (BIT(2))
#define TIMG_WDT_INT_CLR_V 0x1
#define TIMG_WDT_INT_CLR_S 2
/* TIMG_T1_INT_CLR : WO ;bitpos:[1] ;default: 1'h0 ; */
/*description: interrupt when timer1 alarm*/
#define TIMG_T1_INT_CLR (BIT(1))
#define TIMG_T1_INT_CLR_M (BIT(1))
#define TIMG_T1_INT_CLR_V 0x1
#define TIMG_T1_INT_CLR_S 1
/* TIMG_T0_INT_CLR : WO ;bitpos:[0] ;default: 1'h0 ; */
/*description: interrupt when timer0 alarm*/
#define TIMG_T0_INT_CLR (BIT(0))
#define TIMG_T0_INT_CLR_M (BIT(0))
#define TIMG_T0_INT_CLR_V 0x1
#define TIMG_T0_INT_CLR_S 0
#define TIMG_NTIMERS_DATE_REG(i) (REG_TIMG_BASE(i) + 0x00f8)
/* TIMG_NTIMERS_DATE : R/W ;bitpos:[27:0] ;default: 28'h1604290 ; */
/*description: Version of this regfile*/
#define TIMG_NTIMERS_DATE 0x0FFFFFFF
#define TIMG_NTIMERS_DATE_M ((TIMG_NTIMERS_DATE_V)<<(TIMG_NTIMERS_DATE_S))
#define TIMG_NTIMERS_DATE_V 0xFFFFFFF
#define TIMG_NTIMERS_DATE_S 0
#define TIMGCLK_REG(i) (REG_TIMG_BASE(i) + 0x00fc)
/* TIMG_CLK_EN : R/W ;bitpos:[31] ;default: 1'h0 ; */
/*description: Force clock enable for this regfile*/
#define TIMG_CLK_EN (BIT(31))
#define TIMG_CLK_EN_M (BIT(31))
#define TIMG_CLK_EN_V 0x1
#define TIMG_CLK_EN_S 31
#ifdef __cplusplus
}
#endif
```
|
Ešref Jašarević (born 5 February 1951) is a Bosnian retired footballer.
Club career
Born in Gradačac, SR Bosnia-Herzegovina, FPR Yugoslavia, he started playing with local side NK Zvijezda Gradačac, but he would become famous while playing in the Yugoslav First League club FK Sloboda Tuzla where he played a total of 172 first league matches, and scored 13 goals. In 1979, he spent half a season in Turkey with Galatasaray SK, before returning to Yugoslav First League, this time, to FK Napredak Kruševac where he played until 1982.
International career
In 1977, he played two matches for the Yugoslavia national football team. His debut was on January 30, in Bogota, against Colombia (a 1–0 win), and the other match was played on February 1, in León, against Mexico (a 5–1 win).
After retiring, he became a sports director of NK Zvijezda Gradačac.
References
External sources
Profile at Serbian Federation website.
1951 births
Living people
People from Gradačac
Sportspeople from Tuzla Canton
Men's association football midfielders
Yugoslav men's footballers
Yugoslavia men's international footballers
NK Zvijezda Gradačac players
FK Sloboda Tuzla players
Galatasaray S.K. footballers
FK Napredak Kruševac players
Yugoslav First League players
Süper Lig players
Yugoslav Second League players
Yugoslav expatriate men's footballers
Expatriate men's footballers in Turkey
Yugoslav expatriate sportspeople in Turkey
|
The Baden Class III b engines were German steam locomotives designed for hauling passenger trains for the Grand Duchy of Baden State Railway (Großherzoglich Badische Staatseisenbahn).
In this class the Baden state railway grouped rebuilt locomotives of former classes III and III a.
Seventy engines of Baden classes III and III a were rebuilt between 1881 and 1891. They were all given a new, more powerful, three-ring boiler with a larger grate area. Because this was longer than the original boilers, the frame of the Class III had to be extended (on the Class IIIa it was long enough) and a buffer beam built onto the front.
The boiler had a Belpaire firebox with a somewhat higher top surface. A larger steam dome was fitted on top of the rear boiler ring.
Further reading
Lokomotiv-Archiv Baden, Hermann Lohr, Georg Thielmann, transpress Berlin, , 1988, pp 82–88.
See also
Grand Duchy of Baden State Railway
List of Baden locomotives and railbuses
4-4-0 locomotives
03 b
Standard gauge locomotives of Germany
Railway locomotives introduced in 1881
2′B n2 locomotives
Passenger locomotives
|
The FIVB Beach Volleyball World Championships is the double-gender world championship for the sport of beach volleyball organized by the (FIVB) the sport's global governing body. The first official edition of the event was held in Los Angeles, United States in 1997 and tournaments had been awarded every two years since then. Before 1997, ten unofficial championships not organized by the FIVB were all held in Rio de Janeiro, Brazil between 1987 and 1996. The most recent World Championships took place in Mexico in 2023.
Winning the World Championships is considered to be one of the highest honours in international beach volleyball, surpassing the FIVB Beach Volleyball World Tour and being surpassed only by the beach volleyball tournament at the Summer Olympic Games.
Format
The tournament has a 48-team main draw per gender and consists of two stages: the group stage followed by the knockout stage. The prize pool for each gender is US$500,000.
Editions
First Beach Volleyball World Championships were held from 10 to 13 September 1997 in Los Angeles, California (United States). It was the first official edition of this event, after 10 unofficial championships between 1987 and 1996.
Results summary
Men's tournament
Women's tournament
Medals
As of 2023 Beach Volleyball World Championships.
Men
Women
Total
See also
Beach volleyball at the Summer Olympics
FIVB Beach Volleyball World Tour
European Beach Volleyball Tour
FIVB Beach Volleyball U23 World Championships
FIVB Beach Volleyball U21 World Championships
FIVB Beach Volleyball U19 World Championships
FIVB Beach Volleyball U17 World Championships
References
External links
Fédération Internationale de Volleyball – official website
2017 FIVB Beach Volleyball World Championships – official website
Recurring sporting events established in 1997
Biennial sporting events
|
Heliconius metharme is a species of butterfly of the family Nymphalidae. It was described by Wilhelm Ferdinand Erichson in 1849. It is widespread in the Amazon basin, Venezuela and the Guianas. The habitat consists of deep forests.
It is part of a mimicry ring with Heliconius sara, Heliconius wallacei and Heliconius doris.
The larvae are gregarious and feed on Dilkea and Mitostemma species.
Subspecies
Heliconius metharme metharme — Guyana
Heliconius metharme makiritare (Brown & Fernández, 1985) — Venezuela
Heliconius metharme perseis Stichel, 1923 — Colombia
References
Neruda metharme at Insecta.pro
Butterflies described in 1849
Heliconius
|
Muhammad Yar Khuhawar is a Pakistani analytical chemist who is Professor Emeritus at the Institute of Advanced Research Studies in Chemical Sciences, University of Sindh, Jamshoro.
He is working with more than 50 researchers, focusing in the field of chromatography and electrophoresis for analytical method development for metals, non-metals and biological active compounds from a wide variety of samples. His research group has synthesized a number of Schiff base polymers and has developed a new stationary phase for Gas Chromatography and sensitive Liquid Chromatography procedures for the determination of pesticide malathion.
He has also published a number of research papers in the field of environmental monitoring of water resources of Sindh province including: Indus River, canals, lakes, natural springs, sea, underground water resources, municipal sewerage and receiving bodies in terms of pollution and utilities for biological life inhabiting the area.
He has a Ph.D. in Chemistry from Birmingham University, United Kingdom, and was awarded the degree of Doctor of Science (D.Sc.) from University of Birmingham, in 2000. He has more than 250 research publications in national and international journals.
References
External links
http://www.ceacsu.edu.pk/Faculty/Journal.htm
http://www.rsc.org/publishing/journals/AN/article.asp?doi=AN9921701725
http://www.aasci.org/conference/env/2007/Presentation-schedule-EST2007.pdf
https://j-chrom-sci.com/abstracts/2010/april/303-khuhawar.html
https://web.archive.org/web/20130816031319/http://htcrl.usindh.edu.pk/
http://iarscs.usindh.edu.pk/
Pakistani chemists
Living people
1948 births
Academic staff of the University of Sindh
Sindhi people
Alumni of the University of Birmingham
|
```vue
<script lang="ts" setup>
import { ref, computed, watch, onMounted } from 'vue'
import { useEnhancer } from '/@/app/enhancer'
import { getPageURL } from '/@/transforms/url'
import HeaderView from './header.vue'
import FooterView from './footer.vue'
import AsideView from './aside.vue'
const { route } = useEnhancer()
const pageURL = computed(() => getPageURL(route.fullPath))
const isLoaded = ref(false)
const isOpenedAside = ref(false)
const openAside = () => (isOpenedAside.value = true)
const closeAside = () => (isOpenedAside.value = false)
onMounted(() => {
isLoaded.value = true
watch(isOpenedAside, (opened) => {
document.body.style.overflow = opened ? 'hidden' : 'auto'
})
})
</script>
<template>
<div class="mobile-main">
<div class="asider" :class="{ loaded: isLoaded, opened: isOpenedAside }">
<aside-view class="aside" />
</div>
<div class="main" :class="{ opened: isOpenedAside }">
<div v-if="isOpenedAside" class="close-mask" @click="closeAside" />
<header-view :opened="isOpenedAside" @close="closeAside" @open="openAside" />
<main class="main-container">
<router-view name="mobile" v-slot="{ Component, route: r }">
<div class="route-view" v-if="r.meta.responsive">
<component :is="Component" :key="r.name" />
</div>
<div class="fallback" v-else>
<i class="iconfont icon-desktop"></i>
<p class="text">
<i18n zh=" PC " en="This page only support visit on PC" />
</p>
<a class="link" :href="pageURL">{{ pageURL }}</a>
</div>
</router-view>
</main>
<footer-view />
</div>
</div>
</template>
<style lang="scss" scoped>
@import '/src/styles/variables.scss';
@import '/src/styles/mixins.scss';
.mobile-main {
color: $color-text;
background-color: $module-bg-hover;
$aside-width: 66%;
.asider {
width: $aside-width;
height: 100%;
position: fixed;
top: 0;
left: 0;
background-color: $mobile-aside-bg;
transform: translate3d(-100%, 0, 0);
.aside {
opacity: 0;
transition: $mobile-aside-transition;
}
&.loaded {
transition: $mobile-aside-transition;
}
&.opened {
transform: translate3d(0, 0, 0);
overflow: hidden;
-webkit-overflow-scrolling: touch;
.aside {
opacity: 1;
transition: $mobile-aside-transition;
}
}
}
.main {
min-height: 100vh;
position: relative;
transition: $mobile-aside-transition;
&.opened {
transform: translate3d($aside-width, 0, 0);
}
.close-mask {
position: absolute;
width: 100vw;
height: 100%;
top: 0;
right: 0;
z-index: $z-index-top;
background-color: $module-bg-translucent;
@include backdrop-blur(3px);
}
.main-container {
position: relative;
overflow: hidden;
width: 100%;
min-height: calc(100vh - $mobile-header-height);
margin: 0;
padding: $gap-lg $gap;
padding-top: $mobile-header-height + $gap-lg;
transition: width 0.35s;
display: flex;
.fallback {
width: 100%;
padding-bottom: 6rem; // for middle
display: flex;
flex-direction: column;
align-items: center;
justify-content: center;
.iconfont {
color: $color-text-divider;
font-size: 12rem;
}
.text {
font-size: $font-size-h4;
margin-bottom: $gap-lg;
}
.link {
color: $color-text-disabled;
@include text-underline(0.4em);
}
}
.route-view {
width: 100%;
}
}
}
}
</style>
```
|
```go
//
//
// path_to_url
//
// Unless required by applicable law or agreed to in writing, software
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
package monitoredresource
import (
"contrib.go.opencensus.io/exporter/stackdriver/monitoredresource/aws"
"contrib.go.opencensus.io/exporter/stackdriver/monitoredresource/gcp"
)
// GKEContainer represents gke_container type monitored resource.
// For definition refer to
// path_to_url#tag_gke_container
// Deprecated: please use gcp.GKEContainer from "contrib.go.opencensus.io/exporter/stackdriver/monitoredresource/gcp".
type GKEContainer struct {
// ProjectID is the identifier of the GCP project associated with this resource, such as "my-project".
ProjectID string
// InstanceID is the numeric VM instance identifier assigned by Compute Engine.
InstanceID string
// ClusterName is the name for the cluster the container is running in.
ClusterName string
// ContainerName is the name of the container.
ContainerName string
// NamespaceID is the identifier for the cluster namespace the container is running in
NamespaceID string
// PodID is the identifier for the pod the container is running in.
PodID string
// Zone is the Compute Engine zone in which the VM is running.
Zone string
// LoggingMonitoringV2Enabled is the identifier if user enabled V2 logging and monitoring for GKE
LoggingMonitoringV2Enabled bool
}
// MonitoredResource returns resource type and resource labels for GKEContainer
func (gke *GKEContainer) MonitoredResource() (resType string, labels map[string]string) {
gcpGKE := gcp.GKEContainer(*gke)
return gcpGKE.MonitoredResource()
}
// GCEInstance represents gce_instance type monitored resource.
// For definition refer to
// path_to_url#tag_gce_instance
// Deprecated: please use gcp.GCEInstance from "contrib.go.opencensus.io/exporter/stackdriver/monitoredresource/gcp".
type GCEInstance struct {
// ProjectID is the identifier of the GCP project associated with this resource, such as "my-project".
ProjectID string
// InstanceID is the numeric VM instance identifier assigned by Compute Engine.
InstanceID string
// Zone is the Compute Engine zone in which the VM is running.
Zone string
}
// MonitoredResource returns resource type and resource labels for GCEInstance
func (gce *GCEInstance) MonitoredResource() (resType string, labels map[string]string) {
gcpGCE := gcp.GCEInstance(*gce)
return gcpGCE.MonitoredResource()
}
// AWSEC2Instance represents aws_ec2_instance type monitored resource.
// For definition refer to
// path_to_url#tag_aws_ec2_instance
// Deprecated: please use aws.EC2Container from "contrib.go.opencensus.io/exporter/stackdriver/monitoredresource/aws".
type AWSEC2Instance struct {
// AWSAccount is the AWS account number for the VM.
AWSAccount string
// InstanceID is the instance id of the instance.
InstanceID string
// Region is the AWS region for the VM. The format of this field is "aws:{region}",
// where supported values for {region} are listed at
// path_to_url
Region string
}
// MonitoredResource returns resource type and resource labels for AWSEC2Instance
func (ec2 *AWSEC2Instance) MonitoredResource() (resType string, labels map[string]string) {
awsEC2 := aws.EC2Instance(*ec2)
return awsEC2.MonitoredResource()
}
```
|
```java
/*
*
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing, software
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*
* Contributors:
* ohun@live.cn ()
*/
package com.mpush.api.common;
import java.util.concurrent.Executor;
/**
* Created by ohun on 2017/7/15.
*
* @author ohun@live.cn ()
*/
public interface Monitor {
void monitor(String name, Thread thread);
void monitor(String name, Executor executor);
}
```
|
```java
/*
*
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing,
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* specific language governing permissions and limitations
*/
package org.ballerinalang.test.structs;
import io.ballerina.runtime.api.utils.StringUtils;
import io.ballerina.runtime.api.values.BArray;
import io.ballerina.runtime.api.values.BMap;
import io.ballerina.runtime.api.values.BString;
import org.ballerinalang.test.BCompileUtil;
import org.ballerinalang.test.BRunUtil;
import org.ballerinalang.test.CompileResult;
import org.testng.Assert;
import org.testng.annotations.AfterClass;
import org.testng.annotations.BeforeClass;
import org.testng.annotations.Test;
/**
* Test cases for user defined struct types in ballerina.
*/
public class StructTest {
private CompileResult compileResult;
@BeforeClass
public void setup() {
compileResult = BCompileUtil.compile("test-src/structs/struct.bal");
}
@Test(description = "Test Basic struct operations")
public void testBasicStruct() {
BArray returns = (BArray) BRunUtil.invoke(compileResult, "testCreateStruct");
Assert.assertTrue(returns.get(0) instanceof BString);
Assert.assertEquals(returns.get(0).toString(), "Jack");
Assert.assertTrue(returns.get(1) instanceof BMap);
BMap<String, ?> adrsMap = ((BMap) returns.get(1));
Assert.assertEquals(adrsMap.get(StringUtils.fromString("country")), StringUtils.fromString("USA"));
Assert.assertEquals(adrsMap.get(StringUtils.fromString("state")), StringUtils.fromString("CA"));
Assert.assertTrue(returns.get(2) instanceof Long);
Assert.assertEquals(returns.get(2), 25L);
}
@Test(description = "Test using expressions as index for struct arrays")
public void testExpressionAsIndex() {
Object returns = BRunUtil.invoke(compileResult, "testExpressionAsIndex");
Assert.assertTrue(returns instanceof BString);
Assert.assertEquals(returns.toString(), "Jane");
}
/* @Test(description = "Test struct operations inside a connector")
public void testStructInConnector() {
Object returns = Functions.invoke(bLangProgram, "testAction1");
Assert.assertEquals(returns.length, 1);
Assert.assertTrue(returns instanceof BString);
Assert.assertEquals(returns.toString(), "Jack");
}*/
@Test(description = "Test using structs inside structs")
public void testStructOfStructs() {
Object returns = BRunUtil.invoke(compileResult, "testStructOfStruct");
Assert.assertTrue(returns instanceof BString);
Assert.assertEquals(returns.toString(), "USA");
}
@Test(description = "Test returning fields of a struct")
public void testReturnStructAttributes() {
Object returns = BRunUtil.invoke(compileResult, "testReturnStructAttributes");
Assert.assertTrue(returns instanceof BString);
Assert.assertEquals(returns.toString(), "emily");
}
@Test(description = "Test using struct expression as a index in another struct expression")
public void testStructExpressionAsIndex() {
Object returns = BRunUtil.invoke(compileResult, "testStructExpressionAsIndex");
Assert.assertTrue(returns instanceof BString);
Assert.assertEquals(returns.toString(), "emily");
}
@Test(description = "Test default value of a struct field")
public void testDefaultValue() {
BArray returns = (BArray) BRunUtil.invoke(compileResult, "testDefaultVal");
// Check default value of a field where the default value is set
Assert.assertTrue(returns.get(0) instanceof BString);
Assert.assertEquals(returns.get(0).toString(), "default first name");
// Check the default value of a field where the default value is not set
Assert.assertTrue(returns.get(1) instanceof BString);
Assert.assertEquals(returns.get(1).toString(), "");
Assert.assertTrue(returns.get(2) instanceof Long);
Assert.assertEquals(returns.get(2), 999L);
}
@Test(description = "Test default value of a nested struct field")
public void testNestedFieldDefaultValue() {
BArray returns = (BArray) BRunUtil.invoke(compileResult, "testNestedFieldDefaultVal");
Assert.assertTrue(returns.get(0) instanceof BString);
Assert.assertEquals(returns.get(0).toString(), "default first name");
Assert.assertTrue(returns.get(1) instanceof BString);
Assert.assertEquals(returns.get(1).toString(), "Smith");
Assert.assertTrue(returns.get(2) instanceof Long);
Assert.assertEquals(returns.get(2), 999L);
}
@Test(description = "Test default value of a nested struct field")
public void testNestedStructInit() {
Object returns = BRunUtil.invoke(compileResult, "testNestedStructInit");
Assert.assertTrue(returns instanceof BMap);
BMap<String, Object> person = ((BMap<String, Object>) returns);
Assert.assertEquals(person.get(StringUtils.fromString("name")).toString(), "aaa");
Assert.assertEquals((person.get(StringUtils.fromString("age"))), 25L);
Assert.assertTrue(person.get(StringUtils.fromString("parent")) instanceof BMap);
BMap<String, Object> parent = ((BMap<String, Object>) person.get(StringUtils.fromString("parent")));
Assert.assertEquals(parent.get(StringUtils.fromString("name")).toString(), "bbb");
Assert.assertEquals((parent.get(StringUtils.fromString("age"))), 50L);
}
@Test(description = "Test negative default values in struct")
public void testNegativeDefaultValue() {
BArray returns = (BArray) BRunUtil.invoke(compileResult, "getStructNegativeValues");
Assert.assertEquals(returns.size(), 4);
Assert.assertSame(returns.get(0).getClass(), Long.class);
Assert.assertSame(returns.get(1).getClass(), Long.class);
Assert.assertSame(returns.get(2).getClass(), Double.class);
Assert.assertSame(returns.get(3).getClass(), Double.class);
Assert.assertEquals(returns.get(0), -9L);
Assert.assertEquals(returns.get(1), -8L);
Assert.assertEquals(returns.get(2), -88.234);
Assert.assertEquals(returns.get(3), -24.99);
}
@Test(description = "Test negative default values in struct")
public void testStructToString() {
Object returns = BRunUtil.invoke(compileResult, "getStruct");
Assert.assertEquals(returns.toString(), "{\"name\":\"aaa\",\"lname\":\"\",\"adrs\":{},\"age\":25," +
"\"family\":{\"spouse\":\"\",\"noOfChildren\":0,\"children\":[]},\"parent\":{\"name\":\"bbb\"," +
"\"lname\":\"ccc\",\"adrs\":{},\"age\":50,\"family\":{\"spouse\":\"\",\"noOfChildren\":0," +
"\"children\":[]},\"parent\":null}}");
}
@Test
public void testStructLiteral() {
CompileResult compileResult = BCompileUtil.compile(
"test-src/structs/ObjectWithPrivateFieldsTestProject/struct-literals.bal");
Object returns = BRunUtil.invoke(compileResult, "testStructLiteral1");
Assert.assertEquals(returns.toString(), "{\"dptName\":\"\",\"employees\":[],\"manager\":{\"name\":\"default " +
"first name\",\"lname\":\"\",\"adrs\":{},\"age\":999,\"child\":null}}");
returns = BRunUtil.invoke(compileResult, "testStructLiteral2");
Assert.assertEquals(returns.toString(),
"{\"name\":\"default first name\",\"lname\":\"\",\"adrs\":{},\"age\":999,\"child\":null}");
}
@Test
public void testStructLiteralInitFunc() {
CompileResult result = BCompileUtil.compile(
"test-src/structs/ObjectWithPrivateFieldsTestProject/nested-struct-inline-init.bal");
Object returns = BRunUtil.invoke(result, "testCreateStruct");
Assert.assertEquals(returns.toString(),
"{\"name\":\"default first name\",\"fname\":\"\",\"lname\":\"Doe\",\"adrs\":{},\"age\":999," +
"\"family\":{\"spouse\":\"Jane\",\"noOfChildren\":0,\"children\":[\"Alex\",\"Bob\"]}}");
}
@AfterClass
public void tearDown() {
compileResult = null;
}
}
```
|
Richard Strachan (born 18 November 1986) is a British personal trainer, former athlete who competed in sprinting, and former boxer. During his time in athletics, Strachan has represented Great Britain at World and European Championships, winning two European Golds and a European Silver medal. Strachan mainly specialized in the 400 metres distance as seen from his involvement in the British team during the European indoor championships on several occasions.
Following his departure from athletics, Strachan has become a strength coach, coaching YouTubers such as Vikkstar123, TBJZL, and Manny. Strachan owns a YouTube channel with 62,000 subscribers, posting both sprinting and strength training videos with the YouTubers.
Amateur boxing record
References
External links
1986 births
Living people
British male sprinters
|
Best Friend (), also known as Good Neighbor is a 2020 South Korean comedy-drama film directed by Lee Hwan-kyung. The film starring Jung Woo, Oh Dal-soo, Kim Hee-won, Kim Byung-chul, Lee Yu-bi, Jo Hyeon-cheol, Ji Seung-hyun, Kim Sun-kyung and Yeom Hye-ran. It was released theatrically on November 25, 2020.
Cast
Jung Woo as Yoo Dae-kwon
Oh Dal-soo as Lee Ui-sik
Kim Hee-won as Chief Kim
Kim Byung-chul as Dong-sik
Lee Yu-bi as Lee Eun-jin
Jo Hyeon-cheol as Young-cheol
Ji Seung-hyun as Dong-hyuk
Kim Sun-kyung as Young-ja
Yeom Hye-ran
References
External links
2020 comedy-drama films
2020 films
2020s political comedy-drama films
2020s Korean-language films
South Korean comedy-drama films
South Korean political comedy films
South Korean political drama films
Films about democracy
Films about human rights
Films about ideologies
Films about politicians
Films about presidents
Films about security and surveillance
Films set in Seoul
Films set in 1985
Films set in the 1980s
Films shot in North Chungcheong Province
Films shot in South Jeolla Province
Little Big Pictures films
|
A calavera (Spanish – for "skull"), in the context of Day of the Dead, is a representation of a human skull or skeleton. The term is often applied to edible or decorative skulls made (usually with molds) from either sugar (called Alfeñiques) or clay, used in the Mexican celebration of the Day of the Dead () and the Roman Catholic holiday All Souls' Day. Calavera can also refer to any artistic representations of skulls or skeletons, such as those in the prints of José Guadalupe Posada, or to gifts or treats in relation to the Day of the Dead. Some widely known calaveras are created with cane sugar, decorated with items such as colored foil, icing, beads, and sometimes objects such as feathers. They range in multiple colors.
Traditional methods for producing sugar skulls with molds have been in use for a long time, though the first known mention of the sale of skeletal figures dates to the 1740s. The sugar skulls were originally created as gifts, to be eaten by children. They are sometimes now used as offerings to be placed on altars known as ofrendas ("offerings") for Día de Muertos. It has been argued that the tradition has roots in indigenous celebrations, by groups including the Aztec, Mayan, and Toltec commemorations. However, what we now call Day of the Dead is more Catholic than indigenous because the Spanish tried to eradicate indigenous religions Moreover, as Stanley Brandes has argued, these skulls and skeletons have seven characteristics. They are: (1) ephemeral; (2) seasonal; (3) humorous; (4) secular; (5) commercial; (6) made for living people; (7) meant to be played with; (8) small and transportable; (9) made and consumed by an urban population. They are "lighthearted emblems of death." Thus they are not derived from sacred Mesoamerican traditions.
Sugar skulls were not traditionally used on loved ones' ofrendas, though they are now. In Mexico, children who have died, are celebrated on 1 November. Adults are thought to return on 2 November. It is believed that the departed return home to enjoy the offerings on the altar. Some believe that they consume the essence of the food offerings, others believe they merely sense or savor them without consuming them.
In pre-Columbian times, the images of skulls and skeletons were depicted in stone carvings (and sometimes in the form of real skulls) because bones were thought to be important repositories of life energies and power. The Spanish also utilized skulls as memento mori symbols.
During the 19th and 20th centuries, caricaturists, most eminently Manual Manilla and José Guadalupe Posada made influential calaveras, which were accompanied by satirical, rhymed commentaries. The most famous one was Posada's Catrina, who wears a big feathered hat. She was elaborated by Diego Rivera into a full figure with a long dress, and this figure has been reworked by many other artists. Catrina is the most famous figure associated with the Day of the Dead.
During Day of the Dead, skulls and skeletons are created from many materials such as wood, sugar paste, nuts, chocolate, etc. When sugar skulls are purchased or given as gifts, the name of the deceased is often written with icing across the forehead of the skull on colored foil.
Production
Traditional production methods with molds have been used for a long time. The process involves using molds to cast the calaveras. Production can be a lengthy process: a craftsman who creates elaborate calaveras might spend four to six months producing and decorating the skulls for a season. The most elaborately made sugar skulls are considered folk art, and are not meant to be consumed.
The production process is more focused on the aesthetic appeal of the skull than on the taste or food safety of the product. Furthermore, many calaveras feature inedible decorations, such as beads, feathers, and foil. Some skulls are decorated with sombreros, although these designs are not as popular as they were in the 1970s.
The calaveras are traditionally sold at outdoor market stalls beginning days or a couple of weeks before the Day of the Dead. The most famous place to purchase sugar skulls and related confections (chocolate, marzipan, candied vegetables, etc.) is the Alfeñique fair in Toluca, which is near Mexico City.
Some calaveras are produced to be edible. Most are cast as one piece from cane sugar, which can either be left unflavored or else flavored with vanilla. Some calaveras are also made from chocolate. The calaveras are typically colored with vegetable dyes. As with the more decorative calaveras, these will sometimes have names written on the foreheads, as well. Calaveras may be eaten, or kept for a few days and then thrown away.
Clay skulls
Clay toy variations of calaveras also resemble the shape of human skulls. These toys are often painted a metallic silver color, but they may also be found in colors such as white, black, and red. Beaded eyes of many colors may also be added for decoration.
Literary calaveras
Poetry written for the Day of the Dead are known as literary calaveras, and are intended to humorously criticize the living while reminding them of their mortality. The important precedent for this development is the first illustrated paper that was published in Mexico. Called El Calavera, it began publishing in 1847, illustrated with a skull-faced figure, but the paper was quickly suppressed. An "offering" to President Benito Juarez illustrated with skull and bones was published in 1871. Today literary calaveras are given to family members and friends; published and illustrated versions satirize celebrities and famous organizations in the press.
Literary calaveras flourished in the late 19th century, accompanied by drawings that satirized important politicians. Living personalities are depicted as skeletons, exhibiting recognizable traits, making them easily identifiable. Additionally, drawings of fictional dead personalities often contained text elements providing details of the deaths of various individuals.
Face painting
The act of painting a human face to resemble a skull, sometimes known as facepainting, "sugar skull" make-up, or Catrina make-up, is not a traditional practice during Day of the Dead (except for Catrina impersonators). However, it has become popular in recent years, particularly in urban centers.
See also
Calaca
Papel picado
La Calavera Catrina
Santa Muerte
Grim Fandango
References
External links
Information about Jose Guadalupe Posada, Posada Art Foundation
Information about calaveras de azúcar, Decalaveras Foundation
Information about calaveras literarias, Calaveras Foundation
Spanish words and phrases
Mexican culture
Day of the Dead food
|
M. Saravanan may refer to:
Saravanan Murugan, Malaysian politician
M. Saravanan (film producer), Indian film producer
M. Saravanan (film director), Tamil film director
|
```html
<!DOCTYPE html>
<html xmlns="path_to_url"><head><title>Mat (owl.Owl_regression.Make_Embedded.Optimise.Algodiff.Mat)</title><meta charset="utf-8"/><link rel="stylesheet" href="../../../../../../odoc.support/odoc.css"/><meta name="generator" content="odoc 2.4.2"/><meta name="viewport" content="width=device-width,initial-scale=1.0"/><script src="../../../../../../odoc.support/highlight.pack.js"></script><script>hljs.initHighlightingOnLoad();</script></head><body class="odoc"><nav class="odoc-nav"><a href="../index.html">Up</a> <a href="../../../../../index.html">owl</a> » <a href="../../../../index.html">Owl_regression</a> » <a href="../../../index.html">Make_Embedded</a> » <a href="../../index.html">Optimise</a> » <a href="../index.html">Algodiff</a> » Mat</nav><header class="odoc-preamble"><h1>Module <code><span>Algodiff.Mat</span></code></h1></header><div class="odoc-content"><div class="odoc-spec"><div class="spec value anchored" id="val-empty"><a href="#val-empty" class="anchor"></a><code><span><span class="keyword">val</span> empty : <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-zeros"><a href="#val-zeros" class="anchor"></a><code><span><span class="keyword">val</span> zeros : <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-eye"><a href="#val-eye" class="anchor"></a><code><span><span class="keyword">val</span> eye : <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-ones"><a href="#val-ones" class="anchor"></a><code><span><span class="keyword">val</span> ones : <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-uniform"><a href="#val-uniform" class="anchor"></a><code><span><span class="keyword">val</span> uniform : <span><span class="optlabel">?a</span>:<a href="../A/index.html#type-elt">A.elt</a> <span class="arrow">-></span></span> <span><span class="optlabel">?b</span>:<a href="../A/index.html#type-elt">A.elt</a> <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-gaussian"><a href="#val-gaussian" class="anchor"></a><code><span><span class="keyword">val</span> gaussian : <span><span class="optlabel">?mu</span>:<a href="../A/index.html#type-elt">A.elt</a> <span class="arrow">-></span></span> <span><span class="optlabel">?sigma</span>:<a href="../A/index.html#type-elt">A.elt</a> <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-shape"><a href="#val-shape" class="anchor"></a><code><span><span class="keyword">val</span> shape : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> int * int</span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-numel"><a href="#val-numel" class="anchor"></a><code><span><span class="keyword">val</span> numel : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> int</span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-row_num"><a href="#val-row_num" class="anchor"></a><code><span><span class="keyword">val</span> row_num : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> int</span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-col_num"><a href="#val-col_num" class="anchor"></a><code><span><span class="keyword">val</span> col_num : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> int</span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-reset"><a href="#val-reset" class="anchor"></a><code><span><span class="keyword">val</span> reset : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> unit</span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-reshape"><a href="#val-reshape" class="anchor"></a><code><span><span class="keyword">val</span> reshape : <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-get"><a href="#val-get" class="anchor"></a><code><span><span class="keyword">val</span> get : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-set"><a href="#val-set" class="anchor"></a><code><span><span class="keyword">val</span> set : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-row"><a href="#val-row" class="anchor"></a><code><span><span class="keyword">val</span> row : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-mean"><a href="#val-mean" class="anchor"></a><code><span><span class="keyword">val</span> mean : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-add"><a href="#val-add" class="anchor"></a><code><span><span class="keyword">val</span> add : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-sub"><a href="#val-sub" class="anchor"></a><code><span><span class="keyword">val</span> sub : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-mul"><a href="#val-mul" class="anchor"></a><code><span><span class="keyword">val</span> mul : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-div"><a href="#val-div" class="anchor"></a><code><span><span class="keyword">val</span> div : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-dot"><a href="#val-dot" class="anchor"></a><code><span><span class="keyword">val</span> dot : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-map_by_row"><a href="#val-map_by_row" class="anchor"></a><code><span><span class="keyword">val</span> map_by_row : <span><span>(<span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a>)</span> <span class="arrow">-></span></span> <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-of_arrays"><a href="#val-of_arrays" class="anchor"></a><code><span><span class="keyword">val</span> of_arrays : <span><span><span><a href="../A/index.html#type-elt">A.elt</a> array</span> array</span> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-init_2d"><a href="#val-init_2d" class="anchor"></a><code><span><span class="keyword">val</span> init_2d : <span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <span><span>(<span>int <span class="arrow">-></span></span> <span>int <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a>)</span> <span class="arrow">-></span></span> <a href="../index.html#type-t">t</a></span></code></div></div><div class="odoc-spec"><div class="spec value anchored" id="val-print"><a href="#val-print" class="anchor"></a><code><span><span class="keyword">val</span> print : <span><a href="../index.html#type-t">t</a> <span class="arrow">-></span></span> unit</span></code></div></div></div></body></html>
```
|
Polyeucte drama by Pierre Corneille
Polyeucte (opera), Charles Gounod 1878
Polyeucte, ouverture by Paul Dukas 1891
Polyeucte, film Camille de Morlhon 1910
|
Suctoria are ciliates that become sessile in their developed stage and then lose their redundant cilia. They feed by extracellular digestion. They were originally thought to feed by suction – hence their name. In fact, they use specialized microtubules to ensnare and manipulate their prey. They live in both freshwater and marine environments, including some that live on the surface of aquatic animals, and typically feed on other ciliates. Instead of a single cytostome, each cell feeds by means of several specialized tentacles. These are supported by microtubules and phyllae, and have toxic extrusomes called haptocysts at the tip, which they attach to prey. They then suck the prey's cytoplasm directly into a food vacuole inside the cell, where they digest and absorb its contents. Most suctoria are around 15-30 μm in size, with a non-contractile stalk and often a lorica or shell.
Suctoria reproduce primarily by budding, producing swarmers that lack both tentacles and stalks but have cilia. They may also reproduce through conjugation, which is peculiar in involving cells of different size and often involves total fusion. The way buds form is the primary distinction between different orders of suctoria. Among the Exogenida, including common genera like Podophrya and Sphaerophrya, they appear directly on the cell surface. Among the Endogenida, for instance Tokophrya and Acineta, they form in an internal pouch and escape through an opening—and among the Evaginogenida, they form in a pouch that inverts before they are released.
Once the swarmers have found a place to attach themselves, they quickly develop stalks and tentacles. The cilia are lost, but the underlying infraciliature persists throughout the entire life-cycle. This has a structure that, together with other ultrastructural similarities, places the suctoria within the class Phyllopharyngea.
References
External links
Suctoria pictures and videos
Phyllopharyngea
SAR supergroup subclasses
|
```javascript
/**
* @license Apache-2.0
*
*
*
* path_to_url
*
* Unless required by applicable law or agreed to in writing, software
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*/
'use strict';
var randu = require( '@stdlib/random/base/randu' );
var incrnansumabs2 = require( './../lib' );
var accumulator;
var sum;
var v;
var i;
// Initialize an accumulator:
accumulator = incrnansumabs2();
// For each simulated datum, update the sum...
console.log( '\nValue\tSum\n' );
for ( i = 0; i < 100; i++ ) {
if ( randu() < 0.2 ) {
v = NaN;
} else {
v = ( randu()*100.0 ) - 50.0;
}
sum = accumulator( v );
console.log( '%d\t%d', v.toFixed( 3 ), ( sum === null ) ? NaN : sum.toFixed( 3 ) );
}
console.log( '\nFinal sum: %d\n', accumulator() );
```
|
Wonder World is a natural theme park located in San Marcos, Texas. The park's primary attraction is Wonder Cave, an ancient earthquake cave and Recorded Texas Historic Landmark. Other attractions include an anti-gravity house, a trackless motor train, and a wildlife park.
References
External links
Official Website
Buildings and structures in San Marcos, Texas
Recorded Texas Historic Landmarks
San Marcos, Texas
Tourist attractions in Hays County, Texas
|
János Czetz (June 8, 1822 – September 6, 1904) was a prominent Hungarian freedom fighter of Armenian and Hungarian-Székely origin, a military commander during the Hungarian Revolution of 1848, and the organizer of Argentina's first national military academy.
From 1870, Cztetz was one of the main organizers and first director of the National Military College, an important contributor to the military government of Argentine president Domingo Faustino Sarmiento. The collaboration helped Argentina become one of the most advanced armies outside of Europe. The strict discipline installed at the College and indirectly in the Army, helped bring an end to the long period of the Argentine civil wars. He published a "Treatise on permanent and temporary fortification".
References
1822 births
1904 deaths
People from Covasna County
Argentine people of Armenian descent
Hungarian people of Armenian descent
Hungarian emigrants to Argentina
Burials at La Recoleta Cemetery
People of the Hungarian Revolution of 1848
Theresian Military Academy alumni
|
Donington and Albrighton is a local nature reserve being a valuable site for wildlife and recreation in Shropshire, it has a historical reference as it is the site of St Cuthberts Well which was believed to hold powers to cure the blind.
Location
This site is in Albrighton Village in Humphreston Brook Valley, in Shropshire, close to the Church. Travelling to the site can be done by train, car or bus, with a carpark for the reserve signposted by Donington Church. Postcode: WV7 3EP. Grid reference: SJ 809046.
Background
This 5.4 hectare Nature reserve site in the villages of Donington and Albrighton is located precisely in the valley of the Humphreston Brook.
Wildlife
St. Cuthbert's Meadow is the main part of the reserve, and is mostly used as a recreational area by the locals.
The St. Cuthbert's well is surrounded by older trees like Beech, Hazel, but most noticeably Old Yew. These mature trees make it an important area for nesting birds.
The Donington Pool itself is a valuable area for breeding wildfowl, some fishing is allowed on this pool but is done through the Royal British Legion.
At the far side of Donington Pool is a willow carr woodland area, with streams and marshes. This woodland is not open to the public due to conflicting issues with Health and Safety, but is still an important area for wildlife.
More than ninety different bird types have been found either in, or flying over the site, and 30 bird species are known to breed in this reserve.
History of the site
Historically the margin marking the boundary between Donington and Albrighton is shown by the Humphreston Brook. John Talbot, a miller in the early 17th century, made Donington Pool by damming the brook at Rectory Road, in order to power his mill by providing a non-stop water supply.
However because of this damming he was fined by Donington, but not by Albrighton due to permission for the mill already being approved.
The original pool created was bigger but due to stages of Hydrosere Succession it is gradually shrinking and will eventually turn into marshland, and wet woodland.
The famous well of St. Cuthberts is believed to hold miraculous water that can cure the blind.
Art and the Local Nature Reserve
The fishing pool in the LNR has been awarded £50,000 to be spent on improving access to the public and those with disabilities.
These improvements in Donington Pool, Albrighton, include new fishing pegs and an access platform to allow all visitors to get to the water's edge.
An art project funded by the arts council has recently occurred on this LNR site.
This involved Richard Taylor, an artist who with the help of the local people designed the centre piece. This centre piece was placed in the pool, and is made up of copper sculptured fish, with open mouths that spray water.
To represent the wildflowers present on the site, metal and glass sculptures were placed in the meadow, and brass rubbings were used to include the history of St. Cuthberts well in the design.
As the area needs to be maintained, RAF Cosford Trainees were asked to help. This work involved repairing a pathway, and clearing vegetation.
Gallery
References
External links
http://www.dalnr.org/
Nature reserves in Shropshire
|
"I Do" is a song by South Korean girl group (G)I-dle. It was released on July 14, 2023, through Cube and 88rising, as the group's first original English single from their first English-language EP Heat. It was written by its producers Rogét Chahayed and Imad Royal with Blaise Railey and Drew Love. Musically, it is a pop song that utilizes 1980s synth pop sound, while its lyrics revolve around a story about a lost love and the bitterness that comes with it.
Commercially, "I Do" peaked at number 27 on the Billboard US Pop Airplay and at number 39 on the New Zealand Hot Singles RMNZ. An accompanying music video was uploaded onto 88rising's YouTube channel simultaneously with the single's release and has surpassed 19 million views on the platform. On July 24, an alternative music video titled "I Do: A Love Story in Bangkok," starring Tontawan Tantivejakul and Jirawat Sutivanichsak, was posted on 88rising's YouTube account.
Background and release
On May 15, 2023, the group released their sixth Korean-language I Feel, which was a commercial success debuting at number one on the Circle Chart with one million copies sold in the first week and receiving positive reviews from domestic critics. On July 10, Billboard reported a partnership between the group's parent company Cube Entertainment and the American record label 88rising, where the group would release their first English-language extended play titled Heat, with "I Do" serving as the lead single, released on July 13, 2023, on streaming platforms.
Composition
"I Do" is a pop song inspired by 1980s synth-pop elements. It is characterized by "the harmonization of relaxed tempos" and electronic elements "that give it a dreamy feel," while its lyrics revolve around a "lost love of the past and the bitterness that accompanies it."
Music video
An accompanying music video for "I Do" was preceded by five concept images and two teaser videos. In the video the members take on the role of a shapeshifting alien who crash lands on Earth and meets a "kind stranger who takes her in, and the pair soon fall in love." However, the duo are chased down by "mysterious men in hazmat suits," and are later cornered in a deserted warehouse. In order to save her love, the alien uses her magic powers and sacrifices herself.
An alternative music video titled "I Do: A Love Story in Bangkok," starring Tontawan Tantivejakul and Jirawat Sutivanichsak but with no appearances from (G)I-dle, was posted on 88rising's YouTube account on July 24. The video depicts Tontawan and Jirawat as students falling in love in high school before an emotional break up at prom. The video ends with the two reunited, presumably years later, after a chance encounter at a park.
Credits and personnel
(G)I-dle – vocals
Imad Royal – songwriter, producer
Rogét Chahayed – songwriter, producer
Blaise Railey – songwriter
Drew Love – songwriter
Choi Ye-Ji – recording engineer
Tristan Hoogland – mixing engineer
Dale Becker – mastering engineer
Katie Harvey – assistant mastering engineer
Charts
Release history
References
External links
(G)I-dle songs
2023 singles
2023 songs
88rising singles
Cube Entertainment singles
South Korean pop songs
South Korean synth-pop songs
English-language South Korean songs
Songs written by Rogét Chahayed
|
Geordie (released in the United States as Wee Geordie) is a 1955 British film directed and co-produced by Frank Launder, with Bill Travers in the title role as a Scotsman who becomes an athlete and competes at the 1956 Summer Olympics in Melbourne.
The film is based on David Walker's 1950 novel of the same title, adapted for the screen by Launder and his co-producer Sidney Gilliat.
Plot
The story begins in a small Highland school classroom. Geordie MacTaggart is a "wee" (small) Scottish schoolboy, and the son of a gamekeeper. Although his best friend Jean does not mind his height, after he sees a newspaper advertisement for a bodybuilding correspondence course offered by Henry Samson, he sends for the course and embarks diligently on Samson's fitness programme. By the time Geordie turns 21, he has grown into a tall, fit man who continues to follow Samson's long-distance instructions. Jean, however, disapproves of the amount of time he spends training.
Geordie works as assistant to his father, the local laird's head gamekeeper. One day, when they are out together in a storm, his father becomes ill. Geordie carries him home many miles, but his father develops pneumonia and dies. The laird (Alistair Sim) makes Geordie the new gamekeeper.
One day, he gets a letter from Samson, who suggests he take up hammer throwing. On his first attempt, he almost hits the laird, who then tries to show him how it is done. However, the laird's own hammer throw almost hits the local minister, who is passing by on his bike. It turns out that the minister is knowledgeable about the sport, and he trains Geordie. At the minister's urging, Geordie reluctantly enters a Highland games event. He initially makes two bad throws, but after the unexpected appearance (and encouragement) of Jean, he wins with his final throw.
Two members of the Olympics selection committee visit him and invite him to join the British team for the Melbourne Olympic Games in Australia. Geordie is once again reluctant, as he does not particularly care to compete against others, but finally agrees. He takes the train to London, where he finally gets to meet Henry Samson, who has come to see him off when he boards the ship for Australia.
Unhappy to be away from home, Geordie finds it difficult to be enthusiastic about training on board ship. However, Helga, a Danish female shot putter, takes a shine to Geordie and talks him out of his mood, though Geordie remains oblivious to the fact that she is attracted to him. When they reach Melbourne, Geordie goes sightseeing with Helga before the games, buying a highly unusual hat for Jean. An accident occurs nearby, and a man is pinned underneath a car. After several men working together are unable to lift the car, Geordie manages to do it all by himself. His feat is reported in the newspapers, and he becomes very popular.
A problem arises: Geordie insists on wearing his late father's Black Watch kilt in the opening ceremony, something he had promised his mother he would do. When he is told that he must wear the same uniform as the rest of the team, he states "no kilt, no performance!" Not having received a reply from London, Lord Pauceton, the head of the British team, gives in. After Geordie comes out last in the opening parade of athletes in his kilt, Pauceton receives a telegram emphatically ordering him not to let Geordie wear his kilt, but he ignores it.
During the competition, a listless, dispirited Geordie fails with his first two throws. Then, before his third and final throw, he finds inspiration by recalling Jean's encouragement at his first competition. He then makes a world record throw and wins the competition. However, Jean hears on the radio that Helga has rushed up, embracing and kissing Geordie in front of everyone in the stadium, and she is heartbroken.
On Geordie's return, there is no one to meet him at the station apart from his mother and a driver for the trap. On the way home, they encounter the laird, who tells him that many think his actions have brought scandal to the glen. Geordie spots Jean fishing and goes to her. They argue, then fall in the stream. After they get out, Geordie shows her the hat he bought for her; she pretends to think it is "braw" (fine), and they kiss and make up.
Main cast
Alastair Sim as The Laird
Bill Travers as Geordie
Paul Young as Young Geordie
Anna Ferguson as Young Jean
Norah Gorsen as Jean
Molly Urquhart as Geordie's mother
Francis de Wolff as Henry Samson
Jack Radcliffe as The Minister
Brian Reece as Olympic Selector
Raymond Huntley as Olympic Selector
Miles Malleson as Lord Paunceton
Jameson Clark as Geordie's father
Doris Goddard as Helga
Stanley Baxter as Postman
Duncan Macrae as Schoolmaster
Michael Ripper as Australian journalist
Production
The film was going to be directed by Sidney Gilliat but he was caught up with post production problems on The Constant Husband due to colour stock so Launder stepped in.
Reception
The film premiered at the Plaza in London on 2 September 1955. The Times' poetic reviewer found the film to have "gracious and decorous atmosphere; the steep hills have stamped themselves on everyone's imagination, so that the story-telling is not flat, the dialogue never airless. The earth is beneath us, the sky above, and to receive messages from both and to interpret them is the splendidly authentic figure of Mr. Alastair Sim's Laird".
In real life, the hammer throw at the Melbourne Olympics was won by American Hal Connolly.
Box office
According to the National Film Finance Corporation, the film made a comfortable profit. According to Kinematograph Weekly, it was a "money maker" at the British box office in 1955.
See also
List of films about the sport of athletics
References
External links
1955 films
1950s sports comedy-drama films
British sports comedy-drama films
Athletics films
Films about the 1956 Summer Olympics
Films based on Canadian novels
Films set in 1956
Films set in Scotland
Films directed by Frank Launder
Films with screenplays by Frank Launder and Sidney Gilliat
Films scored by William Alwyn
1955 comedy films
1955 drama films
1950s English-language films
1950s British films
|
Alf Bjørseth (born 24 November 1941) is a Norwegian chemist and businessperson in the energy sector.
He hails from Gamle Oslo. After finishing his secondary education in 1960, he took the cand.real. degree at the University of Oslo in 1969. He was subsequently employed at SI and took his dr.philos. degree in 1979.
After spending the years 1980 to 1982 as research director at SI, he started a career in the industry sector. He was vice executive of Deminex from 1982 to 1984, director of research at Norsk Hydro from 1984 to 1990 and director of technology at Elkem from 1990 to 1994. In 1994 he founded the company ScanWafer, later REC. In 2005 he founded Norsun, then sold his shares in REC, and founded Scatec.
Bjørseth was decorated with Norwegian Chemical Society's Guldberg and Waage Medal in 2003 and became Commander of the Order of St. Olav in 2006. He is a member of the Norwegian Academy of Science and Letters and the Norwegian Academy of Technological Sciences.
References
1941 births
Living people
Scientists from Oslo
Norwegian chemists
University of Oslo alumni
Businesspeople from Oslo
Norsk Hydro people
Businesspeople in energy
People associated with solar power
Members of the Norwegian Academy of Science and Letters
Members of the Norwegian Academy of Technological Sciences
|
```javascript
/**
* Graphology Connected Caveman Graph Generator
* =============================================
*
* Function generating connected caveman graphs.
*
* [Article]:
* Watts, D. J. 'Networks, Dynamics, and the Small-World Phenomenon.'
* Amer. J. Soc. 105, 493-527, 1999.
*/
var isGraphConstructor = require('graphology-utils/is-graph-constructor'),
empty = require('../classic/empty.js');
/**
* Function returning a connected caveman graph with desired properties.
*
* @param {Class} GraphClass - The Graph Class to instantiate.
* @param {number} l - The number of cliques in the graph.
* @param {number} k - Size of the cliques.
* @return {Graph}
*/
module.exports = function connectedCaveman(GraphClass, l, k) {
if (!isGraphConstructor(GraphClass))
throw new Error(
'graphology-generators/community/connected-caveman: invalid Graph constructor.'
);
var m = l * k;
var graph = empty(GraphClass, m);
if (k < 2) return graph;
var i, j, s;
for (i = 0; i < m; i += k) {
for (j = i; j < i + k; j++) {
for (s = j + 1; s < i + k; s++) {
if (j !== i || j !== s - 1) graph.addEdge(j, s);
}
}
if (i > 0) graph.addEdge(i, (i - 1) % m);
}
graph.addEdge(0, m - 1);
return graph;
};
```
|
The name Cristobal has been used for four tropical cyclones in the Atlantic Ocean. Cristobal replaced the name Cesar after the 1996 season.
Tropical Storm Cristobal (2002), a relatively weak tropical storm causing only minor damage in Bermuda
Tropical Storm Cristobal (2008), formed near the South Carolina coast causing minimal damage
Hurricane Cristobal (2014), a Category 1 hurricane that affected Caribbean islands, Bermuda, and the United States East Coast
Tropical Storm Cristobal (2020), formed over the Bay of Campeche from the remnants of Tropical Storm Amanda from the East Pacific, earliest third named storm in the Atlantic basin; made landfall in Mexico, then slowly turned north into the Gulf of Mexico and made a second landfall in Louisiana as a moderately-strong tropical storm.
Atlantic hurricane set index articles
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.