text stringlengths 12 14.7k |
|---|
Proof of identity (blockchain consensus) : Proof of identity (PoID) is a consensus protocol for permission-less blockchains, in which each uniquely identified individual receives one equal unit of voting power and associated rewards (minting token). The protocol is based on biometric identification, humanity identifica... |
Proof of identity (blockchain consensus) : The proof of identity protocol combines state of the art of 3D mapping face recognition technologies, the attendance to humanity identification parties, and the decentralized (stored in the blockchain) supervision of an AI engine that randomly forms parties and carries on orga... |
Proof of identity (blockchain consensus) : The proof of identity consensus protocol has the prerogatives of being: Non-quantitative: once the network is sufficiently populated, any blockchain/cryptocurrency can join the network and have its blocks validated by the PoID. Permanently secured against Sybil attacks: the Po... |
Proof of personhood : Proof of personhood (PoP) is a means of resisting malicious attacks on peer to peer networks, particularly, attacks that utilize multiple fake identities, otherwise known as a Sybil attack. Decentralized online platforms are particularly vulnerable to such attacks by their very nature, as notional... |
Proof of personhood : The problem of Sybil attacks using many virtual identities has been recognized for decades as a fundamental challenge for distributed systems that expect each human user to have only one account or identity. CAPTCHAs attempt to rate-limit automated Sybil attacks by using automated Turing tests to ... |
Proof of personhood : A variety of approaches to implementing proof of personhood have been proposed, some in experimental deployment. |
Proof of personhood : One proposed use for proof of personhood is to ensure that voting power in permissionless consensus algorithms is widely distributed, and to avoid the re-centralization that has been observed in proof of work mining pools, and predicted in proof of stake systems. Another proposed use is to facilit... |
Proof of personhood : Proof of authority Proof of space == References == |
Reverse Turing test : A reverse Turing test is a Turing test in which failure suggests that the test-taker is human, while success suggests the test-taker is automated. Conventionally, the Turing test is conceived as having a few computer AI subjects communicate with each other and one human subject which attempts to a... |
Reverse Turing test : Arguably the standard form of the reverse Turing test is one in which the subjects attempt to appear to be a computer rather than a human. A formal reverse Turing test follows the same format as a Turing test. Human subjects attempt to imitate the conversational style of a conversation program. Do... |
Reverse Turing test : Since Turing test judges are sometimes presented with genuinely human subjects, as a control, it inevitably occurs that a small proportion of such control subjects are judged to be computers. This is considered humorous and often embarrassing for the subject. This situation may be described litera... |
Reverse Turing test : The term "reverse Turing test" has also been applied to a Turing test (test of humanity) that is administered by a computer. In other words, a computer administers a test to determine if the subject is or is not human. Such procedures, called CAPTCHAs, are used in some anti-spam systems to prevent... |
Reverse Turing test : An alternative conception of a Reverse Turing Test is to use the test to determine whether sufficient information is being transmitted between the tester and the subject. For example, if the information sent by the tester is insufficient for the human doctor to perform diagnosis accurately, then a... |
Reverse Turing test : Wizard of Oz experiment Turing test |
Reverse Turing test : How I failed the Turing test (PDF) by Jason Striegel "Hello, Are You Human?" Cocktail hour inversion of the Turing Test "The Blurring Test/MrMind" MrMind administers The Blurring Test by asking, "Can you convince me that you are human?" |
The Turing Test (video game) : The Turing Test is a first-person puzzle video game developed by Bulkhead Interactive and published by Square Enix Collective. The game was released for Microsoft Windows and Xbox One in August 2016, for PlayStation 4 in January 2017, for Nintendo Switch in February 2020, and for Stadia i... |
The Turing Test (video game) : The Turing Test is a puzzle video game with gameplay experienced from a first-person perspective. The player assumes the role of Ava Turing, an International Space Agency (ISA) engineer working at a research station located on Europa, one of Jupiter's moons. The puzzles involve distributi... |
The Turing Test (video game) : In the far future, engineer Ava Turing is one of several members of a research team sent via the International Space Agency (ISA) to excavate Jupiter's moon Europa. While Ava remains in cryogenic slumber, the other team members are woken and travel to the moon to set up their base and beg... |
The Turing Test (video game) : In 2015, UK-based studios Deco Digital and Bevel Studios released their first-person puzzle game, Pneuma: Breath of Life. Towards the end of the year, the two studios merged to form Bulkhead Interactive, the developer of The Turing Test. The game was developed using Unreal Engine 4. The T... |
The Turing Test (video game) : The Turing Test received "generally favourable" reviews, according to review aggregator Metacritic. |
Visual Turing Test : The Visual Turing Test is “an operator-assisted device that produces a stochastic sequence of binary questions from a given test image”. The query engine produces a sequence of questions that have unpredictable answers given the history of questions. The test is only about vision and does not requi... |
Visual Turing Test : Research in computer vision dates back to the 1960s when Seymour Papert first attempted to solve the problem. This unsuccessful attempt was referred to as the Summer Vision Project. The reason why it was not successful was because computer vision is more complicated than what people think. The comp... |
Visual Turing Test : A large number of datasets have been annotated and generalised to benchmark performances of difference classes of algorithms to assess different vision tasks (e.g., object detection/recognition) on some image domain (e.g., scene images). One of the most famous datasets in computer vision is ImageNe... |
Visual Turing Test : The Visual Turing Test (VTT) unlike the Turing test has a query engine system which interrogates a computer vision system in the presence of a human co-ordinator. It is a system that generates a random sequence of binary questions specific to the test image, such that the answer to any question k i... |
Visual Turing Test : To select the next question in the sequence, VTT has to estimate the predictability of every proposed question. This is done using the annotated training set of Images. Each Image is annotated with bounding box around the objects and labelled with the attributes, and pairs of objects are labelled w... |
Visual Turing Test : Detailed example sequences can be found here. |
Visual Turing Test : The Images considered for the Geman et al. work are that of ‘Urban street scenes’ dataset, which has scenes of streets from different cities across the world. This why the types of objects are constrained to people and vehicles for this experiment. Another dataset introduced by the Max Planck Insti... |
Visual Turing Test : This is a very recent work published on March 9, 2015, in the journal Proceedings of the National Academy of Sciences, by researchers from Brown University and Johns Hopkins University. It evaluates how the computer vision systems understand the Images as compared to humans. Currently the test is w... |
Winograd schema challenge : The Winograd schema challenge (WSC) is a test of machine intelligence proposed in 2012 by Hector Levesque, a computer scientist at the University of Toronto. Designed to be an improvement on the Turing test, it is a multiple-choice test that employs questions of a very specific structure: th... |
Winograd schema challenge : The Winograd Schema Challenge was proposed in the spirit of the Turing test. Proposed by Alan Turing in 1950, the Turing test plays a central role in the philosophy of artificial intelligence. Turing proposed that, instead of debating whether a machine can think, the science of AI should be ... |
Winograd schema challenge : The key factor in the WSC is the special format of its questions, which are derived from Winograd schemas. Questions of this form may be tailored to require knowledge and commonsense reasoning in a variety of domains. They must also be carefully written not to betray their answers by selecti... |
Winograd schema challenge : In 2016 and 2018, Nuance Communications sponsored a competition, offering a grand prize of $25,000 for the top scorer above 90% (for comparison, humans correctly answer to 92–96% of WSC questions). However, nobody came close to winning the prize in 2016 and the 2018 competition was cancelled... |
Winograd schema challenge : Website for the contest sponsored by Nuance Communications Kocijan, Vid; Davis, Ernest; Lukasiewicz, Thomas; Marcus, Gary; Morgenstern, Leora (1 December 2023). "The defeat of the Winograd Schema Challenge". Artificial Intelligence. 325: 103971. arXiv:2201.02387. doi:10.1016/j.artint.2023.10... |
AI Safety Institute : An AI Safety Institute (AISI), in general, is a state-backed institute aiming to evaluate and ensure the safety of the most advanced artificial intelligence (AI) models, also called frontier AI models. AI safety gained prominence in 2023, notably with public declarations about potential existentia... |
AI Safety Institute : In 2023, Rishi Sunak, the Prime Minister of the United Kingdom, expressed his intention to "make the U.K. not just the intellectual home but the geographical home of global AI safety regulation" and unveiled plans for an AI Safety Summit. He emphasized the need for independent safety evaluations, ... |
AI Safety Institute : The United Kingdom founded in April 2023 a safety organisation called Frontier AI Taskforce, with an initial budget of £100 million. In November 2023, it evolved into the UK AISI, and continued to be led by Ian Hogarth. The AISI is part of the United Kingdom's Department for Science, Innovation an... |
AI Safety Institute : The US AISI was founded in November 2023 as part of the NIST. This happened the day after the signature of the Executive Order 14110. In February 2024, Joe Biden's former economic policy adviser Elizabeth Kelly was appointed to lead it. In February 2024, the US government created the US AI Safety ... |
AI Safety Institute : The Ministry of Electronics and Information Technology held consultations with Meta Platforms, Google, Microsoft, IBM, OpenAI, NASSCOM, Broadband India Forum, Software Alliance, Indian Institutes of Technology, The Quantum Hub, Digital Empowerment Foundation, and Access Now on October 7, 2024, in ... |
AI Safety Institute : Alignment Research Center Foundation model Regulation of artificial intelligence |
AI Safety Institute : European AI Office Japan AI Safety Institute UK AI Safety Institute US AI Safety Institute KR AI Safety Institute |
AI Safety Summit : The AI Safety Summit was an international conference discussing the safety and regulation of artificial intelligence. It was held at Bletchley Park, Milton Keynes, United Kingdom, on 1–2 November 2023. It was the first ever global summit on artificial intelligence, and is planned to become a recurrin... |
AI Safety Summit : The prime minister of the United Kingdom at the time, Rishi Sunak, made AI one of the priorities of his government, announcing that the UK would host a global AI Safety conference in autumn 2023. |
AI Safety Summit : Bletchley Park was a World War II codebreaking facility established by the British government on the site of a Victorian manor and is in the British city of Milton Keynes. It has played an important role in the history of computing, with some of the first modern computers being built at the facility. |
AI Safety Summit : 28 countries at the summit, including the United States, China, Australia, and the European Union, have issued an agreement known as the Bletchley Declaration, calling for international co-operation to manage the challenges and risks of artificial intelligence. The Bletchley Declaration affirms that ... |
AI Safety Summit : The following individuals attended the summit: Rishi Sunak, Prime Minister of the United Kingdom Kamala Harris, Vice President of the United States Charles III, King of the United Kingdom (attending virtually) Elon Musk, CEO of Tesla, owner of X, SpaceX, Neuralink, and xAI Giorgia Meloni, Prime Minis... |
AI Safety Summit : AI Summits continued in subsequent years, including: AI Seoul Summit: hosted by South Korea on 21–22 May 2024. AI Action Summit: held in France on 10–11 February 2025. == References == |
Existential risk from artificial intelligence : Existential risk from artificial intelligence refers to the idea that substantial progress in artificial general intelligence (AGI) could lead to human extinction or an irreversible global catastrophe. One argument for the importance of this risk references how human bein... |
Existential risk from artificial intelligence : One of the earliest authors to express serious concern that highly advanced machines might pose existential risks to humanity was the novelist Samuel Butler, who wrote in his 1863 essay Darwin among the Machines: The upshot is simply a question of time, but that the time ... |
Existential risk from artificial intelligence : An existential risk is "one that threatens the premature extinction of Earth-originating intelligent life or the permanent and drastic destruction of its potential for desirable future development". Besides extinction risk, there is the risk that the civilization gets per... |
Existential risk from artificial intelligence : The alignment problem is the research problem of how to reliably assign objectives, preferences or ethical principles to AIs. |
Existential risk from artificial intelligence : A December 2024 study by Apollo Research found that advanced LLMs like OpenAI o1 sometimes deceive in order to accomplish their goal, to prevent them from being changed, or to ensure their deployment. Forms of deception observed in the experimental environment and scenari... |
Existential risk from artificial intelligence : Some scholars have proposed hypothetical scenarios to illustrate some of their concerns. |
Existential risk from artificial intelligence : The thesis that AI could pose an existential risk provokes a wide range of reactions in the scientific community and in the public at large, but many of the opposing viewpoints share common ground. Observers tend to agree that AI has significant potential to improve socie... |
Existential risk from artificial intelligence : Many scholars concerned about AGI existential risk believe that extensive research into the "control problem" is essential. This problem involves determining which safeguards, algorithms, or architectures can be implemented to increase the likelihood that a recursively-im... |
Existential risk from artificial intelligence : Clark, Jack (2015a). "Musk-Backed Group Probes Risks Behind Artificial Intelligence". Bloomberg.com. Archived from the original on 30 October 2015. Retrieved 30 October 2015. |
Death of Elaine Herzberg : The death of Elaine Herzberg (August 2, 1968 – March 18, 2018) was the first recorded case of a pedestrian fatality involving a self-driving car, after a collision that occurred late in the evening of March 18, 2018. Herzberg was pushing a bicycle across a four-lane road in Tempe, Arizona, Un... |
Death of Elaine Herzberg : Herzberg was crossing Mill Avenue (North) from west to east, approximately 360 feet (110 m) south of the intersection with Curry Road, outside the designated pedestrian crosswalk, close to the Red Mountain Freeway. She was pushing a bicycle laden with shopping bags, and had crossed at least t... |
Death of Elaine Herzberg : The county district attorney's office recused itself from the investigation, due to a prior joint partnership with Uber promoting their services as an alternative to driving under the influence of alcohol. Accounts of the crash have been conflicting in terms of the speed limit at the place of... |
Death of Elaine Herzberg : Prior to the fatal incident, Arizona Governor Doug Ducey had encouraged Uber to enter the state. He signed Executive Order 2015-09 on August 25, 2015, entitled "Self-Driving Vehicle Testing and Piloting in the State of Arizona; Self-Driving Vehicle Oversight Committee", establishing a welcomi... |
Death of Elaine Herzberg : Mary Ward, the first person known to have been killed by an automobile, 1869 Bridget Driscoll, the first pedestrian death by automobile in Great Britain, 1896 Henry H. Bliss, the first automobile death in the Americas, 1899 Thomas Selfridge, the first person to die in an airplane crash, 1908 ... |
Death of Elaine Herzberg : NTSB investigation of Uber crash, Accident No. HWY18FH010 Dashcam video related to crash, via BBC Davies, Alex (June 22, 2018). "The unavoidable folly of making humans train self-driving cars". Wired. Retrieved June 26, 2018. Video illustrating issue of speed and Sight Distance PBS NOVA: Look... |
Regulation of artificial intelligence : Regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI). It is part of the broader regulation of algorithms. The regulatory and policy landscape for AI is an emerging issue in jurisdicti... |
Regulation of artificial intelligence : According to Stanford University's 2023 AI Index, the annual number of bills mentioning "artificial intelligence" passed in 127 surveyed countries jumped from one in 2016 to 37 in 2022. In 2017, Elon Musk called for regulation of AI development. According to NPR, the Tesla CEO wa... |
Regulation of artificial intelligence : The regulation of artificial intelligences is the development of public sector policies and laws for promoting and regulating AI. Regulation is now generally considered necessary to both encourage AI and manage associated risks. Public administration and policy considerations gen... |
Regulation of artificial intelligence : The development of a global governance board to regulate AI development was suggested at least as early as 2017. In December 2018, Canada and France announced plans for a G7-backed International Panel on Artificial Intelligence, modeled on the International Panel on Climate Chang... |
Regulation of artificial intelligence : The regulatory and policy landscape for AI is an emerging issue in regional and national jurisdictions globally, for example in the European Union and Russia. Since early 2016, many national, regional and international authorities have begun adopting strategies, actions plans and... |
Regulation of artificial intelligence : Legal questions related to lethal autonomous weapons systems (LAWS), in particular compliance with the laws of armed conflict, have been under discussion at the United Nations since 2013, within the context of the Convention on Certain Conventional Weapons. Notably, informal meet... |
Regulation of artificial intelligence : AI alignment Algorithmic accountability Algorithmic bias Artificial intelligence Artificial intelligence and elections Artificial intelligence arms race Artificial intelligence in government Ethics of artificial intelligence Government by algorithm Legal informatics Regulation of... |
Representational harm : Systems cause representational harm when they misrepresent a group of people in a negative manner. Representational harms include perpetuating harmful stereotypes about or minimizing the existence of a social group, such as a racial, ethnic, gender, or religious group. Machine learning algorithm... |
Representational harm : Three prominent types of representational harm include stereotyping, denigration, and misrecognition. These subcategories present many dangers to individuals and groups. Stereotypes are oversimplified and usually undesirable representations of a specific group of people, usually by race and gend... |
Representational harm : As the dangers of representational harm have become better understood, some researchers have developed methods to measure representational harm in algorithms. Modeling stereotyping is one way to identify representational harm. Representational stereotyping can be quantified by comparing the pred... |
Representational harm : One of the most notorious examples of representational harm was committed by Google in 2015 when an algorithm in Google Photos classified Black people as gorillas. Developers at Google said that the problem was caused because there were not enough faces of Black people in the training dataset fo... |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : The Safe and Secure Innovation for Frontier Artificial Intelligence Models Act, or SB 1047, was a failed 2024 California bill intended to "mitigate the risk of catastrophic harms from AI models so advanced that they are not yet known to exist"... |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : The rapid increase in capabilities of AI systems in the 2020s, including the release of ChatGPT in November 2022, caused some researchers and members of the public to become concerned existential risks associated with increasingly powerful AI ... |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : SB 1047 would have covered AI models with training compute over 1026 integer or floating-point operations and a cost of over $100 million. If a covered model is fine-tuned using more than $10 million, the resulting model would also have been c... |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : The Artificial Intelligence Policy Institute, a pro-regulation AI think tank, ran three polls of California respondents on whether they supported or opposed SB 1047. The third poll asked the question "Some policy makers are proposing a law in ... |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : Artificial general intelligence Regulation of AI in the United States Regulation of artificial intelligence |
Safe and Secure Innovation for Frontier Artificial Intelligence Models Act : Bill tracker CalMatters Supporting website Economic Security California Action, Center for AI Safety Action Fund, and Encode Justice Opposing website Andreessen Horowitz |
Semantic reasoner : A semantic reasoner, reasoning engine, rules engine, or simply a reasoner, is a piece of software able to infer logical consequences from a set of asserted facts or axioms. The notion of a semantic reasoner generalizes that of an inference engine, by providing a richer set of mechanisms to work with... |
Semantic reasoner : Notable semantic reasoners and related software: |
Semantic reasoner : Business rules engine Doxastic logic Expert systems Logic programming Method of analytic tableaux Solver |
Semantic reasoner : OWL 2 Reasoners listed on W3C SW Working Group homepage SPARQL Query Language for RDF Marko Luther, Thorsten Liebig, Sebastian Böhm, Olaf Noppens: Who the Heck Is the Father of Bob?. ESWC 2009: 66-80 Jurgen Bock, Peter Haase, Qiu Ji, Raphael Volz. Benchmarking OWL Reasoners. Mirror available. In ARe... |
BRFplus : BRFplus (Business Rule Framework plus) is a business rule management system (BRMS) offered by SAP AG. BRFplus is part of the SAP NetWeaver ABAP stack. Therefore, all SAP applications that are based on SAP NetWeaver can access BRFplus within the boundaries of an SAP system. However, it is also possible to gene... |
BRFplus : BRFplus offers a unified modeling and runtime environment for business rules that addresses both technical users (programmers, system administrators) as well as business users who take care of operational business processes (like procurement, bidding, tax form validation, etc.). The different requirements and... |
BRFplus : BRFplus is positioned as the successor product of an older business rule solution known as BRF (Business Rule Framework). For a longer transition phase, both solutions exist in parallel. However, an increasing number of SAP applications that used to be based on BRF are migrating to BRFplus. While BRFplus supp... |
BRFplus : BRFplus has been delivered to the public with SAP NetWeaver 7.0 Enhancement Package 1 for the first time. Being part of SAP NetWeaver, the usage of BRFplus is covered by the "SAP NetWeaver Foundation for Third Party Applications" license, with no additional costs. |
BRFplus : Carsten Ziegler, Thomas Albrecht: BRFplus – Business Rule Management for ABAP Applications. Galileo Press 2011. ISBN 978-1-59229-293-6 |
BRFplus : BRFplus on SAP Developer's Network (SDN) BRFplus online documentation on SAP Help Portal BRF Plus-A real time example |
Business rule management system : A BRMS or business rule management system is a software system used to define, deploy, execute, monitor and maintain the variety and complexity of decision logic that is used by operational systems within an organization or enterprise. This logic, also referred to as business rules, in... |
Business rule management system : A BRMS includes, at minimum: A repository, allowing decision logic to be externalized from core application code Tools, allowing both technical developers and business experts to define and manage decision logic A runtime environment, allowing applications to invoke decision logic mana... |
Business rule management system : In a BRMS, a representation of business rules maps to a software system for execution. A BRMS therefore relates to model-driven engineering, such as the model-driven architecture (MDA) of the Object Management Group (OMG). It is no coincidence that many of the related standards come un... |
Business rule management system : The OMG Decision Model and Notation standard is designed to standardize elements of business rules development, specially decision table representations. There is also a standard for a Java Runtime API for rule engines JSR-94. OMG Business Motivation Model (BMM): A model of how strateg... |
Business rule management system : BPMS DBMS RDMS Business rules Business rules approach Business rules engine Decision Model and Notation |
Business rule management system : Workshop summary paper: Six Views on the Business Rule Management System |
Business rules engine : A business rules engine is a software system that executes one or more business rules in a runtime production environment. The rules might come from legal regulation ("An employee can be fired for any reason or no reason but not for an illegal reason"), company policy ("All customers that spend ... |
Business rules engine : In any IT application, business rules can change more frequently than other parts of the application code. Rules engines or inference engines serve as pluggable software components which execute business rules that a business rules approach has externalized or separated from application code. Th... |
Business rules engine : An article in Computerworld traces rules engines to the early 1990s and to products from the likes of Pegasystems, Fair Isaac Corp, ILOG and eMerge from Sapiens. |
Business rules engine : Many organizations' rules efforts combine aspects of what is generally considered workflow design with traditional rule design. This failure to separate the two approaches can lead to problems with the ability to re-use and control both business rules and workflows. Design approaches that avoid ... |
Business rules engine : There are a number of different types of rule engines. These types (generally) differ in how Rules are scheduled for execution. Most rules engines used by businesses are forward chaining, which can be further divided into two classes: The first class processes so-called production/inference rule... |
Business rules engine : One common use case for rules engines is standardized access control to applications. OASIS defines a rules engine architecture and standard dedicated to access control called XACML (eXtensible Access Control Markup Language). One key difference between a XACML rule engine and a business rule en... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.