text
stringlengths
12
14.7k
Imageability : Holahan, Charles J.; Sorenson, Paul F. (1985-09-01). "The role of figural organization in city imageability: An information processing analysis". Journal of Environmental Psychology. Smolík Filip (2019-05-21). "Imageability and Neighborhood Density Facilitate the Age of Word Acquisition in Czech". Journa...
IMS VDEX : IMS Vocabulary Definition Exchange (IMS VDEX) is a mark-up language or grammar for controlled vocabularies developed by IMS Global as an open specification, with the Final Specification being approved in February 2004. IMS VDEX allows the exchange and expression of simple machine-readable lists of human lang...
IMS VDEX : VDEX was designed to supplement other IMS specifications and the IEEE LOM standard by giving additional semantic control to tool developers. IMS VDEX could be used for the following purposes. It is used in practice for other purposes as well. Interfaces providing pre-defined choices – providing radio buttons...
IMS VDEX : The VDEX Information Model is represented in the diagram. A VDEX file describing a vocabulary comprises a number of information elements, most of which are relatively simple, such as a string representation of the default (human) language or a URI identifying the value domain (or vocabulary). Some of the ele...
IMS VDEX : ALOHA Metadata Tagging Tool — Java-based software project that can read IMS VDEX files. IVIMEDS 1G v1.0 Archived 2006-06-22 at the Wayback Machine – from The International Virtual Medical School – includes VDEX instances in curriculum maps. Partners can create their own maps in VDEX format and use these to h...
IMS VDEX : IMS Global Learning object metadata
IMS VDEX : ^ Marc van Coillie Using IMS VDEX for the EDS AP – EIfEL ^ Antonio Sarasa, Jose Manuel Canabal, Juan Carlos Sacristan, Raquel Jimenez Using IMS VDEX in Agrega Archived 2011-10-02 at the Wayback Machine
IMS VDEX : IMS VDEX — official resources by IMS global What is IMS VDEX Archived 2011-07-05 at the Wayback Machine — JISC CETIS CETIS Metadata and Digital Repository Special Interest Group (SIG) Archived 2006-10-11 at the Wayback Machine — mailing list for those in UK Higher and Further Education interested in creating...
Information Coding Classification : The Information Coding Classification (ICC) is a classification system covering almost all extant 6500 knowledge fields (knowledge domains). Its conceptualization goes beyond the scope of the well known library classification systems, such as Dewey Decimal Classification (DDC), Unive...
Information Coding Classification : At the end of 1970, Prof. Alwin Diemer, Univ.of Düsseldorf proposed to Ingetraut Dahlberg to undertake a philosophical dissertation on The universal classification system of knowledge, its ontological, epistemological, and information theoretical foundations. Diemer had in mind an in...
Information Coding Classification : These were published in full length in the book Wissensorganisation. Entwicklung, Aufgabe, Anwendung, Zukunft and the article Information Coding Classification. Geschichtliches, Prinzipien, Inhaltliches, hence it suffices to just mention their topics with some necessary additions. Pr...
Information Coding Classification : The first two levels of ICC can be represented by following matrix. The first hierarchical level of the 9 subject categories results from the first vertical array under codes 1-9. The second hierarchical level of subject categories is structured by the 9 functionally ordered form cat...
Information Coding Classification : Possibility to roughly structure documents, especially bibliographies and reference works. Structuring personal repertories, e.g. a Who's Who in Who's Who in Classification and Indexing Supporting the recollection of statistics by knowledge fields, e.g. also concerning university pro...
Integrated Operations in the High North : Integrated Operations in the High North (IOHN, IO High North or IO in the High North) is a unique collaboration project that during a four-year period starting May 2008 is working on designing, implementing and testing a Digital Platform for what in the upstream oil and gas ind...
Integrated Operations in the High North : The consortium consists of the following 22 participants (in alphabetical order):
Integrated Operations in the High North : Integrated Operations Semantic Web ISO 15926 aka Oil and Gas Ontology, an enabler for the next or second generation of Integrated Operations by integrating data across disciplines and business domains. Petroleum exploration in the Arctic POSC Caesar Association, the custodian o...
Integrated Operations in the High North : Integrated Operations in the High North Archived 2015-12-24 at the Wayback Machine website W3C workshop on Semantic Web in Oil and Gas industry, Houston, December 9–10, 2008. Position papers from several participants in IOHN. Semantic Days 2009 conference, Stavanger, May 18–20,...
Is-a : In knowledge representation and ontology components, including for object-oriented programming and design, is-a (also written as is_a or is a) is a subsumptive relationship between abstractions (e.g., types, classes), wherein one class A is a subclass of another class B (and so B is a superclass of A). In other ...
Is-a : The is-a relationship is to be contrasted with the has-a (has_a or has a) relationship between types (classes); confusing the relations has-a and is-a is a common error when designing a model (e.g., a computer program) of the real-world relationship between an object and its subordinate. The is-a relationship ma...
Is-a : Subtyping enables a given type to be substituted for another type or abstraction. Subtyping is said to establish an is-a relationship between the subtype and some existing abstraction, either implicitly or explicitly, depending on language support. The relationship can be expressed explicitly via inheritance in ...
Is-a : Liskov substitution principle explains a property, "If for each object o1 of type S there is an object o2 of type T such that for all programs P defined in terms of T, the behavior of P is unchanged when o1 is substituted for o2 then S is a subtype of T,". Following example shows a violation of LSP. Here is perh...
Is-a : Inheritance (object-oriented programming) Liskov substitution principle (in object-oriented programming) Subsumption Is-a Hypernymy (and supertype) Hyponymy (and subtype) Has-a Holonymy Meronymy
Is-a : Ronald J. Brachman; "What IS-A is and isn't. An Analysis of Taxonomic Links in Semantic Networks". IEEE Computer, 16 (10); October 1983 Jean-Luc Hainaut, Jean-Marc Hick, Vincent Englebert, Jean Henrard, Didier Roland: Understanding Implementations of IS-A Relations. ER 1996: 42-57
Ishikawa diagram : Ishikawa diagrams (also called fishbone diagrams, herringbone diagrams, cause-and-effect diagrams) are causal diagrams created by Kaoru Ishikawa that show the potential causes of a specific event. Common uses of the Ishikawa diagram are product design and quality defect prevention to identify potenti...
Ishikawa diagram : The defect, or the problem to be solved, is shown as the fish's head, facing to the right, with the causes extending to the left as fishbones; the ribs branch off the backbone for major causes, with sub-branches for root-causes, to as many levels as required. Ishikawa diagrams were popularized in the...
Ishikawa diagram : Root-cause analysis is intended to reveal key relationships among various variables, and the possible causes provide additional insight into process behavior. It shows high-level causes that lead to the problem encountered by providing a snapshot of the current situation. There can be confusion about...
Ishikawa diagram : Seven basic tools of quality Five whys Issue map Issue tree Resource management
Ishikawa diagram : Ishikawa, Kaoru (1990); (Translator: J. H. Loftus); Introduction to Quality Control; 448 p; ISBN 4-906224-61-X OCLC 61341428 Dale, Barrie G. et al. (2007); Managing Quality 5th ed; ISBN 978-1-4051-4279-3 OCLC 288977828 Project Management Institute (2015-01-01). Business Analysis for Practitioners. Pr...
ISO 15926 : The ISO 15926 is a standard for data integration, sharing, exchange, and hand-over between computer systems. The title, "Industrial automation systems and integration—Integration of life-cycle data for process plants including oil and gas production facilities", is regarded too narrow by the present ISO 159...
ISO 15926 : In 1991 a European Union ESPRIT-, named ProcessBase, started. The focus of this research project was to develop a data model for lifecycle information of a facility that would suit the requirements of the process industries. At the time that the project duration had elapsed, a consortium of companies involv...
ISO 15926 : In 2004, the National Institute of Science and Technology (NIST) released a report on the impact of the lack of digital interoperability in the capital projects industry. They pegged the cost of inadequate interoperability to be $5.8 billion per year. The full report is over 200 pages.
ISO 15926 : ISO 15926 has thirteen parts (as of February 2022): Part 1 - Overview and fundamental principles Part 2 - Data model Part 3 - Reference data for geometry and topology Part 4 - Reference Data, the terms used within facilities for the process industry Part 6 - Methodology for the development and validation of...
ISO 15926 : There are a number of projects working on the extension of the ISO 15926 standard in different application areas.
ISO 15926 : One of the main requirements was (and still is) that the scope of the data model covers the entire lifecycle of a facility (e.g. oil refinery) and its components (e.g. pipes, pumps and their parts, etc.). Since such a facility over such a long time entails many different types of activities on a myriad of d...
ISO 15926 : 15926.org: A forum for ISO 15926 discussions and team collaboration. iringtoday.com: - An online ISO 15926 thought leadership community geared toward engineering management. .15926 Editor Open-source software to view, edit and verify ISO 15926 data. XMpLant - A translation tool to convert 2D and 3D plant an...
Issue tree : An issue tree, also called logic tree, is a graphical breakdown of a question that dissects it into its different components vertically and that progresses into details as it reads to the right.: 47 Issue trees are useful in problem solving to identify the root causes of a problem as well as to identify it...
Issue tree : According to professor of strategy Arnaud Chevallier, elaborating an approach used at McKinsey & Company, there are two types of issue trees: diagnostic ones and solution ones. Diagnostic trees break down a "why" key question, identifying all the possible root causes for the problem. Solution trees break d...
Issue tree : Four basic rules can help ensure that issue trees are optimal, according to Chevallier: Consistently answer a "why" or a "how" question Progress from the key question to the analysis as it moves to the right Have branches that are mutually exclusive and collectively exhaustive (MECE) Use an insightful brea...
Issue tree : Fault tree analysis Five whys Horizon scanning Ishikawa diagram Root cause analysis Why–because analysis
Issue tree : Chevallier, Arnaud (11 April 2011). "Don't get lost in the terminology". powerful-problem-solving.com. Retrieved 2018-10-06. Issue trees, issue maps, logic trees, how trees, why trees, diagnostic trees, solution trees, decision trees, fact trees, hypothesis trees... How should you call your trees?... Call ...
ITools Resourceome : iTools is a distributed infrastructure for managing, discovery, comparison and integration of computational biology resources. iTools employs Biositemap technology to retrieve and service meta-data about diverse bioinformatics data services, tools, and web-services. iTools is developed by the Natio...
ITools Resourceome : Interactive iTools Server
Keyword AAA : Keyword AAA is a thesaurus created by the State Records Authority of New South Wales, Australia. It is often used to categorise documents in a document management system. The thesaurus is often implemented in terms of ISO 2788.
Keyword AAA : Keyword AAA Overview Developing and implementing a keyword thesaurus
KL-ONE : KL-ONE (pronounced "kay ell won") is a knowledge representation system in the tradition of semantic networks and frames; that is, it is a frame language. The system is an attempt to overcome semantic indistinctness in semantic network representations and to explicitly represent conceptual information as a stru...
KL-ONE : There is a whole family of KL-ONE-like systems. One of the innovations that KL-ONE initiated was the use of a deductive classifier, an automated reasoning engine that can validate a frame ontology and deduce new information about the ontology based on the initial information provided by a domain expert. Frames...
KL-ONE : Ontology language == References ==
Knowledge Engineering Environment : Knowledge Engineering Environment (KEE) is a frame-based development tool for expert systems. It was developed and sold by IntelliCorp, and was first released in 1983. It ran on Lisp machines, and was later ported to Lucid Common Lisp with the CLX library, an X Window System (X11) in...
Knowledge Engineering Environment : Expert system Frame language Inference engine IntelliCorp (software) Knowledge base Knowledge-based system Knowledge representation
Knowledge Engineering Environment : An Assessment of Tools for Building Large Knowledge-Based Systems
Knowledge organization : Knowledge organization (KO), organization of knowledge, organization of information, or information organization is an intellectual discipline concerned with activities such as document description, indexing, and classification that serve to provide systems of representation and order for knowl...
Knowledge processing for robots : KnowRob (Knowledge processing for robots) is a system which combines knowledge representation and reasoning methods to acquire and ground knowledge. This system is the backbone of openEASE. both under developing at the Institute for Artificial Intelligence at the University of Bremen, ...
Knowledge processing for robots : KnowRob can serve as a common sense framework for the integration of knowledge. This knowledge can be static encyclopedic knowledge, common sense knowledge, task descriptions, environment models, object information, observed actions, etc., which can come from different sources, like ma...
Knowledge processing for robots : To represent the knowledge, KnowRob uses the OWL ontology language and an extended first-order logic knowledge representation with computable predicates. To give the order of subactions, KnowRob includes a pair-wise ordering constrain, which gives a partial ordering. KnowRob adopts the...
Knowledge processing for robots : The knowledge is computed by external methods using Prolog queries. In the second version of the KnowRob system, is included a better structure of the packages and documentations. Which includes some extensions from the previous version, as well as a logic based language. For example, ...
Knowledge processing for robots : The goal of the KnowRob framework is to make semantic knowledge available for service robots. It is able to answer queries about missing information in vague instructions for tasks. This is possible with the actions hierarchical representation and information about objects which can be...
Knowledge processing for robots : Project homepage Institute for Artificial Intelligence
Knowledge space : In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in th...
Knowledge space : Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. C...
Knowledge space : Knowledge Space Theory-based models presume that an educational subject S can be modeled as a finite set Q of concepts, skills, or topics. Each feasible state of knowledge about S is then a subset of Q; the set of all such feasible states is K. The precise term for the information (Q, K) depends on th...
Knowledge space : In practice, there exist several methods to construct knowledge spaces. The most frequently used method is querying experts. There exist several querying algorithms that allow one or several experts to construct a knowledge space by answering a sequence of simple questions. Another method is to constr...
Knowledge value chain : A knowledge value chain is a sequence of intellectual tasks by which knowledge workers build their employer's unique competitive advantage and/or social and environmental benefit. As an example, the components of a research and development project form a knowledge value chain. Productivity impro...
Korean Decimal Classification : The Korean Decimal Classification (KDC) is a system of library classification used in South Korea. The structure and main level classes of the KDC are based on the Dewey Decimal Classification. The KDC is maintained and published by the Classification Committee of the Korean Library Asso...
Korean Decimal Classification : Multiple library classification systems had been developed for Korean libraries before the publication of the KDC. These included the Railway Bureau Library Classification(1920), the Korean Decimal Classification edited by Bong-Suk Park(known as KDCP, 1947), the Han-Un Decimal Classifica...
Korean Decimal Classification : The KDC classifies resources primarily by discipline, though some classes are collocated by subject. There are eight auxiliary mnemonic tables used to expand class numbers. The main classes of the KDC are the same as the main classes of the Dewey Decimal Classification, but four of those...
Korean Decimal Classification : KDC is used by a wide range of libraries within Korea, including by the National Library of Korea and most school and public libraries in the country, along with some university libraries, such as the one at Keimyung University. == References ==
Learning object metadata : Learning Object Metadata is a data model, usually encoded in XML, used to describe a learning object and similar digital resources used to support learning. The purpose of learning object metadata is to support the reusability of learning objects, to aid discoverability, and to facilitate the...
Learning object metadata : The IEEE working group that developed the standard defined learning objects, for the purposes of the standard, as being "any entity, digital or non-digital, that may be used for learning, education or training." This definition has struck many commentators as being rather broad in its scope, ...
Learning object metadata : Application profile Content package Dublin Core Learning object Metadata Metadata Standards OAI-PMH SCORM XML m:Learning Object Metadata
Learning object metadata : IMS Global LRMI (Learning Resource Metadata Initiative) cancore.athabascau.ca is a thorough element-by-element guide to implementing the IEEE LOM. www.imsglobal.org: IMS Global Learning Consortium Learning resource meta-data specification. ltsc.ieee.org: XML Binding Specification. www.intrall...
Library classification : A library classification is a system used within a library to organize materials, including books, sound and video recordings, electronic materials, etc., both on shelves and in catalogs and indexes. Each item is typically assigned a call number, which identifies the location of the item within...
Library classification : Library classification is an important and crucial aspect in library and information science. It is distinct from scientific classification in that it has as its goal to provide a useful ordering of documents rather than a theoretical organization of knowledge. Although it has the practical pur...
Library classification : Library classifications were preceded by classifications used by bibliographers such as Conrad Gessner. The earliest library classification schemes organized books in broad subject categories. The earliest known library classification scheme is the Pinakes by Callimachus, a scholar at the Libra...
Library classification : There are many standard systems of library classification in use, and many more have been proposed over the years. However, in general, classification systems can be divided into three types depending on how they are used: Universal schemes Covers all subjects, e.g. the Dewey Decimal Classifica...
Library classification : Classification types denote the classification or categorization according to the form or characteristics or qualities of a classification scheme or schemes. Method and system has similar meaning. Method or methods or system means the classification schemes like Dewey Decimal Classification or ...
Library classification : Library classification is associated with library (descriptive) cataloging under the rubric of cataloging and classification, sometimes grouped together as technical services. The library professional who engages in the process of cataloging and classifying library materials is called a catalog...
Library classification : As a result of differences in notation, history, use of enumeration, hierarchy, and facets, classification systems can differ in the following ways: Type of Notation: Notation can be pure (consisting of only numerals, for example) or mixed (consisting of letters and numerals, or letters, numera...
Library classification : Attribute-value system Categorization Classification (general theory) Decimal classification Document classification Information retrieval Knowledge organization Library management Library of Congress Subject Headings W. C. Berwick Sayers Subject indexing
Library classification : Chan, L. M. (1994). Cataloging and Classification: An Introduction. New York: McGraw-Hill. ISBN 9780070105065.
Linear belief function : Linear belief functions are an extension of the Dempster–Shafer theory of belief functions to the case when variables of interest are continuous. Examples of such variables include financial asset prices, portfolio performance, and other antecedent and consequent variables. The theory was origi...
Linear belief function : A linear belief function intends to represent our belief regarding the location of the true value as follows: We are certain that the truth is on a so-called certainty hyperplane but we do not know its exact location; along some dimensions of the certainty hyperplane, we believe the true value ...
Linear belief function : A linear belief function can represent both logical and probabilistic knowledge for three types of variables: deterministic such as an observable or controllable, random whose distribution is normal, and vacuous on which no knowledge bears. Logical knowledge is represented by linear equations, ...
Linear belief function : There are two basic operations for making inferences in expert systems using linear belief functions: combination and marginalization. Combination corresponds to the integration of knowledge whereas marginalization corresponds to the coarsening of knowledge. Making an inference involves combini...
Linear belief function : We may use an audit problem to illustrate the three types of variables as follows. Suppose we want to audit the ending balance of accounts receivable (E). As we saw earlier, E is equal to the beginning balance (B) plus the sales (S) for the period minus the cash receipts (C) on the sales plus a...
Living graph : In terms of knowledge representation, a living graph (also referred to as a lifeline, living timeline or fortune line) is a graph similar to a chronology timeline which places events along a horizontal axis, while the vertical axis is used to represent factors such as an emotional reaction to those event...
Living graph : Interactive Living Graph Templates in Flash FACE Living Graph Exercise
Logic Programming Associates : Logic Programming Associates (LPA) is a company specializing in logic programming and artificial intelligence software. LPA was founded in 1980 and is widely known for its range of Prolog compilers, the Flex expert system toolkit and most recently, VisiRule. LPA was established to exploit...
Logic Programming Associates : One of the first Prolog implementations made available by LPA was micro-PROLOG which ran on popular 8-bit home computers such as the Sinclair ZX Spectrum and Apple II. The 8-bit micro-PROLOG interpreter was soon followed by micro-PROLOG Professional one of the first Prolog implementations...
Logic Programming Associates : In 1989, in response to the rise of interest in Expert Systems and the emergence of products such as Crystal, GoldWorks, NExpert, LPA developed the Flex expert system toolkit, which incorporated frame-based reasoning with inheritance, rule-based programming and data-driven procedures. Fle...
Logic Programming Associates : In 1992, LPA helped set up the Prolog Vendors Group, a not-for-profit organization whose aim was to help promote Prolog by making people aware of its usage in industry.
Logic Programming Associates : Between 1996 and 1998, based on work co-funded through a DTI Smart award, LPA developed ScaffoldIT, a tool for building dynamic documents and intelligent web sites. This technology, built using the LPA Prolog engine and associated ProWeb Server, was able to generate complex, personalised ...
Logic Programming Associates : In 2004, LPA launched VisiRule a graphical tool for developing knowledge-based and decision support systems. VisiRule was described in IEEE Potentials in 2007 (see Drawing on your knowledge with VisiRule): VisiRule has been used in various sectors, to build legal expert systems, machine d...
Logic Programming Associates : For many years, LPA has worked closely with Valdis Krebs, an American-Latvian researcher, author, and consultant in the field of social and organizational network analysis. Valdis is the founder and chief scientist of Orgnet, and the creator of the popular Inflow software package. LPA Pro...
Logic Programming Associates : LPA home page About LPA on VisiRule web site Micro-PROLOG (in Spanish) Aspects of PROLOG History VisiRule demos VisiRule: a new graphical business rules tool from LPA A flex-based expert system for sewage treatment works support ESSE: An Expert System for Software Evaluation LPA Delivers ...
Logico-linguistic modeling : Logico-linguistic modeling is a method for building knowledge-based systems with a learning capability using conceptual models from soft systems methodology, modal predicate logic, and logic programming languages such as Prolog.
Logico-linguistic modeling : Logico-linguistic modeling is a six-stage method developed primarily for building knowledge-based systems (KBS), but it also has application in manual decision support systems and information source analysis. Logico-linguistic models have a superficial similarity to John F. Sowa's conceptua...
Logico-linguistic modeling : SSM employs three modeling devices i.e. rich pictures, root definitions, and conceptual models of human activity systems. The root definitions and conceptual models are built by stakeholders themselves in an iterative debate organized by a facilitator. The strengths of this method lie, firs...
Logico-linguistic modeling : The logico-linguistic modeling method comprises six stages.
Logico-linguistic modeling : While logico-linguistic modeling overcomes the problems found in SSM's transition from conceptual model to computer code, it does so at the expense of increased stakeholder constructed model complexity. The benefits of this complexity are questionable and this modeling method may be much ha...
Logico-linguistic modeling : Argument map Cognitive map Concept map Fuzzy cognitive map Knowledge representation and reasoning Rhetorical structure theory Semantic network